Stein, A.
1991-01-01
The theory and practical application of techniques of statistical interpolation are studied in this thesis, and new developments in multivariate spatial interpolation and the design of sampling plans are discussed. Several applications to studies in soil science are
Interpolation functors and interpolation spaces
Brudnyi, Yu A
1991-01-01
The theory of interpolation spaces has its origin in the classical work of Riesz and Marcinkiewicz but had its first flowering in the years around 1960 with the pioneering work of Aronszajn, Calderón, Gagliardo, Krein, Lions and a few others. It is interesting to note that what originally triggered off this avalanche were concrete problems in the theory of elliptic boundary value problems related to the scale of Sobolev spaces. Later on, applications were found in many other areas of mathematics: harmonic analysis, approximation theory, theoretical numerical analysis, geometry of Banach spaces, nonlinear functional analysis, etc. Besides this the theory has a considerable internal beauty and must by now be regarded as an independent branch of analysis, with its own problems and methods. Further development in the 1970s and 1980s included the solution by the authors of this book of one of the outstanding questions in the theory of the real method, the K-divisibility problem. In a way, this book harvests the r...
Sixtus, Frederick
2009-01-01
Inhalt: Interpol - Kurzer geschichtlicher Abriss - Interpol heute - Struktur - Die Kernfunktionen Interpols Europol (oder: Europäisches Polizeiamt) - Kurzer geschichtlicher Abriss - Europol heute - Struktur Die Kontrolle Europols - Die Kernaufgaben Europols - Wie arbeiten die internationalen Polizeibehörden tatsächlich? - Vorboten einer Weltpolizei?
SPLINE, Spline Interpolation Function
International Nuclear Information System (INIS)
Allouard, Y.
1977-01-01
1 - Nature of physical problem solved: The problem is to obtain an interpolated function, as smooth as possible, that passes through given points. The derivatives of these functions are continuous up to the (2Q-1) order. The program consists of the following two subprograms: ASPLERQ. Transport of relations method for the spline functions of interpolation. SPLQ. Spline interpolation. 2 - Method of solution: The methods are described in the reference under item 10
Monotone piecewise bicubic interpolation
International Nuclear Information System (INIS)
Carlson, R.E.; Fritsch, F.N.
1985-01-01
In a 1980 paper the authors developed a univariate piecewise cubic interpolation algorithm which produces a monotone interpolant to monotone data. This paper is an extension of those results to monotone script C 1 piecewise bicubic interpolation to data on a rectangular mesh. Such an interpolant is determined by the first partial derivatives and first mixed partial (twist) at the mesh points. Necessary and sufficient conditions on these derivatives are derived such that the resulting bicubic polynomial is monotone on a single rectangular element. These conditions are then simplified to a set of sufficient conditions for monotonicity. The latter are translated to a system of linear inequalities, which form the basis for a monotone piecewise bicubic interpolation algorithm. 4 references, 6 figures, 2 tables
Linear interpolation of histograms
Read, A L
1999-01-01
A prescription is defined for the interpolation of probability distributions that are assumed to have a linear dependence on a parameter of the distributions. The distributions may be in the form of continuous functions or histograms. The prescription is based on the weighted mean of the inverses of the cumulative distributions between which the interpolation is made. The result is particularly elegant for a certain class of distributions, including the normal and exponential distributions, and is useful for the interpolation of Monte Carlo simulation results which are time-consuming to obtain.
Feature displacement interpolation
DEFF Research Database (Denmark)
Nielsen, Mads; Andresen, Per Rønsholt
1998-01-01
Given a sparse set of feature matches, we want to compute an interpolated dense displacement map. The application may be stereo disparity computation, flow computation, or non-rigid medical registration. Also estimation of missing image data, may be phrased in this framework. Since the features...... often are very sparse, the interpolation model becomes crucial. We show that a maximum likelihood estimation based on the covariance properties (Kriging) show properties more expedient than methods such as Gaussian interpolation or Tikhonov regularizations, also including scale......-selection. The computational complexities are identical. We apply the maximum likelihood interpolation to growth analysis of the mandibular bone. Here, the features used are the crest-lines of the object surface....
Extension Of Lagrange Interpolation
Directory of Open Access Journals (Sweden)
Mousa Makey Krady
2015-01-01
Full Text Available Abstract In this paper is to present generalization of Lagrange interpolation polynomials in higher dimensions by using Gramers formula .The aim of this paper is to construct a polynomials in space with error tends to zero.
Calculating SPRT Interpolation Error
Filipe, E.; Gentil, S.; Lóio, I.; Bosma, R.; Peruzzi, A.
2018-02-01
Interpolation error is a major source of uncertainty in the calibration of standard platinum resistance thermometer (SPRT) in the subranges of the International Temperature Scale of 1990 (ITS-90). This interpolation error arises because the interpolation equations prescribed by the ITS-90 cannot perfectly accommodate all the SPRTs natural variations in the resistance-temperature behavior, and generates different forms of non-uniqueness. This paper investigates the type 3 non-uniqueness for fourteen SPRTs of five different manufacturers calibrated over the water-zinc subrange and demonstrates the use of the method of divided differences for calculating the interpolation error. The calculated maximum standard deviation of 0.25 mK (near 100°C) is similar to that observed in previous studies.
Simple monotonic interpolation scheme
International Nuclear Information System (INIS)
Greene, N.M.
1980-01-01
A procedure for presenting tabular data, such as are contained in the ENDF/B files, that is simpler, more general, and potentially much more compact than the present schemes used with ENDF/B is presented. The method has been successfully used for Bondarenko interpolation in a module of the AMPX system. 1 figure, 1 table
Fuzzy Interpolation and Other Interpolation Methods Used in Robot Calibrations
Directory of Open Access Journals (Sweden)
Ying Bai
2012-01-01
Full Text Available A novel interpolation algorithm, fuzzy interpolation, is presented and compared with other popular interpolation methods widely implemented in industrial robots calibrations and manufacturing applications. Different interpolation algorithms have been developed, reported, and implemented in many industrial robot calibrations and manufacturing processes in recent years. Most of them are based on looking for the optimal interpolation trajectories based on some known values on given points around a workspace. However, it is rare to build an optimal interpolation results based on some random noises, and this is one of the most popular topics in industrial testing and measurement applications. The fuzzy interpolation algorithm (FIA reported in this paper provides a convenient and simple way to solve this problem and offers more accurate interpolation results based on given position or orientation errors that are randomly distributed in real time. This method can be implemented in many industrial applications, such as manipulators measurements and calibrations, industrial automations, and semiconductor manufacturing processes.
International Nuclear Information System (INIS)
Blok, M. de; Nationaal Inst. voor Kernfysica en Hoge-Energiefysica
1990-01-01
This report describes a time-interpolator with which time differences can be measured using digital and analog techniques. It concerns a maximum measuring time of 6.4 μs with a resolution of 100 ps. Use is made of Emitter Coupled Logic (ECL) and analogues of high-frequency techniques. The difficulty which accompanies the use of ECL-logic is keeping as short as possible the mutual connections and closing properly the outputs in order to avoid reflections. The digital part of the time-interpolator consists of a continuous running clock and logic which converts an input signal into a start- and stop signal. The analog part consists of a Time to Amplitude Converter (TAC) and an analog to digital converter. (author). 3 refs.; 30 figs
Interpolating string field theories
International Nuclear Information System (INIS)
Zwiebach, B.
1992-01-01
This paper reports that a minimal area problem imposing different length conditions on open and closed curves is shown to define a one-parameter family of covariant open-closed quantum string field theories. These interpolate from a recently proposed factorizable open-closed theory up to an extended version of Witten's open string field theory capable of incorporating on shell closed strings. The string diagrams of the latter define a new decomposition of the moduli spaces of Riemann surfaces with punctures and boundaries based on quadratic differentials with both first order and second order poles
Smooth Phase Interpolated Keying
Borah, Deva K.
2007-01-01
Smooth phase interpolated keying (SPIK) is an improved method of computing smooth phase-modulation waveforms for radio communication systems that convey digital information. SPIK is applicable to a variety of phase-shift-keying (PSK) modulation schemes, including quaternary PSK (QPSK), octonary PSK (8PSK), and 16PSK. In comparison with a related prior method, SPIK offers advantages of better performance and less complexity of implementation. In a PSK scheme, the underlying information waveform that one seeks to convey consists of discrete rectangular steps, but the spectral width of such a waveform is excessive for practical radio communication. Therefore, the problem is to smooth the step phase waveform in such a manner as to maintain power and bandwidth efficiency without incurring an unacceptably large error rate and without introducing undesired variations in the amplitude of the affected radio signal. Although the ideal constellation of PSK phasor points does not cause amplitude variations, filtering of the modulation waveform (in which, typically, a rectangular pulse is converted to a square-root raised cosine pulse) causes amplitude fluctuations. If a power-efficient nonlinear amplifier is used in the radio communication system, the fluctuating-amplitude signal can undergo significant spectral regrowth, thus compromising the bandwidth efficiency of the system. In the related prior method, one seeks to solve the problem in a procedure that comprises two major steps: phase-value generation and phase interpolation. SPIK follows the two-step approach of the related prior method, but the details of the steps are different. In the phase-value-generation step, the phase values of symbols in the PSK constellation are determined by a phase function that is said to be maximally smooth and that is chosen to minimize the spectral spread of the modulated signal. In this step, the constellation is divided into two groups by assigning, to information symbols, phase values
A disposition of interpolation techniques
Knotters, M.; Heuvelink, G.B.M.
2010-01-01
A large collection of interpolation techniques is available for application in environmental research. To help environmental scientists in choosing an appropriate technique a disposition is made, based on 1) applicability in space, time and space-time, 2) quantification of accuracy of interpolated
Fuzzy linguistic model for interpolation
International Nuclear Information System (INIS)
Abbasbandy, S.; Adabitabar Firozja, M.
2007-01-01
In this paper, a fuzzy method for interpolating of smooth curves was represented. We present a novel approach to interpolate real data by applying the universal approximation method. In proposed method, fuzzy linguistic model (FLM) applied as universal approximation for any nonlinear continuous function. Finally, we give some numerical examples and compare the proposed method with spline method
Occlusion-Aware View Interpolation
Directory of Open Access Journals (Sweden)
Ince Serdar
2008-01-01
Full Text Available Abstract View interpolation is an essential step in content preparation for multiview 3D displays, free-viewpoint video, and multiview image/video compression. It is performed by establishing a correspondence among views, followed by interpolation using the corresponding intensities. However, occlusions pose a significant challenge, especially if few input images are available. In this paper, we identify challenges related to disparity estimation and view interpolation in presence of occlusions. We then propose an occlusion-aware intermediate view interpolation algorithm that uses four input images to handle the disappearing areas. The algorithm consists of three steps. First, all pixels in view to be computed are classified in terms of their visibility in the input images. Then, disparity for each pixel is estimated from different image pairs depending on the computed visibility map. Finally, luminance/color of each pixel is adaptively interpolated from an image pair selected by its visibility label. Extensive experimental results show striking improvements in interpolated image quality over occlusion-unaware interpolation from two images and very significant gains over occlusion-aware spline-based reconstruction from four images, both on synthetic and real images. Although improvements are obvious only in the vicinity of object boundaries, this should be useful in high-quality 3D applications, such as digital 3D cinema and ultra-high resolution multiview autostereoscopic displays, where distortions at depth discontinuities are highly objectionable, especially if they vary with viewpoint change.
BIMOND3, Monotone Bivariate Interpolation
International Nuclear Information System (INIS)
Fritsch, F.N.; Carlson, R.E.
2001-01-01
1 - Description of program or function: BIMOND is a FORTRAN-77 subroutine for piecewise bi-cubic interpolation to data on a rectangular mesh, which reproduces the monotonousness of the data. A driver program, BIMOND1, is provided which reads data, computes the interpolating surface parameters, and evaluates the function on a mesh suitable for plotting. 2 - Method of solution: Monotonic piecewise bi-cubic Hermite interpolation is used. 3 - Restrictions on the complexity of the problem: The current version of the program can treat data which are monotone in only one of the independent variables, but cannot handle piecewise monotone data
The research on NURBS adaptive interpolation technology
Zhang, Wanjun; Gao, Shanping; Zhang, Sujia; Zhang, Feng
2017-04-01
In order to solve the problems of Research on NURBS Adaptive Interpolation Technology, such as interpolation time bigger, calculation more complicated, and NURBS curve step error are not easy changed and so on. This paper proposed a study on the algorithm for NURBS adaptive interpolation method of NURBS curve and simulation. We can use NURBS adaptive interpolation that calculates (xi, yi, zi). Simulation results show that the proposed NURBS curve interpolator meets the high-speed and high-accuracy interpolation requirements of CNC systems. The interpolation of NURBS curve should be finished. The simulation results show that the algorithm is correct; it is consistent with a NURBS curve interpolation requirements.
The EH Interpolation Spline and Its Approximation
Directory of Open Access Journals (Sweden)
Jin Xie
2014-01-01
Full Text Available A new interpolation spline with two parameters, called EH interpolation spline, is presented in this paper, which is the extension of the standard cubic Hermite interpolation spline, and inherits the same properties of the standard cubic Hermite interpolation spline. Given the fixed interpolation conditions, the shape of the proposed splines can be adjusted by changing the values of the parameters. Also, the introduced spline could approximate to the interpolated function better than the standard cubic Hermite interpolation spline and the quartic Hermite interpolation splines with single parameter by a new algorithm.
Temporal interpolation in Meteosat images
DEFF Research Database (Denmark)
Larsen, Rasmus; Hansen, Johan Dore; Ersbøll, Bjarne Kjær
in such animated films are perceived as being jerky due to t he low temporal sampling rate in general and missing images in particular. In order to perform a satisfactory temporal interpolation we estimate and use the optical flow corresponding to every image in the sequenc e. The estimation of the optical flow...... a threshold between clouds and land/water. The temperature maps are estimated using observations from the image sequence itself at cloud free pixels and ground temperature measurements from a series of meteor ological observation stations in Europe. The temporal interpolation of the images is bas ed on a path...... of each pixel determined by the estimated optical flow. The performance of the algorithm is illustrated by the interpolation of a sequence of Meteosat infrared images....
INTERPOL's Surveillance Network in Curbing Transnational Terrorism
Gardeazabal, Javier; Sandler, Todd
2015-01-01
This paper investigates the role that INTERPOL surveillance – the Mobile INTERPOL Network Database (MIND) and the Fixed INTERPOL Network Database (FIND) – played in the War on Terror since its inception in 2005. MIND/FIND surveillance allows countries to screen people and documents systematically at border crossings against INTERPOL databases on terrorists, fugitives, and stolen and lost travel documents. Such documents have been used in the past by terrorists to transit borders. By applyi...
Potential problems with interpolating fields
Energy Technology Data Exchange (ETDEWEB)
Birse, Michael C. [The University of Manchester, Theoretical Physics Division, School of Physics and Astronomy, Manchester (United Kingdom)
2017-11-15
A potential can have features that do not reflect the dynamics of the system it describes but rather arise from the choice of interpolating fields used to define it. This is illustrated using a toy model of scattering with two coupled channels. A Bethe-Salpeter amplitude is constructed which is a mixture of the waves in the two channels. The potential derived from this has a strong repulsive core, which arises from the admixture of the closed channel in the wave function and not from the dynamics of the model. (orig.)
Interpolation of rational matrix functions
Ball, Joseph A; Rodman, Leiba
1990-01-01
This book aims to present the theory of interpolation for rational matrix functions as a recently matured independent mathematical subject with its own problems, methods and applications. The authors decided to start working on this book during the regional CBMS conference in Lincoln, Nebraska organized by F. Gilfeather and D. Larson. The principal lecturer, J. William Helton, presented ten lectures on operator and systems theory and the interplay between them. The conference was very stimulating and helped us to decide that the time was ripe for a book on interpolation for matrix valued functions (both rational and non-rational). When the work started and the first partial draft of the book was ready it became clear that the topic is vast and that the rational case by itself with its applications is already enough material for an interesting book. In the process of writing the book, methods for the rational case were developed and refined. As a result we are now able to present the rational case as an indepe...
Evaluation of various interpolants available in DICE
Energy Technology Data Exchange (ETDEWEB)
Turner, Daniel Z. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Reu, Phillip L. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Crozier, Paul [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
2015-02-01
This report evaluates several interpolants implemented in the Digital Image Correlation Engine (DICe), an image correlation software package developed by Sandia. By interpolants we refer to the basis functions used to represent discrete pixel intensity data as a continuous signal. Interpolation is used to determine intensity values in an image at non - pixel locations. It is also used, in some cases, to evaluate the x and y gradients of the image intensities. Intensity gradients subsequently guide the optimization process. The goal of this report is to inform analysts as to the characteristics of each interpolant and provide guidance towards the best interpolant for a given dataset. This work also serves as an initial verification of each of the interpolants implemented.
Analysis of ECT Synchronization Performance Based on Different Interpolation Methods
Directory of Open Access Journals (Sweden)
Yang Zhixin
2014-01-01
Full Text Available There are two synchronization methods of electronic transformer in IEC60044-8 standard: impulsive synchronization and interpolation. When the impulsive synchronization method is inapplicability, the data synchronization of electronic transformer can be realized by using the interpolation method. The typical interpolation methods are piecewise linear interpolation, quadratic interpolation, cubic spline interpolation and so on. In this paper, the influences of piecewise linear interpolation, quadratic interpolation and cubic spline interpolation for the data synchronization of electronic transformer are computed, then the computational complexity, the synchronization precision, the reliability, the application range of different interpolation methods are analyzed and compared, which can serve as guide studies for practical applications.
Differential Interpolation Effects in Free Recall
Petrusic, William M.; Jamieson, Donald G.
1978-01-01
Attempts to determine whether a sufficiently demanding and difficult interpolated task (shadowing, i.e., repeating aloud) would decrease recall for earlier-presented items as well as for more recent items. Listening to music was included as a second interpolated task. Results support views that serial position effects reflect a single process.…
Transfinite C2 interpolant over triangles
International Nuclear Information System (INIS)
Alfeld, P.; Barnhill, R.E.
1984-01-01
A transfinite C 2 interpolant on a general triangle is created. The required data are essentially C 2 , no compatibility conditions arise, and the precision set includes all polynomials of degree less than or equal to eight. The symbol manipulation language REDUCE is used to derive the scheme. The scheme is discretized to two different finite dimensional C 2 interpolants in an appendix
Interpolation of diffusion weighted imaging datasets
DEFF Research Database (Denmark)
Dyrby, Tim B; Lundell, Henrik; Burke, Mark W
2014-01-01
by the interpolation method used should be considered. The results indicate that conventional interpolation methods can be successfully applied to DWI datasets for mining anatomical details that are normally seen only at higher resolutions, which will aid in tractography and microstructural mapping of tissue...
An Improved Rotary Interpolation Based on FPGA
Directory of Open Access Journals (Sweden)
Mingyu Gao
2014-08-01
Full Text Available This paper presents an improved rotary interpolation algorithm, which consists of a standard curve interpolation module and a rotary process module. Compared to the conventional rotary interpolation algorithms, the proposed rotary interpolation algorithm is simpler and more efficient. The proposed algorithm was realized on a FPGA with Verilog HDL language, and simulated by the ModelSim software, and finally verified on a two-axis CNC lathe, which uses rotary ellipse and rotary parabolic as an example. According to the theoretical analysis and practical process validation, the algorithm has the following advantages: firstly, less arithmetic items is conducive for interpolation operation; and secondly the computing time is only two clock cycles of the FPGA. Simulations and actual tests have proved that the high accuracy and efficiency of the algorithm, which shows that it is highly suited for real-time applications.
Analysis of velocity planning interpolation algorithm based on NURBS curve
Zhang, Wanjun; Gao, Shanping; Cheng, Xiyan; Zhang, Feng
2017-04-01
To reduce interpolation time and Max interpolation error in NURBS (Non-Uniform Rational B-Spline) inter-polation caused by planning Velocity. This paper proposed a velocity planning interpolation algorithm based on NURBS curve. Firstly, the second-order Taylor expansion is applied on the numerator in NURBS curve representation with parameter curve. Then, velocity planning interpolation algorithm can meet with NURBS curve interpolation. Finally, simulation results show that the proposed NURBS curve interpolator meet the high-speed and high-accuracy interpolation requirements of CNC systems. The interpolation of NURBS curve should be finished.
Matching interpolation of CT faulted images based on corresponding object
International Nuclear Information System (INIS)
Chen Lingna
2005-01-01
For CT faulted images interpolation this paper presents a corresponding pint matching interpolation algorithm, which is based on object feature. Compared with the traditional interpolation algorithms, the new algorithm improves visual effect and its interpolation error. The computer experiments show that the algorithm can effectively improve the interpolation quality, especially more clear scene at the boundary. (authors)
Interferometric interpolation of sparse marine data
Hanafy, Sherif M.
2013-10-11
We present the theory and numerical results for interferometrically interpolating 2D and 3D marine surface seismic profiles data. For the interpolation of seismic data we use the combination of a recorded Green\\'s function and a model-based Green\\'s function for a water-layer model. Synthetic (2D and 3D) and field (2D) results show that the seismic data with sparse receiver intervals can be accurately interpolated to smaller intervals using multiples in the data. An up- and downgoing separation of both recorded and model-based Green\\'s functions can help in minimizing artefacts in a virtual shot gather. If the up- and downgoing separation is not possible, noticeable artefacts will be generated in the virtual shot gather. As a partial remedy we iteratively use a non-stationary 1D multi-channel matching filter with the interpolated data. Results suggest that a sparse marine seismic survey can yield more information about reflectors if traces are interpolated by interferometry. Comparing our results to those of f-k interpolation shows that the synthetic example gives comparable results while the field example shows better interpolation quality for the interferometric method. © 2013 European Association of Geoscientists & Engineers.
Comparison of interpolation and approximation methods for optical freeform synthesis
Voznesenskaya, Anna; Krizskiy, Pavel
2017-06-01
Interpolation and approximation methods for freeform surface synthesis are analyzed using the developed software tool. Special computer tool is developed and results of freeform surface modeling with piecewise linear interpolation, piecewise quadratic interpolation, cubic spline interpolation, Lagrange polynomial interpolation are considered. The most accurate interpolation method is recommended. Surface profiles are approximated with the square least method. The freeform systems are generated in optical design software.
Traffic volume estimation using network interpolation techniques.
2013-12-01
Kriging method is a frequently used interpolation methodology in geography, which enables estimations of unknown values at : certain places with the considerations of distances among locations. When it is used in transportation field, network distanc...
Revisiting Veerman’s interpolation method
DEFF Research Database (Denmark)
Christiansen, Peter; Bay, Niels Oluf
2016-01-01
for comparison. Bulge testing and tensile testing of aluminium sheets containingelectro-chemically etched circle grids are performed to experimentally determine the forming limit of the sheet material.The forming limit is determined using (a) Veerman’s interpolation method, (b) exact Lagrangian interpolation......This article describes an investigation of Veerman’s interpolation method and its applicability for determining sheet metalformability. The theoretical foundation is established and its mathematical assumptions are clarified. An exact Lagrangianinterpolation scheme is also established...... and (c) FEsimulations. A comparison of the determined forming limits yields insignificant differences in the limit strain obtainedwith Veerman’s method or exact Lagrangian interpolation for the two sheet metal forming processes investigated. Theagreement with the FE-simulations is reasonable....
Kuu plaat : Interpol Antics. Plaadid kauplusest Lasering
2005-01-01
Heliplaatidest: "Interpol Antics", Scooter "Mind the Gap", Slide-Fifty "The Way Ahead", Psyhhoterror "Freddy, löö esimesena!", Riho Sibul "Must", Bossacucanova "Uma Batida Diferente", "Biscantorat - Sound of the spirit from Glenstal Abbey"
Interpol pidas mõttetalguid / Allan Espenberg
Espenberg, Allan
2008-01-01
Maailma kriminaalspetsialistid tulid Venemaal kokku, et valida rahvusvahelisele kriminaalpolitsei organisatsioonile Interpol uus juhtkond ning määrata kindlaks oma lähemad ja kaugemad tööülesanded
NOAA Optimum Interpolation (OI) SST V2
National Oceanic and Atmospheric Administration, Department of Commerce — The optimum interpolation (OI) sea surface temperature (SST) analysis is produced weekly on a one-degree grid. The analysis uses in situ and satellite SST's plus...
Interpolation of uniformly absolutely continuous operators
Czech Academy of Sciences Publication Activity Database
Cobos, F.; Gogatishvili, Amiran; Opic, B.; Pick, L.
2013-01-01
Roč. 286, 5-6 (2013), s. 579-599 ISSN 0025-584X R&D Projects: GA ČR GA201/08/0383 Institutional support: RVO:67985840 Keywords : uniformly absolutely continuous operators * interpolation * type of an interpolation method Subject RIV: BA - General Mathematics Impact factor: 0.658, year: 2013 http://onlinelibrary.wiley.com/doi/10.1002/ mana .201100205/full
Integration and interpolation of sampled waveforms
International Nuclear Information System (INIS)
Stearns, S.D.
1978-01-01
Methods for integrating, interpolating, and improving the signal-to-noise ratio of digitized waveforms are discussed with regard to seismic data from underground tests. The frequency-domain integration method and the digital interpolation method of Schafer and Rabiner are described and demonstrated using test data. The use of bandpass filtering for noise reduction is also demonstrated. With these methods, a backlog of seismic test data has been successfully processed
Interpolation for a subclass of H
Indian Academy of Sciences (India)
|g(zm)| ≤ c |zm − zm |, ∀m ∈ N. Thus it is natural to pose the following interpolation problem for H. ∞. : DEFINITION 4. We say that (zn) is an interpolating sequence in the weak sense for H. ∞ if given any sequence of complex numbers (λn) verifying. |λn| ≤ c ψ(zn,z. ∗ n) |zn − zn |, ∀n ∈ N,. (4) there exists a product fg ∈ H.
Calculation of electromagnetic parameter based on interpolation algorithm
International Nuclear Information System (INIS)
Zhang, Wenqiang; Yuan, Liming; Zhang, Deyuan
2015-01-01
Wave-absorbing material is an important functional material of electromagnetic protection. The wave-absorbing characteristics depend on the electromagnetic parameter of mixed media. In order to accurately predict the electromagnetic parameter of mixed media and facilitate the design of wave-absorbing material, based on the electromagnetic parameters of spherical and flaky carbonyl iron mixture of paraffin base, this paper studied two different interpolation methods: Lagrange interpolation and Hermite interpolation of electromagnetic parameters. The results showed that Hermite interpolation is more accurate than the Lagrange interpolation, and the reflectance calculated with the electromagnetic parameter obtained by interpolation is consistent with that obtained through experiment on the whole. - Highlights: • We use interpolation algorithm on calculation of EM-parameter with limited samples. • Interpolation method can predict EM-parameter well with different particles added. • Hermite interpolation is more accurate than Lagrange interpolation. • Calculating RL based on interpolation is consistent with calculating RL from experiment
Linear Invariant Tensor Interpolation Applied to Cardiac Diffusion Tensor MRI
Gahm, Jin Kyu; Wisniewski, Nicholas; Kindlmann, Gordon; Kung, Geoffrey L.; Klug, William S.; Garfinkel, Alan; Ennis, Daniel B.
2015-01-01
Purpose Various methods exist for interpolating diffusion tensor fields, but none of them linearly interpolate tensor shape attributes. Linear interpolation is expected not to introduce spurious changes in tensor shape. Methods Herein we define a new linear invariant (LI) tensor interpolation method that linearly interpolates components of tensor shape (tensor invariants) and recapitulates the interpolated tensor from the linearly interpolated tensor invariants and the eigenvectors of a linearly interpolated tensor. The LI tensor interpolation method is compared to the Euclidean (EU), affine-invariant Riemannian (AI), log-Euclidean (LE) and geodesic-loxodrome (GL) interpolation methods using both a synthetic tensor field and three experimentally measured cardiac DT-MRI datasets. Results EU, AI, and LE introduce significant microstructural bias, which can be avoided through the use of GL or LI. Conclusion GL introduces the least microstructural bias, but LI tensor interpolation performs very similarly and at substantially reduced computational cost. PMID:23286085
Interpolation effects in tabulated interatomic potentials
Wen, M.; Whalen, S. M.; Elliott, R. S.; Tadmor, E. B.
2015-10-01
Empirical interatomic potentials are widely used in atomistic simulations due to their ability to compute the total energy and interatomic forces quickly relative to more accurate quantum calculations. The functional forms in these potentials are sometimes stored in a tabulated format, as a collection of data points (argument-value pairs), and a suitable interpolation (often spline-based) is used to obtain the function value at an arbitrary point. We explore the effect of these interpolations on the potential predictions by calculating the quasi-harmonic thermal expansion and finite-temperature elastic constant of a one-dimensional chain compared with molecular dynamics simulations. Our results show that some predictions are affected by the choice of interpolation regardless of the number of tabulated data points. Our results clearly indicate that the interpolation must be considered part of the potential definition, especially for lattice dynamics properties that depend on higher-order derivatives of the potential. This is facilitated by the Knowledgebase of Interatomic Models (KIM) project, in which both the tabulated data (‘parameterized model’) and the code that interpolates them to compute energy and forces (‘model driver’) are stored and given unique citeable identifiers. We have developed cubic and quintic spline model drivers for pair functional type models (EAM, FS, EMT) and uploaded them to the OpenKIM repository (https://openkim.org).
INTERPOL's Surveillance Network in Curbing Transnational Terrorism
Gardeazabal, Javier; Sandler, Todd
2015-01-01
Abstract This paper investigates the role that International Criminal Police Organization (INTERPOL) surveillance—the Mobile INTERPOL Network Database (MIND) and the Fixed INTERPOL Network Database (FIND)—played in the War on Terror since its inception in 2005. MIND/FIND surveillance allows countries to screen people and documents systematically at border crossings against INTERPOL databases on terrorists, fugitives, and stolen and lost travel documents. Such documents have been used in the past by terrorists to transit borders. By applying methods developed in the treatment‐effects literature, this paper establishes that countries adopting MIND/FIND experienced fewer transnational terrorist attacks than they would have had they not adopted MIND/FIND. Our estimates indicate that, on average, from 2008 to 2011, adopting and using MIND/FIND results in 0.5 fewer transnational terrorist incidents each year per 100 million people. Thus, a country like France with a population just above 64 million people in 2008 would have 0.32 fewer transnational terrorist incidents per year owing to its use of INTERPOL surveillance. This amounts to a sizeable average proportional reduction of about 30 percent.
2010-07-01
... 28 Judicial Administration 1 2010-07-01 2010-07-01 false Exemption of the INTERPOL-United States National Central Bureau (INTERPOL-USNCB) System. 16.103 Section 16.103 Judicial Administration DEPARTMENT... Privacy Act § 16.103 Exemption of the INTERPOL-United States National Central Bureau (INTERPOL-USNCB...
Interpolation of quasi-Banach spaces
International Nuclear Information System (INIS)
Tabacco Vignati, A.M.
1986-01-01
This dissertation presents a method of complex interpolation for familities of quasi-Banach spaces. This method generalizes the theory for families of Banach spaces, introduced by others. Intermediate spaces in several particular cases are characterized using different approaches. The situation when all the spaces have finite dimensions is studied first. The second chapter contains the definitions and main properties of the new interpolation spaces, and an example concerning the Schatten ideals associated with a separable Hilbert space. The case of L/sup P/ spaces follows from the maximal operator theory contained in Chapter III. Also introduced is a different method of interpolation for quasi-Banach lattices of functions, and conditions are given to guarantee that the two techniques yield the same result. Finally, the last chapter contains a different, and more direct, approach to the case of Hardy spaces
Multiscale empirical interpolation for solving nonlinear PDEs
Calo, Victor M.
2014-12-01
In this paper, we propose a multiscale empirical interpolation method for solving nonlinear multiscale partial differential equations. The proposed method combines empirical interpolation techniques and local multiscale methods, such as the Generalized Multiscale Finite Element Method (GMsFEM). To solve nonlinear equations, the GMsFEM is used to represent the solution on a coarse grid with multiscale basis functions computed offline. Computing the GMsFEM solution involves calculating the system residuals and Jacobians on the fine grid. We use empirical interpolation concepts to evaluate these residuals and Jacobians of the multiscale system with a computational cost which is proportional to the size of the coarse-scale problem rather than the fully-resolved fine scale one. The empirical interpolation method uses basis functions which are built by sampling the nonlinear function we want to approximate a limited number of times. The coefficients needed for this approximation are computed in the offline stage by inverting an inexpensive linear system. The proposed multiscale empirical interpolation techniques: (1) divide computing the nonlinear function into coarse regions; (2) evaluate contributions of nonlinear functions in each coarse region taking advantage of a reduced-order representation of the solution; and (3) introduce multiscale proper-orthogonal-decomposition techniques to find appropriate interpolation vectors. We demonstrate the effectiveness of the proposed methods on several nonlinear multiscale PDEs that are solved with Newton\\'s methods and fully-implicit time marching schemes. Our numerical results show that the proposed methods provide a robust framework for solving nonlinear multiscale PDEs on a coarse grid with bounded error and significant computational cost reduction.
Positivity Preserving Interpolation Using Rational Bicubic Spline
Directory of Open Access Journals (Sweden)
Samsul Ariffin Abdul Karim
2015-01-01
Full Text Available This paper discusses the positivity preserving interpolation for positive surfaces data by extending the C1 rational cubic spline interpolant of Karim and Kong to the bivariate cases. The partially blended rational bicubic spline has 12 parameters in the descriptions where 8 of them are free parameters. The sufficient conditions for the positivity are derived on every four boundary curves network on the rectangular patch. Numerical comparison with existing schemes also has been done in detail. Based on Root Mean Square Error (RMSE, our partially blended rational bicubic spline is on a par with the established methods.
Image coding using adaptive recursive interpolative DPCM.
Gifford, E A; Hunt, B R; Marcellin, M W
1995-01-01
A predictive image coder having minimal decoder complexity is presented. The image coder utilizes recursive interpolative DPCM in conjunction with adaptive classification, entropy-constrained trellis coded quantization, and optimal rate allocation to obtain signal-to-noise ratios (SNRs) in the range of those provided by the most advanced transform coders.
Interpolation of intermolecular potentials using Gaussian processes
Uteva, Elena; Graham, Richard S.; Wilkinson, Richard D.; Wheatley, Richard J.
2017-10-01
A procedure is proposed to produce intermolecular potential energy surfaces from limited data. The procedure involves generation of geometrical configurations using a Latin hypercube design, with a maximin criterion, based on inverse internuclear distances. Gaussian processes are used to interpolate the data, using over-specified inverse molecular distances as covariates, greatly improving the interpolation. Symmetric covariance functions are specified so that the interpolation surface obeys all relevant symmetries, reducing prediction errors. The interpolation scheme can be applied to many important molecular interactions with trivial modifications. Results are presented for three systems involving CO2, a system with a deep energy minimum (HF-HF), and a system with 48 symmetries (CH4-N2). In each case, the procedure accurately predicts an independent test set. Training this method with high-precision ab initio evaluations of the CO2-CO interaction enables a parameter-free, first-principles prediction of the CO2-CO cross virial coefficient that agrees very well with experiments.
Statistical investigation of hydraulic driven circular interpolation ...
Indian Academy of Sciences (India)
2Mechanical Education Department, Gazi University, 06500 Ankara, Turkey. 3Electrical and Electronics Engineering .... PLC (Programmable Logic Controller) set. An incremental type linear encoder with ... realize the CNC basic motions such as linear (G01) and circular interpolation (G02, G03). 2.1 CNC system. The control ...
Interpolation for a subclass of H
Indian Academy of Sciences (India)
Abstract. We introduce and characterize two types of interpolating sequences in the unit disc D of the complex plane for the class of all functions being the product of two analytic functions in D, one bounded and another regular up to the boundary of D, concretely in the Lipschitz class, and at least one of them vanishing at ...
An efficient implementation of reconfigurable interpolation rootraised ...
Indian Academy of Sciences (India)
Hence, multiplexers, shifters, and adders in the multiplier structure are reduced, which results in theimprovement of operating frequency. The number of addition operations is further reduced using programmable adders and an efficient polyphase interpolation structure is implemented to reduce the hardware cost.
Interpolation for a subclass of H∞
Indian Academy of Sciences (India)
We introduce and characterize two types of interpolating sequences in the unit disc D of the complex plane for the class of all functions being the product of two analytic functions in D , one bounded and another regular up to the boundary of D , concretely in the Lipschitz class, and at least one of them vanishing at some ...
Research on an innovative modification algorithm of NURBS curve interpolation
Zhang, Wanjun; Gao, Shanping; Cheng, Xiyan; Zhang, Feng
2017-04-01
in order to solve the problems of modification algorithm of NURBS curve interpolation, Such as interpolation time bigger, NURBS curve step error and chord error are not easy changed, and so on. A novel proposed a modification algorithm of NURBS curve interpolation. The algorithm has merits such as higher interpolation position accuracy, short processing time and so on. In this simulation, an open five-axis CNC platform based on SIEMENS 840D CNC system is developed for verifying the proposed modification algorithm of NURBS curve interpolation experimentally. The simulation results show that the algorithm is correct; it is consistent with a NURBS curve interpolation requirements.
Application of Hardy's multiquadric interpolation to hydrodynamics
Energy Technology Data Exchange (ETDEWEB)
Kansa, E.J.
1985-10-01
Hardy's multiquadric interpolation (MQI) scheme is a global, continuously differentiable interpolation method for solving scattered data interpolation problems. It is capable of producing monotonic, extremely accurate interpolating functions, integrals, and derivatives. Derivative estimates for a variety of one and two-dimensional surfaces were obtained. MQI was then applied to the spherical blast wave problem of von Neumann. The numerical solution agreed extremely well with the exact solution. 17 refs., 3 figs., 2 tabs.
Parametric Integration by Magic Point Empirical Interpolation
Gaß, Maximilian; Glau, Kathrin
2015-01-01
We derive analyticity criteria for explicit error bounds and an exponential rate of convergence of the magic point empirical interpolation method introduced by Barrault et al. (2004). Furthermore, we investigate its application to parametric integration. We find that the method is well-suited to Fourier transforms and has a wide range of applications in such diverse fields as probability and statistics, signal and image processing, physics, chemistry and mathematical finance. To illustrate th...
Parametric Integration by Magic Point Empirical Interpolation
Gaß, M., Glau, K.
2016-01-01
We derive analyticity criteria for explicit error bounds and an exponential rate of convergence of the magic point empirical interpolation method introduced by Barrault et al. (2004). Furthermore, we investigate its application to parametric integration. We find that the method is well-suited to Fourier transforms and has a wide range of applications in such diverse fields as probability and statistics, signal and image processing, physics, chemistry and mathematical finance. To illustrate th...
Some splines produced by smooth interpolation
Czech Academy of Sciences Publication Activity Database
Segeth, Karel
2018-01-01
Roč. 319, 15 February (2018), s. 387-394 ISSN 0096-3003 R&D Projects: GA ČR GA14-02067S Institutional support: RVO:67985840 Keywords : smooth data approximation * smooth data interpolation * cubic spline Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.738, year: 2016 http://www. science direct.com/ science /article/pii/S0096300317302746?via%3Dihub
Some splines produced by smooth interpolation
Czech Academy of Sciences Publication Activity Database
Segeth, Karel
2018-01-01
Roč. 319, 15 February (2018), s. 387-394 ISSN 0096-3003 R&D Projects: GA ČR GA14-02067S Institutional support: RVO:67985840 Keywords : smooth data approximation * smooth data interpolation * cubic spline Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.738, year: 2016 http://www.sciencedirect.com/science/article/pii/S0096300317302746?via%3Dihub
Delimiting areas of endemism through kernel interpolation.
Oliveira, Ubirajara; Brescovit, Antonio D; Santos, Adalberto J
2015-01-01
We propose a new approach for identification of areas of endemism, the Geographical Interpolation of Endemism (GIE), based on kernel spatial interpolation. This method differs from others in being independent of grid cells. This new approach is based on estimating the overlap between the distribution of species through a kernel interpolation of centroids of species distribution and areas of influence defined from the distance between the centroid and the farthest point of occurrence of each species. We used this method to delimit areas of endemism of spiders from Brazil. To assess the effectiveness of GIE, we analyzed the same data using Parsimony Analysis of Endemism and NDM and compared the areas identified through each method. The analyses using GIE identified 101 areas of endemism of spiders in Brazil GIE demonstrated to be effective in identifying areas of endemism in multiple scales, with fuzzy edges and supported by more synendemic species than in the other methods. The areas of endemism identified with GIE were generally congruent with those identified for other taxonomic groups, suggesting that common processes can be responsible for the origin and maintenance of these biogeographic units.
Delimiting areas of endemism through kernel interpolation.
Directory of Open Access Journals (Sweden)
Ubirajara Oliveira
Full Text Available We propose a new approach for identification of areas of endemism, the Geographical Interpolation of Endemism (GIE, based on kernel spatial interpolation. This method differs from others in being independent of grid cells. This new approach is based on estimating the overlap between the distribution of species through a kernel interpolation of centroids of species distribution and areas of influence defined from the distance between the centroid and the farthest point of occurrence of each species. We used this method to delimit areas of endemism of spiders from Brazil. To assess the effectiveness of GIE, we analyzed the same data using Parsimony Analysis of Endemism and NDM and compared the areas identified through each method. The analyses using GIE identified 101 areas of endemism of spiders in Brazil GIE demonstrated to be effective in identifying areas of endemism in multiple scales, with fuzzy edges and supported by more synendemic species than in the other methods. The areas of endemism identified with GIE were generally congruent with those identified for other taxonomic groups, suggesting that common processes can be responsible for the origin and maintenance of these biogeographic units.
Image Interpolation Scheme based on SVM and Improved PSO
Jia, X. F.; Zhao, B. T.; Liu, X. X.; Song, H. P.
2018-01-01
In order to obtain visually pleasing images, a support vector machines (SVM) based interpolation scheme is proposed, in which the improved particle swarm optimization is applied to support vector machine parameters optimization. Training samples are constructed by the pixels around the pixel to be interpolated. Then the support vector machine with optimal parameters is trained using training samples. After the training, we can get the interpolation model, which can be employed to estimate the unknown pixel. Experimental result show that the interpolated images get improvement PNSR compared with traditional interpolation methods, which is agrees with the subjective quality.
Spline interpolations besides wood model widely used in lactation
Korkmaz, Mehmet
2017-04-01
In this study, for lactation curve, spline interpolations, alternative modeling passing through exactly all data points with respect to widely used Wood model applied to lactation data were be discussed. These models are linear spline, quadratic spline and cubic spline. The observed and estimated values according to spline interpolations and Wood model were given with their Error Sum of Squares and also the lactation curves of spline interpolations and widely used Wood model were shown on the same graph. Thus, the differences have been observed. The estimates for some intermediate values were done by using spline interpolations and Wood model. By using spline interpolations, the estimates of intermediate values could be made more precise. Furthermore, by using spline interpolations, the predicted values for missing or incorrect observation were very successful according to the values of Wood model. By using spline interpolations, new ideas and interpretations in addition to the information of the well-known classical analysis were shown to the investigators.
Research progress and hotspot analysis of spatial interpolation
Jia, Li-juan; Zheng, Xin-qi; Miao, Jin-li
2018-02-01
In this paper, the literatures related to spatial interpolation between 1982 and 2017, which are included in the Web of Science core database, are used as data sources, and the visualization analysis is carried out according to the co-country network, co-category network, co-citation network, keywords co-occurrence network. It is found that spatial interpolation has experienced three stages: slow development, steady development and rapid development; The cross effect between 11 clustering groups, the main convergence of spatial interpolation theory research, the practical application and case study of spatial interpolation and research on the accuracy and efficiency of spatial interpolation. Finding the optimal spatial interpolation is the frontier and hot spot of the research. Spatial interpolation research has formed a theoretical basis and research system framework, interdisciplinary strong, is widely used in various fields.
Interpolating of climate data using R
Reinhardt, Katja
2017-04-01
Interpolation methods are used in many different geoscientific areas, such as soil physics, climatology and meteorology. Thereby, unknown values are calculated by using statistical calculation approaches applied on known values. So far, the majority of climatologists have been using computer languages, such as FORTRAN or C++, but there is also an increasing number of climate scientists using R for data processing and visualization. Most of them, however, are still working with arrays and vector based data which is often associated with complex R code structures. For the presented study, I have decided to convert the climate data into geodata and to perform the whole data processing using the raster package, gstat and similar packages, providing a much more comfortable way for data handling. A central goal of my approach is to create an easy to use, powerful and fast R script, implementing the entire geodata processing and visualization into a single and fully automated R based procedure, which allows avoiding the necessity of using other software packages, such as ArcGIS or QGIS. Thus, large amount of data with recurrent process sequences can be processed. The aim of the presented study, which is located in western Central Asia, is to interpolate wind data based on the European reanalysis data Era-Interim, which are available as raster data with a resolution of 0.75˚ x 0.75˚ , to a finer grid. Therefore, various interpolation methods are used: inverse distance weighting, the geostatistical methods ordinary kriging and regression kriging, generalized additve model and the machine learning algorithms support vector machine and neural networks. Besides the first two mentioned methods, the methods are used with influencing factors, e.g. geopotential and topography.
Nuclear data banks generation by interpolation
International Nuclear Information System (INIS)
Castillo M, J. A.
1999-01-01
Nuclear Data Bank generation, is a process in which a great amount of resources is required, both computing and humans. If it is taken into account that at some times it is necessary to create a great amount of those, it is convenient to have a reliable tool that generates Data Banks with the lesser resources, in the least possible time and with a very good approximation. In this work are shown the results obtained during the development of INTPOLBI code, use to generate Nuclear Data Banks employing bicubic polynominal interpolation, taking as independent variables the uranium and gadolinia percents. Two proposal were worked, applying in both cases the finite element method, using one element with 16 nodes to carry out the interpolation. In the first proposals the canonic base was employed, to obtain the interpolating polynomial and later, the corresponding linear equation systems. In the solution of this systems the Gaussian elimination methods with partial pivot was applied. In the second case, the Newton base was used to obtain the mentioned system, resulting in a triangular inferior matrix, which structure, applying elemental operations, to obtain a blocks diagonal matrix, with special characteristics and easier to work with. For the validation tests, a comparison was made between the values obtained with INTPOLBI and INTERTEG (create at the Instituto de Investigaciones Electricas (MX) with the same purpose) codes, and Data Banks created through the conventional process, that is, with nuclear codes normally used. Finally, it is possible to conclude that the Nuclear Data Banks generated with INTPOLBI code constitute a very good approximation that, even though do not wholly replace conventional process, however are helpful in cases when it is necessary to create a great amount of Data Banks
Generation of nuclear data banks through interpolation
International Nuclear Information System (INIS)
Castillo M, J.A.
1999-01-01
Nuclear Data Bank generation, is a process in which a great amount of resources is required, both computing and humans. If it is taken into account that at some times it is necessary to create a great amount of those, it is convenient to have a reliable tool that generates Data Banks with the lesser resources, in the least possible time and with a very good approximation. In this work are shown the results obtained during the development of INTPOLBI code, used to generate Nuclear Data Banks employing bi cubic polynomial interpolation, taking as independent variables the uranium and gadolinium percents. Two proposals were worked, applying in both cases the finite element method, using one element with 16 nodes to carry out the interpolation. In the first proposals the canonic base was employed to obtain the interpolating polynomial and later, the corresponding linear equations system. In the solution of this system the Gaussian elimination method with partial pivot was applied. In the second case, the Newton base was used to obtain the mentioned system, resulting in a triangular inferior matrix, which structure, applying elemental operations, to obtain a blocks diagonal matrix, with special characteristics and easier to work with. For the validations test, a comparison was made between the values obtained with INTPOLBI and INTERTEG (created at the Instituto de Investigaciones Electricas with the same purpose) codes, and Data Banks created through the conventional process, that is, with nuclear codes normally used. Finally, it is possible to conclude that the Nuclear Data Banks generated with INTPOLBI code constitute a very good approximation that, even though do not wholly replace conventional process, however are helpful in cases when it is necessary to create a great amount of Data Banks. (Author)
Calculation of reactivity without Lagrange interpolation
International Nuclear Information System (INIS)
Suescun D, D.; Figueroa J, J. H.; Rodriguez R, K. C.; Villada P, J. P.
2015-09-01
A new method to solve numerically the inverse equation of punctual kinetics without using Lagrange interpolating polynomial is formulated; this method uses a polynomial approximation with N points based on a process of recurrence for simulating different forms of nuclear power. The results show a reliable accuracy. Furthermore, the method proposed here is suitable for real-time measurements of reactivity, with step sizes of calculations greater that Δt = 0.3 s; due to its precision can be used to implement a digital meter of reactivity in real time. (Author)
Topics in multivariate approximation and interpolation
Jetter, Kurt
2005-01-01
This book is a collection of eleven articles, written by leading experts and dealing with special topics in Multivariate Approximation and Interpolation. The material discussed here has far-reaching applications in many areas of Applied Mathematics, such as in Computer Aided Geometric Design, in Mathematical Modelling, in Signal and Image Processing and in Machine Learning, to mention a few. The book aims at giving a comprehensive information leading the reader from the fundamental notions and results of each field to the forefront of research. It is an ideal and up-to-date introduction for gr
Air Quality Assessment Using Interpolation Technique
Directory of Open Access Journals (Sweden)
Awkash Kumar
2016-07-01
Full Text Available Air pollution is increasing rapidly in almost all cities around the world due to increase in population. Mumbai city in India is one of the mega cities where air quality is deteriorating at a very rapid rate. Air quality monitoring stations have been installed in the city to regulate air pollution control strategies to reduce the air pollution level. In this paper, air quality assessment has been carried out over the sample region using interpolation techniques. The technique Inverse Distance Weighting (IDW of Geographical Information System (GIS has been used to perform interpolation with the help of concentration data on air quality at three locations of Mumbai for the year 2008. The classification was done for the spatial and temporal variation in air quality levels for Mumbai region. The seasonal and annual variations of air quality levels for SO2, NOx and SPM (Suspended Particulate Matter have been focused in this study. Results show that SPM concentration always exceeded the permissible limit of National Ambient Air Quality Standard. Also, seasonal trends of pollutant SPM was low in monsoon due rain fall. The finding of this study will help to formulate control strategies for rational management of air pollution and can be used for many other regions.
Size-Dictionary Interpolation for Robot's Adjustment
Directory of Open Access Journals (Sweden)
Morteza eDaneshmand
2015-05-01
Full Text Available This paper describes the classification and size-dictionary interpolation of the three-dimensional data obtained by a laser scanner to be used in a realistic virtual fitting room, where automatic activation of the chosen mannequin robot, while several mannequin robots of different genders and sizes are simultaneously connected to the same computer, is also considered to make it mimic the body shapes and sizes instantly. The classification process consists of two layers, dealing, respectively, with gender and size. The interpolation procedure tries to find out which set of the positions of the biologically-inspired actuators for activation of the mannequin robots could lead to the closest possible resemblance of the shape of the body of the person having been scanned, through linearly mapping the distances between the subsequent size-templates and the corresponding position set of the bioengineered actuators, and subsequently, calculating the control measures that could maintain the same distance proportions, where minimizing the Euclidean distance between the size-dictionary template vectors and that of the desired body sizes determines the mathematical description. In this research work, the experimental results of the implementation of the proposed method on Fits.me's mannequin robots are visually illustrated, and explanation of the remaining steps towards completion of the whole realistic online fitting package is provided.
Monotonicity preserving splines using rational cubic Timmer interpolation
Zakaria, Wan Zafira Ezza Wan; Alimin, Nur Safiyah; Ali, Jamaludin Md
2017-08-01
In scientific application and Computer Aided Design (CAD), users usually need to generate a spline passing through a given set of data, which preserves certain shape properties of the data such as positivity, monotonicity or convexity. The required curve has to be a smooth shape-preserving interpolant. In this paper a rational cubic spline in Timmer representation is developed to generate interpolant that preserves monotonicity with visually pleasing curve. To control the shape of the interpolant three parameters are introduced. The shape parameters in the description of the rational cubic interpolant are subjected to monotonicity constrained. The necessary and sufficient conditions of the rational cubic interpolant are derived and visually the proposed rational cubic Timmer interpolant gives very pleasing results.
Experimental Performance of Spatial Interpolators for Ground Water Salinity
International Nuclear Information System (INIS)
Alsaaran, Nasser A.
2005-01-01
Mapping groundwater qualities requires either sampling on a fine regular grid or spatial interpolation. The latter is usually used because the cost of the former is prohibitive. Experimental performance of five spatial interpolators for groundwater salinity was investigated using cross validation. The methods included ordinary kriging (OK), lognormal kriging, inverse distance, inverse squared distance and inverse cubed distance. The results show that OK outperformed other interpolators in terms of bias. Interpolation accuracy based on mean absolute difference criterion is relatively high for all interpolators with small difference among them. While three-dimensional surfaces produced by all inverse distance based procedures are dominated by isolated peaks and pits, surfaces produced by kriging are free from localized pits and peaks, and show areas of low groundwater salinity as elongated basins and areas of high salinity as ridges, which make regional trends easy to identify. Considering all criteria, OK was judged to be the most suitable spatial interpolator for groundwater salinity in this study. (author)
Multiresolution Motion Estimation for Low-Rate Video Frame Interpolation
Directory of Open Access Journals (Sweden)
Hezerul Abdul Karim
2004-09-01
Full Text Available Interpolation of video frames with the purpose of increasing the frame rate requires the estimation of motion in the image so as to interpolate pixels along the path of the objects. In this paper, the specific challenges of low-rate video frame interpolation are illustrated by choosing one well-performing algorithm for high-frame-rate interpolation (Castango 1996 and applying it to low frame rates. The degradation of performance is illustrated by comparing the original algorithm, the algorithm adapted to low frame rate, and simple averaging. To overcome the particular challenges of low-frame-rate interpolation, two algorithms based on multiresolution motion estimation are developed and compared on objective and subjective basis and shown to provide an elegant solution to the specific challenges of low-frame-rate video interpolation.
Data interpolation using rational cubic Ball spline with three parameters
Karim, Samsul Ariffin Abdul
2016-11-01
Data interpolation is an important task for scientific visualization. This research introduces new rational cubic Ball spline scheme with three parameters. The rational cubic Ball will be used for data interpolation with or without true derivative values. Error estimation show that the proposed scheme works well and is a very good interpolant to approximate the function. All graphical examples are presented by using Mathematica software.
Systems and methods for interpolation-based dynamic programming
Rockwood, Alyn
2013-01-03
Embodiments of systems and methods for interpolation-based dynamic programming. In one embodiment, the method includes receiving an object function and a set of constraints associated with the objective function. The method may also include identifying a solution on the objective function corresponding to intersections of the constraints. Additionally, the method may include generating an interpolated surface that is in constant contact with the solution. The method may also include generating a vector field in response to the interpolated surface.
An Evaluation of Interpol's Cooperative-Based Counterterrorism Linkages
Todd Sandler; Daniel G. Arce; Walter Enders
2011-01-01
This paper evaluates the payback from efforts of the International Criminal Police Organization (Interpol) to coordinate proactive counterterrorism measures by its member countries to arrest terrorists and weaken their ability to conduct operations. We use Interpol arrest data and data on utilization of Interpol resources by member countries to compute counterfactual benefit measurements, which, when matched with costs, yield benefit-cost ratios. The average of these ratios is approximately 2...
Distance-two interpolation for parallel algebraic multigrid
International Nuclear Information System (INIS)
Sterck, H de; Falgout, R D; Nolting, J W; Yang, U M
2007-01-01
In this paper we study the use of long distance interpolation methods with the low complexity coarsening algorithm PMIS. AMG performance and scalability is compared for classical as well as long distance interpolation methods on parallel computers. It is shown that the increased interpolation accuracy largely restores the scalability of AMG convergence factors for PMIS-coarsened grids, and in combination with complexity reducing methods, such as interpolation truncation, one obtains a class of parallel AMG methods that enjoy excellent scalability properties on large parallel computers
Indeterminacy of interpolation problems in the Stieltjes class
International Nuclear Information System (INIS)
Dyukarev, Yu M
2005-01-01
The concept of ordered families of interpolation problems in the Stieltjes class is introduced. Ordered families are used for the introduction of the concept of limiting interpolation problem in the same class. The limiting interpolation problem is proved to be soluble. A criterion for the complete indeterminacy of a limiting interpolation problem in the Stieltjes class is obtained. All solutions in the completely indeterminate case are described in terms of linear fractional transformations. General constructions are illustrated by the examples of the Stieltjes moment problem and the Nevanlinna-Pick problem in the Stieltjes class.
Electronic structure interpolation via atomic orbitals
Energy Technology Data Exchange (ETDEWEB)
Chen Mohan; Guo, G-C; He Lixin, E-mail: helx@ustc.edu.cn [Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei, 230026 (China)
2011-08-17
We present an efficient scheme for accurate electronic structure interpolation based on systematically improvable optimized atomic orbitals. The atomic orbitals are generated by minimizing the spillage value between the atomic basis calculations and the converged plane wave basis calculations on some coarse k-point grid. They are then used to calculate the band structure of the full Brillouin zone using the linear combination of atomic orbitals algorithms. We find that usually 16-25 orbitals per atom can give an accuracy of about 10 meV compared to the full ab initio calculations, and the accuracy can be systematically improved by using more atomic orbitals. The scheme is easy to implement and robust, and works equally well for metallic systems and systems with complicated band structures. Furthermore, the atomic orbitals have much better transferability than Shirley's basis and Wannier functions, which is very useful for perturbation calculations.
The role of interpolation in PVC-induced cardiomyopathy.
Olgun, Hilal; Yokokawa, Miki; Baman, Timir; Kim, Hyungjin Myra; Armstrong, William; Good, Eric; Chugh, Aman; Pelosi, Frank; Crawford, Thomas; Oral, Hakan; Morady, Fred; Bogun, Frank
2011-07-01
Frequent premature ventricular complexes (PVCs) can cause cardiomyopathy. The mechanism is not known and may be multifactorial. This study assessed the role of PVC interpolation in PVC-induced cardiomyopathy. In 51 consecutive patients (14 women, age 49 ± 15 years, ejection fraction (EF) 0.49 ± 0.14) with frequent PVCs, 24-hour Holter recordings were performed. The amount of interpolation was determined and correlated with the presence of PVC-induced cardiomyopathy. In addition, parameters measured during an electrophysiology study were correlated with the Holter findings. Fourteen of the 21 patients (67%) with cardiomyopathy had interpolated PVCs, compared with only 6 of 30 patients (20%) without PVC-induced cardiomyopathy (P PVC burden than patients without interpolation (28% ± 12% vs. 15% ± 15%; P = .002). The burden of interpolated PVCs correlated with the presence of PVC cardiomyopathy (21% ± 30% vs. 4% ± 13%; P = .008). Both PVC burden and interpolation independently predicted PVC-induced cardiomyopathy (odds ratio 1.07, 95% confidence interval 1.01 to 1.13, P = .02; and odds ratio 4.43, 95% confidence interval 1.06 to 18.48, P = .04, respectively). The presence of ventriculoatrial block at a ventricular pacing cycle length of 600 ms correlated with the presence of interpolation (P = .004). Patients with interpolation had a longer mean ventriculoatrial block cycle length than patients without interpolated PVCs (520 ± 110 ms vs. 394 ± 92 ms; P = .01). The presence of interpolated PVCs was predictive of the presence of PVC cardiomyopathy. Interpolation may play an important role in the generation of PVC-induced cardiomyopathy. Copyright © 2011 Heart Rhythm Society. Published by Elsevier Inc. All rights reserved.
Shape Preserving Interpolation Using C2 Rational Cubic Spline
Directory of Open Access Journals (Sweden)
Samsul Ariffin Abdul Karim
2016-01-01
Full Text Available This paper discusses the construction of new C2 rational cubic spline interpolant with cubic numerator and quadratic denominator. The idea has been extended to shape preserving interpolation for positive data using the constructed rational cubic spline interpolation. The rational cubic spline has three parameters αi, βi, and γi. The sufficient conditions for the positivity are derived on one parameter γi while the other two parameters αi and βi are free parameters that can be used to change the final shape of the resulting interpolating curves. This will enable the user to produce many varieties of the positive interpolating curves. Cubic spline interpolation with C2 continuity is not able to preserve the shape of the positive data. Notably our scheme is easy to use and does not require knots insertion and C2 continuity can be achieved by solving tridiagonal systems of linear equations for the unknown first derivatives di, i=1,…,n-1. Comparisons with existing schemes also have been done in detail. From all presented numerical results the new C2 rational cubic spline gives very smooth interpolating curves compared to some established rational cubic schemes. An error analysis when the function to be interpolated is ft∈C3t0,tn is also investigated in detail.
Visualizing and Understanding the Components of Lagrange and Newton Interpolation
Yang, Yajun; Gordon, Sheldon P.
2016-01-01
This article takes a close look at Lagrange and Newton interpolation by graphically examining the component functions of each of these formulas. Although interpolation methods are often considered simply to be computational procedures, we demonstrate how the components of the polynomial terms in these formulas provide insight into where these…
An efficient interpolation filter VLSI architecture for HEVC standard
Zhou, Wei; Zhou, Xin; Lian, Xiaocong; Liu, Zhenyu; Liu, Xiaoxiang
2015-12-01
The next-generation video coding standard of High-Efficiency Video Coding (HEVC) is especially efficient for coding high-resolution video such as 8K-ultra-high-definition (UHD) video. Fractional motion estimation in HEVC presents a significant challenge in clock latency and area cost as it consumes more than 40 % of the total encoding time and thus results in high computational complexity. With aims at supporting 8K-UHD video applications, an efficient interpolation filter VLSI architecture for HEVC is proposed in this paper. Firstly, a new interpolation filter algorithm based on the 8-pixel interpolation unit is proposed in this paper. It can save 19.7 % processing time on average with acceptable coding quality degradation. Based on the proposed algorithm, an efficient interpolation filter VLSI architecture, composed of a reused data path of interpolation, an efficient memory organization, and a reconfigurable pipeline interpolation filter engine, is presented to reduce the implement hardware area and achieve high throughput. The final VLSI implementation only requires 37.2k gates in a standard 90-nm CMOS technology at an operating frequency of 240 MHz. The proposed architecture can be reused for either half-pixel interpolation or quarter-pixel interpolation, which can reduce the area cost for about 131,040 bits RAM. The processing latency of our proposed VLSI architecture can support the real-time processing of 4:2:0 format 7680 × 4320@78fps video sequences.
Some observations on interpolating gauges and non-covariant gauges
Indian Academy of Sciences (India)
We discuss the viability of using interpolating gauges to deﬁne the non-covariant gauges starting from the covariant ones. We draw attention to the need for a very careful treatment of boundary condition deﬁning term. We show that the boundary condition needed to maintain gauge-invariance as the interpolating parameter ...
Algorithm for applying interpolation in digital signal processing ...
African Journals Online (AJOL)
Software-defined radios and test equipment use a variety of digital signal processing techniques to improve system performance. Interpolation is one technique that can be used to increase the sample rate of digital signals. In this work, we illustrated interpolation in the time domain by writing appropriate codes using ...
Interpolation of fuzzy data | Khodaparast | Journal of Fundamental ...
African Journals Online (AJOL)
In the current world and in the field of science and technology, interpolation issues are also of a fuzzy type, it has many scientific applications in developmental work, medical issues, imaging, engineering software and graphics. Therefore, in this article we intend to investigate Interpolation of fuzzy data in order to apply fuzzy ...
Selection of an Appropriate Interpolation Method for Rainfall Data In ...
African Journals Online (AJOL)
There are many interpolation methods in use with various limitations and likelihood of errors. This study applied five interpolation methods to existing rainfall data in central Nigeria to determine the most appropriate method that returned the best prediction of rainfall at an ungauged site. The methods include the inverse ...
Optimal interpolation schemes for particle tracking in turbulence
van Hinsberg, M.A.T.; ten Thije Boonkkamp, J.H.M.; Toschi, F.; Clercx, H.J.H.
2013-01-01
An important aspect in numerical simulations of particle-laden turbulent flows is the interpolation of the flow field needed for the computation of the Lagrangian trajectories. The accuracy of the interpolation method has direct consequences for the acceleration spectrum of the fluid particles and
EFEKTIFITAS PERANAN INTERPOL DALAM MENANGULANGI JARINGAN NARKOTIKA DI INDONESIA
RICHARD LIU, VINSENSIUS
2013-01-01
Penelitian ini bertujuan untuk mengetahui dan menjelaskan efektifitas peran Interpol didalam penanggulangan jaringan narkotika di Indonesia, Startegi Interpol dalam menangani jaringan narkotika internasional di Indonesia, dan sikap Pemerintah Indonesia dalam menanggulangi jaringan narkotika Internasional. Penulis membatasi penelitian ini dalam kurun waktu 3 tahun yaitu 2009-2011 Tipe penelitian yang penulis gunakan untuk mencapai tujuan penelitian adalah tipe penelitian deskriptif. Tekn...
A FRACTAL-BASED STOCHASTIC INTERPOLATION SCHEME IN SUBSURFACE HYDROLOGY
The need for a realistic and rational method for interpolating sparse data sets is widespread. Real porosity and hydraulic conductivity data do not vary smoothly over space, so an interpolation scheme that preserves irregularity is desirable. Such a scheme based on the properties...
Input variable selection for interpolating high-resolution climate ...
African Journals Online (AJOL)
Although the primary input data of climate interpolations are usually meteorological data, other related (independent) variables are frequently incorporated in the interpolation process. One such variable is elevation, which is known to have a strong influence on climate. This research investigates the potential of 4 additional ...
Steady State Stokes Flow Interpolation for Fluid Control
DEFF Research Database (Denmark)
Bhatacharya, Haimasree; Nielsen, Michael Bang; Bridson, Robert
2012-01-01
Fluid control methods often require surface velocities interpolated throughout the interior of a shape to use the velocity as a feedback force or as a boundary condition. Prior methods for interpolation in computer graphics — velocity extrapolation in the normal direction and potential flow...
The Use of Wavelets in Image Interpolation: Possibilities and Limitations
Directory of Open Access Journals (Sweden)
M. Grgic
2007-12-01
Full Text Available Discrete wavelet transform (DWT can be used in various applications, such as image compression and coding. In this paper we examine how DWT can be used in image interpolation. Afterwards proposed method is compared with two other traditional interpolation methods. For the case of magnified image achieved by interpolation, original image is unknown and there is no perfect way to judge the magnification quality. Common approach is to start with an original image, generate a lower resolution version of original image by downscaling, and then use different interpolation methods to magnify low resolution image. After that original and magnified images are compared to evaluate difference between them using different picture quality measures. Our results show that comparison of image interpolation methods depends on downscaling technique, image contents and quality metric. For fair comparison all these parameters need to be considered.
Interpolation from Grid Lines: Linear, Transfinite and Weighted Method
DEFF Research Database (Denmark)
Lindberg, Anne-Sofie Wessel; Jørgensen, Thomas Martini; Dahl, Vedrana Andersen
2017-01-01
When two sets of line scans are acquired orthogonal to each other, intensity values are known along the lines of a grid. To view these values as an image, intensities need to be interpolated at regularly spaced pixel positions. In this paper we evaluate three methods for interpolation from grid...... of transfinite method close to grid lines, and the stability of the linear method. We perform an extensive evaluation of the three interpolation methods across a range of upsampling rates for two data sets. Depending on the upsampling rate, we show significant difference in the performance of the three methods....... We find that the transfinite interpolation works well for small upsampling rates and the proposed weighted interpolation method performs very well for all relevant upsampling rates....
Scalable Intersample Interpolation Architecture for High-channel-count Beamformers
DEFF Research Database (Denmark)
Tomov, Borislav Gueorguiev; Nikolov, Svetoslav I; Jensen, Jørgen Arendt
2011-01-01
Modern ultrasound scanners utilize digital beamformers that operate on sampled and quantized echo signals. Timing precision is of essence for achieving good focusing. The direct way to achieve it is through the use of high sampling rates, but that is not economical, so interpolation between echo...... samples is used. This paper presents a beamformer architecture that combines a band-pass filter-based interpolation algorithm with the dynamic delay-and-sum focusing of a digital beamformer. The reduction in the number of multiplications relative to a linear perchannel interpolation and band-pass per......-channel interpolation architecture is respectively 58 % and 75 % beamformer for a 256-channel beamformer using 4-tap filters. The approach allows building high channel count beamformers while maintaining high image quality due to the use of sophisticated intersample interpolation....
Interpolator for numerically controlled machine tools
Bowers, Gary L.; Davenport, Clyde M.; Stephens, Albert E.
1976-01-01
A digital differential analyzer circuit is provided that depending on the embodiment chosen can carry out linear, parabolic, circular or cubic interpolation. In the embodiment for parabolic interpolations, the circuit provides pulse trains for the X and Y slide motors of a two-axis machine to effect tool motion along a parabolic path. The pulse trains are generated by the circuit in such a way that parabolic tool motion is obtained from information contained in only one block of binary input data. A part contour may be approximated by one or more parabolic arcs. Acceleration and initial velocity values from a data block are set in fixed bit size registers for each axis separately but simultaneously and the values are integrated to obtain the movement along the respective axis as a function of time. Integration is performed by continual addition at a specified rate of an integrand value stored in one register to the remainder temporarily stored in another identical size register. Overflows from the addition process are indicative of the integral. The overflow output pulses from the second integration may be applied to motors which position the respective machine slides according to a parabolic motion in time to produce a parabolic machine tool motion in space. An additional register for each axis is provided in the circuit to allow "floating" of the radix points of the integrand registers and the velocity increment to improve position accuracy and to reduce errors encountered when the acceleration integrand magnitudes are small when compared to the velocity integrands. A divider circuit is provided in the output of the circuit to smooth the output pulse spacing and prevent motor stall, because the overflow pulses produced in the binary addition process are spaced unevenly in time. The divider has the effect of passing only every nth motor drive pulse, with n being specifiable. The circuit inputs (integrands, rates, etc.) are scaled to give exactly n times the
Functions with disconnected spectrum sampling, interpolation, translates
Olevskii, Alexander M
2016-01-01
The classical sampling problem is to reconstruct entire functions with given spectrum S from their values on a discrete set L. From the geometric point of view, the possibility of such reconstruction is equivalent to determining for which sets L the exponential system with frequencies in L forms a frame in the space L^2(S). The book also treats the problem of interpolation of discrete functions by analytic ones with spectrum in S and the problem of completeness of discrete translates. The size and arithmetic structure of both the spectrum S and the discrete set L play a crucial role in these problems. After an elementary introduction, the authors give a new presentation of classical results due to Beurling, Kahane, and Landau. The main part of the book focuses on recent progress in the area, such as construction of universal sampling sets, high-dimensional and non-analytic phenomena. The reader will see how methods of harmonic and complex analysis interplay with various important concepts in different areas, ...
Spatiotemporal video deinterlacing using control grid interpolation
Venkatesan, Ragav; Zwart, Christine M.; Frakes, David H.; Li, Baoxin
2015-03-01
With the advent of progressive format display and broadcast technologies, video deinterlacing has become an important video-processing technique. Numerous approaches exist in the literature to accomplish deinterlacing. While most earlier methods were simple linear filtering-based approaches, the emergence of faster computing technologies and even dedicated video-processing hardware in display units has allowed higher quality but also more computationally intense deinterlacing algorithms to become practical. Most modern approaches analyze motion and content in video to select different deinterlacing methods for various spatiotemporal regions. We introduce a family of deinterlacers that employs spectral residue to choose between and weight control grid interpolation based spatial and temporal deinterlacing methods. The proposed approaches perform better than the prior state-of-the-art based on peak signal-to-noise ratio, other visual quality metrics, and simple perception-based subjective evaluations conducted by human viewers. We further study the advantages of using soft and hard decision thresholds on the visual performance.
Rainfall variation by geostatistical interpolation method
Directory of Open Access Journals (Sweden)
Glauber Epifanio Loureiro
2013-08-01
Full Text Available This article analyses the variation of rainfall in the Tocantins-Araguaia hydrographic region in the last two decades, based upon the rain gauge stations of the ANA (Brazilian National Water Agency HidroWeb database for the years 1983, 1993 and 2003. The information was systemized and treated with Hydrologic methods such as method of contour and interpolation for ordinary kriging. The treatment considered the consistency of the data, the density of the space distribution of the stations and the periods of study. The results demonstrated that the total volume of water precipitated annually did not change significantly in the 20 years analyzed. However, a significant variation occurred in its spatial distribution. By analyzing the isohyet it was shown that there is a displacement of the precipitation at Tocantins Baixo (TOB of approximately 10% of the total precipitated volume. This displacement can be caused by global change, by anthropogenic activities or by regional natural phenomena. However, this paper does not explore possible causes of the displacement.
Spatial interpolation mthods for integrating Newton's equation
International Nuclear Information System (INIS)
Gueron, S.; Shalloway, D.
1996-01-01
Numerical integration of Newton's equation in multiple dimensions plays an important role in many fields such as biochemistry and astrophysics. Currently, some of the most important practical questions in these areas cannot be addressed because the large dimensionality of the variable space and complexity of the required force evaluations precludes integration over sufficiently large time intervals. Improving the efficiency of algorithms for this purpose is therefore of great importance. Standard numerical integration schemes (e.g., leap-frog and Runge-Kutta) ignore the special structure of Newton's equation that, for conservative systems, constrains the force to be the gradient of a scalar potential. We propose a new class of open-quotes spatial interpolationclose quotes (SI) integrators that exploit this property by interpolating the force in space rather than (as with standard methods) in time. Since the force is usually a smoother function of space than of time, this can improve algorithmic efficiency and accuracy. In particular, an SI integrator solves the one- and two-dimensional harmonic oscillators exactly with one force evaluation per step. A simple type of time-reversible SI algorithm is described and tested. Significantly improved performance is achieved on one- and multi-dimensional benchmark problems. 19 refs., 4 figs., 1 tab
Clustering metagenomic sequences with interpolated Markov models
Directory of Open Access Journals (Sweden)
Kelley David R
2010-11-01
Full Text Available Abstract Background Sequencing of environmental DNA (often called metagenomics has shown tremendous potential to uncover the vast number of unknown microbes that cannot be cultured and sequenced by traditional methods. Because the output from metagenomic sequencing is a large set of reads of unknown origin, clustering reads together that were sequenced from the same species is a crucial analysis step. Many effective approaches to this task rely on sequenced genomes in public databases, but these genomes are a highly biased sample that is not necessarily representative of environments interesting to many metagenomics projects. Results We present SCIMM (Sequence Clustering with Interpolated Markov Models, an unsupervised sequence clustering method. SCIMM achieves greater clustering accuracy than previous unsupervised approaches. We examine the limitations of unsupervised learning on complex datasets, and suggest a hybrid of SCIMM and supervised learning method Phymm called PHYSCIMM that performs better when evolutionarily close training genomes are available. Conclusions SCIMM and PHYSCIMM are highly accurate methods to cluster metagenomic sequences. SCIMM operates entirely unsupervised, making it ideal for environments containing mostly novel microbes. PHYSCIMM uses supervised learning to improve clustering in environments containing microbial strains from well-characterized genera. SCIMM and PHYSCIMM are available open source from http://www.cbcb.umd.edu/software/scimm.
Research of Cubic Bezier Curve NC Interpolation Signal Generator
Directory of Open Access Journals (Sweden)
Shijun Ji
2014-08-01
Full Text Available Interpolation technology is the core of the computer numerical control (CNC system, and the precision and stability of the interpolation algorithm directly affect the machining precision and speed of CNC system. Most of the existing numerical control interpolation technology can only achieve circular arc interpolation, linear interpolation or parabola interpolation, but for the numerical control (NC machining of parts with complicated surface, it needs to establish the mathematical model and generate the curved line and curved surface outline of parts and then discrete the generated parts outline into a large amount of straight line or arc to carry on the processing, which creates the complex program and a large amount of code, so it inevitably introduce into the approximation error. All these factors affect the machining accuracy, surface roughness and machining efficiency. The stepless interpolation of cubic Bezier curve controlled by analog signal is studied in this paper, the tool motion trajectory of Bezier curve can be directly planned out in CNC system by adjusting control points, and then these data were put into the control motor which can complete the precise feeding of Bezier curve. This method realized the improvement of CNC trajectory controlled ability from the simple linear and circular arc to the complex project curve, and it provides a new way for economy realizing the curve surface parts with high quality and high efficiency machining.
Spatial interpolation of monthly mean air temperature data for Latvia
Aniskevich, Svetlana
2016-04-01
Temperature data with high spatial resolution are essential for appropriate and qualitative local characteristics analysis. Nowadays the surface observation station network in Latvia consists of 22 stations recording daily air temperature, thus in order to analyze very specific and local features in the spatial distribution of temperature values in the whole Latvia, a high quality spatial interpolation method is required. Until now inverse distance weighted interpolation was used for the interpolation of air temperature data at the meteorological and climatological service of the Latvian Environment, Geology and Meteorology Centre, and no additional topographical information was taken into account. This method made it almost impossible to reasonably assess the actual temperature gradient and distribution between the observation points. During this project a new interpolation method was applied and tested, considering auxiliary explanatory parameters. In order to spatially interpolate monthly mean temperature values, kriging with external drift was used over a grid of 1 km resolution, which contains parameters such as 5 km mean elevation, continentality, distance from the Gulf of Riga and the Baltic Sea, biggest lakes and rivers, population density. As the most appropriate of these parameters, based on a complex situation analysis, mean elevation and continentality was chosen. In order to validate interpolation results, several statistical indicators of the differences between predicted values and the values actually observed were used. Overall, the introduced model visually and statistically outperforms the previous interpolation method and provides a meteorologically reasonable result, taking into account factors that influence the spatial distribution of the monthly mean temperature.
[An Improved Spectral Quaternion Interpolation Method of Diffusion Tensor Imaging].
Xu, Yonghong; Gao, Shangce; Hao, Xiaofei
2016-04-01
Diffusion tensor imaging(DTI)is a rapid development technology in recent years of magnetic resonance imaging.The diffusion tensor interpolation is a very important procedure in DTI image processing.The traditional spectral quaternion interpolation method revises the direction of the interpolation tensor and can preserve tensors anisotropy,but the method does not revise the size of tensors.The present study puts forward an improved spectral quaternion interpolation method on the basis of traditional spectral quaternion interpolation.Firstly,we decomposed diffusion tensors with the direction of tensors being represented by quaternion.Then we revised the size and direction of the tensor respectively according to different situations.Finally,we acquired the tensor of interpolation point by calculating the weighted average.We compared the improved method with the spectral quaternion method and the Log-Euclidean method by the simulation data and the real data.The results showed that the improved method could not only keep the monotonicity of the fractional anisotropy(FA)and the determinant of tensors,but also preserve the tensor anisotropy at the same time.In conclusion,the improved method provides a kind of important interpolation method for diffusion tensor image processing.
Trivariate Local Lagrange Interpolation and Macro Elements of Arbitrary Smoothness
Matt, Michael Andreas
2012-01-01
Michael A. Matt constructs two trivariate local Lagrange interpolation methods which yield optimal approximation order and Cr macro-elements based on the Alfeld and the Worsey-Farin split of a tetrahedral partition. The first interpolation method is based on cubic C1 splines over type-4 cube partitions, for which numerical tests are given. The second is the first trivariate Lagrange interpolation method using C2 splines. It is based on arbitrary tetrahedral partitions using splines of degree nine. The author constructs trivariate macro-elements based on the Alfeld split, where each tetrahedron
INTERPOL DVI best-practice standards--An overview.
Sweet, David
2010-09-10
A description of the International Criminal Police Organization and its role in disaster victim identification is provided along with a summary of the standards developed and circulated to responders in INTERPOL member countries (188 throughout the world) to insure evidence-based DVI practices. Following the INTERPOL-mediated DVI response in 2005 to the SE Asia tsunami, many lessons learned have been recorded. Based on these current standards, INTERPOL's approach to DVI reflects a modern approach and philosophy. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.
Four-Point n-Ary Interpolating Subdivision Schemes
Directory of Open Access Journals (Sweden)
Ghulam Mustafa
2013-01-01
Full Text Available We present an efficient and simple algorithm to generate 4-point n-ary interpolating schemes. Our algorithm is based on three simple steps: second divided differences, determination of position of vertices by using second divided differences, and computation of new vertices. It is observed that 4-point n-ary interpolating schemes generated by completely different frameworks (i.e., Lagrange interpolant and wavelet theory can also be generated by the proposed algorithm. Furthermore, we have discussed continuity, Hölder regularity, degree of polynomial generation, polynomial reproduction, and approximation order of the schemes.
MODIS Snow Cover Recovery Using Variational Interpolation
Tran, H.; Nguyen, P.; Hsu, K. L.; Sorooshian, S.
2017-12-01
Cloud obscuration is one of the major problems that limit the usages of satellite images in general and in NASA's Moderate Resolution Imaging Spectroradiometer (MODIS) global Snow-Covered Area (SCA) products in particular. Among the approaches to resolve the problem, the Variational Interpolation (VI) algorithm method, proposed by Xia et al., 2012, obtains cloud-free dynamic SCA images from MODIS. This method is automatic and robust. However, computational deficiency is a main drawback that degrades applying the method for larger scales (i.e., spatial and temporal scales). To overcome this difficulty, this study introduces an improved version of the original VI. The modified VI algorithm integrates the MINimum RESidual (MINRES) iteration (Paige and Saunders., 1975) to prevent the system from breaking up when applied to much broader scales. An experiment was done to demonstrate the crash-proof ability of the new algorithm in comparison with the original VI method, an ability that is obtained when maintaining the distribution of the weights set after solving the linear system. After that, the new VI algorithm was applied to the whole Contiguous United States (CONUS) over four winter months of 2016 and 2017, and validated using the snow station network (SNOTEL). The resulting cloud free images have high accuracy in capturing the dynamical changes of snow in contrast with the MODIS snow cover maps. Lastly, the algorithm was applied to create a Cloud free images dataset from March 10, 2000 to February 28, 2017, which is able to provide an overview of snow trends over CONUS for nearly two decades. ACKNOWLEDGMENTSWe would like to acknowledge NASA, NOAA Office of Hydrologic Development (OHD) National Weather Service (NWS), Cooperative Institute for Climate and Satellites (CICS), Army Research Office (ARO), ICIWaRM, and UNESCO for supporting this research.
National Research Council Canada - National Science Library
Ingel, R
1999-01-01
... (which require derivative information) interpolation functions as well as standard Lagrangian functions, which can be linear, quadratic or cubic, have been used to construct the interpolation windows...
Efficient Algorithms and Design for Interpolation Filters in Digital Receiver
Directory of Open Access Journals (Sweden)
Xiaowei Niu
2014-05-01
Full Text Available Based on polynomial functions this paper introduces a generalized design method for interpolation filters. The polynomial-based interpolation filters can be implemented efficiently by using a modified Farrow structure with an arbitrary frequency response, the filters allow many pass- bands and stop-bands, and for each band the desired amplitude and weight can be set arbitrarily. The optimization coefficients of the interpolation filters in time domain are got by minimizing the weighted mean squared error function, then converting to solve the quadratic programming problem. The optimization coefficients in frequency domain are got by minimizing the maxima (MiniMax of the weighted mean squared error function. The degree of polynomials and the length of interpolation filter can be selected arbitrarily. Numerical examples verified the proposed design method not only can reduce the hardware cost effectively but also guarantee an excellent performance.
Analysis of Spatial Interpolation in the Material-Point Method
DEFF Research Database (Denmark)
Andersen, Søren; Andersen, Lars
2010-01-01
This paper analyses different types of spatial interpolation for the material-point method The interpolations include quadratic elements and cubic splines in addition to the standard linear shape functions usually applied. For the small-strain problem of a vibrating bar, the best results...... are obtained using quadratic elements. It is shown that for more complex problems, the use of partially negative shape functions is inconsistent with the material-point method in its current form, necessitating other types of interpolation such as cubic splines in order to obtain smoother representations...... of field quantities The properties of different interpolation functions are analysed using numerical examples, including the classical cantil-evered beam problem....
[Multimodal medical image registration using cubic spline interpolation method].
He, Yuanlie; Tian, Lianfang; Chen, Ping; Wang, Lifei; Ye, Guangchun; Mao, Zongyuan
2007-12-01
Based on the characteristic of the PET-CT multimodal image series, a novel image registration and fusion method is proposed, in which the cubic spline interpolation method is applied to realize the interpolation of PET-CT image series, then registration is carried out by using mutual information algorithm and finally the improved principal component analysis method is used for the fusion of PET-CT multimodal images to enhance the visual effect of PET image, thus satisfied registration and fusion results are obtained. The cubic spline interpolation method is used for reconstruction to restore the missed information between image slices, which can compensate for the shortage of previous registration methods, improve the accuracy of the registration, and make the fused multimodal images more similar to the real image. Finally, the cubic spline interpolation method has been successfully applied in developing 3D-CRT (3D Conformal Radiation Therapy) system.
Nonlinear interpolation fractal classifier for multiple cardiac arrhythmias recognition
Energy Technology Data Exchange (ETDEWEB)
Lin, C.-H. [Department of Electrical Engineering, Kao-Yuan University, No. 1821, Jhongshan Rd., Lujhu Township, Kaohsiung County 821, Taiwan (China); Institute of Biomedical Engineering, National Cheng-Kung University, Tainan 70101, Taiwan (China)], E-mail: eechl53@cc.kyu.edu.tw; Du, Y.-C.; Chen Tainsong [Institute of Biomedical Engineering, National Cheng-Kung University, Tainan 70101, Taiwan (China)
2009-11-30
This paper proposes a method for cardiac arrhythmias recognition using the nonlinear interpolation fractal classifier. A typical electrocardiogram (ECG) consists of P-wave, QRS-complexes, and T-wave. Iterated function system (IFS) uses the nonlinear interpolation in the map and uses similarity maps to construct various data sequences including the fractal patterns of supraventricular ectopic beat, bundle branch ectopic beat, and ventricular ectopic beat. Grey relational analysis (GRA) is proposed to recognize normal heartbeat and cardiac arrhythmias. The nonlinear interpolation terms produce family functions with fractal dimension (FD), the so-called nonlinear interpolation function (NIF), and make fractal patterns more distinguishing between normal and ill subjects. The proposed QRS classifier is tested using the Massachusetts Institute of Technology-Beth Israel Hospital (MIT-BIH) arrhythmia database. Compared with other methods, the proposed hybrid methods demonstrate greater efficiency and higher accuracy in recognizing ECG signals.
PSPLINE: Princeton Spline and Hermite cubic interpolation routines
McCune, Doug
2017-10-01
PSPLINE is a collection of Spline and Hermite interpolation tools for 1D, 2D, and 3D datasets on rectilinear grids. Spline routines give full control over boundary conditions, including periodic, 1st or 2nd derivative match, or divided difference-based boundary conditions on either end of each grid dimension. Hermite routines take the function value and derivatives at each grid point as input, giving back a representation of the function between grid points. Routines are provided for creating Hermite datasets, with appropriate boundary conditions applied. The 1D spline and Hermite routines are based on standard methods; the 2D and 3D spline or Hermite interpolation functions are constructed from 1D spline or Hermite interpolation functions in a straightforward manner. Spline and Hermite interpolation functions are often much faster to evaluate than other representations using e.g. Fourier series or otherwise involving transcendental functions.
Some observations on interpolating gauges and non-covariant gauges
Indian Academy of Sciences (India)
covariant gauges starting from the covariant ones. We draw attention to the need for a very careful treatment of boundary condition defining term. We show that the boundary condition needed to maintain gauge- invariance as the interpolating ...
Comparing interpolation schemes in dynamic receive ultrasound beamforming
DEFF Research Database (Denmark)
Kortbek, Jacob; Andresen, Henrik; Nikolov, Svetoslav
2005-01-01
conventional B-mode imaging and linear interpolation, the difference in mean SLMLR is 6.2 dB. With polynomial interpolation the ratio is in the range 6.2 dB to 0.3 dB using 2nd to 5th order polynomials, and with FIR interpolation the ratio is in the range 5.8 dB to 0.1 dB depending on the filter design....... The SNR is between 21 dB and 45 dB with the polynomial interpolation and between 37 dB and 43 dB with FIR filtering. In the synthetic aperture imaging modality the difference in mean SLMLRrangesfrom14dBto33dBand6dBto31dBforthe polynomial and FIR filtering schemes respectively. By using a proper...
A Meshfree Quasi-Interpolation Method for Solving Burgers’ Equation
Directory of Open Access Journals (Sweden)
Mingzhu Li
2014-01-01
Full Text Available The main aim of this work is to consider a meshfree algorithm for solving Burgers’ equation with the quartic B-spline quasi-interpolation. Quasi-interpolation is very useful in the study of approximation theory and its applications, since it can yield solutions directly without the need to solve any linear system of equations and overcome the ill-conditioning problem resulting from using the B-spline as a global interpolant. The numerical scheme is presented, by using the derivative of the quasi-interpolation to approximate the spatial derivative of the dependent variable and a low order forward difference to approximate the time derivative of the dependent variable. Compared to other numerical methods, the main advantages of our scheme are higher accuracy and lower computational complexity. Meanwhile, the algorithm is very simple and easy to implement and the numerical experiments show that it is feasible and valid.
Rhie-Chow interpolation in strong centrifugal fields
Bogovalov, S. V.; Tronin, I. V.
2015-10-01
Rhie-Chow interpolation formulas are derived from the Navier-Stokes and continuity equations. These formulas are generalized to gas dynamics in strong centrifugal fields (as high as 106 g) occurring in gas centrifuges.
Interpolating and sampling sequences in finite Riemann surfaces
Ortega-Cerda, Joaquim
2007-01-01
We provide a description of the interpolating and sampling sequences on a space of holomorphic functions on a finite Riemann surface, where a uniform growth restriction is imposed on the holomorphic functions.
Interpol: An R package for preprocessing of protein sequences.
Heider, Dominik; Hoffmann, Daniel
2011-06-17
Most machine learning techniques currently applied in the literature need a fixed dimensionality of input data. However, this requirement is frequently violated by real input data, such as DNA and protein sequences, that often differ in length due to insertions and deletions. It is also notable that performance in classification and regression is often improved by numerical encoding of amino acids, compared to the commonly used sparse encoding. The software "Interpol" encodes amino acid sequences as numerical descriptor vectors using a database of currently 532 descriptors (mainly from AAindex), and normalizes sequences to uniform length with one of five linear or non-linear interpolation algorithms. Interpol is distributed with open source as platform independent R-package. It is typically used for preprocessing of amino acid sequences for classification or regression. The functionality of Interpol widens the spectrum of machine learning methods that can be applied to biological sequences, and it will in many cases improve their performance in classification and regression.
Application Of Laplace Interpolation In The Analysis Of Geopotential ...
African Journals Online (AJOL)
difference) method can be applied to regions of high data gradients without distortions and smoothing. However, by itself, this method is not convenient for the interpolation of geophysical data, which often consists of regions of widely variable ...
Comparison Searching Process of Linear, Binary and Interpolation Algorithm
Rahim, Robbi; Nurarif, Saiful; Ramadhan, Mukhlis; Aisyah, Siti; Purba, Windania
2017-12-01
Searching is a process that cannot be issued for a transaction and communication process, many search algorithms that can be used to facilitate the search, linear, binary, and interpolation algorithms are some searching algorithms that can be utilized, the comparison of the three algorithms is performed by testing to search data with different length with pseudo process approach, and the result achieved that the interpolation algorithm is slightly faster than the other two algorithms.
Considerations Related to Interpolation of Experimental Data Using Piecewise Functions
Directory of Open Access Journals (Sweden)
Stelian Alaci
2016-12-01
Full Text Available The paper presents a method for experimental data interpolation by means of a piecewise function, the points where the form of the function changes being found simultaneously with the other parameters utilized in an optimization criterion. The optimization process is based on defining the interpolation function using a single expression founded on the Heaviside function and regarding the optimization function as a generalised infinitely derivable function. The exemplification of the methodology is made via a tangible example.
Data interpolation in the definition of management zones
Schenatto, Kelyn; Universidade Tecnológica Federal do Paraná; Souza, Eduardo Godoy; Universidade Estadual do Oeste do Paraná; Bazzi, Claudio Leones; Universidade Tecnológica Federal do Paraná; Bier, Vanderlei Arthur; Universidade Estadual do Oeste do Paraná; Betzek, Nelson Miguel; Universidade Tecnológica Federal do Paraná; Gavioli, Alan; Universidade Tecnológica Federal do Paraná
2016-01-01
Precision agriculture (PA) comprises the use of management zones (MZs). Sample data are usually interpolated to define MZs. Current research checks whether there is a need for data interpolation by evaluating the quality of MZs by five indices – variance reduction (VR), fuzzy performance index (FPI), modified partition entropy index (MPE), Kappa index and the cluster validation index (CVI), of which the latter has been focused in current assay. Soil texture, soil resistance to penetration, el...
Survey: interpolation methods for whole slide image processing.
Roszkowiak, L; Korzynska, A; Zak, J; Pijanowska, D; Swiderska-Chadaj, Z; Markiewicz, T
2017-02-01
Evaluating whole slide images of histological and cytological samples is used in pathology for diagnostics, grading and prognosis . It is often necessary to rescale whole slide images of a very large size. Image resizing is one of the most common applications of interpolation. We collect the advantages and drawbacks of nine interpolation methods, and as a result of our analysis, we try to select one interpolation method as the preferred solution. To compare the performance of interpolation methods, test images were scaled and then rescaled to the original size using the same algorithm. The modified image was compared to the original image in various aspects. The time needed for calculations and results of quantification performance on modified images were also compared. For evaluation purposes, we used four general test images and 12 specialized biological immunohistochemically stained tissue sample images. The purpose of this survey is to determine which method of interpolation is the best to resize whole slide images, so they can be further processed using quantification methods. As a result, the interpolation method has to be selected depending on the task involving whole slide images. © 2016 The Authors Journal of Microscopy © 2016 Royal Microscopical Society.
Analysis of Interpolation Methods in the Image Reconstruction Tasks
Directory of Open Access Journals (Sweden)
V. T. Nguyen
2017-01-01
Full Text Available The article studies the interpolation methods used for image reconstruction. These methods were also implemented and tested with several images to estimate their effectiveness.The considered interpolation methods are a nearest-neighbor method, linear method, a cubic B-spline method, a cubic convolution method, and a Lanczos method. For each method were presented an interpolation kernel (interpolation function and a frequency response (Fourier transform.As a result of the experiment, the following conclusions were drawn:- the nearest neighbor algorithm is very simple and often used. With using this method, the reconstructed images contain artifacts (blurring and haloing;- the linear method is quickly and easily performed. It also reduces some visual distortion caused by changing image size. Despite the advantages using this method causes a large amount of interpolation artifacts, such as blurring and haloing;- cubic B-spline method provides smoothness of reconstructed images and eliminates apparent ramp phenomenon. But in the interpolation process a low-pass filter is used, and a high frequency component is suppressed. This will lead to fuzzy edge and false artificial traces;- cubic convolution method offers less distortion interpolation. But its algorithm is more complicated and more execution time is required as compared to the nearest-neighbor method and the linear method;- using the Lanczos method allows us to achieve a high-definition image. In spite of the great advantage the method requires more execution time as compared to the other methods of interpolation.The result obtained not only shows a comparison of the considered interpolation methods for various aspects, but also enables users to select an appropriate interpolation method for their applications.It is advisable to study further the existing methods and develop new ones using a number of methods
5-D interpolation with wave-front attributes
Xie, Yujiang; Gajewski, Dirk
2017-11-01
Most 5-D interpolation and regularization techniques reconstruct the missing data in the frequency domain by using mathematical transforms. An alternative type of interpolation methods uses wave-front attributes, that is, quantities with a specific physical meaning like the angle of emergence and wave-front curvatures. In these attributes structural information of subsurface features like dip and strike of a reflector are included. These wave-front attributes work on 5-D data space (e.g. common-midpoint coordinates in x and y, offset, azimuth and time), leading to a 5-D interpolation technique. Since the process is based on stacking next to the interpolation a pre-stack data enhancement is achieved, improving the signal-to-noise ratio (S/N) of interpolated and recorded traces. The wave-front attributes are determined in a data-driven fashion, for example, with the Common Reflection Surface (CRS method). As one of the wave-front-attribute-based interpolation techniques, the 3-D partial CRS method was proposed to enhance the quality of 3-D pre-stack data with low S/N. In the past work on 3-D partial stacks, two potential problems were still unsolved. For high-quality wave-front attributes, we suggest a global optimization strategy instead of the so far used pragmatic search approach. In previous works, the interpolation of 3-D data was performed along a specific azimuth which is acceptable for narrow azimuth acquisition but does not exploit the potential of wide-, rich- or full-azimuth acquisitions. The conventional 3-D partial CRS method is improved in this work and we call it as a wave-front-attribute-based 5-D interpolation (5-D WABI) as the two problems mentioned above are addressed. Data examples demonstrate the improved performance by the 5-D WABI method when compared with the conventional 3-D partial CRS approach. A comparison of the rank-reduction-based 5-D seismic interpolation technique with the proposed 5-D WABI method is given. The comparison reveals that
Spatial interpolation methods for monthly rainfalls and temperatures in Basilicata
Directory of Open Access Journals (Sweden)
Ferrara A
2008-12-01
Full Text Available Spatial interpolated climatic data on grids are important as input in forest modeling because climate spatial variability has a direct effect on productivity and forest growth. Maps of climatic variables can be obtained by different interpolation methods depending on data quality (number of station, spatial distribution, missed data etc. and topographic and climatic features of study area. In this paper four methods are compared to interpolate monthly rainfall at regional scale: 1 inverse distance weighting (IDW; 2 regularized spline with tension (RST; 3 ordinary kriging (OK; 4 universal kriging (UK. Besides, an approach to generate monthly surfaces of temperatures over regions of complex terrain and with limited number of stations is presented. Daily data were gathered from 1976 to 2006 period and then gaps in the time series were filled in order to obtain monthly mean temperatures and cumulative precipitation. Basic statistics of monthly dataset and analysis of relationship of temperature and precipitation to elevation were performed. A linear relationship was found between temperature and altitude, while no relationship was found between rainfall and elevation. Precipitations were then interpolated without taking into account elevation. Based on root mean squared error for each month the best method was ranked. Results showed that universal kriging (UK is the best method in spatial interpolation of rainfall in study area. Then cross validation was used to compare prediction performance of tree different variogram model (circular, spherical, exponential using UK algorithm in order to produce final maps of monthly precipitations. Before interpolating temperatures were referred to see level using the calculated lapse rate and a digital elevation model (DEM. The result of interpolation with RST was then set to originally elevation with an inverse procedure. To evaluate the quality of interpolated surfaces a comparison between interpolated and
A FAST MORPHING-BASED INTERPOLATION FOR MEDICAL IMAGES: APPLICATION TO CONFORMAL RADIOTHERAPY
Directory of Open Access Journals (Sweden)
Hussein Atoui
2011-05-01
Full Text Available A method is presented for fast interpolation between medical images. The method is intended for both slice and projective interpolation. It allows offline interpolation between neighboring slices in tomographic data. Spatial correspondence between adjacent images is established using a block matching algorithm. Interpolation of image intensities is then carried out by morphing between the images. The morphing-based method is compared to standard linear interpolation, block-matching-based interpolation and registrationbased interpolation in 3D tomographic data sets. Results show that the proposed method scored similar performance in comparison to registration-based interpolation, and significantly outperforms both linear and block-matching-based interpolation. This method is applied in the context of conformal radiotherapy for online projective interpolation between Digitally Reconstructed Radiographs (DRRs.
Data interpolation in the definition of management zones
Directory of Open Access Journals (Sweden)
Kelyn Schenatto
2016-01-01
Full Text Available Precision agriculture (PA comprises the use of management zones (MZs. Sample data are usually interpolated to define MZs. Current research checks whether there is a need for data interpolation by evaluating the quality of MZs by five indices – variance reduction (VR, fuzzy performance index (FPI, modified partition entropy index (MPE, Kappa index and the cluster validation index (CVI, of which the latter has been focused in current assay. Soil texture, soil resistance to penetration, elevation and slope in an experimental area of 15.5 ha were employed as attributes to the generation of MZ, correlating them with data of soybean yield from 2011-2012 and 2012-2013 harvests. Data interpolation prior to MZs generation is important to achieve MZs as a smoother contour and for a greater reduction in data variance. The Kriging interpolator had the best performance. CVI index proved to be efficient in choosing MZs, with a less subjective decision on the best interpolator or number of MZs.
Research on the DDA Precision Interpolation Algorithm for Continuity of Speed and Acceleration
Directory of Open Access Journals (Sweden)
Kai Sun
2014-05-01
Full Text Available The interpolation technology is critical to performance of CNC and industrial robots; this paper proposes a new precision interpolation algorithm based on analysis of root cause in speed and acceleration. To satisfy continuity of speed and acceleration in interpolation process, this paper describes, respectively, variable acceleration precision interpolation of two stages and three sections. Testing shows that CNC system can be enhanced significantly by using the new fine interpolation algorithm in this paper.
Importance of interpolation and coincidence errors in data fusion
Ceccherini, Simone; Carli, Bruno; Tirelli, Cecilia; Zoppetti, Nicola; Del Bianco, Samuele; Cortesi, Ugo; Kujanpää, Jukka; Dragani, Rossana
2018-02-01
The complete data fusion (CDF) method is applied to ozone profiles obtained from simulated measurements in the ultraviolet and in the thermal infrared in the framework of the Sentinel 4 mission of the Copernicus programme. We observe that the quality of the fused products is degraded when the fusing profiles are either retrieved on different vertical grids or referred to different true profiles. To address this shortcoming, a generalization of the complete data fusion method, which takes into account interpolation and coincidence errors, is presented. This upgrade overcomes the encountered problems and provides products of good quality when the fusing profiles are both retrieved on different vertical grids and referred to different true profiles. The impact of the interpolation and coincidence errors on number of degrees of freedom and errors of the fused profile is also analysed. The approach developed here to account for the interpolation and coincidence errors can also be followed to include other error components, such as forward model errors.
Interpolation of vector fields from human cardiac DT-MRI
Yang, F.; Zhu, Y. M.; Rapacchi, S.; Luo, J. H.; Robini, M.; Croisille, P.
2011-03-01
There has recently been increased interest in developing tensor data processing methods for the new medical imaging modality referred to as diffusion tensor magnetic resonance imaging (DT-MRI). This paper proposes a method for interpolating the primary vector fields from human cardiac DT-MRI, with the particularity of achieving interpolation and denoising simultaneously. The method consists of localizing the noise-corrupted vectors using the local statistical properties of vector fields, removing the noise-corrupted vectors and reconstructing them by using the thin plate spline (TPS) model, and finally applying global TPS interpolation to increase the resolution in the spatial domain. Experiments on 17 human hearts show that the proposed method allows us to obtain higher resolution while reducing noise, preserving details and improving direction coherence (DC) of vector fields as well as fiber tracking. Moreover, the proposed method perfectly reconstructs azimuth and elevation angle maps.
Interpolant tree automata and their application in Horn clause verification
DEFF Research Database (Denmark)
Kafle, Bishoksan; Gallagher, John Patrick
2016-01-01
This paper investigates the combination of abstract interpretation over the domain of convex polyhedra with interpolant tree automata, in an abstraction-refinement scheme for Horn clause verification. These techniques have been previously applied separately, but are combined in a new way in this ......This paper investigates the combination of abstract interpretation over the domain of convex polyhedra with interpolant tree automata, in an abstraction-refinement scheme for Horn clause verification. These techniques have been previously applied separately, but are combined in a new way...... clause verification problems indicates that the combination of interpolant tree automaton with abstract interpretation gives some increase in the power of the verification tool, while sometimes incurring a performance overhead....
Interpolation of vector fields from human cardiac DT-MRI
International Nuclear Information System (INIS)
Yang, F; Zhu, Y M; Rapacchi, S; Robini, M; Croisille, P; Luo, J H
2011-01-01
There has recently been increased interest in developing tensor data processing methods for the new medical imaging modality referred to as diffusion tensor magnetic resonance imaging (DT-MRI). This paper proposes a method for interpolating the primary vector fields from human cardiac DT-MRI, with the particularity of achieving interpolation and denoising simultaneously. The method consists of localizing the noise-corrupted vectors using the local statistical properties of vector fields, removing the noise-corrupted vectors and reconstructing them by using the thin plate spline (TPS) model, and finally applying global TPS interpolation to increase the resolution in the spatial domain. Experiments on 17 human hearts show that the proposed method allows us to obtain higher resolution while reducing noise, preserving details and improving direction coherence (DC) of vector fields as well as fiber tracking. Moreover, the proposed method perfectly reconstructs azimuth and elevation angle maps.
Discrete Sine Transform-Based Interpolation Filter for Video Compression
Directory of Open Access Journals (Sweden)
MyungJun Kim
2017-11-01
Full Text Available Fractional pixel motion compensation in high-efficiency video coding (HEVC uses an 8-point filter and a 7-point filter, which are based on the discrete cosine transform (DCT, for the 1/2-pixel and 1/4-pixel interpolations, respectively. In this paper, discrete sine transform (DST-based interpolation filters (DST-IFs are proposed for fractional pixel motion compensation in terms of coding efficiency improvement. Firstly, a performance of the DST-based interpolation filters (DST-IFs using 8-point and 7-point filters for the 1/2-pixel and 1/4-pixel interpolations is compared with that of the DCT-based IFs (DCT-IFs using 8-point and 7-point filters for the 1/2-pixel and 1/4-pixel interpolations, respectively, for fractional pixel motion compensation. Finally, the DST-IFs using 12-point and 11-point filters for the 1/2-pixel and 1/4-pixel interpolations, respectively, are proposed only for bi-directional motion compensation in terms of the coding efficiency. The 8-point and 7-point DST-IF methods showed average Bjøntegaard Delta (BD-rate reductions of 0.7% and 0.3% in the random access (RA and low delay B (LDB configurations, respectively, in HEVC. The 12-point and 11-point DST-IF methods showed average BD-rate reductions of 1.4% and 1.2% in the RA and LDB configurations for the Luma component, respectively, in HEVC.
Gribov ambiguities at the Landau-maximal Abelian interpolating gauge
Energy Technology Data Exchange (ETDEWEB)
Pereira, Antonio D.; Sobreiro, Rodrigo F. [UFF-Universidade Federal Fluminense, Instituto de Fisica, Niteroi, RJ (Brazil)
2014-08-15
In a previous work, we presented a new method to account for the Gribov ambiguities in non-Abelian gauge theories. The method consists on the introduction of an extra constraint which directly eliminates the infinitesimal Gribov copies without the usual geometric approach. Such strategy allows one to treat gauges with non-hermitian Faddeev-Popov operator. In this work, we apply this method to a gauge which interpolates among the Landau and maximal Abelian gauges. The result is a local and power counting renormalizable action, free of infinitesimal Gribov copies. Moreover, the interpolating tree-level gluon propagator is derived. (orig.)
An adaptive interpolation scheme for molecular potential energy surfaces
Kowalewski, Markus; Larsson, Elisabeth; Heryudono, Alfa
2016-08-01
The calculation of potential energy surfaces for quantum dynamics can be a time consuming task—especially when a high level of theory for the electronic structure calculation is required. We propose an adaptive interpolation algorithm based on polyharmonic splines combined with a partition of unity approach. The adaptive node refinement allows to greatly reduce the number of sample points by employing a local error estimate. The algorithm and its scaling behavior are evaluated for a model function in 2, 3, and 4 dimensions. The developed algorithm allows for a more rapid and reliable interpolation of a potential energy surface within a given accuracy compared to the non-adaptive version.
Scientific data interpolation with low dimensional manifold model
International Nuclear Information System (INIS)
Zhu, Wei; Wang, Bao; Barnard, Richard C.; Hauck, Cory D.
2017-01-01
Here, we propose to apply a low dimensional manifold model to scientific data interpolation from regular and irregular samplings with a significant amount of missing information. The low dimensionality of the patch manifold for general scientific data sets has been used as a regularizer in a variational formulation. The problem is solved via alternating minimization with respect to the manifold and the data set, and the Laplace–Beltrami operator in the Euler–Lagrange equation is discretized using the weighted graph Laplacian. Various scientific data sets from different fields of study are used to illustrate the performance of the proposed algorithm on data compression and interpolation from both regular and irregular samplings.
Scientific data interpolation with low dimensional manifold model
Zhu, Wei; Wang, Bao; Barnard, Richard; Hauck, Cory D.; Jenko, Frank; Osher, Stanley
2018-01-01
We propose to apply a low dimensional manifold model to scientific data interpolation from regular and irregular samplings with a significant amount of missing information. The low dimensionality of the patch manifold for general scientific data sets has been used as a regularizer in a variational formulation. The problem is solved via alternating minimization with respect to the manifold and the data set, and the Laplace-Beltrami operator in the Euler-Lagrange equation is discretized using the weighted graph Laplacian. Various scientific data sets from different fields of study are used to illustrate the performance of the proposed algorithm on data compression and interpolation from both regular and irregular samplings.
A Note on Interpolation of Stable Processes | Nassiuma | Journal of ...
African Journals Online (AJOL)
Interpolation procedures tailored for gaussian processes may not be applied to infinite variance stable processes. Alternative techniques suitable for a limited set of stable case with index α∈(1,2] were initially studied by Pourahmadi (1984) for harmonizable processes. This was later extended to the ARMA stable process ...
The Grand Tour via Geodesic Interpolation of 2-frames
Asimov, Daniel; Buja, Andreas
1994-01-01
Grand tours are a class of methods for visualizing multivariate data, or any finite set of points in n-space. The idea is to create an animation of data projections by moving a 2-dimensional projection plane through n-space. The path of planes used in the animation is chosen so that it becomes dense, that is, it comes arbitrarily close to any plane. One of the original inspirations for the grand tour was the experience of trying to comprehend an abstract sculpture in a museum. One tends to walk around the sculpture, viewing it from many different angles. A useful class of grand tours is based on the idea of continuously interpolating an infinite sequence of randomly chosen planes. Visiting randomly (more precisely: uniformly) distributed planes guarantees denseness of the interpolating path. In computer implementations, 2-dimensional orthogonal projections are specified by two 1-dimensional projections which map to the horizontal and vertical screen dimensions, respectively. Hence, a grand tour is specified by a path of pairs of orthonormal projection vectors. This paper describes an interpolation scheme for smoothly connecting two pairs of orthonormal vectors, and thus for constructing interpolating grand tours. The scheme is optimal in the sense that connecting paths are geodesics in a natural Riemannian geometry.
Functional Commutant Lifting and Interpolation on Generalized Analytic Polyhedra
Czech Academy of Sciences Publication Activity Database
Ambrozie, Calin-Grigore
2008-01-01
Roč. 34, č. 2 (2008), s. 519-543 ISSN 0362-1588 R&D Projects: GA ČR(CZ) GA201/06/0128 Institutional research plan: CEZ:AV0Z10190503 Keywords : intertwining lifting * interpolation * analytic functions Subject RIV: BA - General Mathematics Impact factor: 0.327, year: 2008
Geometries and interpolations for symmetric positive definite matrices
DEFF Research Database (Denmark)
Feragen, Aasa; Fuster, Andrea
2017-01-01
In this survey we review classical and recently proposed Riemannian metrics and interpolation schemes on the space of symmetric positive definite (SPD) matrices. We perform simulations that illustrate the problem of tensor fattening not only in the usually avoided Frobenius metric, but also...
Robust control, multidimensional systems and multivariable Nevanlinna-Pick interpolation
Ball, J.A.; ter Horst, S.
2010-01-01
The connection between the standard $H^\\infty$-problem in control theory and Nevanlinna-Pick interpolation in operator theory was established in the 1980s, and has led to a fruitful cross-pollination between the two fields since. In the meantime, research in $H^\\infty$-control theory has moved on to
Quantitative analysis of the reconstruction performance of interpolants
Lansing, Donald L.; Park, Stephen K.
1987-01-01
The analysis presented provides a quantitative measure of the reconstruction or interpolation performance of linear, shift-invariant interpolants. The performance criterion is the mean square error of the difference between the sampled and reconstructed functions. The analysis is applicable to reconstruction algorithms used in image processing and to many types of splines used in numerical analysis and computer graphics. When formulated in the frequency domain, the mean square error clearly separates the contribution of the interpolation method from the contribution of the sampled data. The equations provide a rational basis for selecting an optimal interpolant; that is, one which minimizes the mean square error. The analysis has been applied to a selection of frequently used data splines and reconstruction algorithms: parametric cubic and quintic Hermite splines, exponential and nu splines (including the special case of the cubic spline), parametric cubic convolution, Keys' fourth-order cubic, and a cubic with a discontinuous first derivative. The emphasis in this paper is on the image-dependent case in which no a priori knowledge of the frequency spectrum of the sampled function is assumed.
LIP: The Livermore Interpolation Package, Version 1.6
Energy Technology Data Exchange (ETDEWEB)
Fritsch, F. N. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
2016-01-04
This report describes LIP, the Livermore Interpolation Package. LIP was totally rewritten from the package described in [1]. In particular, the independent variables are now referred to as x and y, since it is a general-purpose package that need not be restricted to equation of state data, which uses variables ρ (density) and T (temperature).
Interpolation decoding method with variable parameters for fractal image compression
International Nuclear Information System (INIS)
He Chuanjiang; Li Gaoping; Shen Xiaona
2007-01-01
The interpolation fractal decoding method, which is introduced by [He C, Yang SX, Huang X. Progressive decoding method for fractal image compression. IEE Proc Vis Image Signal Process 2004;3:207-13], involves generating progressively the decoded image by means of an interpolation iterative procedure with a constant parameter. It is well-known that the majority of image details are added at the first steps of iterations in the conventional fractal decoding; hence the constant parameter for the interpolation decoding method must be set as a smaller value in order to achieve a better progressive decoding. However, it needs to take an extremely large number of iterations to converge. It is thus reasonable for some applications to slow down the iterative process at the first stages of decoding and then to accelerate it afterwards (e.g., at some iteration as we need). To achieve the goal, this paper proposed an interpolation decoding scheme with variable (iteration-dependent) parameters and proved the convergence of the decoding process mathematically. Experimental results demonstrate that the proposed scheme has really achieved the above-mentioned goal
Approximating Exponential and Logarithmic Functions Using Polynomial Interpolation
Gordon, Sheldon P.; Yang, Yajun
2017-01-01
This article takes a closer look at the problem of approximating the exponential and logarithmic functions using polynomials. Either as an alternative to or a precursor to Taylor polynomial approximations at the precalculus level, interpolating polynomials are considered. A measure of error is given and the behaviour of the error function is…
Improved Interpolation Kernels for Super-resolution Algorithms
DEFF Research Database (Denmark)
Rasti, Pejman; Orlova, Olga; Tamberg, Gert
2016-01-01
Super resolution (SR) algorithms are widely used in forensics investigations to enhance the resolution of images captured by surveillance cameras. Such algorithms usually use a common interpolation algorithm to generate an initial guess for the desired high resolution (HR) image. This initial gue...
On some interpolation properties in locally convex spaces
Energy Technology Data Exchange (ETDEWEB)
Pater, Flavius [Department of Mathematics, Politehnica University of Timişoara, 300004 Timişoara (Romania)
2015-03-10
The aim of this paper is to introduce the notion of interpolation between locally convex spaces, the real method, and to present some elementary results in this setting. This represents a generalization from the Banach spaces framework to the locally convex spaces sequentially complete one, where the operators acting on them are locally bounded.
Some observations on interpolating gauges and non-covariant gauges
Indian Academy of Sciences (India)
tion that are not normally taken into account in the BRST formalism that ignores the ε-term, and that they are characteristic of the way the singularities in propagators are handled. We argue that a prescription, in general, will require renormalization; if at all it is to be viable. Keywords. Non-covariant gauges; interpolating ...
Twitch interpolation technique in testing of maximal muscle strength
DEFF Research Database (Denmark)
Bülow, P M; Nørregaard, J; Danneskiold-Samsøe, B
1993-01-01
The aim was to study the methodological aspects of the muscle twitch interpolation technique in estimating the maximal force of contraction in the quadriceps muscle utilizing commercial muscle testing equipment. Six healthy subjects participated in seven sets of experiments testing the effects...
Voluntary activation of trapezius measured with twitch interpolation
DEFF Research Database (Denmark)
Taylor, Janet L; Olsen, Henrik Baare; Sjøgaard, Gisela
2009-01-01
This study investigated the feasibility of measuring voluntary activation of the trapezius muscle with twitch interpolation. Subjects (n=8) lifted the right shoulder or both shoulders against fixed force transducers. Stimulation of the accessory nerve in the neck was used to evoke maximal twitche...
Spatial interpolation quality assessments for soil sensor transect datasets
Near-ground geophysical soil sensors provide extremely valuable information for precision agriculture applications. Indeed, their readings can be used as proxy for many soil parameters. Typically, leave-one-out (loo) cross-validation (CV) of spatial interpolation of sensor data returns overly optimi...
Interpolation on sparse Gauss-Chebyshev grids in higher dimensions
F. Sprengel
1998-01-01
textabstractIn this paper, we give a unified approach to error estimates for interpolation on sparse Gauss--Chebyshev grids for multivariate functions from Besov--type spaces with dominating mixed smoothness properties. The error bounds obtained for this method are almost optimal for the considered
Multivariable operator-valued Nevanlinna-Pick interpolation: a survey
Ball, J.A.; ter Horst, S.|info:eu-repo/dai/nl/298809877
2010-01-01
The theory of Nevanlinna-Pick and Carathéodory-Fejér interpolation for matrix- and operator-valued Schur class functions on the unit disk is now well established. Recent work has produced extensions of the theory to a variety of multivariable settings, including the ball and the polydisk (both
Interpolation solution of the single-impurity Anderson model
International Nuclear Information System (INIS)
Kuzemsky, A.L.
1990-10-01
The dynamical properties of the single-impurity Anderson model (SIAM) is studied using a novel Irreducible Green's Function method (IGF). The new solution for one-particle GF interpolating between the strong and weak correlation limits is obtained. The unified concept of relevant mean-field renormalizations is indispensable for strong correlation limit. (author). 21 refs
Blind Authentication Using Periodic Properties ofInterpolation
Czech Academy of Sciences Publication Activity Database
Mahdian, Babak; Saic, Stanislav
2008-01-01
Roč. 3, č. 3 (2008), s. 529-538 ISSN 1556-6013 R&D Projects: GA ČR GA102/08/0470 Institutional research plan: CEZ:AV0Z10750506 Keywords : image forensics * digital forgery * image tampering * interpolation detection * resampling detection Subject RIV: IN - Informatics, Computer Science Impact factor: 2.230, year: 2008
Limiting reiteration for real interpolation with slowly varying functions
Czech Academy of Sciences Publication Activity Database
Gogatishvili, Amiran; Opic, Bohumír; Trebels, W.
2005-01-01
Roč. 278, 1-2 (2005), s. 86-107 ISSN 0025-584X R&D Projects: GA ČR(CZ) GA201/01/0333 Institutional research plan: CEZ:AV0Z10190503 Keywords : real interpolation * K-functional * limiting reiteration Subject RIV: BA - General Mathematics Impact factor: 0.465, year: 2005
Interpol: An R package for preprocessing of protein sequences
Directory of Open Access Journals (Sweden)
Heider Dominik
2011-06-01
Full Text Available Abstract Background Most machine learning techniques currently applied in the literature need a fixed dimensionality of input data. However, this requirement is frequently violated by real input data, such as DNA and protein sequences, that often differ in length due to insertions and deletions. It is also notable that performance in classification and regression is often improved by numerical encoding of amino acids, compared to the commonly used sparse encoding. Results The software "Interpol" encodes amino acid sequences as numerical descriptor vectors using a database of currently 532 descriptors (mainly from AAindex, and normalizes sequences to uniform length with one of five linear or non-linear interpolation algorithms. Interpol is distributed with open source as platform independent R-package. It is typically used for preprocessing of amino acid sequences for classification or regression. Conclusions The functionality of Interpol widens the spectrum of machine learning methods that can be applied to biological sequences, and it will in many cases improve their performance in classification and regression.
Accuracy of stream habitat interpolations across spatial scales
Sheehan, Kenneth R.; Welsh, Stuart A.
2013-01-01
Stream habitat data are often collected across spatial scales because relationships among habitat, species occurrence, and management plans are linked at multiple spatial scales. Unfortunately, scale is often a factor limiting insight gained from spatial analysis of stream habitat data. Considerable cost is often expended to collect data at several spatial scales to provide accurate evaluation of spatial relationships in streams. To address utility of single scale set of stream habitat data used at varying scales, we examined the influence that data scaling had on accuracy of natural neighbor predictions of depth, flow, and benthic substrate. To achieve this goal, we measured two streams at gridded resolution of 0.33 × 0.33 meter cell size over a combined area of 934 m2 to create a baseline for natural neighbor interpolated maps at 12 incremental scales ranging from a raster cell size of 0.11 m2 to 16 m2 . Analysis of predictive maps showed a logarithmic linear decay pattern in RMSE values in interpolation accuracy for variables as resolution of data used to interpolate study areas became coarser. Proportional accuracy of interpolated models (r2 ) decreased, but it was maintained up to 78% as interpolation scale moved from 0.11 m2 to 16 m2 . Results indicated that accuracy retention was suitable for assessment and management purposes at various scales different from the data collection scale. Our study is relevant to spatial modeling, fish habitat assessment, and stream habitat management because it highlights the potential of using a single dataset to fulfill analysis needs rather than investing considerable cost to develop several scaled datasets.
Spatial interpolation schemes of daily precipitation for hydrologic modeling
Hwang, Y.; Clark, M.R.; Rajagopalan, B.; Leavesley, G.
2012-01-01
Distributed hydrologic models typically require spatial estimates of precipitation interpolated from sparsely located observational points to the specific grid points. We compare and contrast the performance of regression-based statistical methods for the spatial estimation of precipitation in two hydrologically different basins and confirmed that widely used regression-based estimation schemes fail to describe the realistic spatial variability of daily precipitation field. The methods assessed are: (1) inverse distance weighted average; (2) multiple linear regression (MLR); (3) climatological MLR; and (4) locally weighted polynomial regression (LWP). In order to improve the performance of the interpolations, the authors propose a two-step regression technique for effective daily precipitation estimation. In this simple two-step estimation process, precipitation occurrence is first generated via a logistic regression model before estimate the amount of precipitation separately on wet days. This process generated the precipitation occurrence, amount, and spatial correlation effectively. A distributed hydrologic model (PRMS) was used for the impact analysis in daily time step simulation. Multiple simulations suggested noticeable differences between the input alternatives generated by three different interpolation schemes. Differences are shown in overall simulation error against the observations, degree of explained variability, and seasonal volumes. Simulated streamflows also showed different characteristics in mean, maximum, minimum, and peak flows. Given the same parameter optimization technique, LWP input showed least streamflow error in Alapaha basin and CMLR input showed least error (still very close to LWP) in Animas basin. All of the two-step interpolation inputs resulted in lower streamflow error compared to the directly interpolated inputs. ?? 2011 Springer-Verlag.
Time-Frequency Signal Representations Using Interpolations in Joint-Variable Domains
2016-06-14
frequently encountered in various radar applications. Data interpolators play a unique role in TF signal representations under missing samples. When...applied in the instantaneous autocorrelation domain over the time variable, the low-pass filter characteristic underlying linear interpolators lends...itself to cross-terms reduction in the ambiguity domain. This is in contrast to interpolation performed over the lag variable or a direct interpolation
LINTAB, Linear Interpolable Tables from any Continuous Variable Function
International Nuclear Information System (INIS)
1988-01-01
1 - Description of program or function: LINTAB is designed to construct linearly interpolable tables from any function. The program will start from any function of a single continuous variable... FUNKY(X). By user input the function can be defined, (1) Over 1 to 100 X ranges. (2) Within each X range the function is defined by 0 to 50 constants. (3) At boundaries between X ranges the function may be continuous or discontinuous (depending on the constants used to define the function within each X range). 2 - Method of solution: LINTAB will construct a table of X and Y values where the tabulated (X,Y) pairs will be exactly equal to the function (Y=FUNKY(X)) and linear interpolation between the tabulated pairs will be within any user specified fractional uncertainty of the function for all values of X within the requested X range
Interpolation strategies for reducing IFOV artifacts in microgrid polarimeter imagery.
Ratliff, Bradley M; LaCasse, Charles F; Tyo, J Scott
2009-05-25
Microgrid polarimeters are composed of an array of micro-polarizing elements overlaid upon an FPA sensor. In the past decade systems have been designed and built in all regions of the optical spectrum. These systems have rugged, compact designs and the ability to obtain a complete set of polarimetric measurements during a single image capture. However, these systems acquire the polarization measurements through spatial modulation and each measurement has a varying instantaneous field-of-view (IFOV). When these measurements are combined to estimate the polarization images, strong edge artifacts are present that severely degrade the estimated polarization imagery. These artifacts can be reduced when interpolation strategies are first applied to the intensity data prior to Stokes vector estimation. Here we formally study IFOV error and the performance of several bilinear interpolation strategies used for reducing it.
Comparison of interpolation methods for raster images scaling
Directory of Open Access Journals (Sweden)
Trubakov A.O.
2017-03-01
Full Text Available The article is devoted to the problem of efficient scaling of raster images. We consider some negative effects, related with scaling of raster images. Besides, we consider an analysis of several methods that are used to increase sizes of ras-ter images. Among them are nearest neighbor algorithm, bilinear interpolation, bicubic interpolation. We consider our research methodology, and then we tell about result of algorithms comparison. We use two criteria: quality of output images and performance of algorithms. Due to this research we can tell some recommendations on the choice of algo-rithms for increment of raster images. It is useful because there is no single universal algorithm for efficient solution to the problem.
Runoff Interpolation and Budyko Framework over 300 Catchments across China
Qiu, Ning
2017-04-01
The Budyko hypothesis illustrates that mean annual evapotranspiration is largely determined by precipitation and potential evapotranspiration, which can be adopted to estimate mean annual actual evapotranspiration. In this study Fu's equation derived from the Budyko hypothesis is firstly tested by using mean annual streamflow and meteorological data of over 300 hydrological stations from ten main basins in China. Result shows that significant differences yield in the application of Fu's equation among basins. Secondly, the relationship between the single parameterωin Fu's equation and climatic and human factors was built to reveal the time variation of it. Meanwhile, the spacial structure characteristic of the regionalized variable ω was analyzed including spatial autocorrelation and locality. Then a stochastic interpolation scheme based on geostatistical interpolation, adding a constraint of global water balance in river system, is developed to mapping ω and runoff, aimed to predict runoff of elements of target partition of main basins and compare to the results computed by using Budyko hypothesis.
Dynamic Stability Analysis Using High-Order Interpolation
Directory of Open Access Journals (Sweden)
Juarez-Toledo C.
2012-10-01
Full Text Available A non-linear model with robust precision for transient stability analysis in multimachine power systems is proposed. The proposed formulation uses the interpolation of Lagrange and Newton's Divided Difference. The High-Order Interpolation technique developed can be used for evaluation of the critical conditions of the dynamic system.The technique is applied to a 5-area 45-machine model of the Mexican interconnected system. As a particular case, this paper shows the application of the High-Order procedure for identifying the slow-frequency mode for a critical contingency. Numerical examples illustrate the method and demonstrate the ability of the High-Order technique to isolate and extract temporal modal behavior.
Interpolation spaces in the resolution of ill-posed problems
International Nuclear Information System (INIS)
Logon, T.B.
1995-11-01
A number of applied problems connected with the interpretation of geophysical data leads to the resolution of ill-posed problems of the form A x = y δ , where A is an integral operator and y δ - some measurements. In the resolution of these problems by the Tikhonov's variational method, the choice of the stabilizing functional is crucial and needs some a-priori informations about the exact solution. Here the norm of the interpolation spaces X θ,q, which depends on two parameters 0 < θ < 1, 1 ≤ q < ∞ is proposed as a stabilizing functional. The a-priori information about the exact solution is characterized by its membership in one of the interpolation spaces. (author). 9 refs
Interpolated Sounding and Gridded Sounding Value-Added Products
Energy Technology Data Exchange (ETDEWEB)
Jensen, M. P. [Brookhaven National Laboratory (BNL), Upton, NY (United States); Toto, T. [Brookhaven National Laboratory (BNL), Upton, NY (United States)
2016-03-01
Standard Atmospheric Radiation Measurement (ARM) Climate Research Facility sounding files provide atmospheric state data in one dimension of increasing time and height per sonde launch. Many applications require a quick estimate of the atmospheric state at higher time resolution. The INTERPOLATEDSONDE (i.e., Interpolated Sounding) Value-Added Product (VAP) transforms sounding data into continuous daily files on a fixed time-height grid, at 1-minute time resolution, on 332 levels, from the surface up to a limit of approximately 40 km. The grid extends that high so the full height of soundings can be captured; however, most soundings terminate at an altitude between 25 and 30 km, above which no data is provided. Between soundings, the VAP linearly interpolates atmospheric state variables in time for each height level. In addition, INTERPOLATEDSONDE provides relative humidity scaled to microwave radiometer (MWR) observations.
A parameterization of observer-based controllers: Bumpless transfer by covariance interpolation
DEFF Research Database (Denmark)
Stoustrup, Jakob; Komareji, Mohammad
2009-01-01
This paper presents an algorithm to interpolate between two observer-based controllers for a linear multivariable system such that the closed loop system remains stable throughout the interpolation. The method interpolates between the inverse Lyapunov functions for the two original state feedbacks...
2010-01-01
... Executive Order 12425 Designating Interpol as a Public International Organization Entitled To Enjoy Certain... Order 13524 of December 16, 2009 EO 13524 Amending Executive Order 12425 Designating Interpol as a... Organization (INTERPOL), it is hereby ordered that Executive Order 12425 of June 16, 1983, as amended, is...
Direct Trajectory Interpolation on the Surface using an Open CNC
Beudaert , Xavier; Lavernhe , Sylvain; Tournier , Christophe
2014-01-01
International audience; Free-form surfaces are used for many industrial applications from aeronautical parts, to molds or biomedical implants. In the common machining process, computer-aided manufacturing (CAM) software generates approximated tool paths because of the limitation induced by the input tool path format of the industrial CNC. Then, during the tool path interpolation, marks on finished surfaces can appear induced by non smooth feedrate planning. Managing the geometry of the tool p...
Interpolated pressure laws in two-fluid simulations and hyperbolicity
Helluy, Philippe; Jung, Jonathan
2014-01-01
We consider a two-fluid compressible flow. Each fluid obeys a stiffened gas pressure law. The continuous model is well defined without considering mixture regions. However, for numerical applications it is often necessary to consider artificial mixtures, because the two-fluid interface is diffused by the numerical scheme. We show that classic pressure law interpolations lead to a non-convex hyperbolicity domain and failure of well-known numerical schemes. We propose a physically relevant pres...
The modal surface interpolation method for damage localization
Pina Limongelli, Maria
2017-05-01
The Interpolation Method (IM) has been previously proposed and successfully applied for damage localization in plate like structures. The method is based on the detection of localized reductions of smoothness in the Operational Deformed Shapes (ODSs) of the structure. The IM can be applied to any type of structure provided the ODSs are estimated accurately in the original and in the damaged configurations. If the latter circumstance fails to occur, for example when the structure is subjected to an unknown input(s) or if the structural responses are strongly corrupted by noise, both false and missing alarms occur when the IM is applied to localize a concentrated damage. In order to overcome these drawbacks a modification of the method is herein investigated. An ODS is the deformed shape of a structure subjected to a harmonic excitation: at resonances the ODS are dominated by the relevant mode shapes. The effect of noise at resonance is usually lower with respect to other frequency values hence the relevant ODS are estimated with higher reliability. Several methods have been proposed to reliably estimate modal shapes in case of unknown input. These two circumstances can be exploited to improve the reliability of the IM. In order to reduce or eliminate the drawbacks related to the estimation of the ODSs in case of noisy signals, in this paper is investigated a modified version of the method based on a damage feature calculated considering the interpolation error relevant only to the modal shapes and not to all the operational shapes in the significant frequency range. Herein will be reported the comparison between the results of the IM in its actual version (with the interpolation error calculated summing up the contributions of all the operational shapes) and in the new proposed version (with the estimation of the interpolation error limited to the modal shapes).
Image interpolation via graph-based Bayesian label propagation.
Xianming Liu; Debin Zhao; Jiantao Zhou; Wen Gao; Huifang Sun
2014-03-01
In this paper, we propose a novel image interpolation algorithm via graph-based Bayesian label propagation. The basic idea is to first create a graph with known and unknown pixels as vertices and with edge weights encoding the similarity between vertices, then the problem of interpolation converts to how to effectively propagate the label information from known points to unknown ones. This process can be posed as a Bayesian inference, in which we try to combine the principles of local adaptation and global consistency to obtain accurate and robust estimation. Specially, our algorithm first constructs a set of local interpolation models, which predict the intensity labels of all image samples, and a loss term will be minimized to keep the predicted labels of the available low-resolution (LR) samples sufficiently close to the original ones. Then, all of the losses evaluated in local neighborhoods are accumulated together to measure the global consistency on all samples. Moreover, a graph-Laplacian-based manifold regularization term is incorporated to penalize the global smoothness of intensity labels, such smoothing can alleviate the insufficient training of the local models and make them more robust. Finally, we construct a unified objective function to combine together the global loss of the locally linear regression, square error of prediction bias on the available LR samples, and the manifold regularization term. It can be solved with a closed-form solution as a convex optimization problem. Experimental results demonstrate that the proposed method achieves competitive performance with the state-of-the-art image interpolation algorithms.
Construction of fractal surfaces by recurrent fractal interpolation curves
International Nuclear Information System (INIS)
Yun, Chol-hui; O, Hyong-chol; Choi, Hui-chol
2014-01-01
A method to construct fractal surfaces by recurrent fractal curves is provided. First we construct fractal interpolation curves using a recurrent iterated functions system (RIFS) with function scaling factors and estimate their box-counting dimension. Then we present a method of construction of wider class of fractal surfaces by fractal curves and Lipschitz functions and calculate the box-counting dimension of the constructed surfaces. Finally, we combine both methods to have more flexible constructions of fractal surfaces
Accurate interpolation of 3D fields in charged particle optics.
Horák, Michal; Badin, Viktor; Zlámal, Jakub
2018-03-29
Standard 3D interpolation polynomials often suffer from numerical errors of the calculated field and lack of node points in the 3D solution. We introduce a novel method for accurate and smooth interpolation of arbitrary electromagnetic fields in the vicinity of the optical axis valid up to 90% of the bore radius. Our method combines Fourier analysis and Gaussian wavelet interpolation and provides the axial multipole field functions and their derivatives analytically. The results are accurate and noiseless, usually up to the 5th derivative. This is very advantageous for further applications, such as accurate particle tracing, and evaluation of aberration coefficients and other optical properties. The proposed method also enables studying the strength and orientation of all multipole field components. To illustrate the capabilities of the proposed algorithm, we present three examples: a magnetic lens with a hole in the polepiece, a saturated magnetic lens with an elliptic polepiece, and an electrostatic 8-electrode multipole. Copyright © 2018 Elsevier B.V. All rights reserved.
Importance of interpolation and coincidence errors in data fusion
Directory of Open Access Journals (Sweden)
S. Ceccherini
2018-02-01
Full Text Available The complete data fusion (CDF method is applied to ozone profiles obtained from simulated measurements in the ultraviolet and in the thermal infrared in the framework of the Sentinel 4 mission of the Copernicus programme. We observe that the quality of the fused products is degraded when the fusing profiles are either retrieved on different vertical grids or referred to different true profiles. To address this shortcoming, a generalization of the complete data fusion method, which takes into account interpolation and coincidence errors, is presented. This upgrade overcomes the encountered problems and provides products of good quality when the fusing profiles are both retrieved on different vertical grids and referred to different true profiles. The impact of the interpolation and coincidence errors on number of degrees of freedom and errors of the fused profile is also analysed. The approach developed here to account for the interpolation and coincidence errors can also be followed to include other error components, such as forward model errors.
A Direct Coarray Interpolation Approach for Direction Finding
Directory of Open Access Journals (Sweden)
Tao Chen
2017-09-01
Full Text Available Sparse arrays have gained considerable attention in recent years because they can resolve more sources than the number of sensors. The coprime array can resolve O ( M N sources with only O ( M + N sensors, and is a popular sparse array structure due to its closed-form expressions for array configuration and the reduction of the mutual coupling effect. However, because of the existence of holes in its coarray, the performance of subspace-based direction of arrival (DOA estimation algorithms such as MUSIC and ESPRIT is limited. Several coarray interpolation approaches have been proposed to address this issue. In this paper, a novel DOA estimation approach via direct coarray interpolation is proposed. By using the direct coarray interpolation, the reshaping and spatial smoothing operations in coarray-based DOA estimation are not needed. Compared with existing approaches, the proposed approach can achieve a better accuracy with lower complexity. In addition, an improved angular resolution capability is obtained by using the proposed approach. Numerical simulations are conducted to validate the effectiveness of the proposed approach.
Interpolation of daily rainfall using spatiotemporal models and clustering
Militino, A. F.
2014-06-11
Accumulated daily rainfall in non-observed locations on a particular day is frequently required as input to decision-making tools in precision agriculture or for hydrological or meteorological studies. Various solutions and estimation procedures have been proposed in the literature depending on the auxiliary information and the availability of data, but most such solutions are oriented to interpolating spatial data without incorporating temporal dependence. When data are available in space and time, spatiotemporal models usually provide better solutions. Here, we analyse the performance of three spatiotemporal models fitted to the whole sampled set and to clusters within the sampled set. The data consists of daily observations collected from 87 manual rainfall gauges from 1990 to 2010 in Navarre, Spain. The accuracy and precision of the interpolated data are compared with real data from 33 automated rainfall gauges in the same region, but placed in different locations than the manual rainfall gauges. Root mean squared error by months and by year are also provided. To illustrate these models, we also map interpolated daily precipitations and standard errors on a 1km2 grid in the whole region. © 2014 Royal Meteorological Society.
International Nuclear Information System (INIS)
Garcia-Santos, J. M.; Cejudo, J.
2002-01-01
In contrast to conventional computed tomography (CT), helical CT requires the application of interpolators to achieve image reconstruction. This is because the projections processed by the computer are not situated in the same plane. Since the introduction of helical CT. a number of interpolators have been designed in the attempt to maintain the thickness of the reconstructed section as close as possible to the thickness of the X-ray beam. The purpose of this article is to discuss the function of these interpolators, stressing the advantages and considering the possible inconveniences of high-grade curved interpolators with respect to standard linear interpolators. (Author) 7 refs
Study on the algorithm for Newton-Rapson iteration interpolation of NURBS curve and simulation
Zhang, Wanjun; Gao, Shanping; Cheng, Xiyan; Zhang, Feng
2017-04-01
In order to solve the problems of Newton-Rapson iteration interpolation method of NURBS Curve, Such as interpolation time bigger, calculation more complicated, and NURBS curve step error are not easy changed and so on. This paper proposed a study on the algorithm for Newton-Rapson iteration interpolation method of NURBS curve and simulation. We can use Newton-Rapson iterative that calculate (xi, yi, zi). Simulation results show that the proposed NURBS curve interpolator meet the high-speed and high-accuracy interpolation requirements of CNC systems. The interpolation of NURBS curve should be finished. The simulation results show that the algorithm is correct; it is consistent with a NURBS curve interpolation requirements.
Evaluation of Nonlinear Methods for Interpolation of Catchment-Scale
Coleman, M. L.; Niemann, J. D.
2008-12-01
Soil moisture acts as a key state variable in interactions between the atmosphere and land surface, strongly influencing radiation and precipitation partitioning and thus many components of the hydrologic cycle. Despite its importance as a state variable, measuring soil moisture patterns with adequate spatial resolutions over useful spatial extents remains a significant challenge due to both physical and economic constraints. For this reason, ancillary data, such as topographic attributes, have been employed as process proxies and predictor variables for soil moisture. Most methods that have been used to estimate soil moisture from ancillary variables assume that soil moisture is linearly dependent on these variables. However, unsaturated zone water transport is typically modeled as a nonlinear function of the soil moisture state. While that fact does not necessarily imply nonlinear relationships with the ancillary variables, there is some evidence suggesting nonlinear methods may be more efficient than linear methods for interpolating soil moisture from ancillary data. Therefore, this work investigates the value of nonlinear estimation techniques, namely conditional density estimation, support vector machines, and a spatial artificial neural network, for interpolating soil moisture patterns from sparse measurements and ancillary data. The set of candidate predictor variables in this work includes simple and compound terrain attributes calculated from digital elevation models and, in some cases, soil texture data. The initial task in the interpolation procedure is the selection of the most effective predictor variables. Given the possibility of nonlinear relationships, mutual information is used to quantify relationships between candidate variables and soil moisture and ultimately to select the most efficient ancillary data as predictor variables. After selecting a subset of the potential ancillary data variables for use, the nonlinear estimation techniques are
Motion compensated frame interpolation with a symmetric optical flow constraint
DEFF Research Database (Denmark)
Rakêt, Lars Lau; Roholm, Lars; Bruhn, Andrés
2012-01-01
We consider the problem of interpolating frames in an image sequence. For this purpose accurate motion estimation can be very helpful. We propose to move the motion estimation from the surrounding frames directly to the unknown frame by parametrizing the optical flow objective function...... with current state-of-the-art methods. Finally we show that the scheme can be implemented on graphics hardware such that it be- comes possible to double the frame rate of 640 × 480 video footage at 30 fps, i.e. to perform frame doubling in realtime....
Data mining techniques in sensor networks summarization, interpolation and surveillance
Appice, Annalisa; Fumarola, Fabio; Malerba, Donato
2013-01-01
Sensor networks comprise of a number of sensors installed across a spatially distributed network, which gather information and periodically feed a central server with the measured data. The server monitors the data, issues possible alarms and computes fast aggregates. As data analysis requests may concern both present and past data, the server is forced to store the entire stream. But the limited storage capacity of a server may reduce the amount of data stored on the disk. One solution is to compute summaries of the data as it arrives, and to use these summaries to interpolate the real data.
Optimal interpolation method for intercomparison of atmospheric measurements.
Ridolfi, Marco; Ceccherini, Simone; Carli, Bruno
2006-04-01
Intercomparison of atmospheric measurements is often a difficult task because of the different spatial response functions of the experiments considered. We propose a new method for comparison of two atmospheric profiles characterized by averaging kernels with different vertical resolutions. The method minimizes the smoothing error induced by the differences in the averaging kernels by exploiting an optimal interpolation rule to map one profile into the retrieval grid of the other. Compared with the techniques published so far, this method permits one to retain the vertical resolution of the less-resolved profile involved in the intercomparison.
Interpolation in numerical optimization. [by cubic spline generation
Hall, K. R.; Hull, D. G.
1975-01-01
The present work discusses the generation of the cubic-spline interpolator in numerical optimization methods which use a variable-step integrator with step size control based on local relative truncation error. An algorithm for generating the cubic spline with successive over-relaxation is presented which represents an improvement over that given by Ralston and Wilf (1967). Rewriting the code reduces the number of N-vectors from eight to one. The algorithm is formulated in such a way that the solution of the linear system set up yields the first derivatives at the nodal points. This method is as accurate as other schemes but requires the minimum amount of storage.
A Bidirectional Flow Joint Sobolev Gradient for Image Interpolation
Directory of Open Access Journals (Sweden)
Yi Zhan
2013-01-01
Full Text Available An energy functional with bidirectional flow is presented to sharpen image by reducing its edge width, which performs a forward diffusion in brighter lateral on edge ramp and backward diffusion that proceeds in darker lateral. We first consider the diffusion equations as L2 gradient flows on integral functionals and then modify the inner product from L2 to a Sobolev inner product. The experimental results demonstrate that our model efficiently reconstructs the real image, leading to a natural interpolation with reduced blurring, staircase artifacts and preserving better the texture features of image.
Timescape: a simple space-time interpolation geostatistical Algorithm
Ciolfi, Marco; Chiocchini, Francesca; Gravichkova, Olga; Pisanelli, Andrea; Portarena, Silvia; Scartazza, Andrea; Brugnoli, Enrico; Lauteri, Marco
2016-04-01
Environmental sciences include both time and space variability in their datasets. Some established tools exist for both spatial interpolation and time series analysis alone, but mixing space and time variability calls for compromise: Researchers are often forced to choose which is the main source of variation, neglecting the other. We propose a simple algorithm, which can be used in many fields of Earth and environmental sciences when both time and space variability must be considered on equal grounds. The algorithm has already been implemented in Java language and the software is currently available at https://sourceforge.net/projects/timescapeglobal/ (it is published under GNU-GPL v3.0 Free Software License). The published version of the software, Timescape Global, is focused on continent- to Earth-wide spatial domains, using global longitude-latitude coordinates for samples localization. The companion Timescape Local software is currently under development ad will be published with an open license as well; it will use projected coordinates for a local to regional space scale. The basic idea of the Timescape Algorithm consists in converting time into a sort of third spatial dimension, with the addition of some causal constraints, which drive the interpolation including or excluding observations according to some user-defined rules. The algorithm is applicable, as a matter of principle, to anything that can be represented with a continuous variable (a scalar field, technically speaking). The input dataset should contain position, time and observed value of all samples. Ancillary data can be included in the interpolation as well. After the time-space conversion, Timescape follows basically the old-fashioned IDW (Inverse Distance Weighted) interpolation Algorithm, although users have a wide choice of customization options that, at least partially, overcome some of the known issues of IDW. The three-dimensional model produced by the Timescape Algorithm can be
Hörmander spaces, interpolation, and elliptic problems
Mikhailets, Vladimir A; Malyshev, Peter V
2014-01-01
The monograph gives a detailed exposition of the theory of general elliptic operators (scalar and matrix) and elliptic boundary value problems in Hilbert scales of Hörmander function spaces. This theory was constructed by the authors in a number of papers published in 2005-2009. It is distinguished by a systematic use of the method of interpolation with a functional parameter of abstract Hilbert spaces and Sobolev inner product spaces. This method, the theory and their applications are expounded for the first time in the monographic literature. The monograph is written in detail and in a
Stochastic interpolation model of the medial superior olive neural circuit
Czech Academy of Sciences Publication Activity Database
Šanda, Pavel; Maršálek, P.
2012-01-01
Roč. 1434, JAN 24 (2012), s. 257-265 ISSN 0006-8993. [International Workshop on Neural Coding. Limassol, 29.10.2010-03.11.2010] R&D Projects: GA ČR(CZ) GAP103/11/0282 Grant - others:GA MPO(CZ) FR-TI3/869 Institutional research plan: CEZ:AV0Z50110509 Keywords : coincidence detection * directional hearing * interaural time delay * sound azimuth * interpolation model Subject RIV: FH - Neurology Impact factor: 2.879, year: 2012
Trends in Continuity and Interpolation for Computer Graphics.
Gonzalez Garcia, Francisco
2015-01-01
In every computer graphics oriented application today, it is a common practice to texture 3D models as a way to obtain realistic material. As part of this process, mesh texturing, deformation, and visualization are all key parts of the computer graphics field. This PhD dissertation was completed in the context of these three important and related fields in computer graphics. The article presents techniques that improve on existing state-of-the-art approaches related to continuity and interpolation in texture space (texturing), object space (deformation), and screen space (rendering).
Gravity Aided Navigation Precise Algorithm with Gauss Spline Interpolation
Directory of Open Access Journals (Sweden)
WEN Chaobin
2015-01-01
Full Text Available The gravity compensation of error equation thoroughly should be solved before the study on gravity aided navigation with high precision. A gravity aided navigation model construction algorithm based on research the algorithm to approximate local grid gravity anomaly filed with the 2D Gauss spline interpolation is proposed. Gravity disturbance vector, standard gravity value error and Eotvos effect are all compensated in this precision model. The experiment result shows that positioning accuracy is raised by 1 times, the attitude and velocity accuracy is raised by 1～2 times and the positional error is maintained from 100~200 m.
Construction of Large Period Symplectic Maps by Interpolative Methods
Energy Technology Data Exchange (ETDEWEB)
Warnock, Robert; Cai, Yunhai; /SLAC; Ellison, James A.; /New Mexico U.
2009-12-17
The goal is to construct a symplectic evolution map for a large section of an accelerator, say a full turn of a large ring or a long wiggler. We start with an accurate tracking algorithm for single particles, which is allowed to be slightly non-symplectic. By tracking many particles for a distance S one acquires sufficient data to construct the mixed-variable generator of a symplectic map for evolution over S, given in terms of interpolatory functions. Two ways to find the generator are considered: (1) Find its gradient from tracking data, then the generator itself as a line integral. (2) Compute the action integral on many orbits. A test of method (1) has been made in a difficult example: a full turn map for an electron ring with strong nonlinearity near the dynamic aperture. The method succeeds at fairly large amplitudes, but there are technical difficulties near the dynamic aperture due to oddly shaped interpolation domains. For a generally applicable algorithm we propose method (2), realized with meshless interpolation methods.
Image re-sampling detection through a novel interpolation kernel.
Hilal, Alaa
2018-03-27
Image re-sampling involved in re-size and rotation transformations is an essential element block in a typical digital image alteration. Fortunately, traces left from such processes are detectable, proving that the image has gone a re-sampling transformation. Within this context, we present in this paper two original contributions. First, we propose a new re-sampling interpolation kernel. It depends on five independent parameters that controls its amplitude, angular frequency, standard deviation, and duration. Then, we demonstrate its capacity to imitate the same behavior of the most frequent interpolation kernels used in digital image re-sampling applications. Secondly, the proposed model is used to characterize and detect the correlation coefficients involved in re-sampling transformations. The involved process includes a minimization of an error function using the gradient method. The proposed method is assessed over a large database of 11,000 re-sampled images. Additionally, it is implemented within an algorithm in order to assess images that had undergone complex transformations. Obtained results demonstrate better performance and reduced processing time when compared to a reference method validating the suitability of the proposed approaches. Copyright © 2018 Elsevier B.V. All rights reserved.
3D Interpolation Method for CT Images of the Lung
Directory of Open Access Journals (Sweden)
Noriaki Asada
2003-06-01
Full Text Available A 3-D image can be reconstructed from numerous CT images of the lung. The procedure reconstructs a solid from multiple cross section images, which are collected during pulsation of the heart. Thus the motion of the heart is a special factor that must be taken into consideration during reconstruction. The lung exhibits a repeating transformation synchronized to the beating of the heart as an elastic body. There are discontinuities among neighboring CT images due to the beating of the heart, if no special techniques are used in taking CT images. The 3-D heart image is reconstructed from numerous CT images in which both the heart and the lung are taken. Although the outline shape of the reconstructed 3-D heart is quite unnatural, the envelope of the 3-D unnatural heart is fit to the shape of the standard heart. The envelopes of the lung in the CT images are calculated after the section images of the best fitting standard heart are located at the same positions of the CT images. Thus the CT images are geometrically transformed to the optimal CT images fitting best to the standard heart. Since correct transformation of images is required, an Area oriented interpolation method proposed by us is used for interpolation of transformed images. An attempt to reconstruct a 3-D lung image by a series of such operations without discontinuity is shown. Additionally, the same geometrical transformation method to the original projection images is proposed as a more advanced method.
THE EFFECT OF STIMULUS ANTICIPATION ON THE INTERPOLATED TWITCH TECHNIQUE
Directory of Open Access Journals (Sweden)
Duane C. Button
2008-12-01
Full Text Available The objective of this study was to investigate the effect of expected and unexpected interpolated stimuli (IT during a maximum voluntary contraction on quadriceps force output and activation. Two groups of male subjects who were either inexperienced (MI: no prior experience with IT tests or experienced (ME: previously experienced 10 or more series of IT tests received an expected or unexpected IT while performing quadriceps isometric maximal voluntary contractions (MVCs. Measurements included MVC force, quadriceps and hamstrings electromyographic (EMG activity, and quadriceps inactivation as measured by the interpolated twitch technique (ITT. When performing MVCs with the expectation of an IT, the knowledge or lack of knowledge of an impending IT occurring during a contraction did not result in significant overall differences in force, ITT inactivation, quadriceps or hamstrings EMG activity. However, the expectation of an IT significantly (p < 0.0001 reduced MVC force (9.5% and quadriceps EMG activity (14.9% when compared to performing MVCs with prior knowledge that stimulation would not occur. While ME exhibited non-significant decreases when expecting an IT during a MVC, MI force and EMG activity significantly decreased 12.4% and 20.9% respectively. Overall, ME had significantly (p < 0.0001 higher force (14.5% and less ITT inactivation (10.4% than MI. The expectation of the noxious stimuli may account for the significant decrements in force and activation during the ITT
An interpolation boundary treatment for the Lattice Boltzmann method
Deladisma, Marnico D.; Smith, Marc K.
2003-11-01
A new boundary condition for the Lattice Boltzmann method based on bounce-back and spatial interpolations is presented. The boundary condition allows for the placement of a boundary at any position between nodes and tracks the exact position of that boundary. Multi-dimensional interpolation of streaming and bounce-back particle distribution functions from surrounding boundary nodes is used to solve for new distribution values. This allows more information from surrounding nodes to be incorporated into the boundary treatment calculation. Calculations of flow within a 2D rotating annulus (with and without an obstacle placed in the flow) using the present boundary condition are compared with calculations done with the commercial CFD solver Fluent. Results show that the boundary condition is accurate and robust for these cases. The boundary condition also allows for moving boundaries and is easily extended to 3D, which facilitates the simulation of moving 3D particles. The new boundary condition will allow a Lattice Boltzmann simulation of a rotating wall vessel bioreactor with freely suspended tissue constructs whose length scale is about 1 cm.
Color Orchestra: Ordering Color Palettes for Interpolation and Prediction.
Phan, Huy; Fu, Hongbo; Chan, Antoni
2017-04-25
Color theme or color palette can deeply influence the quality and the feeling of a photograph or a graphical design. Although color palettes may come from different sources such as online crowd-sourcing, photographs and graphical designs, in this paper, we consider color palettes extracted from fine art collections, which we believe to be an abundant source of stylistic and unique color themes. We aim to capture color styles embedded in these collections by means of statistical models and to build practical applications upon these models. As artists often use their personal color themes in their paintings, making these palettes appear frequently in the dataset, we employed density estimation to capture the characteristics of palette data. Via density estimation, we carried out various predictions and interpolations on palettes, which led to promising applications such as photo-style exploration, real-time color suggestion, and enriched photo recolorization. It was, however, challenging to apply density estimation to palette data as palettes often come as unordered sets of colors, which make it difficult to use conventional metrics on them. To this end, we developed a divide-and-conquer sorting algorithm to rearrange the colors in the palettes in a coherent order, which allows meaningful interpolation between color palettes. To confirm the performance of our model, we also conducted quantitative experiments on datasets of digitized paintings collected from the Internet and received favorable results.
Interpolated Sounding and Gridded Sounding Value-Added Products
Energy Technology Data Exchange (ETDEWEB)
Toto, T. [Brookhaven National Lab. (BNL), Upton, NY (United States); Jensen, M. [Brookhaven National Lab. (BNL), Upton, NY (United States)
2016-03-01
Standard Atmospheric Radiation Measurement (ARM) Climate Research Facility sounding files provide atmospheric state data in one dimension of increasing time and height per sonde launch. Many applications require a quick estimate of the atmospheric state at higher time resolution. The INTERPOLATEDSONDE (i.e., Interpolated Sounding) Value-Added Product (VAP) transforms sounding data into continuous daily files on a fixed time-height grid, at 1-minute time resolution, on 332 levels, from the surface up to a limit of approximately 40 km. The grid extends that high so the full height of soundings can be captured; however, most soundings terminate at an altitude between 25 and 30 km, above which no data is provided. Between soundings, the VAP linearly interpolates atmospheric state variables in time for each height level. In addition, INTERPOLATEDSONDE provides relative humidity scaled to microwave radiometer (MWR) observations.The INTERPOLATEDSONDE VAP, a continuous time-height grid of relative humidity-corrected sounding data, is intended to provide input to higher-order products, such as the Merged Soundings (MERGESONDE; Troyan 2012) VAP, which extends INTERPOLATEDSONDE by incorporating model data. The INTERPOLATEDSONDE VAP also is used to correct gaseous attenuation of radar reflectivity in products such as the KAZRCOR VAP.
MAGIC: A Tool for Combining, Interpolating, and Processing Magnetograms
Allred, Joel
2012-01-01
Transients in the solar coronal magnetic field are ultimately the source of space weather. Models which seek to track the evolution of the coronal field require magnetogram images to be used as boundary conditions. These magnetograms are obtained by numerous instruments with different cadences and resolutions. A tool is required which allows modelers to fmd all available data and use them to craft accurate and physically consistent boundary conditions for their models. We have developed a software tool, MAGIC (MAGnetogram Interpolation and Composition), to perform exactly this function. MAGIC can manage the acquisition of magneto gram data, cast it into a source-independent format, and then perform the necessary spatial and temporal interpolation to provide magnetic field values as requested onto model-defined grids. MAGIC has the ability to patch magneto grams from different sources together providing a more complete picture of the Sun's field than is possible from single magneto grams. In doing this, care must be taken so as not to introduce nonphysical current densities along the seam between magnetograms. We have designed a method which minimizes these spurious current densities. MAGIC also includes a number of post-processing tools which can provide additional information to models. For example, MAGIC includes an interface to the DA VE4VM tool which derives surface flow velocities from the time evolution of surface magnetic field. MAGIC has been developed as an application of the KAMELEON data formatting toolkit which has been developed by the CCMC.
Interpolation methods for creating a scatter radiation exposure map
Energy Technology Data Exchange (ETDEWEB)
Gonçalves, Elicardo A. de S., E-mail: elicardo.goncalves@ifrj.edu.br [Instituto Federal do Rio de Janeiro (IFRJ), Paracambi, RJ (Brazil); Gomes, Celio S.; Lopes, Ricardo T. [Coordenacao de Pos-Graduacao e Pesquisa de Engenharia (PEN/COPPE/UFRJ), Rio de Janeiro, RJ (Brazil). Programa de Engenharia Nuclear; Oliveira, Luis F. de; Anjos, Marcelino J. dos; Oliveira, Davi F. [Universidade do Estado do Rio de Janeiro (UFRJ), RJ (Brazil). Instituto de Física
2017-07-01
A well know way for best comprehension of radiation scattering during a radiography is to map exposure over the space around the source and sample. This map is done measuring exposure in points regularly spaced, it means, measurement will be placed in localization chosen by increasing a regular steps from a starting point, along the x, y and z axes or even radial and angular coordinates. However, it is not always possible to maintain the accuracy of the steps throughout the entire space, or there will be regions of difficult access where the regularity of the steps will be impaired. This work intended to use some interpolation techniques that work with irregular steps, and to compare their results and their limits. It was firstly done angular coordinates, and tested in lack of some points. Later, in the same data was performed the Delaunay tessellation interpolation ir order to compare. Computational and graphic treatments was done with the GNU OCTAVE software and its image-processing package. Real data was acquired from a bunker where a 6 MeV betatron can be used to produce radiation scattering. (author)
Sparsity-Based Spatial Interpolation in Wireless Sensor Networks
Directory of Open Access Journals (Sweden)
Yan Yao
2011-02-01
Full Text Available In wireless sensor networks, due to environmental limitations or bad wireless channel conditions, not all sensor samples can be successfully gathered at the sink. In this paper, we try to recover these missing samples without retransmission. The missing samples estimation problem is mathematically formulated as a 2-D spatial interpolation. Assuming the 2-D sensor data can be sparsely represented by a dictionary, a sparsity-based recovery approach by solving for l1 norm minimization is proposed. It is shown that these missing samples can be reasonably recovered based on the null space property of the dictionary. This property also points out the way to choose an appropriate sparsifying dictionary to further reduce the recovery errors. The simulation results on synthetic and real data demonstrate that the proposed approach can recover the missing data reasonably well and that it outperforms the weighted average interpolation methods when the data change relatively fast or blocks of samples are lost. Besides, there exists a range of missing rates where the proposed approach is robust to missing block sizes.
Sparsity-based spatial interpolation in wireless sensor networks.
Guo, Di; Qu, Xiaobo; Huang, Lianfen; Yao, Yan
2011-01-01
In wireless sensor networks, due to environmental limitations or bad wireless channel conditions, not all sensor samples can be successfully gathered at the sink. In this paper, we try to recover these missing samples without retransmission. The missing samples estimation problem is mathematically formulated as a 2-D spatial interpolation. Assuming the 2-D sensor data can be sparsely represented by a dictionary, a sparsity-based recovery approach by solving for l(1) norm minimization is proposed. It is shown that these missing samples can be reasonably recovered based on the null space property of the dictionary. This property also points out the way to choose an appropriate sparsifying dictionary to further reduce the recovery errors. The simulation results on synthetic and real data demonstrate that the proposed approach can recover the missing data reasonably well and that it outperforms the weighted average interpolation methods when the data change relatively fast or blocks of samples are lost. Besides, there exists a range of missing rates where the proposed approach is robust to missing block sizes.
Interpolation methods for creating a scatter radiation exposure map
International Nuclear Information System (INIS)
Gonçalves, Elicardo A. de S.; Gomes, Celio S.; Lopes, Ricardo T.; Oliveira, Luis F. de; Anjos, Marcelino J. dos; Oliveira, Davi F.
2017-01-01
A well know way for best comprehension of radiation scattering during a radiography is to map exposure over the space around the source and sample. This map is done measuring exposure in points regularly spaced, it means, measurement will be placed in localization chosen by increasing a regular steps from a starting point, along the x, y and z axes or even radial and angular coordinates. However, it is not always possible to maintain the accuracy of the steps throughout the entire space, or there will be regions of difficult access where the regularity of the steps will be impaired. This work intended to use some interpolation techniques that work with irregular steps, and to compare their results and their limits. It was firstly done angular coordinates, and tested in lack of some points. Later, in the same data was performed the Delaunay tessellation interpolation ir order to compare. Computational and graphic treatments was done with the GNU OCTAVE software and its image-processing package. Real data was acquired from a bunker where a 6 MeV betatron can be used to produce radiation scattering. (author)
A New Interpolation Approach for Linearly Constrained Convex Optimization
Espinoza, Francisco
2012-08-01
In this thesis we propose a new class of Linearly Constrained Convex Optimization methods based on the use of a generalization of Shepard\\'s interpolation formula. We prove the properties of the surface such as the interpolation property at the boundary of the feasible region and the convergence of the gradient to the null space of the constraints at the boundary. We explore several descent techniques such as steepest descent, two quasi-Newton methods and the Newton\\'s method. Moreover, we implement in the Matlab language several versions of the method, particularly for the case of Quadratic Programming with bounded variables. Finally, we carry out performance tests against Matab Optimization Toolbox methods for convex optimization and implementations of the standard log-barrier and active-set methods. We conclude that the steepest descent technique seems to be the best choice so far for our method and that it is competitive with other standard methods both in performance and empirical growth order.
Chen, Xiangdong; He, Liwen; Jeon, Gwanggil; Jeong, Jechang
2014-05-01
In this paper, we present a novel color image demosaicking algorithm based on a directional weighted interpolation method and gradient inverse-weighted filter-based refinement method. By applying a directional weighted interpolation method, the missing center pixel is interpolated, and then using the nearest neighboring pixels of the pre-interpolated pixel within the same color channel, the accuracy of interpolation is refined using a five-point gradient inverse weighted filtering method we proposed. The refined interpolated pixel values can be used to estimate the other missing pixel values successively according to the correlation inter-channels. Experimental analysis of images revealed that our proposed algorithm provided superior performance in terms of both objective and subjective image quality compared to conventional state-of-the-art demosaicking algorithms. Our implementation has very low complexity and is therefore well suited for real-time applications.
Study on the Algorithm for Real-time Interpolation of NURBS Curve and Simulation
Hui Jizhuang; Wei Fangsheng; Gao Kai
2013-01-01
In the paper, In order to meet the needs of high-speed and high- accuracy computerized numerical control machining and guarantee the smooth running in the interpolation processing, A NURBS curve calculation based on adaptive acceleration and deceleration control of look-ahead s-shaped for the real-time interpolation is presented in this paper. The algorithm has merits such as higher position accuracy, short processing time, no variation and so on. Through dynamic path simulation and interpol...
Ho, Yuk-Fan; Ling, Wing-Kuen; Reiss, Joshua; Yu, Xinghuo
2011-01-01
It is well known that second order lowpass interpolative sigma delta modulators (SDMs) may suffer from instability and limit cycle problems when the magnitudes of the input signals are at large and at intermediate levels, respectively. In order to solve these problems, we propose to replace the second order lowpass interpolative SDMs to a specific class of second order bandpass interpolative SDMs with the natural frequencies of the loop filters very close to zero. The global stability propert...
ANGELO-LAMBDA, Covariance matrix interpolation and mathematical verification
International Nuclear Information System (INIS)
Kodeli, Ivo
2007-01-01
1 - Description of program or function: The codes ANGELO-2.3 and LAMBDA-2.3 are used for the interpolation of the cross section covariance data from the original to a user defined energy group structure, and for the mathematical tests of the matrices, respectively. The LAMBDA-2.3 code calculates the eigenvalues of the matrices (both for the original or the converted) and lists them accordingly into positive and negative matrices. This verification is strongly recommended before using any covariance matrices. These versions of the two codes are the extended versions of the previous codes available in the Packages NEA-1264 - ZZ-VITAMIN-J/COVA. They were specifically developed for the purposes of the OECD LWR UAM benchmark, in particular for the processing of the ZZ-SCALE5.1/COVA-44G cross section covariance matrix library retrieved from the SCALE-5.1 package. Either the original SCALE-5.1 libraries or the libraries separated into several files by Nuclides can be (in principle) processed by ANGELO/LAMBDA codes, but the use of the one-nuclide data is strongly recommended. Due to large deviations of the correlation matrix terms from unity observed in some SCALE5.1 covariance matrices, the previous more severe acceptance condition in the ANGELO2.3 code was released. In case the correlation coefficients exceed 1.0, only a warning message is issued, and coefficients are replaced by 1.0. 2 - Methods: ANGELO-2.3 interpolates the covariance matrices to a union grid using flat weighting. LAMBDA-2.3 code includes the mathematical routines to calculate the eigenvalues of the covariance matrices. 3 - Restrictions on the complexity of the problem: The algorithm used in ANGELO is relatively simple, therefore the interpolations involving energy group structure which are very different from the original (e.g. large difference in the number of energy groups) may not be accurate. In particular in the case of the MT=1018 data (fission spectra covariances) the algorithm may not be
Directory of Open Access Journals (Sweden)
Mathieu Lepot
2017-10-01
Full Text Available A thorough review has been performed on interpolation methods to fill gaps in time-series, efficiency criteria, and uncertainty quantifications. On one hand, there are numerous available methods: interpolation, regression, autoregressive, machine learning methods, etc. On the other hand, there are many methods and criteria to estimate efficiencies of these methods, but uncertainties on the interpolated values are rarely calculated. Furthermore, while they are estimated according to standard methods, the prediction uncertainty is not taken into account: a discussion is thus presented on the uncertainty estimation of interpolated/extrapolated data. Finally, some suggestions for further research and a new method are proposed.
Compressive Parameter Estimation for Sparse Translation-Invariant Signals Using Polar Interpolation
DEFF Research Database (Denmark)
Fyhn, Karsten; Duarte, Marco F.; Jensen, Søren Holdt
2015-01-01
We propose new compressive parameter estimation algorithms that make use of polar interpolation to improve the estimator precision. Our work extends previous approaches involving polar interpolation for compressive parameter estimation in two aspects: (i) we extend the formulation from real non...... to attain good estimation precision and keep the computational complexity low. Our numerical experiments show that the proposed algorithms outperform existing approaches that either leverage polynomial interpolation or are based on a conversion to a frequency-estimation problem followed by a super...... interpolation increases the estimation precision....
Directory of Open Access Journals (Sweden)
Mingjian Sun
2015-01-01
Full Text Available Photoacoustic imaging is an innovative imaging technique to image biomedical tissues. The time reversal reconstruction algorithm in which a numerical model of the acoustic forward problem is run backwards in time is widely used. In the paper, a time reversal reconstruction algorithm based on particle swarm optimization (PSO optimized support vector machine (SVM interpolation method is proposed for photoacoustics imaging. Numerical results show that the reconstructed images of the proposed algorithm are more accurate than those of the nearest neighbor interpolation, linear interpolation, and cubic convolution interpolation based time reversal algorithm, which can provide higher imaging quality by using significantly fewer measurement positions or scanning times.
Perbaikan Metode Penghitungan Debit Sungai Menggunakan Cubic Spline Interpolation
Directory of Open Access Journals (Sweden)
Budi I. Setiawan
2007-09-01
Full Text Available Makalah ini menyajikan perbaikan metode pengukuran debit sungai menggunakan fungsi cubic spline interpolation. Fungi ini digunakan untuk menggambarkan profil sungai secara kontinyu yang terbentuk atas hasil pengukuran jarak dan kedalaman sungai. Dengan metoda baru ini, luas dan perimeter sungai lebih mudah, cepat dan tepat dihitung. Demikian pula, fungsi kebalikannnya (inverse function tersedia menggunakan metode. Newton-Raphson sehingga memudahkan dalam perhitungan luas dan perimeter bila tinggi air sungai diketahui. Metode baru ini dapat langsung menghitung debit sungaimenggunakan formula Manning, dan menghasilkan kurva debit (rating curve. Dalam makalah ini dikemukaan satu canton pengukuran debit sungai Rudeng Aceh. Sungai ini mempunyai lebar sekitar 120 m dan kedalaman 7 m, dan pada saat pengukuran mempunyai debit 41 .3 m3/s, serta kurva debitnya mengikuti formula: Q= 0.1649 x H 2.884 , dimana Q debit (m3/s dan H tinggi air dari dasar sungai (m.
Spatial Interpolation of Historical Seasonal Rainfall Indices over Peninsular Malaysia
Hassan, Zulkarnain; Haidir, Ahmad; Saad, Farah Naemah Mohd; Ayob, Afizah; Rahim, Mustaqqim Abdul; Ghazaly, Zuhayr Md.
2018-03-01
The inconsistency in inter-seasonal rainfall due to climate change will cause a different pattern in the rainfall characteristics and distribution. Peninsular Malaysia is not an exception for this inconsistency, in which it is resulting extreme events such as flood and water scarcity. This study evaluates the seasonal patterns in rainfall indices such as total amount of rainfall, the frequency of wet days, rainfall intensity, extreme frequency, and extreme intensity in Peninsular Malaysia. 40 years (1975-2015) data records have been interpolated using Inverse Distance Weighted method. The results show that the formation of rainfall characteristics are significance during the Northeast monsoon (NEM), as compared to Southwest monsoon (SWM). Also, there is a high rainfall intensity and frequency related to extreme over eastern coasts of Peninsula during the NEM season.
Interpolation function for approximating knee joint behavior in human gait
Toth-Taşcǎu, Mirela; Pater, Flavius; Stoia, Dan Ioan
2013-10-01
Starting from the importance of analyzing the kinematic data of the lower limb in gait movement, especially the angular variation of the knee joint, the paper propose an approximation function that can be used for processing the correlation among a multitude of knee cycles. The approximation of the raw knee data was done by Lagrange polynomial interpolation on a signal acquired using Zebris Gait Analysis System. The signal used in approximation belongs to a typical subject extracted from a lot of ten investigated subjects, but the function domain of definition belongs to the entire group. The study of the knee joint kinematics plays an important role in understanding the kinematics of the gait, this articulation having the largest range of motion in whole joints, in gait. The study does not propose to find an approximation function for the adduction-abduction movement of the knee, this being considered a residual movement comparing to the flexion-extension.
Finite element analysis of rotating beams physics based interpolation
Ganguli, Ranjan
2017-01-01
This book addresses the solution of rotating beam free-vibration problems using the finite element method. It provides an introduction to the governing equation of a rotating beam, before outlining the solution procedures using Rayleigh-Ritz, Galerkin and finite element methods. The possibility of improving the convergence of finite element methods through a judicious selection of interpolation functions, which are closer to the problem physics, is also addressed. The book offers a valuable guide for students and researchers working on rotating beam problems – important engineering structures used in helicopter rotors, wind turbines, gas turbines, steam turbines and propellers – and their applications. It can also be used as a textbook for specialized graduate and professional courses on advanced applications of finite element analysis.
Plasma simulation with the Differential Algebraic Cubic Interpolated Propagation scheme
Energy Technology Data Exchange (ETDEWEB)
Utsumi, Takayuki [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment
1998-03-01
A computer code based on the Differential Algebraic Cubic Interpolated Propagation scheme has been developed for the numerical solution of the Boltzmann equation for a one-dimensional plasma with immobile ions. The scheme advects the distribution function and its first derivatives in the phase space for one time step by using a numerical integration method for ordinary differential equations, and reconstructs the profile in phase space by using a cubic polynomial within a grid cell. The method gives stable and accurate results, and is efficient. It is successfully applied to a number of equations; the Vlasov equation, the Boltzmann equation with the Fokker-Planck or the Bhatnagar-Gross-Krook (BGK) collision term and the relativistic Vlasov equation. The method can be generalized in a straightforward way to treat cases such as problems with nonperiodic boundary conditions and higher dimensional problems. (author)
Twitch interpolation technique in testing of maximal muscle strength
DEFF Research Database (Denmark)
Bülow, P M; Nørregaard, J; Danneskiold-Samsøe, B
1993-01-01
The aim was to study the methodological aspects of the muscle twitch interpolation technique in estimating the maximal force of contraction in the quadriceps muscle utilizing commercial muscle testing equipment. Six healthy subjects participated in seven sets of experiments testing the effects...... on twitch size of potentiation, time lag after potentiation, magnitude of voluntary force, stimulus amplitude, stimulus duration, angle of the knee, and angle of the hip. In addition, the consequences of submaximal potentiation on the estimation of maximal force from twitch sizes were studied in five...... healthy subjects. We found an increase in twitch size with increasing levels of potentiation and twitch size decreased exponentially following potentiation. We found a curvilinear relationship between twitch size and voluntary force, and these properties were more obvious when the stimulation intensity...
Interpolation Grid for Local Area of Iasi City
Directory of Open Access Journals (Sweden)
Mihalache Raluca Maria
2014-05-01
Full Text Available Definitive transition to GNSS technology of achieving geodetic networks for cadastre implementation in cities and municipalities, enforce establishing a unique way of linking between current measurements and existing geodetic data, with a sufficient accuracy proper to urban cadastre standards. Regarding city of Iasi, is presented a different method of transformation which consist in an interpolation grid for heights system. The Romanian national height system is „Black Sea-1975” normal heights system. Founded in 1945 by Molodenski, this system uses the quasigeoid as reference surface, being in relation with the ellipsoid through the height anomalies sizes in each point. The unitary transformation between the ETRS- 89 ellipsoidal height system and the normal one, at national level is provided through the „TransdatRo” program developed by NACLR (National Agency for Cadastre and Land Registration.
Interpolation properties of Besov spaces defined on metric spaces
Czech Academy of Sciences Publication Activity Database
Gogatishvili, Amiran; Koskela, P.; Shanmugalingam, N.
2010-01-01
Roč. 283, č. 2 (2010), s. 215-231 ISSN 0025-584X R&D Projects: GA ČR GA201/05/2033; GA ČR GA201/08/0383 Institutional research plan: CEZ:AV0Z10190503 Keywords : Besov spaces * Sobolev spaces * real interpolation method * K-functional * metric measure space * doubling measure space * embedding theorems Subject RIV: BA - General Mathematics Impact factor: 0.653, year: 2010 http://onlinelibrary.wiley.com/doi/10.1002/ mana .200810242/abstract;jsessionid=144DA4B489B3CA8F4C6A08EF8BD172FE.f03t04
Estimating Frequency by Interpolation Using Least Squares Support Vector Regression
Directory of Open Access Journals (Sweden)
Changwei Ma
2015-01-01
Full Text Available Discrete Fourier transform- (DFT- based maximum likelihood (ML algorithm is an important part of single sinusoid frequency estimation. As signal to noise ratio (SNR increases and is above the threshold value, it will lie very close to Cramer-Rao lower bound (CRLB, which is dependent on the number of DFT points. However, its mean square error (MSE performance is directly proportional to its calculation cost. As a modified version of support vector regression (SVR, least squares SVR (LS-SVR can not only still keep excellent capabilities for generalizing and fitting but also exhibit lower computational complexity. In this paper, therefore, LS-SVR is employed to interpolate on Fourier coefficients of received signals and attain high frequency estimation accuracy. Our results show that the proposed algorithm can make a good compromise between calculation cost and MSE performance under the assumption that the sample size, number of DFT points, and resampling points are already known.
Basis set approach in the constrained interpolation profile method
International Nuclear Information System (INIS)
Utsumi, T.; Koga, J.; Yabe, T.; Ogata, Y.; Matsunaga, E.; Aoki, T.; Sekine, M.
2003-07-01
We propose a simple polynomial basis-set that is easily extendable to any desired higher-order accuracy. This method is based on the Constrained Interpolation Profile (CIP) method and the profile is chosen so that the subgrid scale solution approaches the real solution by the constraints from the spatial derivative of the original equation. Thus the solution even on the subgrid scale becomes consistent with the master equation. By increasing the order of the polynomial, this solution quickly converges. 3rd and 5th order polynomials are tested on the one-dimensional Schroedinger equation and are proved to give solutions a few orders of magnitude higher in accuracy than conventional methods for lower-lying eigenstates. (author)
Spatial Interpolation of Historical Seasonal Rainfall Indices over Peninsular Malaysia
Directory of Open Access Journals (Sweden)
Hassan Zulkarnain
2018-01-01
Full Text Available The inconsistency in inter-seasonal rainfall due to climate change will cause a different pattern in the rainfall characteristics and distribution. Peninsular Malaysia is not an exception for this inconsistency, in which it is resulting extreme events such as flood and water scarcity. This study evaluates the seasonal patterns in rainfall indices such as total amount of rainfall, the frequency of wet days, rainfall intensity, extreme frequency, and extreme intensity in Peninsular Malaysia. 40 years (1975-2015 data records have been interpolated using Inverse Distance Weighted method. The results show that the formation of rainfall characteristics are significance during the Northeast monsoon (NEM, as compared to Southwest monsoon (SWM. Also, there is a high rainfall intensity and frequency related to extreme over eastern coasts of Peninsula during the NEM season.
Diabat Interpolation for Polymorph Free-Energy Differences.
Kamat, Kartik; Peters, Baron
2017-02-02
Existing methods to compute free-energy differences between polymorphs use harmonic approximations, advanced non-Boltzmann bias sampling techniques, and/or multistage free-energy perturbations. This work demonstrates how Bennett's diabat interpolation method ( J. Comput. Phys. 1976, 22, 245 ) can be combined with energy gaps from lattice-switch Monte Carlo techniques ( Phys. Rev. E 2000, 61, 906 ) to swiftly estimate polymorph free-energy differences. The new method requires only two unbiased molecular dynamics simulations, one for each polymorph. To illustrate the new method, we compute the free-energy difference between face-centered cubic and body-centered cubic polymorphs for a Gaussian core solid. We discuss the justification for parabolic models of the free-energy diabats and similarities to methods that have been used in studies of electron transfer.
On the exact interpolating function in ABJ theory
Energy Technology Data Exchange (ETDEWEB)
Cavaglià, Andrea [Dipartimento di Fisica and INFN, Università di Torino,Via P. Giuria 1, 10125 Torino (Italy); Gromov, Nikolay [Mathematics Department, King’s College London,The Strand, London WC2R 2LS (United Kingdom); St. Petersburg INP,Gatchina, 188 300, St.Petersburg (Russian Federation); Levkovich-Maslyuk, Fedor [Mathematics Department, King’s College London,The Strand, London WC2R 2LS (United Kingdom); Nordita, KTH Royal Institute of Technology and Stockholm University,Roslagstullsbacken 23, SE-106 91 Stockholm (Sweden)
2016-12-16
Based on the recent indications of integrability in the planar ABJ model, we conjecture an exact expression for the interpolating function h(λ{sub 1},λ{sub 2}) in this theory. Our conjecture is based on the observation that the integrability structure of the ABJM theory given by its Quantum Spectral Curve is very rigid and does not allow for a simple consistent modification. Under this assumption, we revised the previous comparison of localization results and exact all loop integrability calculations done for the ABJM theory by one of the authors and Grigory Sizov, fixing h(λ{sub 1},λ{sub 2}). We checked our conjecture against various weak coupling expansions, at strong coupling and also demonstrated its invariance under the Seiberg-like duality. This match also gives further support to the integrability of the model. If our conjecture is correct, it extends all the available integrability results in the ABJM model to the ABJ model.
INTERPOL, EUROPOL ve Uluslararası Terörizm
KAYA, Sezgin
2015-01-01
Günümüz dünyasında uluslararası terörizm, genellikle devletlerin güvenliklerine yönelik en önemli tehditlerden biri olarak kabul edilmektedir. Bu nedenle devletler, ilgili tehditle mücadele amacıyla çeşitli biçimlerde ve değişik alanlarda is birliği yapmaktadırlar. Bu balgamda, uluslararası terörizme karsı polisiye ve adli meselelere ilişkin olarak is birliği yapmak da oldukça önemlidir. Çalışmada, bu amaçla oluşturulmuş olan INTERPOL ve EUROPOL'ün uluslararası terörizm karsısındaki roll...
Differential maps, difference maps, interpolated maps, and long term prediction
International Nuclear Information System (INIS)
Talman, R.
1988-06-01
Mapping techniques may be thought to be attractive for the long term prediction of motion in accelerators, especially because a simple map can approximately represent an arbitrarily complicated lattice. The intention of this paper is to develop prejudices as to the validity of such methods by applying them to a simple, exactly solveable, example. It is shown that a numerical interpolation map, such as can be generated in the accelerator tracking program TEAPOT, predicts the evolution more accurately than an analytically derived differential map of the same order. Even so, in the presence of ''appreciable'' nonlinearity, it is shown to be impractical to achieve ''accurate'' prediction beyond some hundreds of cycles of oscillation. This suggests that the value of nonlinear maps is restricted to the parameterization of only the ''leading'' deviation from linearity. 41 refs., 6 figs
The interpolation damage detection method for frames under seismic excitation
Limongelli, M. P.
2011-10-01
In this paper a new procedure, addressed as Interpolation Damage Detecting Method (IDDM), is investigated as a possible mean for early detection and location of light damage in a structure struck by an earthquake. Damage is defined in terms of the accuracy of a spline function in interpolating the operational mode shapes (ODS) of the structure. At a certain location a decrease (statistically meaningful) of accuracy, with respect to a reference configuration, points out a localized variation of the operational shapes thus revealing the existence of damage. In this paper, the proposed method is applied to a numerical model of a multistory frame, simulating a damaged condition through a reduction of the story stiffness. Several damage scenarios have been considered and the results indicate the effectiveness of the method to assess and localize damage for the case of concentrated damage and for low to medium levels of noise in the recorded signals. The main advantage of the proposed algorithm is that it does not require a numerical model of the structure as well as an intense data post-processing or user interaction. The ODS are calculated from Frequency Response Functions hence responses recorded on the structure can be directly used without the need of modal identification. Furthermore, the local character of the feature chosen to detect damage makes the IDDM less sensitive to noise and to environmental changes with respect to other damage detection methods. For these reasons the IDDM appears as a valid option for automated post-earthquake damage assessment, able to provide after an earthquake, reliable information about the location of damage.
Lepot, M.J.; Aubin, Jean Baptiste; Clemens, F.H.L.R.
2017-01-01
A thorough review has been performed on interpolation methods to fill gaps in time-series, efficiency criteria, and uncertainty quantifications. On one hand, there are numerous available methods: interpolation, regression, autoregressive, machine learning methods, etc. On the other hand, there are
Application of Time-Frequency Domain Transform to Three-Dimensional Interpolation of Medical Images.
Lv, Shengqing; Chen, Yimin; Li, Zeyu; Lu, Jiahui; Gao, Mingke; Lu, Rongrong
2017-11-01
Medical image three-dimensional (3D) interpolation is an important means to improve the image effect in 3D reconstruction. In image processing, the time-frequency domain transform is an efficient method. In this article, several time-frequency domain transform methods are applied and compared in 3D interpolation. And a Sobel edge detection and 3D matching interpolation method based on wavelet transform is proposed. We combine wavelet transform, traditional matching interpolation methods, and Sobel edge detection together in our algorithm. What is more, the characteristics of wavelet transform and Sobel operator are used. They deal with the sub-images of wavelet decomposition separately. Sobel edge detection 3D matching interpolation method is used in low-frequency sub-images under the circumstances of ensuring high frequency undistorted. Through wavelet reconstruction, it can get the target interpolation image. In this article, we make 3D interpolation of the real computed tomography (CT) images. Compared with other interpolation methods, our proposed method is verified to be effective and superior.
Interpolation of unevenly spaced data using a parabolic leapfrog correction method and cubic splines
Julio L. Guardado; William T. Sommers
1977-01-01
The technique proposed allows interpolation of data recorded at unevenly spaced sites to a regular grid or to other sites. Known data are interpolated to an initial guess field grid of unevenly spaced rows and columns by a simple distance weighting procedure. The initial guess field is then adjusted by using a parabolic leapfrog correction and the known data. The final...
Kriging interpolation in seismic attribute space applied to the South Arne Field, North Sea
DEFF Research Database (Denmark)
Hansen, Thomas Mejer; Mosegaard, Klaus; Schiøtt, Christian
2010-01-01
Seismic attributes can be used to guide interpolation in-between and extrapolation away from well log locations using for example linear regression, neural networks, and kriging. Kriging-based estimation methods (and most other types of interpolation/extrapolation techniques) are intimately linke...
The Neville-Aitken formula for rational interpolants with prescribed poles
Carstensen, C.; Mühlbach, G.
1992-12-01
Using a polynomial description of rational interpolation with prescribed poles a simple purely algebraic proof of a Neville-Aitken recurrence formula for rational interpolants with prescribed poles is presented. It is used to compute the general Cauchy-Vandermonde determinant explicitly in terms of the nodes and poles involved.
Muhammad, Sayyid
2015-01-01
Kejahatan-kejahatan yang diatur dalam konvensi internasional pada dasarnya memiliki tiga karakteristik yaitu: kejahatan yang membahayakan umat manusia, kejahatan yang mana pelakunya dapat diekstradisi, dan kejahatan yang dianggap bukan kejahatan politik. Permasalahan dalam penelitian ini adalah Bagaimana kedudukan Interpol sebagai organisasi internasional.Bagaimana perspektif hukum internasional dalam mengatur perdagangan manusia (human trafficking).Bagaimana peranan Interpol dalam memberanta...
The twitch interpolation technique for study of fatigue of human quadriceps muscle
DEFF Research Database (Denmark)
Bülow, P M; Nørregaard, J; Mehlsen, J
1995-01-01
The aim of the study was to examine if the twitch interpolation technique could be used to objectively measure fatigue in the quadriceps muscle in subjects performing submaximally. The 'true' maximum isometric quadriceps torque was determined in 21 healthy subject using the twitch interpolation...
Okounkov's BC-Type Interpolation Macdonald Polynomials and Their q=1 Limit
Koornwinder, T.H.
2015-01-01
This paper surveys eight classes of polynomials associated with A-type and BC-type root systems: Jack, Jacobi, Macdonald and Koornwinder polynomials and interpolation (or shifted) Jack and Macdonald polynomials and their BC-type extensions. Among these the BC-type interpolation Jack polynomials were
Abstract interpolation in vector-valued de Branges-Rovnyak spaces
Ball, J.A.; Bolotnikov, V.; ter Horst, S.
2011-01-01
Following ideas from the Abstract Interpolation Problem of Katsnelson et al. (Operators in spaces of functions and problems in function theory, vol 146, pp 83–96, Naukova Dumka, Kiev, 1987) for Schur class functions, we study a general metric constrained interpolation problem for functions from a
Conformal Interpolating Algorithm Based on Cubic NURBS in Aspheric Ultra-Precision Machining
International Nuclear Information System (INIS)
Li, C G; Zhang, Q R; Cao, C G; Zhao, S L
2006-01-01
Numeric control machining and on-line compensation for aspheric surface are key techniques in ultra-precision machining. In this paper, conformal cubic NURBS interpolating curve is applied to fit the character curve of aspheric surface. Its algorithm and process are also proposed and imitated by Matlab7.0 software. To evaluate the performance of the conformal cubic NURBS interpolation, we compare it with the linear interpolations. The result verifies this method can ensure smoothness of interpolating spline curve and preserve original shape characters. The surface quality interpolated by cubic NURBS is higher than by line. The algorithm is benefit to increasing the surface form precision of workpieces in ultra-precision machining
Empirical performance of interpolation techniques in risk-neutral density (RND) estimation
Bahaludin, H.; Abdullah, M. H.
2017-03-01
The objective of this study is to evaluate the empirical performance of interpolation techniques in risk-neutral density (RND) estimation. Firstly, the empirical performance is evaluated by using statistical analysis based on the implied mean and the implied variance of RND. Secondly, the interpolation performance is measured based on pricing error. We propose using the leave-one-out cross-validation (LOOCV) pricing error for interpolation selection purposes. The statistical analyses indicate that there are statistical differences between the interpolation techniques:second-order polynomial, fourth-order polynomial and smoothing spline. The results of LOOCV pricing error shows that interpolation by using fourth-order polynomial provides the best fitting to option prices in which it has the lowest value error.
Interpolation Algorithm and Mathematical Model in Automated Welding of Saddle-Shaped Weld
Directory of Open Access Journals (Sweden)
Lianghao Xue
2018-01-01
Full Text Available This paper presents welding torch pose model and interpolation algorithm of trajectory control of saddle-shaped weld formed by intersection of two pipes; the working principle, interpolation algorithm, welding experiment, and simulation result of the automatic welding system of the saddle-shaped weld are described. A variable angle interpolation method is used to control the trajectory and pose of the welding torch, which guarantees the constant linear terminal velocity. The mathematical model of the trajectory and pose of welding torch are established. Simulation and experiment have been carried out to verify the effectiveness of the proposed algorithm and mathematical model. The results demonstrate that the interpolation algorithm is well within the interpolation requirements of the saddle-shaped weld and ideal feed rate stability.
International co-operation through the Interpol system to counter illicit drug trafficking.
Leamy, W J
1983-01-01
The International Criminal Police Organization (ICPO/Interpol), whose main aim is the prevention and suppression of ordinary crime, has 135 member countries. The Government of each of these countries has designated an Interpol National Central Bureau to co-operate and liaise within the framework of Interpol. The Drugs Sub-Division of Interpol's General Secretariat monitors and responds to incoming communications on drug enforcement matters, conducts intelligence analysis of information and produces tactical and strategic intelligence reports as well as statistical and other specialized reports. It received 33,181 and dispatched 6,741 drug-enforcement-related communications in 1982, which was over 60 per cent of the entire communications of the General Secretariat. The Drugs Sub-Division participates in drug training and drug strategy seminars world-wide. Interpol also carries out drug liaison officer programmes in five regions of the world.
Interpolation/penalization applied for strength design of 3D thermoelastic structures
DEFF Research Database (Denmark)
Pedersen, Pauli; Pedersen, Niels L.
2012-01-01
illustrated with examples. In compliance minimization for thermoelastic structures it may be advantageous to decrease the total volume, but for strength maximization it is argued to keep the total permissible volume. Linear interpolation (no penalization) may to a certain extent be argued for 2D thickness...... optimized designs, but for 3D design problems interpolation must be included and not only from the penalization point of view to obtain 0-1 designs. Three interpolation types are presented in a uniform manner, including the well known one parameter penalizations, named SIMP and RAMP. An alternative two...... parameter interpolation in explicit form is preferred, and the influence of interpolation on compliance sensitivity analysis is included. For direct strength maximization the sensitivity analysis of local von Mises stresses is demanding. An applied recursive procedure to obtain uniform energy density...
Identification method for digital image forgery and filtering region through interpolation.
Hwang, Min Gu; Har, Dong Hwan
2014-09-01
Because of the rapidly increasing use of digital composite images, recent studies have identified digital forgery and filtering regions. This research has shown that interpolation, which is used to edit digital images, is an effective way to analyze digital images for composite regions. Interpolation is widely used to adjust the size of the image of a composite target, making the composite image seem natural by rotating or deforming. As a result, many algorithms have been developed to identify composite regions by detecting a trace of interpolation. However, many limitations have been found in detection maps developed to identify composite regions. In this study, we analyze the pixel patterns of noninterpolation and interpolation regions. We propose a detection map algorithm to separate the two regions. To identify composite regions, we have developed an improved algorithm using minimum filer, Laplacian operation and maximum filters. Finally, filtering regions that used the interpolation operation are analyzed using the proposed algorithm. © 2014 American Academy of Forensic Sciences.
Wind Resource Mapping Using Landscape Roughness and Spatial Interpolation Methods
Directory of Open Access Journals (Sweden)
Samuel Van Ackere
2015-08-01
Full Text Available Energy saving, reduction of greenhouse gasses and increased use of renewables are key policies to achieve the European 2020 targets. In particular, distributed renewable energy sources, integrated with spatial planning, require novel methods to optimise supply and demand. In contrast with large scale wind turbines, small and medium wind turbines (SMWTs have a less extensive impact on the use of space and the power system, nevertheless, a significant spatial footprint is still present and the need for good spatial planning is a necessity. To optimise the location of SMWTs, detailed knowledge of the spatial distribution of the average wind speed is essential, hence, in this article, wind measurements and roughness maps were used to create a reliable annual mean wind speed map of Flanders at 10 m above the Earth’s surface. Via roughness transformation, the surface wind speed measurements were converted into meso- and macroscale wind data. The data were further processed by using seven different spatial interpolation methods in order to develop regional wind resource maps. Based on statistical analysis, it was found that the transformation into mesoscale wind, in combination with Simple Kriging, was the most adequate method to create reliable maps for decision-making on optimal production sites for SMWTs in Flanders (Belgium.
Shape Designing of Engineering Images Using Rational Spline Interpolation
Directory of Open Access Journals (Sweden)
Muhammad Sarfraz
2015-01-01
Full Text Available In modern days, engineers encounter a remarkable range of different engineering problems like study of structure, structure properties, and designing of different engineering images, for example, automotive images, aerospace industrial images, architectural designs, shipbuilding, and so forth. This paper purposes an interactive curve scheme for designing engineering images. The purposed scheme furnishes object designing not just in the area of engineering, but it is equally useful for other areas including image processing (IP, Computer Graphics (CG, Computer-Aided Engineering (CAE, Computer-Aided Manufacturing (CAM, and Computer-Aided Design (CAD. As a method, a piecewise rational cubic spline interpolant, with four shape parameters, has been purposed. The method provides effective results together with the effects of derivatives and shape parameters on the shape of the curves in a local and global manner. The spline method, due to its most generalized description, recovers various existing rational spline methods and serves as an alternative to various other methods including v-splines, gamma splines, weighted splines, and beta splines.
Spatial frequency interference effects and interpolation in vernier acuity.
Morgan, M J; Watt, R J
1984-01-01
Discrete spatial sampling of sinusoidal gratings has a detrimental effect upon vernier acuity for such stimuli if the sampling rate is less than 20 c/deg. We have examined the possibility that interpolation failure is due to masking by spatial frequency components near to the signal frequency. In an explicit masking paradigm, vernier acuity for a sine wave grating was found to be adversely affected by the presence of a masking grating falling within a critical band of frequencies near the test target. In the sampled stimulus, removal of sampling replicas similar in frequency to the test improved acuity, but the extent of the residual interference by high frequency components could not be predicted from the masking data. The high frequency interference effect depended on fixed phase relations between frequency components, creating narrow bright bars in the spatial luminance profile, and was greatly reduced by phase randomizing the sampling replicas. Various explanations of this nonlinear behaviour are discussed, including the Marr-Hildreth theory of edge detection.
Xu, Zhuo; Sopher, Daniel; Juhlin, Christopher; Han, Liguo; Gong, Xiangbo
2018-04-01
In towed marine seismic data acquisition, a gap between the source and the nearest recording channel is typical. Therefore, extrapolation of the missing near-offset traces is often required to avoid unwanted effects in subsequent data processing steps. However, most existing interpolation methods perform poorly when extrapolating traces. Interferometric interpolation methods are one particular method that have been developed for filling in trace gaps in shot gathers. Interferometry-type interpolation methods differ from conventional interpolation methods as they utilize information from several adjacent shot records to fill in the missing traces. In this study, we aim to improve upon the results generated by conventional time-space domain interferometric interpolation by performing interferometric interpolation in the Radon domain, in order to overcome the effects of irregular data sampling and limited source-receiver aperture. We apply both time-space and Radon-domain interferometric interpolation methods to the Sigsbee2B synthetic dataset and a real towed marine dataset from the Baltic Sea with the primary aim to improve the image of the seabed through extrapolation into the near-offset gap. Radon-domain interferometric interpolation performs better at interpolating the missing near-offset traces than conventional interferometric interpolation when applied to data with irregular geometry and limited source-receiver aperture. We also compare the interferometric interpolated results with those obtained using solely Radon transform (RT) based interpolation and show that interferometry-type interpolation performs better than solely RT-based interpolation when extrapolating the missing near-offset traces. After data processing, we show that the image of the seabed is improved by performing interferometry-type interpolation, especially when Radon-domain interferometric interpolation is applied.
Linear and Quadratic Interpolators Using Truncated-Matrix Multipliers and Squarers
Directory of Open Access Journals (Sweden)
E. George Walters III
2015-11-01
Full Text Available This paper presents a technique for designing linear and quadratic interpolators for function approximation using truncated multipliers and squarers. Initial coefficient values are found using a Chebyshev-series approximation and then adjusted through exhaustive simulation to minimize the maximum absolute error of the interpolator output. This technique is suitable for any function and any precision up to 24 bits (IEEE single precision. Designs for linear and quadratic interpolators that implement the 1/x, 1/ √ x, log2(1+2x, log2(x and 2x functions are presented and analyzed as examples. Results show that a proposed 24-bit interpolator computing 1/x with a design specification of ±1 unit in the last place of the product (ulp error uses 16.4% less area and 15.3% less power than a comparable standard interpolator with the same error specification. Sixteen-bit linear interpolators for other functions are shown to use up to 17.3% less area and 12.1% less power, and 16-bit quadratic interpolators are shown to use up to 25.8% less area and 24.7% less power.
Restoration of rotational motion blurred image based on Chebyshev polynomial interpolations
Zhang, Tianxu; Hong, Hanyu; Yan, Luxin; Zhang, Xiuhua
2009-10-01
The restoration of rotational motion blurred image involves a lot of interpolations operators in rectangular-to-polar transformation and its inversion of polar-to-rectangular. The technique of interpolation determines the quality of restoration and computational complexity. In this paper, we incorporate orthogonal chebyshev polynomials interpolations into the processing of restoration of rotational motion blurred image, in which the space-variant blurs are decomposed into a series of space-invariant blurs along the blurring paths, and the blurred gray-values of the discrete pixels of the blurring paths are calculated by using of orthogonal chebyshev polynomials' interpolations and the space-variant blurs can be removed along the blurring paths in the polar system. At same way, we use orthogonal chebyshev polynomials' interpolations to perform polar-to-rectangular transformation to put the restored image back to its original rectangular format. In order to overcome the interference of noise, an optimization restoration algorithm based on regularizations is presented, in which non-negative and edge-preserving smoothing are incorporated into the process of restoration. A series of experiments have been performed to test the proposed interpolation method, which show that the proposed interpolations are effective to preserve edges.
Pearce, Mark A
2015-08-01
EBSDinterp is a graphic user interface (GUI)-based MATLAB® program to perform microstructurally constrained interpolation of nonindexed electron backscatter diffraction data points. The area available for interpolation is restricted using variations in pattern quality or band contrast (BC). Areas of low BC are not available for interpolation, and therefore cannot be erroneously filled by adjacent grains "growing" into them. Points with the most indexed neighbors are interpolated first and the required number of neighbors is reduced with each successive round until a minimum number of neighbors is reached. Further iterations allow more data points to be filled by reducing the BC threshold. This method ensures that the best quality points (those with high BC and most neighbors) are interpolated first, and that the interpolation is restricted to grain interiors before adjacent grains are grown together to produce a complete microstructure. The algorithm is implemented through a GUI, taking advantage of MATLAB®'s parallel processing toolbox to perform the interpolations rapidly so that a variety of parameters can be tested to ensure that the final microstructures are robust and artifact-free. The software is freely available through the CSIRO Data Access Portal (doi:10.4225/08/5510090C6E620) as both a compiled Windows executable and as source code.
STUDY OF BLOCKING EFFECT ELIMINATION METHODS BY MEANS OF INTRAFRAME VIDEO SEQUENCE INTERPOLATION
Directory of Open Access Journals (Sweden)
I. S. Rubina
2015-01-01
Full Text Available The paper deals with image interpolation methods and their applicability to eliminate some of the artifacts related to both the dynamic properties of objects in video sequences and algorithms used in the order of encoding steps. The main drawback of existing methods is the high computational complexity, unacceptable in video processing. Interpolation of signal samples for blocking - effect elimination at the output of the convertion encoding is proposed as a part of the study. It was necessary to develop methods for improvement of compression ratio and quality of the reconstructed video data by blocking effect elimination on the borders of the segments by intraframe interpolating of video sequence segments. The main point of developed methods is an adaptive recursive algorithm application with adaptive-sized interpolation kernel both with and without the brightness gradient consideration at the boundaries of objects and video sequence blocks. Within theoretical part of the research, methods of information theory (RD-theory and data redundancy elimination, methods of pattern recognition and digital signal processing, as well as methods of probability theory are used. Within experimental part of the research, software implementation of compression algorithms with subsequent comparison of the implemented algorithms with the existing ones was carried out. Proposed methods were compared with the simple averaging algorithm and the adaptive algorithm of central counting interpolation. The advantage of the algorithm based on the adaptive kernel size selection interpolation is in compression ratio increasing by 30%, and the advantage of the modified algorithm based on the adaptive interpolation kernel size selection is in the compression ratio increasing by 35% in comparison with existing algorithms, interpolation and quality of the reconstructed video sequence improving by 3% compared to the one compressed without interpolation. The findings will be
Garzón-Machado, Víctor; Otto, Rüdiger; del Arco Aguilar, Marcelino José
2014-07-01
Different spatial interpolation techniques have been applied to construct objective bioclimatic maps of La Palma, Canary Islands. Interpolation of climatic data on this topographically complex island with strong elevation and climatic gradients represents a challenge. Furthermore, meteorological stations are not evenly distributed over the island, with few stations at high elevations. We carried out spatial interpolations of the compensated thermicity index (Itc) and the annual ombrothermic Index (Io), in order to obtain appropriate bioclimatic maps by using automatic interpolation procedures, and to establish their relation to potential vegetation units for constructing a climatophilous potential natural vegetation map (CPNV). For this purpose, we used five interpolation techniques implemented in a GIS: inverse distance weighting (IDW), ordinary kriging (OK), ordinary cokriging (OCK), multiple linear regression (MLR) and MLR followed by ordinary kriging of the regression residuals. Two topographic variables (elevation and aspect), derived from a high-resolution digital elevation model (DEM), were included in OCK and MLR. The accuracy of the interpolation techniques was examined by the results of the error statistics of test data derived from comparison of the predicted and measured values. Best results for both bioclimatic indices were obtained with the MLR method with interpolation of the residuals showing the highest R 2 of the regression between observed and predicted values and lowest values of root mean square errors. MLR with correction of interpolated residuals is an attractive interpolation method for bioclimatic mapping on this oceanic island since it permits one to fully account for easily available geographic information but also takes into account local variation of climatic data.
International Nuclear Information System (INIS)
Sims, C.S.; Killough, G.G.
1983-01-01
Various segments of the health physics community advocate the use of different sets of neutron fluence-to-dose equivalent conversion factors as a function of energy and different methods of interpolation between discrete points in those data sets. The major data sets and interpolation methods are used to calculate the spectrum average fluence-to-dose equivalent conversion factors for five spectra associated with the various shielded conditions of the Health Physics Research Reactor. The results obtained by use of the different data sets and interpolation methods are compared and discussed. (author)
Blend Shape Interpolation and FACS for Realistic Avatar
Alkawaz, Mohammed Hazim; Mohamad, Dzulkifli; Basori, Ahmad Hoirul; Saba, Tanzila
2015-03-01
The quest of developing realistic facial animation is ever-growing. The emergence of sophisticated algorithms, new graphical user interfaces, laser scans and advanced 3D tools imparted further impetus towards the rapid advancement of complex virtual human facial model. Face-to-face communication being the most natural way of human interaction, the facial animation systems became more attractive in the information technology era for sundry applications. The production of computer-animated movies using synthetic actors are still challenging issues. Proposed facial expression carries the signature of happiness, sadness, angry or cheerful, etc. The mood of a particular person in the midst of a large group can immediately be identified via very subtle changes in facial expressions. Facial expressions being very complex as well as important nonverbal communication channel are tricky to synthesize realistically using computer graphics. Computer synthesis of practical facial expressions must deal with the geometric representation of the human face and the control of the facial animation. We developed a new approach by integrating blend shape interpolation (BSI) and facial action coding system (FACS) to create a realistic and expressive computer facial animation design. The BSI is used to generate the natural face while the FACS is employed to reflect the exact facial muscle movements for four basic natural emotional expressions such as angry, happy, sad and fear with high fidelity. The results in perceiving the realistic facial expression for virtual human emotions based on facial skin color and texture may contribute towards the development of virtual reality and game environment of computer aided graphics animation systems.
Geospatial Interpolation and Mapping of Tropospheric Ozone Pollution Using Geostatistics
Directory of Open Access Journals (Sweden)
Swatantra R. Kethireddy
2014-01-01
Full Text Available Tropospheric ozone (O3 pollution is a major problem worldwide, including in the United States of America (USA, particularly during the summer months. Ozone oxidative capacity and its impact on human health have attracted the attention of the scientific community. In the USA, sparse spatial observations for O3 may not provide a reliable source of data over a geo-environmental region. Geostatistical Analyst in ArcGIS has the capability to interpolate values in unmonitored geo-spaces of interest. In this study of eastern Texas O3 pollution, hourly episodes for spring and summer 2012 were selectively identified. To visualize the O3 distribution, geostatistical techniques were employed in ArcMap. Using ordinary Kriging, geostatistical layers of O3 for all the studied hours were predicted and mapped at a spatial resolution of 1 kilometer. A decent level of prediction accuracy was achieved and was confirmed from cross-validation results. The mean prediction error was close to 0, the root mean-standardized-prediction error was close to 1, and the root mean square and average standard errors were small. O3 pollution map data can be further used in analysis and modeling studies. Kriging results and O3 decadal trends indicate that the populace in Houston-Sugar Land-Baytown, Dallas-Fort Worth-Arlington, Beaumont-Port Arthur, San Antonio, and Longview are repeatedly exposed to high levels of O3-related pollution, and are prone to the corresponding respiratory and cardiovascular health effects. Optimization of the monitoring network proves to be an added advantage for the accurate prediction of exposure levels.
CSIR Research Space (South Africa)
Bogaers, Alfred EJ
2016-10-01
Full Text Available In this paper we outline the use of radial basis function interpolation (RBF) to transfer information across non-matching and nonconforming interface meshes, with particular focus to partitioned fluid-structure interactions (FSI). In general...
Directory of Open Access Journals (Sweden)
Pengyun Chen
2014-01-01
Full Text Available The interpolation-reconstruction of local underwater terrain using the underwater digital terrain map (UDTM is an important step for building an underwater terrain matching unit and directly affects the accuracy of underwater terrain matching navigation. The Kriging method is often used in terrain interpolation, but, with this method, the local terrain features are often lost. Therefore, the accuracy cannot meet the requirements of practical application. Analysis of the geographical features is performed on the basis of the randomness and self-similarity of underwater terrain. We extract the fractal features of local underwater terrain with the fractal Brownian motion model, compensating for the possible errors of the Kriging method with fractal theory. We then put forward an improved Kriging interpolation method based on this fractal compensation. Interpolation-reconstruction tests show that the method can simulate the real underwater terrain features well and that it has good usability.
A Parallel Strategy for High-speed Interpolation of CNC Using Data Space Constraint Method
Directory of Open Access Journals (Sweden)
Shuan-qiang Yang
2013-12-01
Full Text Available A high-speed interpolation scheme using parallel computing is proposed in this paper. The interpolation method is divided into two tasks, namely, the rough task executing in PC and the fine task in the I/O card. During the interpolation procedure, the double buffers are constructed to exchange the interpolation data between the two tasks. Then, the data space constraint method is adapted to ensure the reliable and continuous data communication between the two buffers. Therefore, the proposed scheme can be realized in the common distribution of the operation systems without real-time performance. The high-speed and high-precision motion control can be achieved as well. Finally, an experiment is conducted on the self-developed CNC platform, the test results are shown to verify the proposed method.
Image Super-Resolution Using Adaptive 2-D Gaussian Basis Function Interpolation
National Research Council Canada - National Science Library
Hunt, Terence
2004-01-01
... characteristics to be more effectively represented. The interpolation is constrained to reproduce the original image mean gray level, and the mean basis function variance is determined using the expected image smoothness for the increased resolution...
Efficient charge assignment and back interpolation in multigrid methods for molecular dynamics.
Banerjee, Sanjay; Board, John A
2005-07-15
The assignment of atomic charges to a regular computational grid and the interpolation of forces from the grid back to the original atomic positions are crucial steps in a multigrid approach to the calculation of molecular forces. For purposes of grid assignment, atomic charges are modeled as truncated Gaussian distributions. The charge assignment and back interpolation methods are currently bottlenecks, and take up to one-third the execution time of the multigrid method each. Here, we propose alternative approaches to both charge assignment and back interpolation where convolution is used both to map Gaussian representations of atomic charges onto the grid and to map the forces computed at grid points back to atomic positions. These approaches achieve the same force accuracy with reduced run time. The proposed charge assignment and back interpolation methods scale better than baseline multigrid computations with both problem size and number of processors. (c) 2005 Wiley Periodicals, Inc.
Directory of Open Access Journals (Sweden)
Huaiqing Zhang
2014-01-01
Full Text Available The spectral leakage has a harmful effect on the accuracy of harmonic analysis for asynchronous sampling. This paper proposed a time quasi-synchronous sampling algorithm which is based on radial basis function (RBF interpolation. Firstly, a fundamental period is evaluated by a zero-crossing technique with fourth-order Newton’s interpolation, and then, the sampling sequence is reproduced by the RBF interpolation. Finally, the harmonic parameters can be calculated by FFT on the synchronization of sampling data. Simulation results showed that the proposed algorithm has high accuracy in measuring distorted and noisy signals. Compared to the local approximation schemes as linear, quadric, and fourth-order Newton interpolations, the RBF is a global approximation method which can acquire more accurate results while the time-consuming is about the same as Newton’s.
NOAA Optimum Interpolation 1/4 Degree Daily Sea Surface Temperature (OISST) Analysis, Version 2
National Oceanic and Atmospheric Administration, Department of Commerce — This high-resolution sea surface temperature (SST) analysis product was developed using an optimum interpolation (OI) technique. The SST analysis has a spatial grid...
Building Input Adaptive Parallel Applications: A Case Study of Sparse Grid Interpolation
Murarasu, Alin
2012-12-01
The well-known power wall resulting in multi-cores requires special techniques for speeding up applications. In this sense, parallelization plays a crucial role. Besides standard serial optimizations, techniques such as input specialization can also bring a substantial contribution to the speedup. By identifying common patterns in the input data, we propose new algorithms for sparse grid interpolation that accelerate the state-of-the-art non-specialized version. Sparse grid interpolation is an inherently hierarchical method of interpolation employed for example in computational steering applications for decompressing highdimensional simulation data. In this context, improving the speedup is essential for real-time visualization. Using input specialization, we report a speedup of up to 9x over the nonspecialized version. The paper covers the steps we took to reach this speedup by means of input adaptivity. Our algorithms will be integrated in fastsg, a library for fast sparse grid interpolation. © 2012 IEEE.
An application of gain-scheduled control using state-space interpolation to hydroactive gas bearings
DEFF Research Database (Denmark)
Theisen, Lukas Roy Svane; Camino, Juan F.; Niemann, Hans Henrik
2016-01-01
, it is possible to design a gain-scheduled controller using multiple controllers optimised for a single frequency. Gain-scheduling strategies using the Youla parametrisation can guarantee stability at the cost of increased controller order and performance loss in the interpolation region. This paper contributes...... with a gain-scheduling strategy using state-space interpolation, which avoids both the performance loss and the increase of controller order associated to the Youla parametrisation. The proposed state-space interpolation for gain-scheduling is applied for mass imbalance rejection for a controllable gas...... bearing scheduled in two parameters. Comparisons against the Youla-based scheduling demonstrate the superiority of the state-space interpolation....
Chen, Shyi-Ming; Hsin, Wen-Chyuan
2015-07-01
In this paper, we propose a new weighted fuzzy interpolative reasoning method for sparse fuzzy rule-based systems based on the slopes of fuzzy sets. We also propose a particle swarm optimization (PSO)-based weights-learning algorithm to automatically learn the optimal weights of the antecedent variables of fuzzy rules for weighted fuzzy interpolative reasoning. We apply the proposed weighted fuzzy interpolative reasoning method using the proposed PSO-based weights-learning algorithm to deal with the computer activity prediction problem, the multivariate regression problems, and the time series prediction problems. The experimental results show that the proposed weighted fuzzy interpolative reasoning method using the proposed PSO-based weights-learning algorithm outperforms the existing methods for dealing with the computer activity prediction problem, the multivariate regression problems, and the time series prediction problems.
National Research Council Canada - National Science Library
Ingel, R
1999-01-01
.... Projection operators are employed for the model reduction or condensation process. Interpolation is then introduced over a user defined frequency window, which can have real and imaginary boundaries and be quite large. Hermitian...
National Oceanic and Atmospheric Administration, Department of Commerce — This feature dataset contains the control points used to validate the accuracies of the interpolated water density rasters for the Gulf of Maine. These control...
Interpolation method for the transport theory and its application in fusion-neutronics analysis
International Nuclear Information System (INIS)
Jung, J.
1981-09-01
This report presents an interpolation method for the solution of the Boltzmann transport equation. The method is based on a flux synthesis technique using two reference-point solutions. The equation for the interpolated solution results in a Volterra integral equation which is proved to have a unique solution. As an application of the present method, tritium breeding ratio is calculated for a typical D-T fusion reactor system. The result is compared to that of a variational technique
Reinhardt, Katja; Samimi, Cyrus
2018-01-01
While climatological data of high spatial resolution are largely available in most developed countries, the network of climatological stations in many other regions of the world still constitutes large gaps. Especially for those regions, interpolation methods are important tools to fill these gaps and to improve the data base indispensible for climatological research. Over the last years, new hybrid methods of machine learning and geostatistics have been developed which provide innovative prospects in spatial predictive modelling. This study will focus on evaluating the performance of 12 different interpolation methods for the wind components \\overrightarrow{u} and \\overrightarrow{v} in a mountainous region of Central Asia. Thereby, a special focus will be on applying new hybrid methods on spatial interpolation of wind data. This study is the first evaluating and comparing the performance of several of these hybrid methods. The overall aim of this study is to determine whether an optimal interpolation method exists, which can equally be applied for all pressure levels, or whether different interpolation methods have to be used for the different pressure levels. Deterministic (inverse distance weighting) and geostatistical interpolation methods (ordinary kriging) were explored, which take into account only the initial values of \\overrightarrow{u} and \\overrightarrow{v} . In addition, more complex methods (generalized additive model, support vector machine and neural networks as single methods and as hybrid methods as well as regression-kriging) that consider additional variables were applied. The analysis of the error indices revealed that regression-kriging provided the most accurate interpolation results for both wind components and all pressure heights. At 200 and 500 hPa, regression-kriging is followed by the different kinds of neural networks and support vector machines and for 850 hPa it is followed by the different types of support vector machine and
Yuval, Yuval; Rimon, Yaara; Graber, Ellen R; Furman, Alex
2014-08-01
A large fraction of the fresh water available for human use is stored in groundwater aquifers. Since human activities such as mining, agriculture, industry and urbanisation often result in incursion of various pollutants to groundwater, routine monitoring of water quality is an indispensable component of judicious aquifer management. Unfortunately, groundwater pollution monitoring is expensive and usually cannot cover an aquifer with the spatial resolution necessary for making adequate management decisions. Interpolation of monitoring data is thus an important tool for supplementing monitoring observations. However, interpolating routine groundwater pollution data poses a special problem due to the nature of the observations. The data from a producing aquifer usually includes many zero pollution concentration values from the clean parts of the aquifer but may span a wide range of values (up to a few orders of magnitude) in the polluted areas. This manuscript presents a methodology that can cope with such datasets and use them to produce maps that present the pollution plumes but also delineates the clean areas that are fit for production. A method for assessing the quality of mapping in a way which is suitable to the data's dynamic range of values is also presented. A local variant of inverse distance weighting is employed to interpolate the data. Inclusion zones around the interpolation points ensure that only relevant observations contribute to each interpolated concentration. Using inclusion zones improves the accuracy of the mapping but results in interpolation grid points which are not assigned a value. The inherent trade-off between the interpolation accuracy and coverage is demonstrated using both circular and elliptical inclusion zones. A leave-one-out cross testing is used to assess and compare the performance of the interpolations. The methodology is demonstrated using groundwater pollution monitoring data from the coastal aquifer along the Israeli
Optimum quantization and interpolation of projections in X-ray computerized tomography
International Nuclear Information System (INIS)
Vajnberg, Eh.I.; Fajngojz, M.L.
1984-01-01
Two methods to increase the accuracy of image reconstruction due to optimization of quantization and interpolation of proections with separate reduction of the main types of errors are described and experimentally studied. A high metrological and calculation efficiency of increasing the count frequency in the reconstructed tomogram 2-4 times is found. The optimum structure of interpolation functions of a minimum extent is calculated
Interpolation of the discrete logarithm in a finite field of characteristic two by Boolean functions
DEFF Research Database (Denmark)
Brandstaetter, Nina; Lange, Tanja; Winterhof, Arne
2005-01-01
We obtain bounds on degree, weight, and the maximal Fourier coefficient of Boolean functions interpolating the discrete logarithm in finite fields of characteristic two. These bounds complement earlier results for finite fields of odd characteristic.......We obtain bounds on degree, weight, and the maximal Fourier coefficient of Boolean functions interpolating the discrete logarithm in finite fields of characteristic two. These bounds complement earlier results for finite fields of odd characteristic....
DWITAMA ILYAS, RAMDAN
2016-01-01
2016 RAMDAN DWITAMA ILYAS B111 12 007, Peranan INTERPOL Dalam Ekstradisi Tersangka Korupsi (Studi Kasus Penangkapan Muhammad Nazaruddin di Cartagena, Colombia). Di bombing oleh, S.M. Noor dan Maskun. Tujuan dari penelitian ini adalah ; (1) untuk mengetahui bagaimana kedudukan deportasi dan ekstradisi dalam proses pemulangan Muhammad Nazaruddin sesuai UNCAC. (2) untuk mengetahui peran Interpol dalam proses pemulangan Muhammad Nazaruddin. Metode penelitian yang diguna...
Directory of Open Access Journals (Sweden)
Mauricio Castro Franco
2017-07-01
Full Text Available Context: Interpolating soil properties at field-scale in the Colombian piedmont eastern plains is challenging due to: the highly and complex variable nature of some processes; the effects of the soil; the land use; and the management. While interpolation techniques are being adapted to include auxiliary information of these effects, the soil data are often difficult to predict using conventional techniques of spatial interpolation. Method: In this paper, we evaluated and compared six spatial interpolation techniques: Inverse Distance Weighting (IDW, Spline, Ordinary Kriging (KO, Universal Kriging (UK, Cokriging (Ckg, and Residual Maximum Likelihood-Empirical Best Linear Unbiased Predictor (REML-EBLUP, from conditioned Latin Hypercube as a sampling strategy. The ancillary information used in Ckg and REML-EBLUP was indexes calculated from a digital elevation model (MDE. The “Random forest” algorithm was used for selecting the most important terrain index for each soil properties. Error metrics were used to validate interpolations against cross validation. Results: The results support the underlying assumption that HCLc captured adequately the full distribution of variables of ancillary information in the Colombian piedmont eastern plains conditions. They also suggest that Ckg and REML-EBLUP perform best in the prediction in most of the evaluated soil properties. Conclusions: Mixed interpolation techniques having auxiliary soil information and terrain indexes, provided a significant improvement in the prediction of soil properties, in comparison with other techniques.
Directory of Open Access Journals (Sweden)
Lei Shi
2016-09-01
Full Text Available Tidal datums are key components in NOAA’s Vertical Datum transformation project (VDatum. In this paper, we propose a statistical interpolation method, derived from the variational principle, to calculate tidal datums by blending the modeled and the observed tidal datums. Through the implementation of this statistical interpolation method in the Chesapeake and Delaware Bays, we conclude that the statistical interpolation method for tidal datums has great advantages over the currently used deterministic interpolation method. The foremost, and inherent, advantage of the statistical interpolation is its capability to integrate data from different sources and with different accuracies without concern for their relative spatial locations. The second advantage is that it provides a spatially varying uncertainty for the entire domain in which data is being integrated. The latter is especially helpful for the decision-making process of where new instruments would be most effectively placed. Lastly, the test case results show that the statistical interpolation reduced the bias, maximum absolute error, mean absolute error, and root mean square error in comparison to the current deterministic approach.
Zheng, Jingjing; Frisch, Michael J
2017-12-12
An efficient geometry optimization algorithm based on interpolated potential energy surfaces with iteratively updated Hessians is presented in this work. At each step of geometry optimization (including both minimization and transition structure search), an interpolated potential energy surface is properly constructed by using the previously calculated information (energies, gradients, and Hessians/updated Hessians), and Hessians of the two latest geometries are updated in an iterative manner. The optimized minimum or transition structure on the interpolated surface is used for the starting geometry of the next geometry optimization step. The cost of searching the minimum or transition structure on the interpolated surface and iteratively updating Hessians is usually negligible compared with most electronic structure single gradient calculations. These interpolated potential energy surfaces are often better representations of the true potential energy surface in a broader range than a local quadratic approximation that is usually used in most geometry optimization algorithms. Tests on a series of large and floppy molecules and transition structures both in gas phase and in solutions show that the new algorithm can significantly improve the optimization efficiency by using the iteratively updated Hessians and optimizations on interpolated surfaces.
Mist'e, Gianluigi Alberto; Benini, Ernesto
2012-06-01
Compressor map interpolation is usually performed through the introduction of auxiliary coordinates (β). In this paper, a new analytical bivariate β function definition to be used in compressor map interpolation is studied. The function has user-defined parameters that must be adjusted to properly fit to a single map. The analytical nature of β allows for rapid calculations of the interpolation error estimation, which can be used as a quantitative measure of interpolation accuracy and also as a valid tool to compare traditional β function interpolation with new approaches (artificial neural networks, genetic algorithms, etc.). The quality of the method is analyzed by comparing the error output to the one of a well-known state-of-the-art methodology. This comparison is carried out for two different types of compressor and, in both cases, the error output using the method presented in this paper is found to be consistently lower. Moreover, an optimization routine able to locally minimize the interpolation error by shape variation of the β function is implemented. Further optimization introducing other important criteria is discussed.
Directory of Open Access Journals (Sweden)
Peilu Liu
2017-10-01
Full Text Available In order to improve the accuracy of ultrasonic phased array focusing time delay, analyzing the original interpolation Cascade-Integrator-Comb (CIC filter, an 8× interpolation CIC filter parallel algorithm was proposed, so that interpolation and multichannel decomposition can simultaneously process. Moreover, we summarized the general formula of arbitrary multiple interpolation CIC filter parallel algorithm and established an ultrasonic phased array focusing time delay system based on 8× interpolation CIC filter parallel algorithm. Improving the algorithmic structure, 12.5% of addition and 29.2% of multiplication was reduced, meanwhile the speed of computation is still very fast. Considering the existing problems of the CIC filter, we compensated the CIC filter; the compensated CIC filter’s pass band is flatter, the transition band becomes steep, and the stop band attenuation increases. Finally, we verified the feasibility of this algorithm on Field Programming Gate Array (FPGA. In the case of system clock is 125 MHz, after 8× interpolation filtering and decomposition, time delay accuracy of the defect echo becomes 1 ns. Simulation and experimental results both show that the algorithm we proposed has strong feasibility. Because of the fast calculation, small computational amount and high resolution, this algorithm is especially suitable for applications with high time delay accuracy and fast detection.
Performance of an Interpolated Stochastic Weather Generator in Czechia and Nebraska
Dubrovsky, M.; Trnka, M.; Hayes, M. J.; Svoboda, M. D.; Semeradova, D.; Metelka, L.; Hlavinka, P.
2008-12-01
Met&Roll is a WGEN-like parametric four-variate daily weather generator (WG), with an optional extension allowing the user to generate additional variables (i.e. wind and water vapor pressure). It is designed to produce synthetic weather series representing present and/or future climate conditions to be used as an input into various models (e.g. crop growth and rainfall runoff models). The present contribution will summarize recent experiments, in which we tested the performance of the interpolated WG, with the aim to examine whether the WG may be used to produce synthetic weather series even for sites having no meteorological observations. The experiments being discussed include: (1) the comparison of various interpolation methods where the performance of the candidate methods is compared in terms of the accuracy of the interpolation for selected WG parameters; (2) assessing the ability of the interpolated WG in the territories of Czechia and Nebraska to reproduce extreme temperature and precipitation characteristics; (3) indirect validation of the interpolated WG in terms of the modeled crop yields simulated by STICS crop growth model (in Czechia); and (4) indirect validation of interpolated WG in terms of soil climate regime characteristics simulated by the SoilClim model (Czechia and Nebraska). The experiments are based on observed daily weather series from two regions: Czechia (area = 78864 km2, 125 stations available) and Nebraska (area = 200520 km2, 28 stations available). Even though Nebraska exhibits a much lower density of stations, this is offset by the state's relatively flat topography, which is an advantage in using the interpolated WG. Acknowledgements: The present study is supported by the AMVIS-KONTAKT project (ME 844) and the GAAV Grant Agency (project IAA300420806).
Gandevia, S C; McNeil, C J; Carroll, T J; Taylor, J L
2013-03-01
The assessment of voluntary activation of human muscles usually depends on measurement of the size of the twitch produced by an interpolated nerve or cortical stimulus. In many forms of fatiguing exercise the superimposed twitch increases and thus voluntary activation appears to decline. This is termed 'central' fatigue. Recent studies on isolated mouse muscle suggest that a peripheral mechanism related to intracellular calcium sensitivity increases interpolated twitches. To test whether this problem developed with human voluntary contractions we delivered maximal tetanic stimulation to the ulnar nerve (≥60 s at physiological motoneuronal frequencies, 30 and 15 Hz). During the tetani (at 30 Hz) in which the force declined by 42%, the absolute size of the twitches evoked by interpolated stimuli (delivered regularly or only in the last second of the tetanus) diminished progressively to less than 1%. With stimulation at 30 Hz, there was also a marked reduction in size and area of the interpolated compound muscle action potential (M wave). With a 15 Hz tetanus, a progressive decline in the interpolated twitch force also occurred (to ∼10%) but did so before the area of the interpolated M wave diminished. These results indicate that the increase in interpolated twitch size predicted from the mouse studies does not occur. Diminution in superimposed twitches occurred whether or not the M wave indicated marked impairment at sarcolemmal/t-tubular levels. Consequently, the increase in superimposed twitch, which is used to denote central fatigue in human fatiguing exercise, is likely to reflect low volitional drive to high-threshold motor units, which stop firing or are discharging at low frequencies.
On the Quality of Velocity Interpolation Schemes for Marker-in-Cell Method and Staggered Grids
Pusok, Adina E.; Kaus, Boris J. P.; Popov, Anton A.
2017-03-01
The marker-in-cell method is generally considered a flexible and robust method to model the advection of heterogenous non-diffusive properties (i.e., rock type or composition) in geodynamic problems. In this method, Lagrangian points carrying compositional information are advected with the ambient velocity field on an Eulerian grid. However, velocity interpolation from grid points to marker locations is often performed without considering the divergence of the velocity field at the interpolated locations (i.e., non-conservative). Such interpolation schemes can induce non-physical clustering of markers when strong velocity gradients are present (Journal of Computational Physics 166:218-252, 2001) and this may, eventually, result in empty grid cells, a serious numerical violation of the marker-in-cell method. To remedy this at low computational costs, Jenny et al. (Journal of Computational Physics 166:218-252, 2001) and Meyer and Jenny (Proceedings in Applied Mathematics and Mechanics 4:466-467, 2004) proposed a simple, conservative velocity interpolation scheme for 2-D staggered grid, while Wang et al. (Geochemistry, Geophysics, Geosystems 16(6):2015-2023, 2015) extended the formulation to 3-D finite element methods. Here, we adapt this formulation for 3-D staggered grids (correction interpolation) and we report on the quality of various velocity interpolation methods for 2-D and 3-D staggered grids. We test the interpolation schemes in combination with different advection schemes on incompressible Stokes problems with strong velocity gradients, which are discretized using a finite difference method. Our results suggest that a conservative formulation reduces the dispersion and clustering of markers, minimizing the need of unphysical marker control in geodynamic models.
Bagheri, H.; Sadjadi, S. Y.; Sadeghian, S.
2013-09-01
One of the most significant tools to study many engineering projects is three-dimensional modelling of the Earth that has many applications in the Geospatial Information System (GIS), e.g. creating Digital Train Modelling (DTM). DTM has numerous applications in the fields of sciences, engineering, design and various project administrations. One of the most significant events in DTM technique is the interpolation of elevation to create a continuous surface. There are several methods for interpolation, which have shown many results due to the environmental conditions and input data. The usual methods of interpolation used in this study along with Genetic Algorithms (GA) have been optimised and consisting of polynomials and the Inverse Distance Weighting (IDW) method. In this paper, the Artificial Intelligent (AI) techniques such as GA and Neural Networks (NN) are used on the samples to optimise the interpolation methods and production of Digital Elevation Model (DEM). The aim of entire interpolation methods is to evaluate the accuracy of interpolation methods. Universal interpolation occurs in the entire neighbouring regions can be suggested for larger regions, which can be divided into smaller regions. The results obtained from applying GA and ANN individually, will be compared with the typical method of interpolation for creation of elevations. The resulting had performed that AI methods have a high potential in the interpolation of elevations. Using artificial networks algorithms for the interpolation and optimisation based on the IDW method with GA could be estimated the high precise elevations.
Hittmeir, Sabine; Philipp, Anne; Seibert, Petra
2017-04-01
In discretised form, an extensive variable usually represents an integral over a 3-dimensional (x,y,z) grid cell. In the case of vertical fluxes, gridded values represent integrals over a horizontal (x,y) grid face. In meteorological models, fluxes (precipitation, turbulent fluxes, etc.) are usually written out as temporally integrated values, thus effectively forming 3D (x,y,t) integrals. Lagrangian transport models require interpolation of all relevant variables towards the location in 4D space of each of the computational particles. Trivial interpolation algorithms usually implicitly assume the integral value to be a point value valid at the grid centre. If the integral value would be reconstructed from the interpolated point values, it would in general not be correct. If nonlinear interpolation methods are used, non-negativity cannot easily be ensured. This problem became obvious with respect to the interpolation of precipitation for the calculation of wet deposition FLEXPART (http://flexpart.eu) which uses ECMWF model output or other gridded input data. The presently implemented method consists of a special preprocessing in the input preparation software and subsequent linear interpolation in the model. The interpolated values are positive but the criterion of cell-wise conservation of the integral property is violated; it is also not very accurate as it smoothes the field. A new interpolation algorithm was developed which introduces additional supporting grid points in each time interval with linear interpolation to be applied in FLEXPART later between them. It preserves the integral precipitation in each time interval, guarantees the continuity of the time series, and maintains non-negativity. The function values of the remapping algorithm at these subgrid points constitute the degrees of freedom which can be prescribed in various ways. Combining the advantages of different approaches leads to a final algorithm respecting all the required conditions. To
DrawFromDrawings: 2D Drawing Assistance via Stroke Interpolation with a Sketch Database.
Matsui, Yusuke; Shiratori, Takaaki; Aizawa, Kiyoharu
2017-07-01
We present DrawFromDrawings, an interactive drawing system that provides users with visual feedback for assistance in 2D drawing using a database of sketch images. Following the traditional imitation and emulation training from art education, DrawFromDrawings enables users to retrieve and refer to a sketch image stored in a database and provides them with various novel strokes as suggestive or deformation feedback. Given regions of interest (ROIs) in the user and reference sketches, DrawFromDrawings detects as-long-as-possible (ALAP) stroke segments and the correspondences between user and reference sketches that are the key to computing seamless interpolations. The stroke-level interpolations are parametrized with the user strokes, the reference strokes, and new strokes created by warping the reference strokes based on the user and reference ROI shapes, and the user study indicated that the interpolation could produce various reasonable strokes varying in shapes and complexity. DrawFromDrawings allows users to either replace their strokes with interpolated strokes (deformation feedback) or overlays interpolated strokes onto their strokes (suggestive feedback). The other user studies on the feedback modes indicated that the suggestive feedback enabled drawers to develop and render their ideas using their own stroke style, whereas the deformation feedback enabled them to finish the sketch composition quickly.
Improving GPU-accelerated adaptive IDW interpolation algorithm using fast kNN search.
Mei, Gang; Xu, Nengxiong; Xu, Liangliang
2016-01-01
This paper presents an efficient parallel Adaptive Inverse Distance Weighting (AIDW) interpolation algorithm on modern Graphics Processing Unit (GPU). The presented algorithm is an improvement of our previous GPU-accelerated AIDW algorithm by adopting fast k-nearest neighbors (kNN) search. In AIDW, it needs to find several nearest neighboring data points for each interpolated point to adaptively determine the power parameter; and then the desired prediction value of the interpolated point is obtained by weighted interpolating using the power parameter. In this work, we develop a fast kNN search approach based on the space-partitioning data structure, even grid, to improve the previous GPU-accelerated AIDW algorithm. The improved algorithm is composed of the stages of kNN search and weighted interpolating. To evaluate the performance of the improved algorithm, we perform five groups of experimental tests. The experimental results indicate: (1) the improved algorithm can achieve a speedup of up to 1017 over the corresponding serial algorithm; (2) the improved algorithm is at least two times faster than our previous GPU-accelerated AIDW algorithm; and (3) the utilization of fast kNN search can significantly improve the computational efficiency of the entire GPU-accelerated AIDW algorithm.
Directory of Open Access Journals (Sweden)
Longxiang Li
Full Text Available Effective assessments of air-pollution exposure depend on the ability to accurately predict pollutant concentrations at unmonitored locations, which can be achieved through spatial interpolation. However, most interpolation approaches currently in use are based on the Euclidean distance, which cannot account for the complex nonlinear features displayed by air-pollution distributions in the wind-field. In this study, an interpolation method based on the shortest path distance is developed to characterize the impact of complex urban wind-field on the distribution of the particulate matter concentration. In this method, the wind-field is incorporated by first interpolating the observed wind-field from a meteorological-station network, then using this continuous wind-field to construct a cost surface based on Gaussian dispersion model and calculating the shortest wind-field path distances between locations, and finally replacing the Euclidean distances typically used in Inverse Distance Weighting (IDW with the shortest wind-field path distances. This proposed methodology is used to generate daily and hourly estimation surfaces for the particulate matter concentration in the urban area of Beijing in May 2013. This study demonstrates that wind-fields can be incorporated into an interpolation framework using the shortest wind-field path distance, which leads to a remarkable improvement in both the prediction accuracy and the visual reproduction of the wind-flow effect, both of which are of great importance for the assessment of the effects of pollutants on human health.
Directory of Open Access Journals (Sweden)
Tao Chen
2017-05-01
Full Text Available The spatial distribution of precipitation is an important aspect of water-related research. The use of different interpolation schemes in the same catchment may cause large differences and deviations from the actual spatial distribution of rainfall. Our study analyzes different methods of spatial rainfall interpolation at annual, daily, and hourly time scales to provide a comprehensive evaluation. An improved regression-based scheme is proposed using principal component regression with residual correction (PCRR and is compared with inverse distance weighting (IDW and multiple linear regression (MLR interpolation methods. In this study, the meso-scale catchment of the Fuhe River in southeastern China was selected as a typical region. Furthermore, a hydrological model HEC-HMS was used to calculate streamflow and to evaluate the impact of rainfall interpolation methods on the results of the hydrological model. Results show that the PCRR method performed better than the other methods tested in the study and can effectively eliminate the interpolation anomalies caused by terrain differences between observation points and surrounding areas. Simulated streamflow showed different characteristics based on the mean, maximum, minimum, and peak flows. The results simulated by PCRR exhibited the lowest streamflow error and highest correlation with measured values at the daily time scale. The application of the PCRR method is found to be promising because it considers multicollinearity among variables.
The effect of interpolation methods in temperature and salinity trends in the Western Mediterranean
Directory of Open Access Journals (Sweden)
M. VARGAS-YANEZ
2012-04-01
Full Text Available Temperature and salinity data in the historical record are scarce and unevenly distributed in space and time and the estimation of linear trends is sensitive to different factors. In the case of the Western Mediterranean, previous works have studied the sensitivity of these trends to the use of bathythermograph data, the averaging methods or the way in which gaps in time series are dealt with. In this work, a new factor is analysed: the effect of data interpolation. Temperature and salinity time series are generated averaging existing data over certain geographical areas and also by means of interpolation. Linear trends from both types of time series are compared. There are some differences between both estimations for some layers and geographical areas, while in other cases the results are consistent. Those results which do not depend on the use of interpolated or non-interpolated data, neither are influenced by data analysis methods can be considered as robust ones. Those results influenced by the interpolation process or the factors analysed in previous sensitivity tests are not considered as robust results.
On analysis-based two-step interpolation methods for randomly sampled seismic data
Yang, Pengliang; Gao, Jinghuai; Chen, Wenchao
2013-02-01
Interpolating the missing traces of regularly or irregularly sampled seismic record is an exceedingly important issue in the geophysical community. Many modern acquisition and reconstruction methods are designed to exploit the transform domain sparsity of the few randomly recorded but informative seismic data using thresholding techniques. In this paper, to regularize randomly sampled seismic data, we introduce two accelerated, analysis-based two-step interpolation algorithms, the analysis-based FISTA (fast iterative shrinkage-thresholding algorithm) and the FPOCS (fast projection onto convex sets) algorithm from the IST (iterative shrinkage-thresholding) algorithm and the POCS (projection onto convex sets) algorithm. A MATLAB package is developed for the implementation of these thresholding-related interpolation methods. Based on this package, we compare the reconstruction performance of these algorithms, using synthetic and real seismic data. Combined with several thresholding strategies, the accelerated convergence of the proposed methods is also highlighted.
Cubic spline interpolation of functions with high gradients in boundary layers
Blatov, I. A.; Zadorin, A. I.; Kitaeva, E. V.
2017-01-01
The cubic spline interpolation of grid functions with high-gradient regions is considered. Uniform meshes are proved to be inefficient for this purpose. In the case of widely applied piecewise uniform Shishkin meshes, asymptotically sharp two-sided error estimates are obtained in the class of functions with an exponential boundary layer. It is proved that the error estimates of traditional spline interpolation are not uniform with respect to a small parameter, and the error can increase indefinitely as the small parameter tends to zero, while the number of nodes N is fixed. A modified cubic interpolation spline is proposed, for which O((ln N/N)4) error estimates that are uniform with respect to the small parameter are obtained.
Seismic Experiment at North Arizona To Locate Washington Fault - 3D Data Interpolation
Hanafy, Sherif M.
2008-10-01
The recorded data is interpolated using sinc technique to create the following two data sets 1. Data Set # 1: Here, we interpolated only in the receiver direction to regularize the receiver interval to 1 m, however, the source locations are the same as the original data (2 and 4 m source intervals). Now the data contains 6 lines, each line has 121 receivers and a total of 240 shot gathers. 2. Data Set # 2: Here, we used the result from the previous step, and interpolated only in the shot direction to regularize the shot interval to 1 m. Now, both shot and receivers has 1 m interval. The data contains 6 lines, each line has 121 receivers and a total of 726 shot gathers.
Interpolation of Missing Precipitation Data Using Kernel Estimations for Hydrologic Modeling
Directory of Open Access Journals (Sweden)
Hyojin Lee
2015-01-01
Full Text Available Precipitation is the main factor that drives hydrologic modeling; therefore, missing precipitation data can cause malfunctions in hydrologic modeling. Although interpolation of missing precipitation data is recognized as an important research topic, only a few methods follow a regression approach. In this study, daily precipitation data were interpolated using five different kernel functions, namely, Epanechnikov, Quartic, Triweight, Tricube, and Cosine, to estimate missing precipitation data. This study also presents an assessment that compares estimation of missing precipitation data through Kth nearest neighborhood (KNN regression to the five different kernel estimations and their performance in simulating streamflow using the Soil Water Assessment Tool (SWAT hydrologic model. The results show that the kernel approaches provide higher quality interpolation of precipitation data compared with the KNN regression approach, in terms of both statistical data assessment and hydrologic modeling performance.
Gorji, Taha; Sertel, Elif; Tanik, Aysegul
2017-12-01
Soil management is an essential concern in protecting soil properties, in enhancing appropriate soil quality for plant growth and agricultural productivity, and in preventing soil erosion. Soil scientists and decision makers require accurate and well-distributed spatially continuous soil data across a region for risk assessment and for effectively monitoring and managing soils. Recently, spatial interpolation approaches have been utilized in various disciplines including soil sciences for analysing, predicting and mapping distribution and surface modelling of environmental factors such as soil properties. The study area selected in this research is Tuz Lake Basin in Turkey bearing ecological and economic importance. Fertile soil plays a significant role in agricultural activities, which is one of the main industries having great impact on economy of the region. Loss of trees and bushes due to intense agricultural activities in some parts of the basin lead to soil erosion. Besides, soil salinization due to both human-induced activities and natural factors has exacerbated its condition regarding agricultural land development. This study aims to compare capability of Local Polynomial Interpolation (LPI) and Radial Basis Functions (RBF) as two interpolation methods for mapping spatial pattern of soil properties including organic matter, phosphorus, lime and boron. Both LPI and RBF methods demonstrated promising results for predicting lime, organic matter, phosphorous and boron. Soil samples collected in the field were used for interpolation analysis in which approximately 80% of data was used for interpolation modelling whereas the remaining for validation of the predicted results. Relationship between validation points and their corresponding estimated values in the same location is examined by conducting linear regression analysis. Eight prediction maps generated from two different interpolation methods for soil organic matter, phosphorus, lime and boron parameters
Wavelet-Smoothed Interpolation of Masked Scientific Data for JPEG 2000 Compression
Energy Technology Data Exchange (ETDEWEB)
Brislawn, Christopher M. [Los Alamos National Laboratory
2012-08-13
How should we manage scientific data with 'holes'? Some applications, like JPEG 2000, expect logically rectangular data, but some sources, like the Parallel Ocean Program (POP), generate data that isn't defined on certain subsets. We refer to grid points that lack well-defined, scientifically meaningful sample values as 'masked' samples. Wavelet-smoothing is a highly scalable interpolation scheme for regions with complex boundaries on logically rectangular grids. Computation is based on forward/inverse discrete wavelet transforms, so runtime complexity and memory scale linearly with respect to sample count. Efficient state-of-the-art minimal realizations yield small constants (O(10)) for arithmetic complexity scaling, and in-situ implementation techniques make optimal use of memory. Implementation in two dimensions using tensor product filter banks is straighsorward and should generalize routinely to higher dimensions. No hand-tuning required when the interpolation mask changes, making the method aeractive for problems with time-varying masks. Well-suited for interpolating undefined samples prior to JPEG 2000 encoding. The method outperforms global mean interpolation, as judged by both SNR rate-distortion performance and low-rate artifact mitigation, for data distributions whose histograms do not take the form of sharply peaked, symmetric, unimodal probability density functions. These performance advantages can hold even for data whose distribution differs only moderately from the peaked unimodal case, as demonstrated by POP salinity data. The interpolation method is very general and is not tied to any particular class of applications, could be used for more generic smooth interpolation.
A study of interpolation method in diagnosis of carpal tunnel syndrome
Directory of Open Access Journals (Sweden)
Alireza Ashraf
2013-01-01
Full Text Available Context: The low correlation between the patients′ signs and symptoms of carpal tunnel syndrome (CTS and results of electrodiagnostic tests makes the diagnosis challenging in mild cases. Interpolation is a mathematical method for finding median nerve conduction velocity (NCV exactly at carpal tunnel site. Therefore, it may be helpful in diagnosis of CTS in patients with equivocal test results. Aim: The aim of this study is to evaluate interpolation method as a CTS diagnostic test. Settings and Design: Patients with two or more clinical symptoms and signs of CTS in a median nerve territory with 3.5 ms ≤ distal median sensory latency <4.6 ms from those who came to our electrodiagnostic clinics and also, age matched healthy control subjects were recruited in the study. Materials and Methods: Median compound motor action potential and median sensory nerve action potential latencies were measured by a MEDLEC SYNERGY VIASIS electromyography and conduction velocities were calculated by both routine method and interpolation technique. Statistical Analysis Used: Chi-square and Student′s t-test were used for comparing group differences. Cut-off points were calculated using receiver operating characteristic curve. Results: A sensitivity of 88%, specificity of 67%, positive predictive value (PPV and negative predictive value (NPV of 70.8% and 84.7% were obtained for median motor NCV and a sensitivity of 98.3%, specificity of 91.7%, PPV and NPV of 91.9% and 98.2% were obtained for median sensory NCV with interpolation technique. Conclusions: Median motor interpolation method is a good technique, but it has less sensitivity and specificity than median sensory interpolation method.
Turning Avatar into Realistic Human Expression Using Linear and Bilinear Interpolations
Hazim Alkawaz, Mohammed; Mohamad, Dzulkifli; Rehman, Amjad; Basori, Ahmad Hoirul
2014-06-01
The facial animation in term of 3D facial data has accurate research support of the laser scan and advance 3D tools for complex facial model production. However, the approach still lacks facial expression based on emotional condition. Though, facial skin colour is required to offers an effect of facial expression improvement, closely related to the human emotion. This paper presents innovative techniques for facial animation transformation using the facial skin colour based on linear interpolation and bilinear interpolation. The generated expressions are almost same to the genuine human expression and also enhance the facial expression of the virtual human.
Workload Balancing on Heterogeneous Systems: A Case Study of Sparse Grid Interpolation
Muraraşu, Alin
2012-01-01
Multi-core parallelism and accelerators are becoming common features of today’s computer systems, as they allow for computational power without sacrificing energy efficiency. Due to heterogeneity, tuning for each type of compute unit and adequate load balancing is essential. This paper proposes static and dynamic solutions for load balancing in the context of an application for visualizing high-dimensional simulation data. The application relies on the sparse grid technique for data compression. Its performance critical part is the interpolation routine used for decompression. Results show that our load balancing scheme allows for an efficient acceleration of interpolation on heterogeneous systems containing multi-core CPUs and GPUs.
Modified control grid interpolation for the volumetric reconstruction of fluid flows.
Frakes, David H; Pekkan, Kerem; Dasi, Lakshmi P; Kitajima, Hiroumi D; de Zelicourt, Diane; Leo, Hwa Liang; Carberry, Josie; Sundareswaran, Kartik; Simon, Helene; Yoganathan, Ajit P
2008-12-01
Complex applications in fluid dynamics research often require more highly resolved velocity data than direct measurements or simulations provide. The advent of stereo PIV and PCMR techniques has advanced the state-of-the-art in flow velocity measurement, but 3D spatial resolution remains limited. Here a new technique is proposed for velocity data interpolation to address this problem. The new method performs with higher quality than competing solutions from the literature in terms of accurately interpolating velocities, maintaining fluid structure and domain boundaries, and preserving coherent structures.
Linear, Transﬁnite and Weighted Method for Interpolation from Grid Lines Applied to OCT Images
DEFF Research Database (Denmark)
Lindberg, Anne-Sofie Wessel; Jørgensen, Thomas Martini; Dahl, Vedrana Andersen
2018-01-01
When performing a line scan using optical coherence tomography (OCT), the distance between the successive scan lines is often large compared to the resolution along each scan line. If two sets of such line scans are acquired orthogonal to each other, intensity values are known along the lines...... of a square grid, but are unknown inside each square. To view these values as an image, intensities need to be interpolated at regularly spaced pixel positions. In this paper we evaluate three methods for interpolation from grid lines: linear, transfinite and weighted. The linear method does not preserve...... scans, acquired such that the lines of the second scan are orthogonal to the first....
Interpolation Method Needed for Numerical Uncertainty Analysis of Computational Fluid Dynamics
Groves, Curtis; Ilie, Marcel; Schallhorn, Paul
2014-01-01
Using Computational Fluid Dynamics (CFD) to predict a flow field is an approximation to the exact problem and uncertainties exist. There is a method to approximate the errors in CFD via Richardson's Extrapolation. This method is based off of progressive grid refinement. To estimate the errors in an unstructured grid, the analyst must interpolate between at least three grids. This paper describes a study to find an appropriate interpolation scheme that can be used in Richardson's extrapolation or other uncertainty method to approximate errors. Nomenclature
Interpolation of meteorological data by kriging method for use in forestry
Directory of Open Access Journals (Sweden)
Ivetić Vladan
2010-01-01
Full Text Available Interpolation is a suitable method of computing the values of a spatial variable at the location which is impossible for measurement, based on the data obtained by the measurement of the same variable at the predetermined locations (e.g. weather stations. In this paper, temperature and rainfall values at 39 weather stations in Serbia and neighbouring countries were interpolated aiming at the research in forestry. The study results are presented in the form of an interactive map of Serbia, which allows a fast and simple determination of the analyzed variable at any point within its territory, which is presented by the example of 27 forest sites.
Angular interpolations and splice options for three-dimensional transport computations
International Nuclear Information System (INIS)
Abu-Shumays, I.K.; Yehnert, C.E.
1996-01-01
New, accurate and mathematically rigorous angular Interpolation strategies are presented. These strategies preserve flow and directionality separately over each octant of the unit sphere, and are based on a combination of spherical harmonics expansions and least squares algorithms. Details of a three-dimensional to three-dimensional (3-D to 3-D) splice method which utilizes the new angular interpolations are summarized. The method has been implemented in a multidimensional discrete ordinates transport computer program. Various features of the splice option are illustrated by several applications to a benchmark Dog-Legged Void Neutron (DLVN) streaming and transport experimental assembly
Convergence and norm estimates of Hermite interpolation at zeros of Chevyshev polynomials.
Al-Khaled, Kamel; Alquran, Marwan
2016-01-01
In this paper, we investigate the simultaneous approximation of a function f ( x ) and its derivative [Formula: see text] by Hermite interpolation operator [Formula: see text] based on Chevyshev polynomials. We also establish general theorem on extreme points for Hermite interpolation operator. Some results are considered to be an improvement over those obtained in Al-Khaled and Khalil (Numer Funct Anal Optim 21(5-6): 579-588, 2000), while others agrees with Pottinger's results (Pottinger in Z Agnew Math Mech 56: T310-T311, 1976).
The twitch interpolation technique for study of fatigue of human quadriceps muscle
DEFF Research Database (Denmark)
Bülow, P M; Nørregaard, J; Mehlsen, J
1995-01-01
The aim of the study was to examine if the twitch interpolation technique could be used to objectively measure fatigue in the quadriceps muscle in subjects performing submaximally. The 'true' maximum isometric quadriceps torque was determined in 21 healthy subject using the twitch interpolation...... technique. Then an endurance test was performed in which the subjects made repeated isometric contractions at 50% of the 'true' maximum torque for 4 s, separated by 6 s rest periods. During the test, the force response to single electrical stimulation (twitch amplitude) was measured at 50% and 25......). In conclusion, the twitch technique can be used for objectively measuring fatigue of the quadriceps muscle....
GA Based Rational cubic B-Spline Representation for Still Image Interpolation
Directory of Open Access Journals (Sweden)
Samreen Abbas
2016-12-01
Full Text Available In this paper, an image interpolation scheme is designed for 2D natural images. A local support rational cubic spline with control parameters, as interpolatory function, is being optimized using Genetic Algorithm (GA. GA is applied to determine the appropriate values of control parameter used in the description of rational cubic spline. Three state-of-the-art Image Quality Assessment (IQA models with traditional one are hired for comparison with existing image interpolation schemes and perceptual quality check of resulting images. The results show that the proposed scheme is better than the existing ones in comparison.
Alpay, Daniel; Dijksma, Aad; Langer, Heinz; Wanjala, Gerald
2006-01-01
We define and solve a boundary interpolation problem for generalized Schur functions s(z) on the open unit disk D which have preassigned asymptotics when z from D tends nontangentially to a boundary point z1 ∈ T. The solutions are characterized via a fractional linear parametrization formula. We
DEFF Research Database (Denmark)
Shekarchi, Sayedali; Christensen-Dalsgaard, Jakob; Hallam, John
2015-01-01
A head-related transfer function (HRTF) model employing Legendre polynomials (LPs) is evaluated as an HRTF spatial complexity indicator and interpolation technique in the azimuth plane. LPs are a set of orthogonal functions derived on the sphere which can be used to compress an HRTF dataset...
A multiparametric method of interpolation using WOA05 applied to anthropogenic CO2 in the Atlantic
Directory of Open Access Journals (Sweden)
Anton Velo
2010-11-01
Full Text Available This paper describes the development of a multiparametric interpolation method and its application to anthropogenic carbon (CANT in the Atlantic, calculated by two estimation methods using the CARINA database. The multiparametric interpolation proposed uses potential temperature (θ, salinity, conservative ‘NO’ and ‘PO’ as conservative parameters for the gridding, and the World Ocean Atlas (WOA05 as a reference for the grid structure and the indicated parameters. We thus complement CARINA data with WOA05 database in an attempt to obtain better gridded values by keeping the physical-biogeochemical sea structures. The algorithms developed here also have the prerequisite of being simple and easy to implement. To test the improvements achieved, a comparison between the proposed multiparametric method and a pure spatial interpolation for an independent parameter (O2 was made. As an application case study, CANT estimations by two methods (φCTº and TrOCA were performed on the CARINA database and then gridded by both interpolation methods (spatial and multiparametric. Finally, a calculation of CANT inventories for the whole Atlantic Ocean was performed with the gridded values and using ETOPO2v2 as the sea bottom. Thus, the inventories were between 55.1 and 55.2 Pg-C with the φCTº method and between 57.9 and 57.6 Pg-C with the TrOCA method.
Feng, Shaodong; Wang, Mingjun; Wu, Jigang
2017-11-01
In a compact lensless in-line holographic microscope, the imaging resolution is generally limited by the sensor pixel size because of the short sample-to-sensor distance. To overcome this problem, we propose to use data interpolation based on iteration with only two intensity measurements to enhance the resolution in holographic reconstruction. We did numerical simulations using the U.S. air force target as the sample and showed that data interpolation in the acquired in-line hologram can be used to enhance the reconstruction resolution. The imaging resolution and contrast can be further improved by combining data interpolation with iterative holographic reconstruction using only two hologram measurements acquired by slightly changing the sample-to-sensor distance while recording the in-line holograms. The two in-line hologram intensity measurements were used as a priori constraint in the iteration process according to the Gerchberg-Saxton algorithm for phase retrieval. The iterative reconstruction results showed that the iteration between the sample plane and the sensor planes can refine the interpolated data and thus further improve the resolution as well as the imaging contrast. Besides numerical simulation, we also experimentally demonstrated the enhancement of imaging resolution and contrast by imaging the U.S. air force target and a microscope slide of filamentous algae.
Remark on the Operator-valued Interpolation for Multivariable Bounded Analytic Functions
Czech Academy of Sciences Publication Activity Database
Ambrozie, Calin-Grigore
2004-01-01
Roč. 53, č. 6 (2004), s. 1551-1576 ISSN 0022-2518 R&D Projects: GA ČR(CZ) GA201/03/0041 Institutional research plan: CEZ:AV0Z1019905 Keywords : von Neumann inequality * interpolation * analytic functions Subject RIV: BA - General Mathematics Impact factor: 0.784, year: 2004
Energy Technology Data Exchange (ETDEWEB)
Davenport, C. M.
1977-02-01
The mathematical basis for an ultraprecise digital differential analyzer circuit for use as a parabolic interpolator on numerically controlled machines has been established, and scaling and other error-reduction techniques have been developed. An exact computer model is included, along with typical results showing tracking to within an accuracy of one part per million.
Interpolation between multi-dimensional histograms using a new non-linear moment morphing method
Baak, M.; Gadatsch, S.; Harrington, R.; Verkerke, W.
2015-01-01
A prescription is presented for the interpolation between multi-dimensional distribution templates based on one or multiple model parameters. The technique uses a linear combination of templates, each created using fixed values of the model׳s parameters and transformed according to a specific
Imaging system design and image interpolation based on CMOS image sensor
Li, Yu-feng; Liang, Fei; Guo, Rui
2009-11-01
An image acquisition system is introduced, which consists of a color CMOS image sensor (OV9620), SRAM (CY62148), CPLD (EPM7128AE) and DSP (TMS320VC5509A). The CPLD implements the logic and timing control to the system. SRAM stores the image data, and DSP controls the image acquisition system through the SCCB (Omni Vision Serial Camera Control Bus). The timing sequence of the CMOS image sensor OV9620 is analyzed. The imaging part and the high speed image data memory unit are designed. The hardware and software design of the image acquisition and processing system is given. CMOS digital cameras use color filter arrays to sample different spectral components, such as red, green, and blue. At the location of each pixel only one color sample is taken, and the other colors must be interpolated from neighboring samples. We use the edge-oriented adaptive interpolation algorithm for the edge pixels and bilinear interpolation algorithm for the non-edge pixels to improve the visual quality of the interpolated images. This method can get high processing speed, decrease the computational complexity, and effectively preserve the image edges.
The role of station density for predicting daily runoff by top-kriging interpolation in Austria
Directory of Open Access Journals (Sweden)
Parajka Juraj
2015-09-01
Full Text Available Direct interpolation of daily runoff observations to ungauged sites is an alternative to hydrological model regionalisation. Such estimation is particularly important in small headwater basins characterized by sparse hydrological and climate observations, but often large spatial variability. The main objective of this study is to evaluate predictive accuracy of top-kriging interpolation driven by different number of stations (i.e. station densities in an input dataset. The idea is to interpolate daily runoff for different station densities in Austria and to evaluate the minimum number of stations needed for accurate runoff predictions. Top-kriging efficiency is tested for ten different random samples in ten different stations densities. The predictive accuracy is evaluated by ordinary cross-validation and full-sample crossvalidations. The methodology is tested by using 555 gauges with daily observations in the period 1987-1997. The results of the cross-validation indicate that, in Austria, top-kriging interpolation is superior to hydrological model regionalisation if station density exceeds approximately 2 stations per 1000 km2 (175 stations in Austria. The average median of Nash-Sutcliffe cross-validation efficiency is larger than 0.7 for densities above 2.4 stations/1000 km2. For such densities, the variability of runoff efficiency is very small over ten random samples. Lower runoff efficiency is found for low station densities (less than 1 station/1000 km2 and in some smaller headwater basins.
An Improved Minimum Error Interpolator of CNC for General Curves Based on FPGA
Directory of Open Access Journals (Sweden)
Jiye HUANG
2014-05-01
Full Text Available This paper presents an improved minimum error interpolation algorithm for general curves generation in computer numerical control (CNC. Compared with the conventional interpolation algorithms such as the By-Point Comparison method, the Minimum- Error method and the Digital Differential Analyzer (DDA method, the proposed improved Minimum-Error interpolation algorithm can find a balance between accuracy and efficiency. The new algorithm is applicable for the curves of linear, circular, elliptical and parabolic. The proposed algorithm is realized on a field programmable gate array (FPGA with Verilog HDL language, and simulated by the ModelSim software, and finally verified on a two-axis CNC lathe. The algorithm has the following advantages: firstly, the maximum interpolation error is only half of the minimum step-size; and secondly the computing time is only two clock cycles of the FPGA. Simulations and actual tests have proved that the high accuracy and efficiency of the algorithm, which shows that it is highly suited for real-time applications.
DEFF Research Database (Denmark)
Hurkmans, R.T.W.L.; Bamber, J.L.; Sørensen, Louise Sandberg
2012-01-01
Estimation of ice sheet mass balance from satellite altimetry requires interpolation of point-scale elevation change (dH/dt) data over the area of interest. The largest dH/dt values occur over narrow, fast-flowing outlet glaciers, where data coverage of current satellite altimetry is poorest...
A new stochastic model considering satellite clock interpolation errors in precise point positioning
Wang, Shengli; Yang, Fanlin; Gao, Wang; Yan, Lizi; Ge, Yulong
2018-03-01
Precise clock products are typically interpolated based on the sampling interval of the observational data when they are used for in precise point positioning. However, due to the occurrence of white noise in atomic clocks, a residual component of such noise will inevitable reside within the observations when clock errors are interpolated, and such noise will affect the resolution of the positioning results. In this paper, which is based on a twenty-one-week analysis of the atomic clock noise characteristics of numerous satellites, a new stochastic observation model that considers satellite clock interpolation errors is proposed. First, the systematic error of each satellite in the IGR clock product was extracted using a wavelet de-noising method to obtain the empirical characteristics of atomic clock noise within each clock product. Then, based on those empirical characteristics, a stochastic observation model was structured that considered the satellite clock interpolation errors. Subsequently, the IGR and IGS clock products at different time intervals were used for experimental validation. A verification using 179 stations worldwide from the IGS showed that, compared with the conventional model, the convergence times using the stochastic model proposed in this study were respectively shortened by 4.8% and 4.0% when the IGR and IGS 300-s-interval clock products were used and by 19.1% and 19.4% when the 900-s-interval clock products were used. Furthermore, the disturbances during the initial phase of the calculation were also effectively improved.
Vegter, H.; van den Boogaard, Antonius H.
2006-01-01
An anisotropic plane stress yield function based on interpolation by second order Bézier curves is proposed. The parameters for the model are readily derived by four mechanical tests: a uniaxial, an equi-biaxial and a plane strain tensile test and a shear test. In case of planar anisotropy, this set
Scattering Amplitudes Interpolating Between Instant Form and Front Form of Relativistic Dynamics
Ji, C.R.; Bakker, B.L.G.; Li, Z.
2014-01-01
Among the three forms of relativistic Hamiltonian dynamics proposed by Dirac in 1949, the front form has the largest number of kinematic generators. This distinction provides useful consequences in the analysis of physical observables in hadron physics. Using the method of interpolation between the
Bayesian interpolation in a dynamic sinusoidal model with application to packet-loss concealment
DEFF Research Database (Denmark)
Nielsen, Jesper Kjær; Christensen, Mads Græsbøll; Cemgil, Ali Taylan
2010-01-01
In this paper, we consider Bayesian interpolation and parameter estimation in a dynamic sinusoidal model. This model is more ﬂexible than the static sinusoidal model since it enables the amplitudes and phases of the sinusoids to be time-varying. For the dynamic sinusoidal model, we derive...
Directory of Open Access Journals (Sweden)
COJOCARU ŞTEFANA
2014-03-01
Full Text Available patial interpolation, in the context of spatial analysis, can be defined as the derivation of new data from already known information, a technique frequently used to predict and quantify spatial variation of a certain property or parameter. In this study we compared the performance of Inverse Distance Weighted (IDW, Ordinary Kriging and Natural Neighbor techniques, applied in spatial interpolation of precipitation parameters (pH, electrical conductivity and total dissolved solids. These techniques are often used when the area of interest is relatively small and the sampled locations are regularly spaced. The methods were tested on data collected in Iasi city (Romania between March – May 2013. Spatial modeling was performed on a small dataset, consisting of 7 sample locations and 13 different known values of each analyzed parameter. The precision of the techniques used is directly dependent on sample density as well as data variation, greater fluctuations in values between locations causing a decrease in the accuracy of the methods used. To validate the results and reveal the best method of interpolating rainfall characteristics, leave-one – out cross-validation approach was used. Comparing residues between the known values and the estimated values of pH, electrical conductivity and total dissolved solids, it was revealed that Natural Neighbor stands out as generating the smallest residues for pH and electrical conductivity, whereas IDW presents the smallest error in interpolating total dissolved solids (the parameter with the highest fluctuations in value.
An Energy Conservative Ray-Tracing Method With a Time Interpolation of the Force Field
Energy Technology Data Exchange (ETDEWEB)
Yao, Jin [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
2015-02-10
A new algorithm that constructs a continuous force field interpolated in time is proposed for resolving existing difficulties in numerical methods for ray-tracing. This new method has improved accuracy, but with the same degree of algebraic complexity compared to Kaisers method.
Evaluation of Teeth and Supporting Structures on Digital Radiograms using Interpolation Methods
Energy Technology Data Exchange (ETDEWEB)
Koh, Kwang Joon [Dept. of Oral and Maxillofacial Radiology, School of Dentistry and Institute of Oral Bio Science , Chonbuk National University, Chonju (Korea, Republic of); Chang, Kee Wan [Dept. of Preventive and Community Dentistry, School of Dentistry and Institute of Oral Bio Science, Chonbuk National University, Chonju (Korea, Republic of)
1999-02-15
To determine the effect of interpolation functions when processing the digital periapical images. The digital images were obtained by Digora and CDR system on the dry skull and human subject. 3 oral radiologists evaluated the 3 portions of each processed image using 7 interpolation methods and ROC curves were obtained by trapezoidal methods. The highest Az value(0.96) was obtained with cubic spline method and the lowest Az value(0.03) was obtained with facet model method in Digora system. The highest Az value(0.79) was obtained with gray segment expansion method and the lowest Az value(0.07) was obtained with facet model method in CDR system. There was significant difference of Az value in original image between Digora and CDR system at alpha=0.05 level. There were significant differences of Az values between Digora and CDR images with cubic spline method, facet model method, linear interpolation method and non-linear interpolation method at alpha= 0.1 level.
Improving the visualization of electron-microscopy data through optical flow interpolation
Carata, Lucian
2013-01-01
Technical developments in neurobiology have reached a point where the acquisition of high resolution images representing individual neurons and synapses becomes possible. For this, the brain tissue samples are sliced using a diamond knife and imaged with electron-microscopy (EM). However, the technique achieves a low resolution in the cutting direction, due to limitations of the mechanical process, making a direct visualization of a dataset difficult. We aim to increase the depth resolution of the volume by adding new image slices interpolated from the existing ones, without requiring modifications to the EM image-capturing method. As classical interpolation methods do not provide satisfactory results on this type of data, the current paper proposes a re-framing of the problem in terms of motion volumes, considering the depth axis as a temporal axis. An optical flow method is adapted to estimate the motion vectors of pixels in the EM images, and this information is used to compute and insert multiple new images at certain depths in the volume. We evaluate the visualization results in comparison with interpolation methods currently used on EM data, transforming the highly anisotropic original dataset into a dataset with a larger depth resolution. The interpolation based on optical flow better reveals neurite structures with realistic undistorted shapes, and helps to easier map neuronal connections. © 2011 ACM.
Interpolation Filter Design for Hearing-Aid Audio Class-D Output Stage Application
DEFF Research Database (Denmark)
Pracný, Peter; Bruun, Erik; Llimos Muntal, Pere
2012-01-01
is to compare this design to designs presented in other state-of-the-art works ranging from hi-fi audio to hearing-aids. By performing comparison, trends and tradeoffs in interpolation filter design are indentified and hearing-aid specifications are derived. The possibilities for hardware reduction...
Effects of interpolation and data resolution on methane emission estimates from rice paddies
Bodegom, van P.M.; Verburg, P.H.; Stein, A.; Adiningsih, S.; Denier van der Gon, H.A.C.
2002-01-01
Rice paddies are an important source of the greenhouse gas methane (CH4). Global methane emission estimates are highly uncertain and do not account for effects of interpolation or data resolution errors. This paper determines such scaling effects for the influence of soil properties on calculated
Tan, Bing; Huang, Min; Zhu, Qibing; Guo, Ya; Qin, Jianwei
2017-12-01
Laser-induced breakdown spectroscopy (LIBS) is an analytical technique that has gained increasing attention because of many applications. The production of continuous background in LIBS is inevitable because of factors associated with laser energy, gate width, time delay, and experimental environment. The continuous background significantly influences the analysis of the spectrum. Researchers have proposed several background correction methods, such as polynomial fitting, Lorenz fitting and model-free methods. However, less of them apply these methods in the field of LIBS Technology, particularly in qualitative and quantitative analyses. This study proposes a method based on spline interpolation for detecting and estimating the continuous background spectrum according to its smooth property characteristic. Experiment on the background correction simulation indicated that, the spline interpolation method acquired the largest signal-to-background ratio (SBR) over polynomial fitting, Lorenz fitting and model-free method after background correction. These background correction methods all acquire larger SBR values than that acquired before background correction (The SBR value before background correction is 10.0992, whereas the SBR values after background correction by spline interpolation, polynomial fitting, Lorentz fitting, and model-free methods are 26.9576, 24.6828, 18.9770, and 25.6273 respectively). After adding random noise with different kinds of signal-to-noise ratio to the spectrum, spline interpolation method acquires large SBR value, whereas polynomial fitting and model-free method obtain low SBR values. All of the background correction methods exhibit improved quantitative results of Cu than those acquired before background correction (The linear correlation coefficient value before background correction is 0.9776. Moreover, the linear correlation coefficient values after background correction using spline interpolation, polynomial fitting, Lorentz
A 70 MSample/s 110 mW 8 b CMOS folding interpolating A/D Converter
Nauta, Bram; Venes, A.G.W.
1995-01-01
In bipolar technology the folding and interpolation technique has proven to be successful for high sample rates. This paper investigates the possibilities of this technique in CMOS. The major advantage of folding and interpolation in CMOS lies in the field of high sample rate in combination with low
The Interpolation Method for Estimating the Above-Ground Biomass Using Terrestrial-Based Inventory
Directory of Open Access Journals (Sweden)
I Nengah Surati Jaya
2014-09-01
Full Text Available This paper examined several methods for interpolating biomass on logged-over dry land forest using terrestrial-based forest inventory in Labanan, East Kalimantan and Lamandau, Kota Wringing Barat, Central Kalimantan. The plot-distances examined was 1,000−1,050 m for Labanan and 1,000−899m for Lawanda. The main objective of this study was to obtain the best interpolation method having the most accurate prediction on spatial distribution of forest biomass for dry land forest. Two main interpolation methods were examined: (1 deterministic approach using the IDW method and (2 geo-statistics approach using Kriging with spherical, circular, linear, exponential, and Gaussian models. The study results at both sites consistently showed that the IDW method was better than the Kriging method for estimating the spatial distribution of biomass. The validation results using chi-square test showed that the IDW interpolation provided accurate biomass estimation. Using the percentage of mean deviation value (MD(%, it was also recognized that the IDWs with power parameter (p of 2 provided relatively low value , i.e., only 15% for Labanan, East Kalimantan Province and 17% for Lamandau, Kota Wringing Barat Central Kalimantan Province. In general, IDW interpolation method provided better results than the Kriging, where the Kriging method provided MD(% of about 27% and 21% for Lamandau and Labanan sites, respectively.Keywords: deterministic, geostatistics, IDW, Kriging, above-groung biomass
Directory of Open Access Journals (Sweden)
Mateusz Szcześniak
2015-02-01
Full Text Available Ground-based precipitation data are still the dominant input type for hydrological models. Spatial variability in precipitation can be represented by spatially interpolating gauge data using various techniques. In this study, the effect of daily precipitation interpolation methods on discharge simulations using the semi-distributed SWAT (Soil and Water Assessment Tool model over a 30-year period is examined. The study was carried out in 11 meso-scale (119–3935 km2 sub-catchments lying in the Sulejów reservoir catchment in central Poland. Four methods were tested: the default SWAT method (Def based on the Nearest Neighbour technique, Thiessen Polygons (TP, Inverse Distance Weighted (IDW and Ordinary Kriging (OK. =The evaluation of methods was performed using a semi-automated calibration program SUFI-2 (Sequential Uncertainty Fitting Procedure Version 2 with two objective functions: Nash-Sutcliffe Efficiency (NSE and the adjusted R2 coefficient (bR2. The results show that: (1 the most complex OK method outperformed other methods in terms of NSE; and (2 OK, IDW, and TP outperformed Def in terms of bR2. The median difference in daily/monthly NSE between OK and Def/TP/IDW calculated across all catchments ranged between 0.05 and 0.15, while the median difference between TP/IDW/OK and Def ranged between 0.05 and 0.07. The differences between pairs of interpolation methods were, however, spatially variable and a part of this variability was attributed to catchment properties: catchments characterised by low station density and low coefficient of variation of daily flows experienced more pronounced improvement resulting from using interpolation methods. Methods providing higher precipitation estimates often resulted in a better model performance. The implication from this study is that appropriate consideration of spatial precipitation variability (often neglected by model users that can be achieved using relatively simple interpolation methods can
Directory of Open Access Journals (Sweden)
A. Verworn
2011-02-01
Full Text Available Hydrological modelling of floods relies on precipitation data with a high resolution in space and time. A reliable spatial representation of short time step rainfall is often difficult to achieve due to a low network density. In this study hourly precipitation was spatially interpolated with the multivariate geostatistical method kriging with external drift (KED using additional information from topography, rainfall data from the denser daily networks and weather radar data. Investigations were carried out for several flood events in the time period between 2000 and 2005 caused by different meteorological conditions. The 125 km radius around the radar station Ummendorf in northern Germany covered the overall study region. One objective was to assess the effect of different approaches for estimation of semivariograms on the interpolation performance of short time step rainfall. Another objective was the refined application of the method kriging with external drift. Special attention was not only given to find the most relevant additional information, but also to combine the additional information in the best possible way. A multi-step interpolation procedure was applied to better consider sub-regions without rainfall.
The impact of different semivariogram types on the interpolation performance was low. While it varied over the events, an averaged semivariogram was sufficient overall. Weather radar data were the most valuable additional information for KED for convective summer events. For interpolation of stratiform winter events using daily rainfall as additional information was sufficient. The application of the multi-step procedure significantly helped to improve the representation of fractional precipitation coverage.
Directory of Open Access Journals (Sweden)
Ly, S.
2013-01-01
Full Text Available Watershed management and hydrological modeling require data related to the very important matter of precipitation, often measured using raingages or weather stations. Hydrological models often require a preliminary spatial interpolation as part of the modeling process. The success of spatial interpolation varies according to the type of model chosen, its mode of geographical management and the resolution used. The quality of a result is determined by the quality of the continuous spatial rainfall, which ensues from the interpolation method used. The objective of this article is to review the existing methods for interpolation of rainfall data that are usually required in hydrological modeling. We review the basis for the application of certain common methods and geostatistical approaches used in interpolation of rainfall. Previous studies have highlighted the need for new research to investigate ways of improving the quality of rainfall data and ultimately, the quality of hydrological modeling.
Directory of Open Access Journals (Sweden)
Emile Faye
Full Text Available Bridging the gap between the predictions of coarse-scale climate models and the fine-scale climatic reality of species is a key issue of climate change biology research. While it is now well known that most organisms do not experience the climatic conditions recorded at weather stations, there is little information on the discrepancies between microclimates and global interpolated temperatures used in species distribution models, and their consequences for organisms' performance. To address this issue, we examined the fine-scale spatiotemporal heterogeneity in air, crop canopy and soil temperatures of agricultural landscapes in the Ecuadorian Andes and compared them to predictions of global interpolated climatic grids. Temperature time-series were measured in air, canopy and soil for 108 localities at three altitudes and analysed using Fourier transform. Discrepancies between local temperatures vs. global interpolated grids and their implications for pest performance were then mapped and analysed using GIS statistical toolbox. Our results showed that global interpolated predictions over-estimate by 77.5 ± 10% and under-estimate by 82.1 ± 12% local minimum and maximum air temperatures recorded in the studied grid. Additional modifications of local air temperatures were due to the thermal buffering of plant canopies (from -2.7 °K during daytime to 1.3 °K during night-time and soils (from -4.9 °K during daytime to 6.7 °K during night-time with a significant effect of crop phenology on the buffer effect. This discrepancies between interpolated and local temperatures strongly affected predictions of the performance of an ectothermic crop pest as interpolated temperatures predicted pest growth rates 2.3-4.3 times lower than those predicted by local temperatures. This study provides quantitative information on the limitation of coarse-scale climate data to capture the reality of the climatic environment experienced by living organisms. In highly
Effect of raingage density, position and interpolation on rainfall-discharge modelling
Ly, S.; Sohier, C.; Charles, C.; Degré, A.
2012-04-01
Precipitation traditionally observed using raingages or weather stations, is one of the main parameters that have direct impact on runoff production. Precipitation data require a preliminary spatial interpolation prior to hydrological modeling. The accuracy of modelling result depends on the accuracy of the interpolated spatial rainfall which differs according to different interpolation methods. The accuracy of the interpolated spatial rainfall is usually determined by cross-validation method. The objective of this study is to assess the different interpolation methods of daily rainfall at the watershed scale through hydrological modelling and to explore the best methods that provide a good long term simulation. Four versions of geostatistics: Ordinary Kriging (ORK), Universal Kriging (UNK), Kriging with External Dridft (KED) and Ordinary Cokriging (OCK) and two types of deterministic methods: Thiessen polygon (THI) and Inverse Distance Weighting (IDW) are used to produce 30-year daily rainfall inputs for a distributed physically-based hydrological model (EPIC-GRID). This work is conducted in the Ourthe and Ambleve nested catchments, located in the Ardennes hilly landscape in the Walloon region, Belgium. The total catchment area is 2908 km2, lies between 67 and 693 m in elevation. The multivariate geostatistics (KED and OCK) are also used by incorporating elevation as external data to improve the rainfall prediction. This work also aims at analysing the effect of different raingage densities and position used for interpolation, on the stream flow modelled to get insight in terms of the capability and limitation of the geostatistical methods. The number of raingage varies from 70, 60, 50, 40, 30, 20, 8 to 4 stations located in and surrounding the catchment area. In the latter case, we try to use different positions: around the catchment and only a part of the catchment. The result shows that the simple method like THI fails to capture the rainfall and to produce
Spatial Interpolation of Rain-field Dynamic Time-Space Evolution in Hong Kong
Liu, P.; Tung, Y. K.
2017-12-01
Accurate and reliable measurement and prediction of spatial and temporal distribution of rain-field over a wide range of scales are important topics in hydrologic investigations. In this study, geostatistical treatment of precipitation field is adopted. To estimate the rainfall intensity over a study domain with the sample values and the spatial structure from the radar data, the cumulative distribution functions (CDFs) at all unsampled locations were estimated. Indicator Kriging (IK) was used to estimate the exceedance probabilities for different pre-selected cutoff levels and a procedure was implemented for interpolating CDF values between the thresholds that were derived from the IK. Different interpolation schemes of the CDF were proposed and their influences on the performance were also investigated. The performance measures and visual comparison between the observed rain-field and the IK-based estimation suggested that the proposed method can provide fine results of estimation of indicator variables and is capable of producing realistic image.
Efficient Implementation of an Optimal Interpolator for Large Spatial Data Sets
Memarsadeghi, Nargess; Mount, David M.
2007-01-01
Scattered data interpolation is a problem of interest in numerous areas such as electronic imaging, smooth surface modeling, and computational geometry. Our motivation arises from applications in geology and mining, which often involve large scattered data sets and a demand for high accuracy. The method of choice is ordinary kriging. This is because it is a best unbiased estimator. Unfortunately, this interpolant is computationally very expensive to compute exactly. For n scattered data points, computing the value of a single interpolant involves solving a dense linear system of size roughly n x n. This is infeasible for large n. In practice, kriging is solved approximately by local approaches that are based on considering only a relatively small'number of points that lie close to the query point. There are many problems with this local approach, however. The first is that determining the proper neighborhood size is tricky, and is usually solved by ad hoc methods such as selecting a fixed number of nearest neighbors or all the points lying within a fixed radius. Such fixed neighborhood sizes may not work well for all query points, depending on local density of the point distribution. Local methods also suffer from the problem that the resulting interpolant is not continuous. Meyer showed that while kriging produces smooth continues surfaces, it has zero order continuity along its borders. Thus, at interface boundaries where the neighborhood changes, the interpolant behaves discontinuously. Therefore, it is important to consider and solve the global system for each interpolant. However, solving such large dense systems for each query point is impractical. Recently a more principled approach to approximating kriging has been proposed based on a technique called covariance tapering. The problems arise from the fact that the covariance functions that are used in kriging have global support. Our implementations combine, utilize, and enhance a number of different
The analysis of composite laminated beams using a 2D interpolating meshless technique
Sadek, S. H. M.; Belinha, J.; Parente, M. P. L.; Natal Jorge, R. M.; de Sá, J. M. A. César; Ferreira, A. J. M.
2018-02-01
Laminated composite materials are widely implemented in several engineering constructions. For its relative light weight, these materials are suitable for aerospace, military, marine, and automotive structural applications. To obtain safe and economical structures, the modelling analysis accuracy is highly relevant. Since meshless methods in the recent years achieved a remarkable progress in computational mechanics, the present work uses one of the most flexible and stable interpolation meshless technique available in the literature—the Radial Point Interpolation Method (RPIM). Here, a 2D approach is considered to numerically analyse composite laminated beams. Both the meshless formulation and the equilibrium equations ruling the studied physical phenomenon are presented with detail. Several benchmark beam examples are studied and the results are compared with exact solutions available in the literature and the results obtained from a commercial finite element software. The results show the efficiency and accuracy of the proposed numeric technique.
An Online Method for Interpolating Linear Parametric Reduced-Order Models
Amsallem, David
2011-01-01
A two-step online method is proposed for interpolating projection-based linear parametric reduced-order models (ROMs) in order to construct a new ROM for a new set of parameter values. The first step of this method transforms each precomputed ROM into a consistent set of generalized coordinates. The second step interpolates the associated linear operators on their appropriate matrix manifold. Real-time performance is achieved by precomputing inner products between the reduced-order bases underlying the precomputed ROMs. The proposed method is illustrated by applications in mechanical and aeronautical engineering. In particular, its robustness is demonstrated by its ability to handle the case where the sampled parameter set values exhibit a mode veering phenomenon. © 2011 Society for Industrial and Applied Mathematics.
An efficient coupled polynomial interpolation scheme for shear mode sandwich beam finite element
Directory of Open Access Journals (Sweden)
Litesh N. Sulbhewar
Full Text Available An efficient piezoelectric sandwich beam finite element is presented here. It employs the coupled polynomial field interpolation scheme for field variables which incorporates electromechanical coupling at interpolation level itself; unlike conventional sandwich beam theory (SBT based formulations available in the literature. A variational formulation is used to derive the governing equations, which are used to establish the relationships between field variables. These relations lead to the coupled polynomial field descriptions of variables, unlike conventional SBT formulations which use assumed independent polynomials. The relative axial displacement is expressed only by coupled terms containing contributions from other mechanical and electrical variables, thus eliminating use of the transverse displacement derivative as a degree of freedom. A set of coupled shape function based on these polynomials has shown the improvement in the convergence characteristics of the SBT based formulation. This improvement in the performance is achieved with one nodal degree of freedom lesser than the conventional SBT formulations.
Hoarau, Charlotte; Christophe, Sidonie
2017-05-01
Graphic interfaces of geoportals allow visualizing and overlaying various (visually) heterogeneous geographical data, often by image blending: vector data, maps, aerial imagery, Digital Terrain Model, etc. Map design and geo-visualization may benefit from methods and tools to hybrid, i.e. visually integrate, heterogeneous geographical data and cartographic representations. In this paper, we aim at designing continuous hybrid visualizations between ortho-imagery and symbolized vector data, in order to control a particular visual property, i.e. the photo-realism perception. The natural appearance (colors, textures) and various texture effects are used to drive the control the photo-realism level of the visualization: color and texture interpolation blocks have been developed. We present a global design method that allows to manipulate the behavior of those interpolation blocks on each type of geographical layer, in various ways, in order to provide various cartographic continua.
Evaluation of Interpolants in Their Ability to Fit Seismometric Time Series
Directory of Open Access Journals (Sweden)
Kanadpriya Basu
2015-08-01
Full Text Available This article is devoted to the study of the ASARCO demolition seismic data. Two different classes of modeling techniques are explored: First, mathematical interpolation methods and second statistical smoothing approaches for curve fitting. We estimate the characteristic parameters of the propagation medium for seismic waves with multiple mathematical and statistical techniques, and provide the relative advantages of each approach to address fitting of such data. We conclude that mathematical interpolation techniques and statistical curve fitting techniques complement each other and can add value to the study of one dimensional time series seismographic data: they can be use to add more data to the system in case the data set is not large enough to perform standard statistical tests.
The Smoothness of Fractal Interpolation Functions on ℝ and on p-Series Local Fields
Directory of Open Access Journals (Sweden)
Jing Li
2014-01-01
Full Text Available A fractal interpolation function on a p-series local field Kp is defined, and its p-type smoothness is shown by virtue of the equivalent relationship between the Hölder type space CσKp and the Lipschitz class Lipσ,Kp. The orders of the p-type derivatives and the fractal dimensions of the graphs of Weierstrass type function on local fields are given as an example. The α-fractal function on ℝ is introduced and the conclusion of its smoothness is improved in a more general case; some examples are shown to support the conclusion. Finally, a comparison between the fractal interpolation functions defined on ℝ and Kp is given.
International Nuclear Information System (INIS)
Misawa, T.; Itakura, H.
1995-01-01
The present article focuses on a dynamical simulation of molecular motion in liquids. In the simulation involving diffusion-controlled reaction with discrete time steps, lack of information regarding the trajectory within the time step may result in a failure to count the number of reactions of the particles within the step. In order to rectify this, an interpolated diffusion process is used. The process is derived from a stochastic interpolation formula recently developed by the first author [J. Math. Phys. 34, 775 (1993)]. In this method, the probability that reaction has occurred during the time step given the initial and final positions of the particles is calculated. Some numerical examples confirm that the theoretical result corresponds to an improvement over the Clifford-Green work [Mol. Phys. 57, 123 (1986)] on the same matter
Linearly interpolated sub-symbol optical phase noise suppression in CO-OFDM system.
Hong, Xuezhi; Hong, Xiaojian; He, Sailing
2015-02-23
An optical phase noise suppression algorithm, LI-SCPEC, based on phase linear interpolation and sub-symbol processing is proposed for CO-OFDM system. By increasing the temporal resolution of carrier phase tracking through dividing one symbol into several sub-blocks, i.e., sub-symbols, inter-carrier-interference (ICI) mitigation is achieved in the proposed algorithm. Linear interpolation is employed to obtain a reliable temporal reference for sub-symbol phase estimation. The new algorithm, with only a few number of sub-symbols (N(B) = 4), can provide a considerably larger laser linewidth tolerance than several other ICI mitigation algorithms as demonstrated by Monte-Carlo simulations. Numerical analysis verifies that the best performance is achieved with an optimal and moderate number of sub-symbols. Complexity analysis shows that the required number of complex-valued multiplications is independent of the number of sub-symbols used in the proposed algorithm.
Dupuis, L. R.; Scoggins, J. R.
1979-01-01
Results of analyses revealed that nonlinear changes or differences formed centers or systems, that were mesosynoptic in nature. These systems correlated well in space with upper level short waves, frontal zones, and radar observed convection, and were very systematic in time and space. Many of the centers of differences were well established in the vertical, extending up to the tropopause. Statistical analysis showed that on the average nonlinear changes were larger in convective areas than nonconvective regions. Errors often exceeding 100 percent were made by assuming variables to change linearly through a 12-h period in areas of thunderstorms, indicating that these nonlinear changes are important in the development of severe weather. Linear changes, however, accounted for more and more of an observed change as the time interval (within the 12-h interpolation period) increased, implying that the accuracy of linear interpolation increased over larger time intervals.
Qian, Fang; Wu, Yihui; Hao, Peng
2017-11-01
Baseline correction is a very important part of pre-processing. Baseline in the spectrum signal can induce uneven amplitude shifts across different wavenumbers and lead to bad results. Therefore, these amplitude shifts should be compensated before further analysis. Many algorithms are used to remove baseline, however fully automated baseline correction is convenient in practical application. A fully automated algorithm based on wavelet feature points and segment interpolation (AWFPSI) is proposed. This algorithm finds feature points through continuous wavelet transformation and estimates baseline through segment interpolation. AWFPSI is compared with three commonly introduced fully automated and semi-automated algorithms, using simulated spectrum signal, visible spectrum signal and Raman spectrum signal. The results show that AWFPSI gives better accuracy and has the advantage of easy use.
Interpolation and Inversion - New Features in the Matlab Sesimic Anisotropy Toolbox
Walker, A.; Wookey, J. M.
2015-12-01
A key step in studies of seismic anisotropy in the mantle is often the creation of models designed to explain its physical origin. We previously released MSAT (the Matlab Seismic Anisotropy Toolbox), which includes a range of functions that can be used together to build these models and provide geological or geophysical insight given measurements of, for example, shear-wave splitting. Here we describe some of the new features of MSAT that will be included in a new release timed to coincide with the 2015 Fall Meeting. A critical step in testing models of the origin of seismic anisotropy is the determination of the misfit between shear-wave splitting parameters predicted from a model and measured from seismic observations. Is a model that correctly reproduces the delay time "better" than a model that correctly reproduces the fast polarization? We have introduced several new methods that use both parameters to calculate the misfit in a meaningful way and these can be used as part of an inversion scheme in order to find a model that best matches measured shear wave splitting. Our preferred approach involves the creation, "splitting", and "unsplitting" of a test wavelet. A measure of the misfit is then provided by the normalized second eigenvalue of the covariance matrix of particle motion for the two wavelets in a way similar to that used to find splitting parameters from data. This can be used as part of an inverse scheme to find a model that can reproduce a set of shear-wave splitting observations. A second challenge is the interpolation of elastic constants between two known points. Naive element-by-element interpolation can result in anomalous seismic velocities from the interpolated tensor. We introduce an interpolation technique involving both the orientation (defined in terms of the eigenvectors of the dilatational or Voigt stiffness tensor) and magnitude of the two end-member elastic tensors. This permits changes in symmetry between the end-members and removes
Attenuation correction of emission PET images with average CT: Interpolation from breath-hold CT
Huang, Tzung-Chi; Zhang, Geoffrey; Chen, Chih-Hao; Yang, Bang-Hung; Wu, Nien-Yun; Wang, Shyh-Jen; Wu, Tung-Hsin
2011-05-01
Misregistration resulting from the difference of temporal resolution in PET and CT scans occur frequently in PET/CT imaging, which causes distortion in tumor quantification in PET. Respiration cine average CT (CACT) for PET attenuation correction has been reported to improve the misalignment effectively by several papers. However, the radiation dose to the patient from a four-dimensional CT scan is relatively high. In this study, we propose a method to interpolate respiratory CT images over a respiratory cycle from inhalation and exhalation breath-hold CT images, and use the average CT from the generated CT set for PET attenuation correction. The radiation dose to the patient is reduced using this method. Six cancer patients of various lesion sites underwent routine free-breath helical CT (HCT), respiration CACT, interpolated average CT (IACT), and 18F-FDG PET. Deformable image registration was used to interpolate the middle phases of a respiratory cycle based on the end-inspiration and end-expiration breath-hold CT scans. The average CT image was calculated from the eight interpolated CT image sets of middle respiratory phases and the two original inspiration and expiration CT images. Then the PET images were reconstructed by these three methods for attenuation correction using HCT, CACT, and IACT. Misalignment of PET image using either CACT or IACT for attenuation correction in PET/CT was improved. The difference in standard uptake value (SUV) from tumor in PET images was most significant between the use of HCT and CACT, while the least significant between the use of CACT and IACT. Besides the similar improvement in tumor quantification compared to the use of CACT, using IACT for PET attenuation correction reduces the radiation dose to the patient.
Water Quality Interpolation Using Various In-Stream Distance Weighting Metrics
Saia, S. M.; Walter, T.; Sullivan, P.; Christie, R.
2012-12-01
Interpolation of water quality samples along the reach of a stream can be used to (1) extend point data to un-sampled locations along the stream network, (2) identify spatial patterns in water quality, and (3) understand how natural and human factors shape these patterns. Kriging, one of the most commonly used geospatial interpolation methods, assumes that nearby sites are spatially auto-correlated; sites closer together have more in common than sites further away. Studies have introduced kriging methods that weight in-stream distance metrics with either landscape attributes (i.e. topography, land use, temperature, and various soil properties) or stream order. Here we present a weighting scheme that combines both surrounding landscape attributes with stream order. We use R, an open-source programming language, to interpolate water quality data collected from the Mianus River in Westchester County, New York. As the major drinking water supply for approximately 100,000 people in Connecticut and New York, the Mianus River watershed community values the cleanliness of its water for recreational activities as well as the sustenance of terrestrial and aquatic wildlife. With the in-stream interpolation results, we can gain a better understanding of factors contributing to water quality issues and observed biogeochemical patterns within the watershed. For example, we can help answer questions such as: How can we target landscape stabilization projects to reduce turbidity? If we find that the most powerful weighting is associated with first order streams and cropland, we know conservation efforts should be focused on agricultural head waters.
Lindley, S. J.; Walsh, T.
There are many modelling methods dedicated to the estimation of spatial patterns in pollutant concentrations, each with their distinctive advantages and disadvantages. The derivation of a surface of air quality values from monitoring data alone requires the conversion of point-based data from a limited number of monitoring stations to a continuous surface using interpolation. Since interpolation techniques involve the estimation of data at un-sampled points based on calculated relationships between data measured at a number of known sample points, they are subject to some uncertainty, both in terms of the values estimated and their spatial distribution. These uncertainties, which are incorporated into many empirical and semi-empirical mapping methodologies, could be recognised in any further usage of the data and also in the assessment of the extent of an exceedence of an air quality standard and the degree of exposure this may represent. There is a wide range of available interpolation techniques and the differences in the characteristics of these result in variations in the output surfaces estimated from the same set of input points. The work presented in this paper provides an examination of uncertainties through the application of a number of interpolation techniques available in standard GIS packages to a case study nitrogen dioxide data set for the Greater Manchester conurbation in northern England. The implications of the use of different techniques are discussed through application to hourly concentrations during an air quality episode and annual average concentrations in 2001. Patterns of concentrations demonstrate considerable differences in the estimated spatial pattern of maxima as the combined effects of chemical processes, topography and meteorology. In the case of air quality episodes, the considerable spatial variability of concentrations results in large uncertainties in the surfaces produced but these uncertainties vary widely from area to area
Synthesis of freeform refractive surfaces forming various radiation patterns using interpolation
Voznesenskaya, Anna; Mazur, Iana; Krizskiy, Pavel
2017-09-01
Optical freeform surfaces are very popular today in such fields as lighting systems, sensors, photovoltaic concentrators, and others. The application of such surfaces allows to obtain systems with a new quality with a reduced number of optical components to ensure high consumer characteristics: small size, weight, high optical transmittance. This article presents the methods of synthesis of refractive surface for a given source and the radiation pattern of various shapes using a computer simulation cubic spline interpolation.
EXAMINATION OF THE VISUAL ACUITY ON THE LCD OPTOTYPE WITH WHOLE-LINE AND INTERPOLATION METHOD
Zajíček Tomáš; Veselý Petr; Veselý Petr; Synek Svatopluk; Synek Svatopluk
2012-01-01
The goal of this work is to show the possibility of us using the LCD optotype in common optometrist practice. Furthermore two commonly used methods for measuring visual acuity will be compared. 69 respondents were used for the measurements. The respondents were divided into two groups according to the measured LCD optotype. The visual acuity was measured using the whole-line method on modified Snellen charts as well as the interpolation method on ETDRS charts. Measurements were taken on the S...
Echantillonnage, interpolation et détection. Applications en imagerie satellitaire.
Almansa, Andrés
2002-01-01
This thesis deals with a few of the many issues that arise in the design of a complete computer vision system, from sampling and interpolation, to feature detection and interpretation. The main motivation for addressing these topics was provided by the French center of space studies (CNES), and the design of earth observation satellites, as well as photogrammetry applications and video-surveillance applications at Cognitech Inc.during the final stages of this work, but most subjects are treat...
Interpolating Spline Curve-Based Perceptual Encryption for 3D Printing Models
Directory of Open Access Journals (Sweden)
Giao N. Pham
2018-02-01
Full Text Available With the development of 3D printing technology, 3D printing has recently been applied to many areas of life including healthcare and the automotive industry. Due to the benefit of 3D printing, 3D printing models are often attacked by hackers and distributed without agreement from the original providers. Furthermore, certain special models and anti-weapon models in 3D printing must be protected against unauthorized users. Therefore, in order to prevent attacks and illegal copying and to ensure that all access is authorized, 3D printing models should be encrypted before being transmitted and stored. A novel perceptual encryption algorithm for 3D printing models for secure storage and transmission is presented in this paper. A facet of 3D printing model is extracted to interpolate a spline curve of degree 2 in three-dimensional space that is determined by three control points, the curvature coefficients of degree 2, and an interpolating vector. Three control points, the curvature coefficients, and interpolating vector of the spline curve of degree 2 are encrypted by a secret key. The encrypted features of the spline curve are then used to obtain the encrypted 3D printing model by inverse interpolation and geometric distortion. The results of experiments and evaluations prove that the entire 3D triangle model is altered and deformed after the perceptual encryption process. The proposed algorithm is responsive to the various formats of 3D printing models. The results of the perceptual encryption process is superior to those of previous methods. The proposed algorithm also provides a better method and more security than previous methods.
International Nuclear Information System (INIS)
Glassey, C.R.; Choy, Y.C.
1977-09-01
The problem of estimation of a physical quantity from a set of measurements is considered. We extend Kalman recursive estimation procedure in two ways. First, we explore how to use the latest observation to retrospectively update estimates of past system states. Second, we show how to apply the retrospective update idea to get interpolation estimates between the epochs of observations. We also show application of these ideas for instrument calibration in nuclear accountability systems
Characterization of a rearrangement-invariant hull of a Besov space via interpolation
Czech Academy of Sciences Publication Activity Database
Gogatishvili, Amiran; Pick, L.; Schneider, J.
2012-01-01
Roč. 25, č. 1 (2012), s. 267-283 ISSN 1139-1138 R&D Projects: GA ČR GA201/05/2033; GA ČR GA201/08/0383 Institutional research plan: CEZ:AV0Z10190503 Keywords : Besov spaces * Lorentz spaces * interpolation Subject RIV: BA - General Mathematics Impact factor: 0.377, year: 2012 http://www.springerlink.com/content/6v5hj295n6467176/
Slemp, Wesley C. H.; Kapania, Rakesh K.; Tessler, Alexander
2010-01-01
Computation of interlaminar stresses from the higher-order shear and normal deformable beam theory and the refined zigzag theory was performed using the Sinc method based on Interpolation of Highest Derivative. The Sinc method based on Interpolation of Highest Derivative was proposed as an efficient method for determining through-the-thickness variations of interlaminar stresses from one- and two-dimensional analysis by integration of the equilibrium equations of three-dimensional elasticity. However, the use of traditional equivalent single layer theories often results in inaccuracies near the boundaries and when the lamina have extremely large differences in material properties. Interlaminar stresses in symmetric cross-ply laminated beams were obtained by solving the higher-order shear and normal deformable beam theory and the refined zigzag theory with the Sinc method based on Interpolation of Highest Derivative. Interlaminar stresses and bending stresses from the present approach were compared with a detailed finite element solution obtained by ABAQUS/Standard. The results illustrate the ease with which the Sinc method based on Interpolation of Highest Derivative can be used to obtain the through-the-thickness distributions of interlaminar stresses from the beam theories. Moreover, the results indicate that the refined zigzag theory is a substantial improvement over the Timoshenko beam theory due to the piecewise continuous displacement field which more accurately represents interlaminar discontinuities in the strain field. The higher-order shear and normal deformable beam theory more accurately captures the interlaminar stresses at the ends of the beam because it allows transverse normal strain. However, the continuous nature of the displacement field requires a large number of monomial terms before the interlaminar stresses are computed as accurately as the refined zigzag theory.
International Nuclear Information System (INIS)
Cho, Sanghee; Grazioso, Ron; Zhang Nan; Aykac, Mehmet; Schmand, Matthias
2011-01-01
The main focus of our study is to investigate how the performance of digital timing methods is affected by sampling rate, anti-aliasing and signal interpolation filters. We used the Nyquist sampling theorem to address some basic questions such as what will be the minimum sampling frequencies? How accurate will the signal interpolation be? How do we validate the timing measurements? The preferred sampling rate would be as low as possible, considering the high cost and power consumption of high-speed analog-to-digital converters. However, when the sampling rate is too low, due to the aliasing effect, some artifacts are produced in the timing resolution estimations; the shape of the timing profile is distorted and the FWHM values of the profile fluctuate as the source location changes. Anti-aliasing filters are required in this case to avoid the artifacts, but the timing is degraded as a result. When the sampling rate is marginally over the Nyquist rate, a proper signal interpolation is important. A sharp roll-off (higher order) filter is required to separate the baseband signal from its replicates to avoid the aliasing, but in return the computation will be higher. We demonstrated the analysis through a digital timing study using fast LSO scintillation crystals as used in time-of-flight PET scanners. From the study, we observed that there is no significant timing resolution degradation down to 1.3 Ghz sampling frequency, and the computation requirement for the signal interpolation is reasonably low. A so-called sliding test is proposed as a validation tool checking constant timing resolution behavior of a given timing pick-off method regardless of the source location change. Lastly, the performance comparison for several digital timing methods is also shown.
Zhang, Dai; Hao, Shiqi; Zhao, Qingsong; Zhao, Qi; Wang, Lei; Wan, Xiongfeng
2018-03-01
Existing wavefront reconstruction methods are usually low in resolution, restricted by structure characteristics of the Shack Hartmann wavefront sensor (SH WFS) and the deformable mirror (DM) in the adaptive optics (AO) system, thus, resulting in weak homodyne detection efficiency for free space optical (FSO) communication. In order to solve this problem, we firstly validate the feasibility of liquid crystal spatial light modulator (LC SLM) using in an AO system. Then, wavefront reconstruction method based on wavelet fractal interpolation is proposed after self-similarity analysis of wavefront distortion caused by atmospheric turbulence. Fast wavelet decomposition is operated to multiresolution analyze the wavefront phase spectrum, during which soft threshold denoising is carried out. The resolution of estimated wavefront phase is then improved by fractal interpolation. Finally, fast wavelet reconstruction is taken to recover wavefront phase. Simulation results reflect the superiority of our method in homodyne detection. Compared with minimum variance estimation (MVE) method based on interpolation techniques, the proposed method could obtain superior homodyne detection efficiency with lower operation complexity. Our research findings have theoretical significance in the design of coherent FSO communication system.
Directory of Open Access Journals (Sweden)
Aihua Liu
2017-01-01
Full Text Available A method of direction-of-arrival (DOA estimation using array interpolation is proposed in this paper to increase the number of resolvable sources and improve the DOA estimation performance for coprime array configuration with holes in its virtual array. The virtual symmetric nonuniform linear array (VSNLA of coprime array signal model is introduced, with the conventional MUSIC with spatial smoothing algorithm (SS-MUSIC applied on the continuous lags in the VSNLA; the degrees of freedom (DoFs for DOA estimation are obviously not fully exploited. To effectively utilize the extent of DoFs offered by the coarray configuration, a compressing sensing based array interpolation algorithm is proposed. The compressing sensing technique is used to obtain the coarse initial DOA estimation, and a modified iterative initial DOA estimation based interpolation algorithm (IMCA-AI is then utilized to obtain the final DOA estimation, which maps the sample covariance matrix of the VSNLA to the covariance matrix of a filled virtual symmetric uniform linear array (VSULA with the same aperture size. The proposed DOA estimation method can efficiently improve the DOA estimation performance. The numerical simulations are provided to demonstrate the effectiveness of the proposed method.
Random Model Sampling: Making Craig Interpolation Work When It Should Not
Directory of Open Access Journals (Sweden)
Marat Akhin
2014-01-01
Full Text Available One of the most serious problems when doing program analyses is dealing with function calls. While function inlining is the traditional approach to this problem, it nonetheless suffers from the increase in analysis complexity due to the state space explosion. Craig interpolation has been successfully used in recent years in the context of bounded model checking to do function summarization which allows one to replace the complete function body with its succinct summary and, therefore, reduce the complexity. Unfortunately this technique can be applied only to a pair of unsatisfiable formulae.In this work-in-progress paper we present an approach to function summarization based on Craig interpolation that overcomes its limitation by using random model sampling. It captures interesting input/output relations, strengthening satisfiable formulae into unsatisfiable ones and thus allowing the use of Craig interpolation. Preliminary experiments show the applicability of this approach; in our future work we plan to do a full evaluation on real-world examples.
Homography Propagation and Optimization for Wide-Baseline Street Image Interpolation.
Nie, Yongwei; Zhang, Zhensong; Sun, Hanqiu; Su, Tan; Li, Guiqing
2017-10-01
Wide-baseline street image interpolation is useful but very challenging. Existing approaches either rely on heavyweight 3D reconstruction or computationally intensive deep networks. We present a lightweight and efficient method which uses simple homography computing and refining operators to estimate piecewise smooth homographies between input views. To achieve the goal, we show how to combine homography fitting and homography propagation together based on reliable and unreliable superpixel discrimination. Such a combination, other than using homography fitting only, dramatically increases the accuracy and robustness of the estimated homographies. Then, we integrate the concepts of homography and mesh warping, and propose a novel homography-constrained warping formulation which enforces smoothness between neighboring homographies by utilizing the first-order continuity of the warped mesh. This further eliminates small artifacts of overlapping, stretching, etc. The proposed method is lightweight and flexible, allows wide-baseline interpolation. It improves the state of the art and demonstrates that homography computation suffices for interpolation. Experiments on city and rural datasets validate the efficiency and effectiveness of our method.
Directory of Open Access Journals (Sweden)
Tsugio Fukuchi
2014-06-01
Full Text Available The finite difference method (FDM based on Cartesian coordinate systems can be applied to numerical analyses over any complex domain. A complex domain is usually taken to mean that the geometry of an immersed body in a fluid is complex; here, it means simply an analytical domain of arbitrary configuration. In such an approach, we do not need to treat the outer and inner boundaries differently in numerical calculations; both are treated in the same way. Using a method that adopts algebraic polynomial interpolations in the calculation around near-wall elements, all the calculations over irregular domains reduce to those over regular domains. Discretization of the space differential in the FDM is usually derived using the Taylor series expansion; however, if we use the polynomial interpolation systematically, exceptional advantages are gained in deriving high-order differences. In using the polynomial interpolations, we can numerically solve the Poisson equation freely over any complex domain. Only a particular type of partial differential equation, Poisson's equations, is treated; however, the arguments put forward have wider generality in numerical calculations using the FDM.
Interpolation Routines Assessment in ALS-Derived Digital Elevation Models for Forestry Applications
Directory of Open Access Journals (Sweden)
Antonio Luis Montealegre
2015-07-01
Full Text Available Airborne Laser Scanning (ALS is capable of estimating a variety of forest parameters using different metrics extracted from the normalized heights of the point cloud using a Digital Elevation Model (DEM. In this study, six interpolation routines were tested over a range of land cover and terrain roughness in order to generate a collection of DEMs with spatial resolution of 1 and 2 m. The accuracy of the DEMs was assessed twice, first using a test sample extracted from the ALS point cloud, second using a set of 55 ground control points collected with a high precision Global Positioning System (GPS. The effects of terrain slope, land cover, ground point density and pulse penetration on the interpolation error were examined stratifying the study area with these variables. In addition, a Classification and Regression Tree (CART analysis allowed the development of a prediction uncertainty map to identify in which areas DEMs and Airborne Light Detection and Ranging (LiDAR derived products may be of low quality. The Triangulated Irregular Network (TIN to raster interpolation method produced the best result in the validation process with the training data set while the Inverse Distance Weighted (IDW routine was the best in the validation with GPS (RMSE of 2.68 cm and RMSE of 37.10 cm, respectively.
A fast rank-reduction algorithm for three-dimensional seismic data interpolation
Jia, Yongna; Yu, Siwei; Liu, Lina; Ma, Jianwei
2016-09-01
Rank-reduction methods have been successfully used for seismic data interpolation and noise attenuation. However, highly intense computation is required for singular value decomposition (SVD) in most rank-reduction methods. In this paper, we propose a simple yet efficient interpolation algorithm, which is based on the Hankel matrix, for randomly missing traces. Following the multichannel singular spectrum analysis (MSSA) technique, we first transform the seismic data into a low-rank block Hankel matrix for each frequency slice. Then, a fast orthogonal rank-one matrix pursuit (OR1MP) algorithm is employed to minimize the low-rank constraint of the block Hankel matrix. In the new algorithm, only the left and right top singular vectors are needed to be computed, thereby, avoiding the complexity of computation required for SVD. Thus, we improve the calculation efficiency significantly. Finally, we anti-average the rank-reduction block Hankel matrix and obtain the reconstructed data in the frequency domain. Numerical experiments on 3D seismic data show that the proposed interpolation algorithm provides much better performance than the traditional MSSA algorithm in computational speed, especially for large-scale data processing.
Directory of Open Access Journals (Sweden)
Annalisa Di Piazza
2015-04-01
Full Text Available An exhaustive comparison among different spatial interpolation algorithms was carried out in order to derive annual and monthly air temperature maps for Sicily (Italy. Deterministic, data-driven and geostatistics algorithms were used, in some cases adding the elevation information and other physiographic variables to improve the performance of interpolation techniques and the reconstruction of the air temperature field. The dataset is given by air temperature data coming from 84 stations spread around the island of Sicily. The interpolation algorithms were optimized by using a subset of the available dataset, while the remaining subset was used to validate the results in terms of the accuracy and bias of the estimates. Validation results indicate that univariate methods, which neglect the information from physiographic variables, significantly entail the largest errors, while performances improve when such parameters are taken into account. The best results at the annual scale have been obtained using the the ordinary kriging of residuals from linear regression and from the artificial neural network algorithm, while, at the monthly scale, a Fourier-series algorithm has been used to downscale mean annual temperature to reproduce monthly values in the annual cycle.
Restoring the missing features of the corrupted speech using linear interpolation methods
Rassem, Taha H.; Makbol, Nasrin M.; Hasan, Ali Muttaleb; Zaki, Siti Syazni Mohd; Girija, P. N.
2017-10-01
One of the main challenges in the Automatic Speech Recognition (ASR) is the noise. The performance of the ASR system reduces significantly if the speech is corrupted by noise. In spectrogram representation of a speech signal, after deleting low Signal to Noise Ratio (SNR) elements, the incomplete spectrogram is obtained. In this case, the speech recognizer should make modifications to the spectrogram in order to restore the missing elements, which is one direction. In another direction, speech recognizer should be able to restore the missing elements due to deleting low SNR elements before performing the recognition. This is can be done using different spectrogram reconstruction methods. In this paper, the geometrical spectrogram reconstruction methods suggested by some researchers are implemented as a toolbox. In these geometrical reconstruction methods, the linear interpolation along time or frequency methods are used to predict the missing elements between adjacent observed elements in the spectrogram. Moreover, a new linear interpolation method using time and frequency together is presented. The CMU Sphinx III software is used in the experiments to test the performance of the linear interpolation reconstruction method. The experiments are done under different conditions such as different lengths of the window and different lengths of utterances. Speech corpus consists of 20 males and 20 females; each one has two different utterances are used in the experiments. As a result, 80% recognition accuracy is achieved with 25% SNR ratio.
Farr, W. M.; Mandel, I.; Stevens, D.
2015-06-01
Selection among alternative theoretical models given an observed dataset is an important challenge in many areas of physics and astronomy. Reversible-jump Markov chain Monte Carlo (RJMCMC) is an extremely powerful technique for performing Bayesian model selection, but it suffers from a fundamental difficulty and it requires jumps between model parameter spaces, but cannot efficiently explore both parameter spaces at once. Thus, a naive jump between parameter spaces is unlikely to be accepted in the Markov chain Monte Carlo (MCMC) algorithm and convergence is correspondingly slow. Here, we demonstrate an interpolation technique that uses samples from single-model MCMCs to propose intermodel jumps from an approximation to the single-model posterior of the target parameter space. The interpolation technique, based on a kD-tree data structure, is adaptive and efficient in modest dimensionality. We show that our technique leads to improved convergence over naive jumps in an RJMCMC, and compare it to other proposals in the literature to improve the convergence of RJMCMCs. We also demonstrate the use of the same interpolation technique as a way to construct efficient "global" proposal distributions for single-model MCMCs without prior knowledge of the structure of the posterior distribution, and discuss improvements that permit the method to be used in higher dimensional spaces efficiently.
Directory of Open Access Journals (Sweden)
Jinyang Song
2018-01-01
Full Text Available Many modulated signals exhibit a cyclostationarity property, which can be exploited in direction-of-arrival (DOA estimation to effectively eliminate interference and noise. In this paper, our aim is to integrate the cyclostationarity with the spatial domain and enable the algorithm to estimate more sources than sensors. However, DOA estimation with a sparse array is performed in the coarray domain and the holes within the coarray limit the usage of the complete coarray information. In order to use the complete coarray information to increase the degrees-of-freedom (DOFs, sparsity-aware-based methods and the difference coarray interpolation methods have been proposed. In this paper, the coarray interpolation technique is further explored with cyclostationary signals. Besides the difference coarray model and its corresponding Toeplitz completion formulation, we build up a sum coarray model and formulate a Hankel completion problem. In order to further improve the performance of the structured matrix completion, we define the spatial spectrum sampling operations and the derivative (conjugate correlation subspaces, which can be exploited to construct orthogonal constraints for the autocorrelation vectors in the coarray interpolation problem. Prior knowledge of the source interval can also be incorporated into the problem. Simulation results demonstrate that the additional constraints contribute to a remarkable performance improvement.
Directory of Open Access Journals (Sweden)
S. W. Lyon
2008-10-01
Full Text Available Spatial patterns of water chemistry along stream networks can be quantified using synoptic or "snapshot" sampling. The basic idea is to sample stream water at many points over a relatively short period of time. Even for intense sampling campaigns, the number of sample points is limited and interpolation methods, like kriging, are commonly used to produce continuous maps of water chemistry based on the point observations from the synoptic sampling. Interpolated concentrations are influenced heavily by how distance between points along the stream network is defined. In this study, we investigate different ways to define distance and test these based on data from a snapshot sampling campaign in a 37-km^{2} watershed in the Catskill Mountains region (New York State. Three distance definitions (or metrics were compared: Euclidean or straight-line distance, in-stream distance, and in-stream distance adjusted according characteristics of the local contributing area, i.e., an adjusted in-stream distance. Using the adjusted distance metric resulted in a lower cross-validation error of the interpolated concentrations, i.e., a better agreement of kriging results with measurements, than the other distance definitions. The adjusted distance metric can also be used in an exploratory manner to test which landscape characteristics are most influential for the spatial patterns of stream water chemistry and, thus, to target future investigations to gain process-based understanding of in-stream chemistry dynamics.
Yang, Shun-hua; Zhang, Hai-tao; Guo, Long; Ren, Yan
2015-06-01
Relative elevation and stream power index were selected as auxiliary variables based on correlation analysis for mapping soil organic matter. Geographically weighted regression Kriging (GWRK) and regression Kriging (RK) were used for spatial interpolation of soil organic matter and compared with ordinary Kriging (OK), which acts as a control. The results indicated that soil or- ganic matter was significantly positively correlated with relative elevation whilst it had a significantly negative correlation with stream power index. Semivariance analysis showed that both soil organic matter content and its residuals (including ordinary least square regression residual and GWR resi- dual) had strong spatial autocorrelation. Interpolation accuracies by different methods were esti- mated based on a data set of 98 validation samples. Results showed that the mean error (ME), mean absolute error (MAE) and root mean square error (RMSE) of RK were respectively 39.2%, 17.7% and 20.6% lower than the corresponding values of OK, with a relative-improvement (RI) of 20.63. GWRK showed a similar tendency, having its ME, MAE and RMSE to be respectively 60.6%, 23.7% and 27.6% lower than those of OK, with a RI of 59.79. Therefore, both RK and GWRK significantly improved the accuracy of OK interpolation of soil organic matter due to their in- corporation of auxiliary variables. In addition, GWRK performed obviously better than RK did in this study, and its improved performance should be attributed to the consideration of sample spatial locations.
The multiple interpolation de La Vallée Poussin problem
Directory of Open Access Journals (Sweden)
Valentin V. Napalkov
2015-03-01
Full Text Available This article is concerned with the solving of multiple interpolation de La Vallée Poussin problem for generalized convolution operator. Particular attention is paid to the proving of the sequential sufficiency of the set of solutions of the generalized convolution operator characteristic equation. In the generalized Bargmann–Fock space the adjoint operator of multiplication by the variable $z$ is the generalized differential operator. Using this operator we introduce the generalized shift and generalized convolution operators. Applying the chain of equivalent assertions we obtain the fact that the multiple interpolation de La Vallée Poussin problem is solvable if and only if the composition of generalized convolution operator with multiplication by the fixed entire function $\\psi(z$ is surjective. Zeros of the function $\\psi(z$ are the nodes of interpolation. The surjectivity of composition of the generalized convolution operator with the multiplication comes down to the proof of the sequential sufficiency of the set of zeros of a generalized convolution operator characteristic function in the set of solutions of the generalized convolution operator with the characteristic function $\\psi(z$. In the proof of the sequential sufficiency it became necessary to consider the relation of eigenfunctions for different values of $\\mu_i.$ The eigenfunction with great value of $\\mu_i$ tends to infinity faster than eigenfunction with a lower value for $z$ tends to infinity. The derivative of the eigenfunction of higher order tends to infinity faster than lower-order derivatives with the same values of $\\mu_i$. A significant role is played by the fact that the kernel of the generalized convolution operator with characteristic function $\\psi(z$ is a finite sum of its eigenfunction and its derivatives. Using the Fischer representation, Dieudonne–Schwartz theorem and Michael's theorem on the existence of a continuous right inverse we obtain that
An edge-directed interpolation method for fetal spine MR images.
Yu, Shaode; Zhang, Rui; Wu, Shibin; Hu, Jiani; Xie, Yaoqin
2013-10-10
Fetal spinal magnetic resonance imaging (MRI) is a prenatal routine for proper assessment of fetus development, especially when suspected spinal malformations occur while ultrasound fails to provide details. Limited by hardware, fetal spine MR images suffer from its low resolution.High-resolution MR images can directly enhance readability and improve diagnosis accuracy. Image interpolation for higher resolution is required in clinical situations, while many methods fail to preserve edge structures. Edge carries heavy structural messages of objects in visual scenes for doctors to detect suspicions, classify malformations and make correct diagnosis. Effective interpolation with well-preserved edge structures is still challenging. In this paper, we propose an edge-directed interpolation (EDI) method and apply it on a group of fetal spine MR images to evaluate its feasibility and performance. This method takes edge messages from Canny edge detector to guide further pixel modification. First, low-resolution (LR) images of fetal spine are interpolated into high-resolution (HR) images with targeted factor by bi-linear method. Then edge information from LR and HR images is put into a twofold strategy to sharpen or soften edge structures. Finally a HR image with well-preserved edge structures is generated. The HR images obtained from proposed method are validated and compared with that from other four EDI methods. Performances are evaluated from six metrics, and subjective analysis of visual quality is based on regions of interest (ROI). All these five EDI methods are able to generate HR images with enriched details. From quantitative analysis of six metrics, the proposed method outperforms the other four from signal-to-noise ratio (SNR), peak signal-to-noise ratio (PSNR), structure similarity index (SSIM), feature similarity index (FSIM) and mutual information (MI) with seconds-level time consumptions (TC). Visual analysis of ROI shows that the proposed method maintains
Energy Technology Data Exchange (ETDEWEB)
Pimentel, David A. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Sheppard, Daniel G. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2018-02-01
It was recently demonstrated that EOSPAC 6 continued to incorrectly create and interpolate pre-inverted SESAME data tables after the release of version 6.3.2beta.2. Significant interpolation pathologies were discovered to occur when EOSPAC 6's host software enabled pre-inversion with the EOS_INVERT_AT_SETUP option. This document describes a solution that uses data transformations found in EOSPAC 5 and its predecessors. The numerical results and performance characteristics of both the default and pre-inverted interpolation modes in both EOSPAC 6.3.2beta.2 and the fixed logic of EOSPAC 6.4.0beta.1 are presented herein, and the latter software release is shown to produce significantly-improved numerical results for the pre-inverted interpolation mode.
International Nuclear Information System (INIS)
Gao Wen-Wu; Wang Zhi-Gang
2014-01-01
Based on the multiquadric trigonometric B-spline quasi-interpolant, this paper proposes a meshless scheme for some partial differential equations whose solutions are periodic with respect to the spatial variable. This scheme takes into account the periodicity of the analytic solution by using derivatives of a periodic quasi-interpolant (multiquadric trigonometric B-spline quasi-interpolant) to approximate the spatial derivatives of the equations. Thus, it overcomes the difficulties of the previous schemes based on quasi-interpolation (requiring some additional boundary conditions and yielding unwanted high-order discontinuous points at the boundaries in the spatial domain). Moreover, the scheme also overcomes the difficulty of the meshless collocation methods (i.e., yielding a notorious ill-conditioned linear system of equations for large collocation points). The numerical examples that are presented at the end of the paper show that the scheme provides excellent approximations to the analytic solutions. (general)
Beudaert , Xavier; Lavernhe , Sylvain; Tournier , Christophe
2014-01-01
International audience; In the common machining process of free-form surfaces, CAM software generates approximated tool paths because of the input tool path format of the industrial CNC. Then, marks on finished surfaces may appear due to non smooth feedrate planning during interpolation. The Direct Trajectory Interpolation on the Surface (DTIS) method allows managing the tool path geometry and the kinematical parameters to achieve higher productivity and a better surface quality. Machining ex...
Ducru, Pablo; Josey, Colin; Dibert, Karia; Sobes, Vladimir; Forget, Benoit; Smith, Kord
2017-04-01
This article establishes a new family of methods to perform temperature interpolation of nuclear interactions cross sections, reaction rates, or cross sections times the energy. One of these quantities at temperature T is approximated as a linear combination of quantities at reference temperatures (Tj). The problem is formalized in a cross section independent fashion by considering the kernels of the different operators that convert cross section related quantities from a temperature T0 to a higher temperature T - namely the Doppler broadening operation. Doppler broadening interpolation of nuclear cross sections is thus here performed by reconstructing the kernel of the operation at a given temperature T by means of linear combination of kernels at reference temperatures (Tj). The choice of the L2 metric yields optimal linear interpolation coefficients in the form of the solutions of a linear algebraic system inversion. The optimization of the choice of reference temperatures (Tj) is then undertaken so as to best reconstruct, in the L∞ sense, the kernels over a given temperature range [Tmin ,Tmax ]. The performance of these kernel reconstruction methods is then assessed in light of previous temperature interpolation methods by testing them upon isotope 238U. Temperature-optimized free Doppler kernel reconstruction significantly outperforms all previous interpolation-based methods, achieving 0.1% relative error on temperature interpolation of 238U total cross section over the temperature range [ 300 K , 3000 K ] with only 9 reference temperatures.
Directory of Open Access Journals (Sweden)
Silvio Jorge Coelho Simões
2012-08-01
Full Text Available The reference evapotranspiration is an important hydrometeorological variable; its measurement is scarce in large portions of the Brazilian territory, what demands the search for alternative methods and techniques for its quantification. In this sense, the present work investigated a method for the spatialization of the reference evapotranspiration using the geostatistical method of kriging, in regions with limited data and hydrometeorological stations. The monthly average reference evapotranspiration was calculated by the Penman-Monteith-FAO equation, based on data from three weather stations located in southern Minas Gerais (Itajubá, Lavras and Poços de Caldas, and subsequently interpolated by ordinary point kriging using the approach "calculate and interpolate." The meteorological data for a fourth station (Três Corações located within the area of interpolation were used to validate the reference evapotranspiration interpolated spatially. Due to the reduced number of stations and the consequent impossibility of carrying variographic analyzes, correlation coefficient (r, index of agreement (d, medium bias error (MBE, root mean square error (RMSE and t-test were used for comparison between the calculated and interpolated reference evapotranspiration for the Três Corações station. The results of this comparison indicated that the spatial kriging procedure, even using a few stations, allows to interpolate satisfactorily the reference evapotranspiration, therefore, it is an important tool for agricultural and hydrological applications in regions with lack of data.
Energy Technology Data Exchange (ETDEWEB)
Castillo M, J. A
1999-07-01
Nuclear Data Bank generation, is a process in which a great amount of resources is required, both computing and humans. If it is taken into account that at some times it is necessary to create a great amount of those, it is convenient to have a reliable tool that generates Data Banks with the lesser resources, in the least possible time and with a very good approximation. In this work are shown the results obtained during the development of INTPOLBI code, use to generate Nuclear Data Banks employing bicubic polynominal interpolation, taking as independent variables the uranium and gadolinia percents. Two proposal were worked, applying in both cases the finite element method, using one element with 16 nodes to carry out the interpolation. In the first proposals the canonic base was employed, to obtain the interpolating polynomial and later, the corresponding linear equation systems. In the solution of this systems the Gaussian elimination methods with partial pivot was applied. In the second case, the Newton base was used to obtain the mentioned system, resulting in a triangular inferior matrix, which structure, applying elemental operations, to obtain a blocks diagonal matrix, with special characteristics and easier to work with. For the validation tests, a comparison was made between the values obtained with INTPOLBI and INTERTEG (create at the Instituto de Investigaciones Electricas (MX) with the same purpose) codes, and Data Banks created through the conventional process, that is, with nuclear codes normally used. Finally, it is possible to conclude that the Nuclear Data Banks generated with INTPOLBI code constitute a very good approximation that, even though do not wholly replace conventional process, however are helpful in cases when it is necessary to create a great amount of Data Banks.
Mohammadi, Seyedeh Atefeh; Azadi, Majid; Rahmani, Morteza
2017-08-01
All numerical weather prediction (NWP) models inherently have substantial biases, especially in the forecast of near-surface weather variables. Statistical methods can be used to remove the systematic error based on historical bias data at observation stations. However, many end users of weather forecasts need bias corrected forecasts at locations that scarcely have any historical bias data. To circumvent this limitation, the bias of surface temperature forecasts on a regular grid covering Iran is removed, by using the information available at observation stations in the vicinity of any given grid point. To this end, the running mean error method is first used to correct the forecasts at observation stations, then four interpolation methods including inverse distance squared weighting with constant lapse rate (IDSW-CLR), Kriging with constant lapse rate (Kriging-CLR), gradient inverse distance squared with linear lapse rate (GIDS-LR), and gradient inverse distance squared with lapse rate determined by classification and regression tree (GIDS-CART), are employed to interpolate the bias corrected forecasts at neighboring observation stations to any given location. The results show that all four interpolation methods used do reduce the model error significantly, but Kriging-CLR has better performance than the other methods. For Kriging-CLR, root mean square error (RMSE) and mean absolute error (MAE) were decreased by 26% and 29%, respectively, as compared to the raw forecasts. It is found also, that after applying any of the proposed methods, unlike the raw forecasts, the bias corrected forecasts do not show spatial or temporal dependency.
Second-order sign-preserving conservative interpolation (remapping) on general grids
Margolin, L G
2003-01-01
An accurate conservative interpolation (remapping) algorithm is an essential component of most arbitrary Lagrangian-Eulerian (ALE) methods. In this paper we describe a local remapping algorithm for a positive scalar function. This algorithm is second-order accurate, conservative, and sign preserving. The algorithm is based on estimating the mass exchanged between cells at their common interface, and so is equally applicable to structured and unstructured grids. We construct the algorithm in a series of steps, clearly delineating the assumptions and errors made at each step. We validate our theory with a suite of numerical examples, analyzing the results from the viewpoint of accuracy and order of convergence.
Energy Technology Data Exchange (ETDEWEB)
Suescun D, D.; Figueroa J, J. H. [Pontificia Universidad Javeriana Cali, Departamento de Ciencias Naturales y Matematicas, Calle 18 No. 118-250, Cali, Valle del Cauca (Colombia); Rodriguez R, K. C.; Villada P, J. P., E-mail: dsuescun@javerianacali.edu.co [Universidad del Valle, Departamento de Fisica, Calle 13 No. 100-00, Cali, Valle del Cauca (Colombia)
2015-09-15
A new method to solve numerically the inverse equation of punctual kinetics without using Lagrange interpolating polynomial is formulated; this method uses a polynomial approximation with N points based on a process of recurrence for simulating different forms of nuclear power. The results show a reliable accuracy. Furthermore, the method proposed here is suitable for real-time measurements of reactivity, with step sizes of calculations greater that Δt = 0.3 s; due to its precision can be used to implement a digital meter of reactivity in real time. (Author)
Interpolation of text from the Castilian Macer Floridus in ms. II-3063 of the Real Biblioteca
Directory of Open Access Journals (Sweden)
Jesús Pensado Figueiras
2012-12-01
Full Text Available Textual analysis of codex II-3063 of the Real Biblioteca (Royal Palace, Madrid has located a new Castilian version of several passages from the Latin work on herbalism De viribus herbarum, better known as Macer Floridus. The text of codex II-3063 contains interpolated descriptions of the virtues of six plants outlined in the Latin text and confirms the hypothesis of a single translation from Latin into a peninsular Romance language, from which later versions would be made until the current number of texts of the Macer Floridus in Catalan, Aragonese, and Castilian was reached.
Cubic scaling algorithms for RPA correlation using interpolative separable density fitting
Lu, Jianfeng; Thicke, Kyle
2017-12-01
We present a new cubic scaling algorithm for the calculation of the RPA correlation energy. Our scheme splits up the dependence between the occupied and virtual orbitals in χ0 by use of Cauchy's integral formula. This introduces an additional integral to be carried out, for which we provide a geometrically convergent quadrature rule. Our scheme also uses the newly developed Interpolative Separable Density Fitting algorithm to further reduce the computational cost in a way analogous to that of the Resolution of Identity method.
Maximum Feedrate Interpolator for Multi-axis CNC Machining with Jerk Constraints
Beudaert , Xavier; Lavernhe , Sylvain; Tournier , Christophe
2012-01-01
A key role of the CNC is to perform the feedrate interpolation which means to generate the setpoints for each machine tool axis. The aim of the VPOp algorithm is to make maximum use of the machine tool respecting both tangential and axis jerk on rotary and linear axes. The developed algorithm uses an iterative constraints intersection approach. At each sampling period, all the constraints given by each axis are expressed and by intersecting all of them the allowable interval for the next poin...
Marson, Avishai; Stern, Adrian
2015-05-01
One of the main limitations of horizontal parallax autostereoscopic displays is the horizontal resolution loss due the need to repartition the pixels of the display panel among the multiple views. Recently we have shown that this problem can be alleviated by applying a color sub-pixel rendering technique1. Interpolated views are generated by down-sampling the panel pixels at sub-pixel level, thus increasing the number of views. The method takes advantage of lower acuity of the human eye to chromatic resolution. Here we supply further support of the technique by analyzing the spectra of the subsampled images.
Alpha brain wave production as an interpolated task in a Brown-Peterson paradigm.
Bridges, C C; Rollings, H E
1976-04-01
Rehearsal, backward counting, and production of alpha brain-waves were used as interpolated tasks in a Brown-Peterson paradigm to determine their effect upon verbal retention. A within-subjects design was used in which trained subjects were told on a given trial either to produce alpha rhythm, mentally rehearse, or count backward following presentation of a CCC trigram. Results for the backward-counting condition duplicate, for the retention intervals used, the shape of the classic Peterson and Peterson forgetting curve but indicate little loss of memory in either the rehearsal or alpha conditions. No siginificant difference was found between the alpha production and rehearsal conditions.
International Nuclear Information System (INIS)
Ovchinnikov, V I
2014-01-01
In the paper, a new description of the generalized Lions-Peetre method of means is found, which enables one to evaluate the interpolation orbits of spaces constructed by this method. The list of these spaces includes all Lorentz spaces with functional parameters, Orlicz spaces, and spaces close to them. This leads in turn to new optimal embedding theorems for Sobolev spaces produced using the Lions-Peetre construction in rearrangement invariant spaces. It turns out that the optimal space of the embedding is also a generalized Lions-Peetre space whose parameters are explicitly evaluated. Bibliography: 18 titles
Directory of Open Access Journals (Sweden)
Shu-Cherng Fang
2010-08-01
Full Text Available We compare univariate L1 interpolating splines calculated on 5-point windows, on 7-point windows and on global data sets using four different spline functionals, namely, ones based on the second derivative, the first derivative, the function value and the antiderivative. Computational results indicate that second-derivative-based 5-point-window L1 splines preserve shape as well as or better than the other types of L1 splines. To calculate second-derivative-based 5-point-window L1 splines, we introduce an analysis-based, parallelizable algorithm. This algorithm is orders of magnitude faster than the previously widely used primal affine algorithm.
Kirkpatrick, J. C.
1976-01-01
A tabulation of selected altitude-correlated values of pressure, density, speed of sound, and coefficient of viscosity for each of six models of the atmosphere is presented in block data format. Interpolation for the desired atmospheric parameters is performed by using cubic spline functions. The recursive relations necessary to compute the cubic spline function coefficients are derived and implemented in subroutine form. Three companion subprograms, which form the preprocessor and processor, are also presented. These subprograms, together with the data element, compose the spline fit atmosphere package. Detailed FLOWGM flow charts and FORTRAN listings of the atmosphere package are presented in the appendix.
Natural spline interpolation and exponential parameterization for length estimation of curves
Kozera, R.; Wilkołazka, M.
2017-07-01
This paper tackles the problem of estimating a length of a regular parameterized curve γ from an ordered sample of interpolation points in arbitrary Euclidean space by a natural spline. The corresponding tabular parameters are not given and are approximated by the so-called exponential parameterization (depending on λ ∈ [0, 1]). The respective convergence orders α(λ) for estimating length of γ are established for curves sampled more-or-less uniformly. The numerical experiments confirm a slow convergence orders α(λ) = 2 for all λ ∈ [0, 1) and a cubic order α(1) = 3 once natural spline is used.
Interpolation between multi-dimensional histograms using a new non-linear moment morphing method
Energy Technology Data Exchange (ETDEWEB)
Baak, M., E-mail: max.baak@cern.ch [CERN, CH-1211 Geneva 23 (Switzerland); Gadatsch, S., E-mail: stefan.gadatsch@nikhef.nl [Nikhef, PO Box 41882, 1009 DB Amsterdam (Netherlands); Harrington, R. [School of Physics and Astronomy, University of Edinburgh, Mayfield Road, Edinburgh, EH9 3JZ, Scotland (United Kingdom); Verkerke, W. [Nikhef, PO Box 41882, 1009 DB Amsterdam (Netherlands)
2015-01-21
A prescription is presented for the interpolation between multi-dimensional distribution templates based on one or multiple model parameters. The technique uses a linear combination of templates, each created using fixed values of the model's parameters and transformed according to a specific procedure, to model a non-linear dependency on model parameters and the dependency between them. By construction the technique scales well with the number of input templates used, which is a useful feature in modern day particle physics, where a large number of templates are often required to model the impact of systematic uncertainties.
Interpolation between multi-dimensional histograms using a new non-linear moment morphing method
International Nuclear Information System (INIS)
Baak, M.; Gadatsch, S.; Harrington, R.; Verkerke, W.
2015-01-01
A prescription is presented for the interpolation between multi-dimensional distribution templates based on one or multiple model parameters. The technique uses a linear combination of templates, each created using fixed values of the model's parameters and transformed according to a specific procedure, to model a non-linear dependency on model parameters and the dependency between them. By construction the technique scales well with the number of input templates used, which is a useful feature in modern day particle physics, where a large number of templates are often required to model the impact of systematic uncertainties
Interpolation between multi-dimensional histograms using a new non-linear moment morphing method
Baak, Max; Harrington, Robert; Verkerke, Wouter
2014-01-01
A prescription is presented for the interpolation between multi-dimensional distribution templates based on one or multiple model parameters. The technique uses a linear combination of templates, each created using fixed values of the model's parameters and transformed according to a specific procedure, to model a non-linear dependency on model parameters and the dependency between them. By construction the technique scales well with the number of input templates used, which is a useful feature in modern day particle physics, where a large number of templates is often required to model the impact of systematic uncertainties.
Interpolation between multi-dimensional histograms using a new non-linear moment morphing method
Baak, Max; Harrington, Robert; Verkerke, Wouter
2015-01-01
A prescription is presented for the interpolation between multi-dimensional distribution templates based on one or multiple model parameters. The technique uses a linear combination of templates, each created using fixed values of the model's parameters and transformed according to a specific procedure, to model a non-linear dependency on model parameters and the dependency between them. By construction the technique scales well with the number of input templates used, which is a useful feature in modern day particle physics, where a large number of templates is often required to model the impact of systematic uncertainties.
On Interpolation Functions of the Generalized Twisted (h,q-Euler Polynomials
Directory of Open Access Journals (Sweden)
Kyoung Ho Park
2009-01-01
Full Text Available The aim of this paper is to construct p-adic twisted two-variable Euler-(h,q-L-functions, which interpolate generalized twisted (h,q-Euler polynomials at negative integers. In this paper, we treat twisted (h,q-Euler numbers and polynomials associated with p-adic invariant integral on ℤp. We will construct two-variable twisted (h,q-Euler-zeta function and two-variable (h,q-L-function in Complex s-plane.
Directory of Open Access Journals (Sweden)
Elmira Ashpazzadeh
2018-04-01
Full Text Available A numerical technique based on the Hermite interpolant multiscaling functions is presented for the solution of Convection-diusion equations. The operational matrices of derivative, integration and product are presented for multiscaling functions and are utilized to reduce the solution of linear Convection-diusion equation to the solution of algebraic equations. Because of sparsity of these matrices, this method is computationally very attractive and reduces the CPU time and computer memory. Illustrative examples are included to demonstrate the validity and applicability of the new technique.
Lagrange polynomial interpolation method applied in the calculation of the J({xi},{beta}) function
Energy Technology Data Exchange (ETDEWEB)
Fraga, Vinicius Munhoz; Palma, Daniel Artur Pinheiro [Centro Federal de Educacao Tecnologica de Quimica de Nilopolis, RJ (Brazil)]. E-mails: munhoz.vf@gmail.com; dpalma@cefeteq.br; Martinez, Aquilino Senra [Universidade Federal do Rio de Janeiro (UFRJ), RJ (Brazil). Coordenacao dos Programas de Pos-graduacao de Engenharia (COPPE) (COPPE). Programa de Engenharia Nuclear]. E-mail: aquilino@lmp.ufrj.br
2008-07-01
The explicit dependence of the Doppler broadening function creates difficulties in the obtaining an analytical expression for J function . The objective of this paper is to present a method for the quick and accurate calculation of J function based on the recent advances in the calculation of the Doppler broadening function and on a systematic analysis of its integrand. The methodology proposed, of a semi-analytical nature, uses the Lagrange polynomial interpolation method and the Frobenius formulation in the calculation of Doppler broadening function . The results have proven satisfactory from the standpoint of accuracy and processing time. (author)
Lagrange polynomial interpolation method applied in the calculation of the J(ξ,β) function
International Nuclear Information System (INIS)
Fraga, Vinicius Munhoz; Palma, Daniel Artur Pinheiro; Martinez, Aquilino Senra
2008-01-01
The explicit dependence of the Doppler broadening function creates difficulties in the obtaining an analytical expression for J function . The objective of this paper is to present a method for the quick and accurate calculation of J function based on the recent advances in the calculation of the Doppler broadening function and on a systematic analysis of its integrand. The methodology proposed, of a semi-analytical nature, uses the Lagrange polynomial interpolation method and the Frobenius formulation in the calculation of Doppler broadening function . The results have proven satisfactory from the standpoint of accuracy and processing time. (author)
Vnukov, A. A.; Shershnev, M. B.
2018-01-01
The aim of this work is the software implementation of three image scaling algorithms using parallel computations, as well as the development of an application with a graphical user interface for the Windows operating system to demonstrate the operation of algorithms and to study the relationship between system performance, algorithm execution time and the degree of parallelization of computations. Three methods of interpolation were studied, formalized and adapted to scale images. The result of the work is a program for scaling images by different methods. Comparison of the quality of scaling by different methods is given.
[The dentist on the DVI team--interdisciplinary philosophy of the Interpol DVI team].
De Valck, Eddy
2005-01-01
Recent experiences from mass disasters have shown that a structured and uniform methodology is the only good way to come to quick and correct results in disaster victim identification. It is therefore mandatory that everyone included in such operations is not only conscious of the procedures, but also willing to apply them. This can only be achieved by providing adequate information and training to all personnel involved, tactical and technical. It seems thus not unreasonable that in the future, DVI Interpol would only be willing to work with accredited personnel and DVI teams, in order to optimize the results of disaster victim identification operations.
Franco, Ademir; Thevissen, Patrick; Coudyzer, Walter; Develter, Wim; Van de Voorde, Wim; Oyen, Raymond; Vandermeulen, Dirk; Jacobs, Reinhilde; Willems, Guy
2013-05-01
Virtual autopsy is a medical imaging technique, using full body computed tomography (CT), allowing for a noninvasive and permanent observation of all body parts. For dental identification clinically and radiologically observed ante-mortem (AM) and post-mortem (PM) oral identifiers are compared. The study aimed to verify if a PM dental charting can be performed on virtual reconstructions of full-body CT's using the Interpol dental codes. A sample of 103 PM full-body CT's was collected from the forensic autopsy files of the Department of Forensic Medicine University Hospitals, KU Leuven, Belgium. For validation purposes, 3 of these bodies underwent a complete dental autopsy, a dental radiological and a full-body CT examination. The bodies were scanned in a Siemens Definition Flash CT Scanner (Siemens Medical Solutions, Germany). The images were examined on 8- and 12-bit screen resolution as three-dimensional (3D) reconstructions and as axial, coronal and sagittal slices. InSpace(®) (Siemens Medical Solutions, Germany) software was used for 3D reconstruction. The dental identifiers were charted on pink PM Interpol forms (F1, F2), using the related dental codes. Optimal dental charting was obtained by combining observations on 3D reconstructions and CT slices. It was not feasible to differentiate between different kinds of dental restoration materials. The 12-bit resolution enabled to collect more detailed evidences, mainly related to positions within a tooth. Oral identifiers, not implemented in the Interpol dental coding were observed. Amongst these, the observed (3D) morphological features of dental and maxillofacial structures are important identifiers. The latter can become particularly more relevant towards the future, not only because of the inherent spatial features, yet also because of the increasing preventive dental treatment, and the decreasing application of dental restorations. In conclusion, PM full-body CT examinations need to be implemented in the
Cut Elimination, Identity Elimination, and Interpolation in Super-Belnap Logics
Czech Academy of Sciences Publication Activity Database
Přenosil, Adam
2017-01-01
Roč. 105, č. 6 (2017), s. 1255-1289 ISSN 0039-3215 R&D Projects: GA ČR GBP202/12/G061 EU Projects: European Commission(XE) 689176 - SYSMICS Institutional support: RVO:67985807 Keywords : Super-Belnap logics * Dunn–Belnap logic * Logic of Paradox * Strong Kleene logic * Exactly True Logic * Gentzen calculus * Cut elimination * Identity elimination * Interpolation Subject RIV: BA - General Mathematics OBOR OECD: Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8) Impact factor: 0.589, year: 2016
Garcia, Matthew; Peters-Lidard, Christa D.; Goodrich, David C.
2008-05-01
Inaccuracy in spatially distributed precipitation fields can contribute significantly to the uncertainty of hydrological states and fluxes estimated from land surface models. This paper examines the results of selected interpolation methods for both convective and mixed/stratiform events that occurred during the North American monsoon season over a dense gauge network at the U.S. Department of Agriculture Agricultural Research Service Walnut Gulch Experimental Watershed in the southwestern United States. The spatial coefficient of variation for the precipitation field is employed as an indicator of event morphology, and a gauge clustering factor CF is formulated as a new, scale-independent measure of network organization. We consider that CF 0 (clustering in the gauge network) will produce errors because of reduced areal representation of the precipitation field. Spatial interpolation is performed using both inverse-distance-weighted (IDW) and multiquadric-biharmonic (MQB) methods. We employ ensembles of randomly selected network subsets for the statistical evaluation of interpolation errors in comparison with the observed precipitation. The magnitude of interpolation errors and differences in accuracy between interpolation methods depend on both the density and the geometrical organization of the gauge network. Generally, MQB methods outperform IDW methods in terms of interpolation accuracy under all conditions, but it is found that the order of the IDW method is important to the results and may, under some conditions, be just as accurate as the MQB method. In almost all results it is demonstrated that the inverse-distance-squared method for spatial interpolation, commonly employed in operational analyses and for engineering assessments, is inferior to the ID-cubed method, which is also more computationally efficient than the MQB method in studies of large networks.
Directory of Open Access Journals (Sweden)
Shaofeng Wang
2017-05-01
Full Text Available Mineral reserve estimation and mining design depend on a precise modeling of the mineralized deposit. A multi-step interpolation algorithm, including 1D biharmonic spline estimator for interpolating floor altitudes, 2D nearest neighbor, linear, natural neighbor, cubic, biharmonic spline, inverse distance weighted, simple kriging, and ordinary kriging interpolations for grade distribution on the two vertical sections at roadways, and 3D linear interpolation for grade distribution between sections, was proposed to build a 3D grade distribution model of the mineralized seam in a longwall mining panel with a U-shaped layout having two roadways at both sides. Compared to field data from exploratory boreholes, this multi-step interpolation using a natural neighbor method shows an optimal stability and a minimal difference between interpolation and field data. Using this method, the 97,576 m3 of bauxite, in which the mass fraction of Al2O3 (Wa and the mass ratio of Al2O3 to SiO2 (Wa/s are 61.68% and 27.72, respectively, was delimited from the 189,260 m3 mineralized deposit in the 1102 longwall mining panel in the Wachangping mine, Southwest China. The mean absolute errors, the root mean squared errors and the relative standard deviations of errors between interpolated data and exploratory grade data at six boreholes are 2.544, 2.674, and 32.37% of Wa; and 1.761, 1.974, and 67.37% of Wa/s, respectively. The proposed method can be used for characterizing the grade distribution in a mineralized seam between two roadways at both sides of a longwall mining panel.
Bogdanov, Alexander; Degtyarev, Alexander; Khramushin, Vasily; Shichkina, Yulia
2018-02-01
Stages of direct computational experiments in hydromechanics based on tensor mathematics tools are represented by conditionally independent mathematical models for calculations separation in accordance with physical processes. Continual stage of numerical modeling is constructed on a small time interval in a stationary grid space. Here coordination of continuity conditions and energy conservation is carried out. Then, at the subsequent corpuscular stage of the computational experiment, kinematic parameters of mass centers and surface stresses at the boundaries of the grid cells are used in modeling of free unsteady motions of volume cells that are considered as independent particles. These particles can be subject to vortex and discontinuous interactions, when restructuring of free boundaries and internal rheological states has place. Transition from one stage to another is provided by interpolation operations of tensor mathematics. Such interpolation environment formalizes the use of physical laws for mechanics of continuous media modeling, provides control of rheological state and conditions for existence of discontinuous solutions: rigid and free boundaries, vortex layers, their turbulent or empirical generalizations.
Mapping snow depth return levels: smooth spatial modeling versus station interpolation
Directory of Open Access Journals (Sweden)
J. Blanchet
2010-12-01
Full Text Available For adequate risk management in mountainous countries, hazard maps for extreme snow events are needed. This requires the computation of spatial estimates of return levels. In this article we use recent developments in extreme value theory and compare two main approaches for mapping snow depth return levels from in situ measurements. The first one is based on the spatial interpolation of pointwise extremal distributions (the so-called Generalized Extreme Value distribution, GEV henceforth computed at station locations. The second one is new and based on the direct estimation of a spatially smooth GEV distribution with the joint use of all stations. We compare and validate the different approaches for modeling annual maximum snow depth measured at 100 sites in Switzerland during winters 1965–1966 to 2007–2008. The results show a better performance of the smooth GEV distribution fitting, in particular where the station network is sparser. Smooth return level maps can be computed from the fitted model without any further interpolation. Their regional variability can be revealed by removing the altitudinal dependent covariates in the model. We show how return levels and their regional variability are linked to the main climatological patterns of Switzerland.
Resistor mesh model of a spherical head: part 1: applications to scalp potential interpolation.
Chauveau, N; Morucci, J P; Franceries, X; Celsis, P; Rigaud, B
2005-11-01
A resistor mesh model (RMM) has been implemented to describe the electrical properties of the head and the configuration of the intracerebral current sources by simulation of forward and inverse problems in electroencephalogram/event related potential (EEG/ERP) studies. For this study, the RMM representing the three basic tissues of the human head (brain, skull and scalp) was superimposed on a spherical volume mimicking the head volume: it included 43 102 resistances and 14 123 nodes. The validation was performed with reference to the analytical model by consideration of a set of four dipoles close to the cortex. Using the RMM and the chosen dipoles, four distinct families of interpolation technique (nearest neighbour, polynomial, splines and lead fields) were tested and compared so that the scalp potentials could be recovered from the electrode potentials. The 3D spline interpolation and the inverse forward technique (IFT) gave the best results. The IFT is very easy to use when the lead-field matrix between scalp electrodes and cortex nodes has been calculated. By simple application of the Moore-Penrose pseudo inverse matrix to the electrode cap potentials, a set of current sources on the cortex is obtained. Then, the forward problem using these cortex sources renders all the scalp potentials.
International Nuclear Information System (INIS)
Soycan, Arzu; Soycan, Metin
2009-01-01
GIS (Geographical Information System) is one of the most striking innovation for mapping applications supplied by the developing computer and software technology to users. GIS is a very effective tool which can show visually combination of the geographical and non-geographical data by recording these to allow interpretations and analysis. DEM (Digital Elevation Model) is an inalienable component of the GIS. The existing TM (Topographic Map) can be used as the main data source for generating DEM by amanual digitizing or vectorization process for the contours polylines. The aim of this study is to examine the DEM accuracies, which were obtained by TMs, as depending on the number of sampling points and grid size. For these purposes, the contours of the several 1/1000 scaled scanned topographical maps were vectorized. The different DEMs of relevant area have been created by using several datasets with different numbers of sampling points. We focused on the DEM creation from contour lines using gridding with RBF (Radial Basis Function) interpolation techniques, namely TPS as the surface fitting model. The solution algorithm and a short review of the mathematical model of TPS (Thin Plate Spline) interpolation techniques are given. In the test study, results of the application and the obtained accuracies are drawn and discussed. The initial object of this research is to discuss the requirement of DEM in GIS, urban planning, surveying engineering and the other applications with high accuracy (a few deci meters). (author)
Measurement and tricubic interpolation of the magnetic field for the OLYMPUS experiment
International Nuclear Information System (INIS)
Bernauer, J.C.; Diefenbach, J.; Elbakian, G.; Gavrilov, G.; Goerrissen, N.; Hasell, D.K.; Henderson, B.S.; Holler, Y.; Karyan, G.; Ludwig, J.; Marukyan, H.; Naryshkin, Y.; O'Connor, C.; Russell, R.L.; Schmidt, A.; Schneekloth, U.; Suvorov, K.; Veretennikov, D.
2016-01-01
The OLYMPUS experiment used a 0.3 T toroidal magnetic spectrometer to measure the momenta of outgoing charged particles. In order to accurately determine particle trajectories, knowledge of the magnetic field was needed throughout the spectrometer volume. For that purpose, the magnetic field was measured at over 36,000 positions using a three-dimensional Hall probe actuated by a system of translation tables. We used these field data to fit a numerical magnetic field model, which could be employed to calculate the magnetic field at any point in the spectrometer volume. Calculations with this model were computationally intensive; for analysis applications where speed was crucial, we pre-computed the magnetic field and its derivatives on an evenly spaced grid so that the field could be interpolated between grid points. We developed a spline-based interpolation scheme suitable for SIMD implementations, with a memory layout chosen to minimize space and optimize the cache behavior to quickly calculate field values. This scheme requires only one-eighth of the memory needed to store necessary coefficients compared with a previous scheme (Lekien and Marsden, 2005 [1]). This method was accurate for the vast majority of the spectrometer volume, though special fits and representations were needed to improve the accuracy close to the magnet coils and along the toroidal axis.
Seismic data interpolation and denoising by learning a tensor tight frame
International Nuclear Information System (INIS)
Liu, Lina; Ma, Jianwei; Plonka, Gerlind
2017-01-01
Seismic data interpolation and denoising plays a key role in seismic data processing. These problems can be understood as sparse inverse problems, where the desired data are assumed to be sparsely representable within a suitable dictionary. In this paper, we present a new method based on a data-driven tight frame (DDTF) of Kronecker type (KronTF) that avoids the vectorization step and considers the multidimensional structure of data in a tensor-product way. It takes advantage of the structure contained in all different modes (dimensions) simultaneously. In order to overcome the limitations of a usual tensor-product approach we also incorporate data-driven directionality. The complete method is formulated as a sparsity-promoting minimization problem. It includes two main steps. In the first step, a hard thresholding algorithm is used to update the frame coefficients of the data in the dictionary; in the second step, an iterative alternating method is used to update the tight frame (dictionary) in each different mode. The dictionary that is learned in this way contains the principal components in each mode. Furthermore, we apply the proposed KronTF to seismic interpolation and denoising. Examples with synthetic and real seismic data show that the proposed method achieves better results than the traditional projection onto convex sets method based on the Fourier transform and the previous vectorized DDTF methods. In particular, the simple structure of the new frame construction makes it essentially more efficient. (paper)
Shortcut in DIC error assessment induced by image interpolation used for subpixel shifting
Bornert, Michel; Doumalin, Pascal; Dupré, Jean-Christophe; Poilane, Christophe; Robert, Laurent; Toussaint, Evelyne; Wattrisse, Bertrand
2017-04-01
In order to characterize errors of Digital Image Correlation (DIC) algorithms, sets of virtual images are often generated from a reference image by in-plane sub-pixel translations. This leads to the determination of the well-known S-shaped bias error curves and their corresponding random error curves. As images are usually shifted by using interpolation schemes similar to those used in DIC algorithms, the question of the possible bias in the quantification of measurement uncertainties of DIC softwares is raised and constitutes the main problematic of this paper. In this collaborative work, synthetic numerically shifted images are built from two methods: one based on interpolations of the reference image and the other based on the transformation of an analytic texture function. Images are analyzed using an in-house subset-based DIC software and results are compared and discussed. The effect of image noise is also highlighted. The main result is that the a priori choices to numerically shift the reference image modify DIC results and may lead to wrong conclusions in terms of DIC error assessment.
An Improved DOA Estimation Approach Using Coarray Interpolation and Matrix Denoising
Directory of Open Access Journals (Sweden)
Muran Guo
2017-05-01
Full Text Available Co-prime arrays can estimate the directions of arrival (DOAs of O ( M N sources with O ( M + N sensors, and are convenient to analyze due to their closed-form expression for the locations of virtual lags. However, the number of degrees of freedom is limited due to the existence of holes in difference coarrays if subspace-based algorithms such as the spatial smoothing multiple signal classification (MUSIC algorithm are utilized. To address this issue, techniques such as positive definite Toeplitz completion and array interpolation have been proposed in the literature. Another factor that compromises the accuracy of DOA estimation is the limitation of the number of snapshots. Coarray-based processing is particularly sensitive to the discrepancy between the sample covariance matrix and the ideal covariance matrix due to the finite number of snapshots. In this paper, coarray interpolation based on matrix completion (MC followed by a denoising operation is proposed to detect more sources with a higher accuracy. The effectiveness of the proposed method is based on the capability of MC to fill in holes in the virtual sensors and that of MC denoising operation to reduce the perturbation in the sample covariance matrix. The results of numerical simulations verify the superiority of the proposed approach.
Memory-efficient optimization of Gyrokinetic particle-to-grid interpolation for multicore processors
Energy Technology Data Exchange (ETDEWEB)
Madduri, Kamesh [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Williams, Samuel [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Ethier, Stephane [Princeton Plasma Physics Lab. (PPPL), Princeton, NJ (United States); Oliker, Leonid [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Shalf, John [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Strohmaier, Erich [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Yelicky, Katherine [Univ. of California, Berkeley, CA (United States)
2009-01-01
We present multicore parallelization strategies for the particle-to-grid interpolation step in the Gyrokinetic Toroidal Code (GTC), a 3D particle-in-cell (PIC) application to study turbulent transport in magnetic-confinement fusion devices. Particle-grid interpolation is a known performance bottleneck in several PIC applications. In GTC, this step involves particles depositing charges to a 3D toroidal mesh, and multiple particles may contribute to the charge at a grid point. We design new parallel algorithms for the GTC charge deposition kernel, and analyze their performance on three leading multicore platforms. We implement thirteen different variants for this kernel and identify the best-performing ones given typical PIC parameters such as the grid size, number of particles per cell, and the GTC-specific particle Larmor radius variation. We find that our best strategies can be 2x faster than the reference optimized MPI implementation, and our analysis provides insight into desirable architectural features for high-performance PIC simulation codes.
A new method for reducing DNL in nuclear ADCs using an interpolation technique
International Nuclear Information System (INIS)
Vaidya, P.P.; Gopalakrishnan, K.R.; Pethe, V.A.; Anjaneyulu, T.
1986-01-01
The paper describes a new method for reducing the DNL associated with nuclear ADCs. The method named the ''interpolation technique'' is utilized to derive the quantisation steps corresponding to the last n bits of the digital code by dividing quantisation steps due to higher significant bits of the DAC, using a chain of resistors. Using comparators, these quantisation steps are compared with the analog voltage to be digitized, which is applied as a voltage shift at both ends of this chain. The output states of the comparators define the n bit code. The errors due to offset voltages and bias currents of the comparators are statistically neutralized by changing the polarity of quantisation steps as well as the polarity of analog voltage (corresponding to last n bits) for alternate A/D conversion. The effect of averaging on the channel profile can be minimized. A 12 bit ADC was constructured using this technique which gives DNL of less than +-1% over most of the channels for conversion time of nearly 4.5 μs. Gatti's sliding scale technique can be implemented for further reduction of DNL. The interpolation technique has a promising potential of improving the resolution of existing 12 bit ADCs to 16 bit, without degrading the percentage DNL significantly. (orig.)
Voice Morphing Using 3D Waveform Interpolation Surfaces and Lossless Tube Area Functions
Directory of Open Access Journals (Sweden)
Lavner Yizhar
2005-01-01
Full Text Available Voice morphing is the process of producing intermediate or hybrid voices between the utterances of two speakers. It can also be defined as the process of gradually transforming the voice of one speaker to that of another. The ability to change the speaker's individual characteristics and to produce high-quality voices can be used in many applications. Examples include multimedia and video entertainment, as well as enrichment of speech databases in text-to-speech systems. In this study we present a new technique which enables production of a given number of intermediate voices or of utterances which gradually change from one voice to another. This technique is based on two components: (1 creation of a 3D prototype waveform interpolation (PWI surface from the LPC residual signal, to produce an intermediate excitation signal; (2 a representation of the vocal tract by a lossless tube area function, and an interpolation of the parameters of the two speakers. The resulting synthesized signal sounds like a natural voice lying between the two original voices.
Estimation of missing rainfall data using spatial interpolation and imputation methods
Radi, Noor Fadhilah Ahmad; Zakaria, Roslinazairimah; Azman, Muhammad Az-zuhri
2015-02-01
This study is aimed to estimate missing rainfall data by dividing the analysis into three different percentages namely 5%, 10% and 20% in order to represent various cases of missing data. In practice, spatial interpolation methods are chosen at the first place to estimate missing data. These methods include normal ratio (NR), arithmetic average (AA), coefficient of correlation (CC) and inverse distance (ID) weighting methods. The methods consider the distance between the target and the neighbouring stations as well as the correlations between them. Alternative method for solving missing data is an imputation method. Imputation is a process of replacing missing data with substituted values. A once-common method of imputation is single-imputation method, which allows parameter estimation. However, the single imputation method ignored the estimation of variability which leads to the underestimation of standard errors and confidence intervals. To overcome underestimation problem, multiple imputations method is used, where each missing value is estimated with a distribution of imputations that reflect the uncertainty about the missing data. In this study, comparison of spatial interpolation methods and multiple imputations method are presented to estimate missing rainfall data. The performance of the estimation methods used are assessed using the similarity index (S-index), mean absolute error (MAE) and coefficient of correlation (R).
Directory of Open Access Journals (Sweden)
Nikesh S. Dattani
2012-03-01
Full Text Available One of the most successful methods for calculating reduced density operator dynamics in open quantum systems, that can give numerically exact results, uses Feynman integrals. However, when simulating the dynamics for a given amount of time, the number of time steps that can realistically be used with this method is always limited, therefore one often obtains an approximation of the reduced density operator at a sparse grid of points in time. Instead of relying only on ad hoc interpolation methods (such as splines to estimate the system density operator in between these points, I propose a method that uses physical information to assist with this interpolation. This method is tested on a physically significant system, on which its use allows important qualitative features of the density operator dynamics to be captured with as little as two time steps in the Feynman integral. This method allows for an enormous reduction in the amount of memory and CPU time required for approximating density operator dynamics within a desired accuracy. Since this method does not change the way the Feynman integral itself is calculated, the value of the density operator approximation at the points in time used to discretize the Feynamn integral will be the same whether or not this method is used, but its approximation in between these points in time is considerably improved by this method. A list of ways in which this proposed method can be further improved is presented in the last section of the article.
Roy, Subrata P.
2014-01-28
The method of moments with interpolative closure (MOMIC) for soot formation and growth provides a detailed modeling framework maintaining a good balance in generality, accuracy, robustness, and computational efficiency. This study presents several computational issues in the development and implementation of the MOMIC-based soot modeling for direct numerical simulations (DNS). The issues of concern include a wide dynamic range of numbers, choice of normalization, high effective Schmidt number of soot particles, and realizability of the soot particle size distribution function (PSDF). These problems are not unique to DNS, but they are often exacerbated by the high-order numerical schemes used in DNS. Four specific issues are discussed in this article: the treatment of soot diffusion, choice of interpolation scheme for MOMIC, an approach to deal with strongly oxidizing environments, and realizability of the PSDF. General, robust, and stable approaches are sought to address these issues, minimizing the use of ad hoc treatments such as clipping. The solutions proposed and demonstrated here are being applied to generate new physical insight into complex turbulence-chemistry-soot-radiation interactions in turbulent reacting flows using DNS. © 2014 Copyright Taylor and Francis Group, LLC.
Directory of Open Access Journals (Sweden)
Xihua Yang
2015-01-01
Full Text Available This paper presents spatial interpolation techniques to produce finer-scale daily rainfall data from regional climate modeling. Four common interpolation techniques (ANUDEM, Spline, IDW, and Kriging were compared and assessed against station rainfall data and modeled rainfall. The performance was assessed by the mean absolute error (MAE, mean relative error (MRE, root mean squared error (RMSE, and the spatial and temporal distributions. The results indicate that Inverse Distance Weighting (IDW method is slightly better than the other three methods and it is also easy to implement in a geographic information system (GIS. The IDW method was then used to produce forty-year (1990–2009 and 2040–2059 time series rainfall data at daily, monthly, and annual time scales at a ground resolution of 100 m for the Greater Sydney Region (GSR. The downscaled daily rainfall data have been further utilized to predict rainfall erosivity and soil erosion risk and their future changes in GSR to support assessments and planning of climate change impact and adaptation in local scale.
Zoraghein, H.; Leyk, S.; Balk, D.
2017-12-01
The analysis of changes in urban land and population is important because the majority of future population growth will take place in urban areas. The U.S. Census historically classifies urban land using population density and various land-use criteria. This study analyzes the reliability of census-defined urban lands for delineating the spatial distribution of urban population and estimating its changes over time. To overcome the problem of incompatible enumeration units between censuses, regular areal interpolation methods including Areal Weighting (AW) and Target Density Weighting (TDW), with and without spatial refinement, are implemented. The goal in this study is to estimate urban population in Massachusetts in 1990 and 2000 (source zones), within tract boundaries of the 2010 census (target zones), respectively, to create a consistent time series of comparable urban population estimates from 1990 to 2010. Spatial refinement is done using ancillary variables such as census-defined urban areas, the National Land Cover Database (NLCD) and the Global Human Settlement Layer (GHSL) as well as different combinations of them. The study results suggest that census-defined urban areas alone are not necessarily the most meaningful delineation of urban land. Instead it appears that alternative combinations of the above-mentioned ancillary variables can better depict the spatial distribution of urban land, and thus make it possible to reduce the estimation error in transferring the urban population from source zones to target zones when running spatially-refined temporal areal interpolation.
Nguyen, Hoai-Nam
2014-01-01
A comprehensive development of interpolating control, this monograph demonstrates the reduced computational complexity of a ground-breaking technique compared with the established model predictive control. The text deals with the regulation problem for linear, time-invariant, discrete-time uncertain dynamical systems having polyhedral state and control constraints, with and without disturbances, and under state or output feedback. For output feedback a non-minimal state-space representation is used with old inputs and outputs as state variables. Constrained Control of Uncertain, Time-Varying, Discrete-time Systems details interpolating control in both its implicit and explicit forms. In the former at most two linear-programming or one quadratic-programming problem are solved on-line at each sampling instant to yield the value of the control variable. In the latter the control law is shown to be piecewise affine in the state, and so the state space is partitioned into polyhedral cells so that at each sampling ...
Joint seismic data denoising and interpolation with double-sparsity dictionary learning
Zhu, Lingchen; Liu, Entao; McClellan, James H.
2017-08-01
Seismic data quality is vital to geophysical applications, so that methods of data recovery, including denoising and interpolation, are common initial steps in the seismic data processing flow. We present a method to perform simultaneous interpolation and denoising, which is based on double-sparsity dictionary learning. This extends previous work that was for denoising only. The original double-sparsity dictionary learning algorithm is modified to track the traces with missing data by defining a masking operator that is integrated into the sparse representation of the dictionary. A weighted low-rank approximation algorithm is adopted to handle the dictionary updating as a sparse recovery optimization problem constrained by the masking operator. Compared to traditional sparse transforms with fixed dictionaries that lack the ability to adapt to complex data structures, the double-sparsity dictionary learning method learns the signal adaptively from selected patches of the corrupted seismic data, while preserving compact forward and inverse transform operators. Numerical experiments on synthetic seismic data indicate that this new method preserves more subtle features in the data set without introducing pseudo-Gibbs artifacts when compared to other directional multi-scale transform methods such as curvelets.
Fusing Satellite-Derived Irradiance and Point Measurements through Optimal Interpolation
Lorenzo, A.; Morzfeld, M.; Holmgren, W.; Cronin, A.
2016-12-01
Satellite-derived irradiance is widely used throughout the design and operation of a solar power plant. While satellite-derived estimates cover a large area, they also have large errors compared to point measurements from sensors on the ground. We describe an optimal interpolation routine that fuses the broad spatial coverage of satellite-derived irradiance with the high accuracy of point measurements. The routine can be applied to any satellite-derived irradiance and point measurement datasets. Unique aspects of this work include the fact that information is spread using cloud location and thickness and that a number of point measurements are collected from rooftop PV systems. The routine is sensitive to errors in the satellite image geolocation, so care must be taken to adjust the cloud locations based on the solar and satellite geometries. Analysis of the optimal interpolation routine over Tucson, AZ, with 20 point measurements shows a significant improvement in the irradiance estimate for two distinct satellite image to irradiance algorithms. Improved irradiance estimates can be used for resource assessment, distributed generation production estimates, and irradiance forecasts.
Batch orographic interpolation of monthly precipitation based on free-of-charge geostatistical tools
Ledvinka, Ondrej
2017-11-01
The effects of possible climate change on water resources in prescribed areas (e.g. river basins) are intensively studied in hydrology. These resources are highly dependent on precipitation totals. When focusing on long-term changes in climate variables, one has to rely on station measurements. However, hydrologists need the information on spatial distribution of precipitation over the areas. For this purpose, the spatial interpolation techniques must be employed. In Czechia, where the addition of elevation co-variables proved to be a good choice, several GIS tools exist that are able to produce time series necessary for climate change analyses. Nevertheless, these tools are exclusively based on commercial software and there is a lack of free-of-charge tools that could be used by everyone. Here, selected free-of-charge geostatistical tools were utilized in order to produce monthly precipitation time series representing six river basins in the Ore Mountains located in NW Bohemia, Czechia and SE Saxony, Germany. The produced series span from January 1961 to December 2012. Rain-gauge data from both Czechia and Germany were used. The universal kriging technique was employed where a multiple linear regression (based on elevation and coordinates) residuals were interpolated. The final time series seem to be homogeneous.
Interpolation problems in meteor radar analysis, part 7.6A
Tetenbaum, D.; Avery, S. K.
1984-01-01
Meteor echoes come from random points in the observation volume, and are irregularly spaced in time. This precludes the use of fast fourier transformations (FFT) techniques on the raw data to give the spectrum of waves that are present. One way around this obstacle is to restrict our interest to a particular class of waves, and fit a corresponding model to the raw data. It is assumed that there is no horizontal variation across the observation volume for tides, but in the vertical this is certainly not the case. If, in addition, we are interested in other types of waves which may be present and whose periods are unknown, then examining the raw line-of-sight velocities does not tell us how to modify the model, since the line-of-sight direction is not fixed. This is the motivation for interpolation. Interpolation takes a temporal series of line-of-sight velocities, and transforms it to a temporal series of wind velocities for each orthogonal direction. The velocities along a given direction can then be examined readily for any waves in addition to tides.
Karimi, Davood; Ward, Rabab K.
2016-03-01
Reducing the radiation dose in computed tomography (CT) requires reducing the number or the energy of the photons that pass through the patient's body. An image reconstructed from such noisy or undersampled measurements will contain much noise and artifacts that can significantly reduce the diagnostic value of the image. Effective sinogram denoising or interpolation can reduce these noise and artifacts. In this paper, we present a novel approach to sinogram smoothing and interpolation. The proposed method iteratively estimates the local slope and curvature of the sinogam and forces the sinogram to follow the estimated slope and curvature. This is performed by projection onto the set of constraints that define the slope and the curvature. The constraints on the slope and curvature correspond to very simple convex sets. Projection onto these sets have simple analytical solutions. Moreover, these operations are highly parallelizable because the equations defining the slope and curvature constraints for all the points in a sinogram can be summarized as five convex sets, regardless of the length of the sinogram. We apply the proposed method on simulated and real data and examine its effect on the quality of the reconstructed image. Our results show that the proposed method is highly effective and can lead to a substantial improvement in the quality of the images reconstructed from noisy sinogram measurements. A comparison with the K-SVD denoising algorithm shows that the proposed algorithm achieves better results. We suggest that the proposed method can be a useful tool for low-dose CT.
Climate applications for NOAA 1/4° Daily Optimum Interpolation Sea Surface Temperature
Boyer, T.; Banzon, P. V. F.; Liu, G.; Saha, K.; Wilson, C.; Stachniewicz, J. S.
2015-12-01
Few sea surface temperature (SST) datasets from satellites have the long temporal span needed for climate studies. The NOAA Daily Optimum Interpolation Sea Surface Temperature (DOISST) on a 1/4° grid, produced at National Centers for Environmental Information, is based primarily on SSTs from the Advanced Very High Resolution Radiometer (AVHRR), available from 1981 to the present. AVHRR data can contain biases, particularly when aerosols are present. Over the three decade span, the largest departure of AVHRR SSTs from buoy temperatures occurred during the Mt Pinatubo and El Chichon eruptions. Therefore, in DOISST, AVHRR SSTs are bias-adjusted to match in situ SSTs prior to interpolation. This produces a consistent time series of complete SST fields that is suitable for modelling and investigating local climate phenomena like El Nino or the Pacific warm blob in a long term context. Because many biological processes and animal distributions are temperature dependent, there are also many ecological uses of DOISST (e.g., coral bleaching thermal stress, fish and marine mammal distributions), thereby providing insights into resource management in a changing ocean. The advantages and limitations of using DOISST for different applications will be discussed.
Remote Sensing of Fractional Green Vegetation Cover Using Spatially-Interpolated Endmembers
Directory of Open Access Journals (Sweden)
Ryutaro Tateishi
2012-09-01
Full Text Available Fractional green vegetation cover (FVC is a useful parameter for many environmental and climate-related applications. A common approach for estimating FVC involves the linear unmixing of two spectral endmembers in a remote sensing image; bare soil and green vegetation. The spectral properties of these two endmembers are typically determined based on field measurements, estimated using additional data sources (e.g., soil databases or land cover maps, or extracted directly from the imagery. Most FVC estimation approaches do not consider that the spectral properties of endmembers may vary across space. However, due to local differences in climate, soil type, vegetation species, etc., the spectral characteristics of soil and green vegetation may exhibit positive spatial autocorrelation. When this is the case, it may be useful to take these local variations into account for estimating FVC. In this study, spatial interpolation (Inverse Distance Weighting and Ordinary Kriging was used to predict variations in the spectral characteristics of bare soil and green vegetation across space. When the spatially-interpolated values were used in place of scene-invariant endmember values to estimate FVC in an Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER image, the accuracy of FVC estimates increased, providing evidence that it may be useful to consider the effects of spatial autocorrelation for spectral mixture analysis.
A Web-Based Tool to Interpolate Nitrogen Loading Using a Genetic Algorithm
Directory of Open Access Journals (Sweden)
Youn Shik Park
2014-09-01
Full Text Available Water quality data may not be collected at a high frequency, nor over the range of streamflow data. For instance, water quality data are often collected monthly, biweekly, or weekly, since collecting and analyzing water quality samples are costly compared to streamflow data. Regression models are often used to interpolate pollutant loads from measurements made intermittently. Web-based Load Interpolation Tool (LOADIN was developed to provide user-friendly interfaces and to allow use of streamflow and water quality data from U.S. Geological Survey (USGS via web access. LOADIN has a regression model assuming that instantaneous load is comprised of the pollutant load based on streamflow and the pollutant load variation within the period. The regression model has eight coefficients determined by a genetic algorithm with measured water quality data. LOADIN was applied to eleven water quality datasets from USGS gage stations located in Illinois, Indiana, Michigan, Minnesota, and Wisconsin states with drainage areas from 44 km2 to 1,847,170 km2. Measured loads were calculated by multiplying nitrogen data by streamflow data associated with measured nitrogen data. The estimated nitrogen loads and measured loads were evaluated using Nash-Sutcliffe Efficiency (NSE and coefficient of determination (R2. NSE ranged from 0.45 to 0.91, and R2 ranged from 0.51 to 0.91 for nitrogen load estimation.
Directory of Open Access Journals (Sweden)
Kurt James Werner
2016-10-01
Full Text Available The magnitude of the Discrete Fourier Transform (DFT of a discrete-time signal has a limited frequency definition. Quadratic interpolation over the three DFT samples surrounding magnitude peaks improves the estimation of parameters (frequency and amplitude of resolved sinusoids beyond that limit. Interpolating on a rescaled magnitude spectrum using a logarithmic scale has been shown to improve those estimates. In this article, we show how to heuristically tune a power scaling parameter to outperform linear and logarithmic scaling at an equivalent computational cost. Although this power scaling factor is computed heuristically rather than analytically, it is shown to depend in a structured way on window parameters. Invariance properties of this family of estimators are studied and the existence of a bias due to noise is shown. Comparing to two state-of-the-art estimators, we show that an optimized power scaling has a lower systematic bias and lower mean-squared-error in noisy conditions for ten out of twelve common windowing functions.
Directory of Open Access Journals (Sweden)
Maheswari Subramanian
2018-01-01
Full Text Available Information hiding techniques have a significant role in recent application areas. Steganography is the embedding of information within an innocent cover work in a way which cannot be detected by any person without accessing the steganographic key. The proposed work uses a steganographic scheme for useful information with the help of human skin tone regions as cover image. The proposed algorithm has undergone Lagrange interpolation encryption for enhancement of the security of the hidden information. First, the skin tone regions are identified by using YCbCr color space which can be used as a cover image. Image pixels which belong to the skin regions are used to carry more secret bits, and the secret information is hidden in both horizontal and vertical sequences of the skin areas of the cover image. The secret information will hide behind the human skin regions rather than other objects in the same image because the skin pixels have high intensity value. The performance of embedding is done and is quite invisible by the vector discrete wavelet transformation (VDWT technique. A new Lagrange interpolation-based encryption method is introduced to achieve high security of the hidden information with higher payload and better visual quality.
Spatio-Temporal Interpolation of Cloudy SST Fields Using Conditional Analog Data Assimilation
Directory of Open Access Journals (Sweden)
Ronan Fablet
2018-02-01
Full Text Available The ever increasing geophysical data streams pouring from earth observation satellite missions and numerical simulations along with the development of dedicated big data infrastructure advocate for truly exploiting the potential of these datasets, through novel data-driven strategies, to deliver enhanced satellite-derived gapfilled geophysical products from partial satellite observations. We here demonstrate the relevance of the analog data assimilation (AnDA for an application to the reconstruction of cloud-free level-4 gridded Sea Surface Temperature (SST. We propose novel AnDA models which exploit auxiliary variables such as sea surface currents and significantly reduce the computational complexity of AnDA. Numerical experiments benchmark the proposed models with respect to state-of-the-art interpolation techniques such as optimal interpolation and EOF-based schemes. We report relative improvement up to 40%/50% in terms of RMSE and also show a good parallelization performance, which supports the feasibility of an upscaling on a global scale.
High-Dimensional Intrinsic Interpolation Using Gaussian Process Regression and Diffusion Maps
International Nuclear Information System (INIS)
Thimmisetty, Charanraj A.; Ghanem, Roger G.; White, Joshua A.; Chen, Xiao
2017-01-01
This article considers the challenging task of estimating geologic properties of interest using a suite of proxy measurements. The current work recast this task as a manifold learning problem. In this process, this article introduces a novel regression procedure for intrinsic variables constrained onto a manifold embedded in an ambient space. The procedure is meant to sharpen high-dimensional interpolation by inferring non-linear correlations from the data being interpolated. The proposed approach augments manifold learning procedures with a Gaussian process regression. It first identifies, using diffusion maps, a low-dimensional manifold embedded in an ambient high-dimensional space associated with the data. It relies on the diffusion distance associated with this construction to define a distance function with which the data model is equipped. This distance metric function is then used to compute the correlation structure of a Gaussian process that describes the statistical dependence of quantities of interest in the high-dimensional ambient space. The proposed method is applicable to arbitrarily high-dimensional data sets. Here, it is applied to subsurface characterization using a suite of well log measurements. The predictions obtained in original, principal component, and diffusion space are compared using both qualitative and quantitative metrics. Considerable improvement in the prediction of the geological structural properties is observed with the proposed method.
Interpolating gain-scheduled H∞ loop shaping design for high speed ball screw feed drives.
Dong, Liang; Tang, WenCheng; Bao, DaFei
2015-03-01
This paper presents a method to design servo controllers for flexible ball screw drives with time-varying dynamics, which are mainly due to the time-varying table position and the workpiece mass. A gain-scheduled H∞ loop shaping controller is designed to achieve high tracking performance against the dynamic variations. H∞ loop shaping design procedure incorporates open loop shaping by a set of compensators to obtain performance/robust stability tradeoffs. The interpolating gain-scheduled controller is obtained by interpolating the state space model of the linear time-invariant (LTI) controllers estimated for fixed values of the scheduling parameters and a linear least squares problem can be solved. The proposed controller has been compared with P/PI with velocity and acceleration feedforward and adaptive backstepping sliding mode control experimentally. The experimental results indicate that the tracking performance has been improved and the robustness for time-varying dynamics has been achieved with the proposed scheme. Copyright © 2014 ISA. Published by Elsevier Ltd. All rights reserved.
Measurement and tricubic interpolation of the magnetic field for the OLYMPUS experiment
Energy Technology Data Exchange (ETDEWEB)
Bernauer, J.C. [Massachusetts Institute of Technology, Laboratory for Nuclear Science, Cambridge, MA (United States); Diefenbach, J. [Hampton University, Hampton, VA (United States); Elbakian, G. [Alikhanyan National Science Laboratory (Yerevan Physics Institute), Yerevan (Armenia); Gavrilov, G. [Petersburg Nuclear Physics Institute, Gatchina (Russian Federation); Goerrissen, N. [Deutsches Elektronen-Synchrotron DESY, Hamburg (Germany); Hasell, D.K.; Henderson, B.S. [Massachusetts Institute of Technology, Laboratory for Nuclear Science, Cambridge, MA (United States); Holler, Y. [Deutsches Elektronen-Synchrotron DESY, Hamburg (Germany); Karyan, G. [Alikhanyan National Science Laboratory (Yerevan Physics Institute), Yerevan (Armenia); Ludwig, J. [Deutsches Elektronen-Synchrotron DESY, Hamburg (Germany); Marukyan, H. [Alikhanyan National Science Laboratory (Yerevan Physics Institute), Yerevan (Armenia); Naryshkin, Y. [Petersburg Nuclear Physics Institute, Gatchina (Russian Federation); O' Connor, C.; Russell, R.L.; Schmidt, A. [Massachusetts Institute of Technology, Laboratory for Nuclear Science, Cambridge, MA (United States); Schneekloth, U. [Deutsches Elektronen-Synchrotron DESY, Hamburg (Germany); Suvorov, K.; Veretennikov, D. [Petersburg Nuclear Physics Institute, Gatchina (Russian Federation)
2016-07-01
The OLYMPUS experiment used a 0.3 T toroidal magnetic spectrometer to measure the momenta of outgoing charged particles. In order to accurately determine particle trajectories, knowledge of the magnetic field was needed throughout the spectrometer volume. For that purpose, the magnetic field was measured at over 36,000 positions using a three-dimensional Hall probe actuated by a system of translation tables. We used these field data to fit a numerical magnetic field model, which could be employed to calculate the magnetic field at any point in the spectrometer volume. Calculations with this model were computationally intensive; for analysis applications where speed was crucial, we pre-computed the magnetic field and its derivatives on an evenly spaced grid so that the field could be interpolated between grid points. We developed a spline-based interpolation scheme suitable for SIMD implementations, with a memory layout chosen to minimize space and optimize the cache behavior to quickly calculate field values. This scheme requires only one-eighth of the memory needed to store necessary coefficients compared with a previous scheme (Lekien and Marsden, 2005 [1]). This method was accurate for the vast majority of the spectrometer volume, though special fits and representations were needed to improve the accuracy close to the magnet coils and along the toroidal axis.
Spatial and temporal interpolation of DInSAR data at different scales
Tessitore, Serena; Fiaschi, Simone; Achilli, Vladimiro; Ahmed, Ahmed; Calcaterra, Domenico; Di Martire, Diego; Guardiola-Albert, Carolina; Meisina, Claudia; Ramondini, Massimo; Floris, Mario
2015-04-01
The present study regards the utilization of multi-pass DInSAR algorithms to the ground displacements monitoring at small and large scales. An integration of in situ and DInSAR data to the elaboration of 2D maps of deformation is proposed. A geo-statistical method for "radar-gauge combination" called Ordinary Kriging of Radar Errors (OKRE) has been used. This algorithm uses the punctual values of a primary variable that is represented by measurements of true deformations, whereas radar is comprised as auxiliary information on the spatial distribution (Erdin, 2013). According to this method, is possible to obtain the interpolated map of deformations by subtracting a radar error map from the original interpolated radar map. In particular, the radar error map is carried out by interpolating the differences between radar and in situ data with the OK interpolator. To this aim, in the present work the available standard spirit levelling and GPS data have been used. Moreover, DInSAR data achieved through two different approaches have been taken into account for the spatial analysis and the error map computation at different scales. Specifically, the Persistent Scatterer Technique (PS-InSAR) and the Small BAseline Subset approach (SBAS) have been used to process the ENVISAT SAR images acquired in the period 2002-2010. In the SBAS processing chain, it is possible to activate the Disconnected Blocks tool and perform the SAR data "temporal interpolation". Since the estimation of the results in the processing takes into account the coherence threshold on the input images stack and their connection criteria, only the pixels above the threshold that are fully connected in all the images are solved. By activating the Disconnect Blocks tool, the results are estimated also for those pixels that respect the threshold criteria at least in the 60% of the images even in a not fully connected stack. In this way, the spatial coverage is higher but the reliability of the results is has to
The interpolation method of stochastic functions and the stochastic variational principle
International Nuclear Information System (INIS)
Liu Xianbin; Chen Qiu
1993-01-01
-order stochastic finite element equations are not very reasonable. On the other hand, Galerkin Method is hopeful, along with the method, the projection principle had been advanced to solve the stochastic operator equations. In Galerkin Method, by means of projecting the stochastic solution functions into the subspace of the solution function space, the treatment of the stochasticity of the structural physical properties and the loads is reasonable. However, the construction or the selection of the subspace of the solution function space which is a Hilbert Space of stochastic functions is difficult, and furthermore it is short of a reasonable rule to measure whether the approximation of the subspace to the solution function space is fine or not. In stochastic finite element method, the discretization of stochastic functions in space and time shows a very importance, so far, the discrete patterns consist of Local Average Theory, Interpolation Method and Orthogonal Expansion Method. Although the Local Average Theory has already been a success in the stationary random fields, it is not suitable for the non-stationary ones as well. For the general stochastic functions, whether it is stationary or not, interpolation method is available. In the present paper, the authors have shown that the error between the true solution function and its approximation, its projection in the subspace, depends continuously on the errors between the stochastic functions and their interpolation functions, the latter rely continuously on the scales of the discrete elements; so a conclusion can be obtained that the Interpolation method of stochastic functions is convergent. That is to say that the approximation solution functions would limit to the true solution functions when the scales of the discrete elements goes smaller and smaller. Using the Interpolation method, a basis of subspace of the solution function space is constructed in this paper, and by means of combining the projection principle and
2016-02-11
INVESTIGATION OF BACK-OFF BASED INTERPOLATION BETWEEN RECURRENT NEURAL NETWORK AND N-GRAM LANGUAGE MODELS X. Chen, X. Liu, M. J. F. Gales, and P. C...weighting based linear interpolation in state-of-the-art ASR systems. However, previous work doesn’t fully exploit the difference of mod- elling power of the...back-off based compact representation of n-gram dependent interpolation weights is pro- posed in this paper. This approach allows weight parameters to
Cui, Jiwen; Zhao, Shiyuan; Yang, Di; Ding, Zhenyang
2018-02-20
We use a spectrum interpolation technique to improve the distributed strain measurement accuracy in a Rayleigh-scatter-based optical frequency domain reflectometry sensing system. We demonstrate that strain accuracy is not limited by the "uncertainty principle" that exists in the time-frequency analysis. Different interpolation methods are investigated and used to improve the accuracy of peak position of the cross-correlation and, therefore, improve the accuracy of the strain. Interpolation implemented by padding zeros on one side of the windowed data in the spatial domain, before the inverse fast Fourier transform, is found to have the best accuracy. Using this method, the strain accuracy and resolution are both improved without decreasing the spatial resolution. The strain of 3 μϵ within the spatial resolution of 1 cm at the position of 21.4 m is distinguished, and the measurement uncertainty is 3.3 μϵ.
Fitzpatrick, Benjamin R; Lamb, David W; Mengersen, Kerrie
2016-01-01
Modern soil mapping is characterised by the need to interpolate point referenced (geostatistical) observations and the availability of large numbers of environmental characteristics for consideration as covariates to aid this interpolation. Modelling tasks of this nature also occur in other fields such as biogeography and environmental science. This analysis employs the Least Angle Regression (LAR) algorithm for fitting Least Absolute Shrinkage and Selection Operator (LASSO) penalized Multiple Linear Regressions models. This analysis demonstrates the efficiency of the LAR algorithm at selecting covariates to aid the interpolation of geostatistical soil carbon observations. Where an exhaustive search of the models that could be constructed from 800 potential covariate terms and 60 observations would be prohibitively demanding, LASSO variable selection is accomplished with trivial computational investment.
Rufo, Montaña; Antolín, Alicia; Paniagua, Jesús M; Jiménez, Antonio
2018-04-01
A comparative study was made of three methods of interpolation - inverse distance weighting (IDW), spline and ordinary kriging - after optimization of their characteristic parameters. These interpolation methods were used to represent the electric field levels for three emission frequencies (774kHz, 900kHz, and 1107kHz) and for the electrical stimulation quotient, Q E , characteristic of complex electromagnetic environments. Measurements were made with a spectrum analyser in a village in the vicinity of medium-wave radio broadcasting antennas. The accuracy of the models was quantified by comparing their predictions with levels measured at the control points not used to generate the models. The results showed that optimizing the characteristic parameters of each interpolation method allows any of them to be used. However, the best results in terms of the regression coefficient between each model's predictions and the actual control point field measurements were for the IDW method. Copyright © 2018 Elsevier Inc. All rights reserved.
Lamb, David W.; Mengersen, Kerrie
2016-01-01
Modern soil mapping is characterised by the need to interpolate point referenced (geostatistical) observations and the availability of large numbers of environmental characteristics for consideration as covariates to aid this interpolation. Modelling tasks of this nature also occur in other fields such as biogeography and environmental science. This analysis employs the Least Angle Regression (LAR) algorithm for fitting Least Absolute Shrinkage and Selection Operator (LASSO) penalized Multiple Linear Regressions models. This analysis demonstrates the efficiency of the LAR algorithm at selecting covariates to aid the interpolation of geostatistical soil carbon observations. Where an exhaustive search of the models that could be constructed from 800 potential covariate terms and 60 observations would be prohibitively demanding, LASSO variable selection is accomplished with trivial computational investment. PMID:27603135
Directory of Open Access Journals (Sweden)
Kresno Wikan Sadono
2016-12-01
Full Text Available Persamaan differensial banyak digunakan untuk menggambarkan berbagai fenomena dalam bidang sains dan rekayasa. Berbagai masalah komplek dalam kehidupan sehari-hari dapat dimodelkan dengan persamaan differensial dan diselesaikan dengan metode numerik. Salah satu metode numerik, yaitu metode meshfree atau meshless berkembang akhir-akhir ini, tanpa proses pembuatan elemen pada domain. Penelitian ini menggabungkan metode meshless yaitu radial basis point interpolation method (RPIM dengan integrasi waktu discontinuous Galerkin method (DGM, metode ini disebut RPIM-DGM. Metode RPIM-DGM diaplikasikan pada advection equation pada satu dimensi. RPIM menggunakan basis function multiquadratic function (MQ dan integrasi waktu diturunkan untuk linear-DGM maupun quadratic-DGM. Hasil simulasi menunjukkan, metode ini mendekati hasil analitis dengan baik. Hasil simulasi numerik dengan RPIM DGM menunjukkan semakin banyak node dan semakin kecil time increment menunjukkan hasil numerik semakin akurat. Hasil lain menunjukkan, integrasi numerik dengan quadratic-DGM untuk suatu time increment dan jumlah node tertentu semakin meningkatkan akurasi dibandingkan dengan linear-DGM. [Title: Numerical solution of advection equation with radial basis interpolation method and discontinuous Galerkin method for time integration] Differential equation is widely used to describe a variety of phenomena in science and engineering. A variety of complex issues in everyday life can be modeled with differential equations and solved by numerical method. One of the numerical methods, the method meshfree or meshless developing lately, without making use of the elements in the domain. The research combines methods meshless, i.e. radial basis point interpolation method with discontinuous Galerkin method as time integration method. This method is called RPIM-DGM. The RPIM-DGM applied to one dimension advection equation. The RPIM using basis function multiquadratic function and time
Directory of Open Access Journals (Sweden)
F. F. Asal
2012-07-01
Full Text Available Digital elevation data obtained from different Engineering Surveying techniques is utilized in generating Digital Elevation Model (DEM, which is employed in many Engineering and Environmental applications. This data is usually in discrete point format making it necessary to utilize an interpolation approach for the creation of DEM. Quality assessment of the DEM is a vital issue controlling its use in different applications; however this assessment relies heavily on statistical methods with neglecting the visual methods. The research applies visual analysis investigation on DEMs generated using IDW interpolator of varying powers in order to examine their potential in the assessment of the effects of the variation of the IDW power on the quality of the DEMs. Real elevation data has been collected from field using total station instrument in a corrugated terrain. DEMs have been generated from the data at a unified cell size using IDW interpolator with power values ranging from one to ten. Visual analysis has been undertaken using 2D and 3D views of the DEM; in addition, statistical analysis has been performed for assessment of the validity of the visual techniques in doing such analysis. Visual analysis has shown that smoothing of the DEM decreases with the increase in the power value till the power of four; however, increasing the power more than four does not leave noticeable changes on 2D and 3D views of the DEM. The statistical analysis has supported these results where the value of the Standard Deviation (SD of the DEM has increased with increasing the power. More specifically, changing the power from one to two has produced 36% of the total increase (the increase in SD due to changing the power from one to ten in SD and changing to the powers of three and four has given 60% and 75% respectively. This refers to decrease in DEM smoothing with the increase in the power of the IDW. The study also has shown that applying visual methods supported
Xiao, Yong; Gu, Xiaomin; Yin, Shiyang; Shao, Jingli; Cui, Yali; Zhang, Qiulan; Niu, Yong
2016-01-01
Based on the geo-statistical theory and ArcGIS geo-statistical module, datas of 30 groundwater level observation wells were used to estimate the decline of groundwater level in Beijing piedmont. Seven different interpolation methods (inverse distance weighted interpolation, global polynomial interpolation, local polynomial interpolation, tension spline interpolation, ordinary Kriging interpolation, simple Kriging interpolation and universal Kriging interpolation) were used for interpolating groundwater level between 2001 and 2013. Cross-validation, absolute error and coefficient of determination (R(2)) was applied to evaluate the accuracy of different methods. The result shows that simple Kriging method gave the best fit. The analysis of spatial and temporal variability suggest that the nugget effects from 2001 to 2013 were increasing, which means the spatial correlation weakened gradually under the influence of human activities. The spatial variability in the middle areas of the alluvial-proluvial fan is relatively higher than area in top and bottom. Since the changes of the land use, groundwater level also has a temporal variation, the average decline rate of groundwater level between 2007 and 2013 increases compared with 2001-2006. Urban development and population growth cause over-exploitation of residential and industrial areas. The decline rate of the groundwater level in residential, industrial and river areas is relatively high, while the decreasing of farmland area and development of water-saving irrigation reduce the quantity of water using by agriculture and decline rate of groundwater level in agricultural area is not significant.
Energy Technology Data Exchange (ETDEWEB)
Penteado, Miguel Suarez Xavier [Pos-Graduacao em Agronomia - Energia na Agricultura, FCA UNESP - Botucatu, SP (Brazil), Dept. de Recursos Naturais], e-mail: miguel_penteado@fca.unesp.br; Escobedo, Joao Francisco [Dept. de Recursos Naturais, FCA/UNESP, Botucatu, SP (Brazil)], e-mail: escobedo@fca.unesp.br; Dal Pai, Alexandre [Faculdade de Tecnologia de Botucatu - FATEC, Botucatu, SP (Brazil)], e-mail: adalpai@fatecbt.edu.br
2011-07-01
This work explores the suitability of the Lagrange interpolating polynomial as a tool to estimate and correct solar databases. From the knowledge of the irradiance distribution over a day, portion of it was removed for applying Lagrange interpolation polynomial. After generation of the estimates by interpolation, the assessment was made by MBE and Rms statistical indicators. The application of Lagrange interpolating generated the following results: underestimation of 0.27% (MBE = -1.83 W/m{sup 2}) and scattering of 0.51% (Rms = 3.48 W/m{sup 2}). (author)
Cheng, Liantao; Zhang, Fenghui; Kang, Xiaoyu; Wang, Lang
2018-05-01
In evolutionary population synthesis (EPS) models, we need to convert stellar evolutionary parameters into spectra via interpolation in a stellar spectral library. For theoretical stellar spectral libraries, the spectrum grid is homogeneous on the effective-temperature and gravity plane for a given metallicity. It is relatively easy to derive stellar spectra. For empirical stellar spectral libraries, stellar parameters are irregularly distributed and the interpolation algorithm is relatively complicated. In those EPS models that use empirical stellar spectral libraries, different algorithms are used and the codes are often not released. Moreover, these algorithms are often complicated. In this work, based on a radial basis function (RBF) network, we present a new spectrum interpolation algorithm and its code. Compared with the other interpolation algorithms that are used in EPS models, it can be easily understood and is highly efficient in terms of computation. The code is written in MATLAB scripts and can be used on any computer system. Using it, we can obtain the interpolated spectra from a library or a combination of libraries. We apply this algorithm to several stellar spectral libraries (such as MILES, ELODIE-3.1 and STELIB-3.2) and give the integrated spectral energy distributions (ISEDs) of stellar populations (with ages from 1 Myr to 14 Gyr) by combining them with Yunnan-III isochrones. Our results show that the differences caused by the adoption of different EPS model components are less than 0.2 dex. All data about the stellar population ISEDs in this work and the RBF spectrum interpolation code can be obtained by request from the first author or downloaded from http://www1.ynao.ac.cn/˜zhangfh.
International Nuclear Information System (INIS)
Michieli, I.
1998-01-01
Recently, a new buildup factors approximation formula based on the expanded polynomial set (E-P function) was successfully introduced (Michieli 1994.) with the maximum approximation error below 4% throughout the standard data domain. Buildup factors interpolation in E-P function parameters for arbitrary source energies, near the K-edge in lead, was satisfactory. Maximum interpolation error, for lead, lays within 12% what appears to be acceptable for most Point Kernel application. 1991. Harima at. al., showed that, near the K-edge, fluctuation in energy of exposure rate attenuation factors i.e.: D(E)B(E, μ E r)exp(-μ E r), given as a function of penetration depth (r) in ordinary length units (not mfps.), is not nearly as great as that of buildup factors. That phenomenon leads to the recommendation (ANSI/ANS-6.4.3) that interpolations in that energy range should be made in the attenuation factors B(E, μ E r)exp(-μ E r) rather than in the buildup factors alone. In present article, such interpolation approach is investigated by applying it to the attenuation factors in lead, with E-P function representation of exposure buildup factors. Simple form of the E-P function leads to strait calculation of new function parameters for arbitrary source energy near the K-edge and thus allowing the same representation form of buildup factors as in the standard interpolation procedure. results of the interpolation are discussed and compared with those from standard approach. (author)
A high-resolution time interpolator based on a delay locked loop and an RC delay line
Mota, M
1999-01-01
An architecture for a time interpolation circuit with an rms error of ~25 ps has been developed in a 0.7- mu m CMOS technology. It is based on a delay locked loop (DLL) driven by a 160-MHz reference clock and a passive RC delay line controlled by an autocalibration circuit. Start-up calibration of the RC delay line is performed using code density tests (CDT). The very small temperature/voltage dependence of R and C parameters and the self calibrating DLL results in a low- power, high-resolution time interpolation circuit in a standard digital CMOS technology. (11 refs).
International Nuclear Information System (INIS)
Li-Min, Ma; Zong-Min, Wu
2010-01-01
In this paper, we use a kind of univariate multiquadric quasi-interpolation to solve a parabolic equation with overspecified data, which has arisen in many physical phenomena. We obtain the numerical scheme by using the derivative of the quasi-interpolation to approximate the spatial derivative of the dependent variable and a simple forward difference to approximate the temporal derivative of the dependent variable. The advantage of the presented scheme is that the algorithm is very simple so it is very easy to implement. The results of the numerical experiment are presented and are compared with the exact solution to confirm the good accuracy of the presented scheme. (general)
Directory of Open Access Journals (Sweden)
Shulun Liu
2018-01-01
Full Text Available Rain gauges are widely used to obtain temporally continuous point rainfall records, which are then interpolated into spatially continuous data to force hydrological models. However, rainfall measurements and interpolation procedure are subject to various uncertainties, which can be reduced by applying quality control and selecting appropriate spatial interpolation approaches. Consequently, the integrated impact of rainfall quality control and interpolation on streamflow simulation has attracted increased attention but not been fully addressed. This study applies a quality control procedure to the hourly rainfall measurements obtained in the Warwick catchment in eastern Australia. The grid-based daily precipitation from the Australian Water Availability Project was used as a reference. The Pearson correlation coefficient between the daily accumulation of gauged rainfall and the reference data was used to eliminate gauges with significant quality issues. The unrealistic outliers were censored based on a comparison between gauged rainfall and the reference. Four interpolation methods, including the inverse distance weighting (IDW, nearest neighbors (NN, linear spline (LN, and ordinary Kriging (OK, were implemented. The four methods were firstly assessed through a cross-validation using the quality-controlled rainfall data. The impacts of the quality control and interpolation on streamflow simulation were then evaluated through a semi-distributed hydrological model. The results showed that the Nash–Sutcliffe model efficiency coefficient (NSE and Bias of the streamflow simulations were significantly improved after quality control. In the cross-validation, the IDW and OK methods resulted in good interpolation rainfall, while the NN led to the worst result. In terms of the impact on hydrological prediction, the IDW led to the most consistent streamflow predictions with the observations, according to the validation at five streamflow-gauged locations
Comparison of interpolation methods for sparse data: Application to wind and concentration fields
International Nuclear Information System (INIS)
Goodin, W.R.; McRae, G.J.; Seinfield, J.H.
1979-01-01
in order to produce gridded fields of pollutant concentration data and surface wind data for use in an air quality model, a number of techniques for interpolating sparse data values are compared. The techniques are compared using three data sets. One is an idealized concentration distribution to which the exact solution is known, the second is a potential flow field, while the third consists of surface ozone concentrations measured in the Los Angeles Basin on a particular day. The results of the study indicate that fitting a second-degree polynomial to each subregion (triangle) in the plane with each data point weighted according to its distance form the subregion provides a good compromise between accuracy and computational cost
Axial resolution and the value of interpolating scan in multislice positron computed tomography.
Senda, M; Yonekura, Y; Tamaki, N; Tanaka, Y; Komori, M; Minato, K; Konishi, J; Torizuka, K
1985-01-01
We have calculated the aperture function of a positron computed tomograph (PCT) with computer simulation, and evaluated the axial resolution of a multislice PCT, Positologica III, both theoretically and experimentally. The axial point spread function (PSF) was approximately a triangle at or near the center of the field, and the sensitivity for the slice decreased significantly as the source moved away off the image plane. Accordingly, there were low sensitivity areas between an in-plane and the adjacent cross-plane. This invisible region was clinically significant if the object was thin enough in the z-axis. In order to fill up the gaps between adjacent slices, it is valuable to move the patient half the slice interval in the z-axis and perform an " interpolating scan."
Error estimates of Lagrange interpolation and orthonormal expansions for Freud weights
Kwon, K. H.; Lee, D. W.
2001-08-01
Let Sn[f] be the nth partial sum of the orthonormal polynomials expansion with respect to a Freud weight. Then we obtain sufficient conditions for the boundedness of Sn[f] and discuss the speed of the convergence of Sn[f] in weighted Lp space. We also find sufficient conditions for the boundedness of the Lagrange interpolation polynomial Ln[f], whose nodal points are the zeros of orthonormal polynomials with respect to a Freud weight. In particular, if W(x)=e-(1/2)x2 is the Hermite weight function, then we obtain sufficient conditions for the inequalities to hold:andwhere and k=0,1,2...,r.
International Nuclear Information System (INIS)
Ng, H.P.; Foong, K.W.C.; Ong, S.H.; Liu, J.; Nowinski, W.L.; Goh, P.S.
2007-01-01
The masseter plays a critical role in the mastication system. A hybrid method to shape-based interpolation is used to build the masseter model from magnetic resonance (MR) data sets. The main contribution here is the localizing of determinative slices in the data sets where clinicians are required to perform manual segmentations in order for an accurate model to be built. Shape-based criteria were used to locate the candidates for determinative slices and fuzzy-c-means (FCM) clustering technique was used to establish the determinative slices. Five masseter models were built in our work and the average overlap indices (κ) achieved is 85.2%. This indicates that there is good agreement between the models and the manual contour tracings. In addition, the time taken, as compared to manually segmenting all the slices, is significantly lesser. (orig.)
Polynomial-interpolation algorithm for van der Pauw Hall measurement in a metal hydride film
Koon, D. W.; Ares, J. R.; Leardini, F.; Fernández, J. F.; Ferrer, I. J.
2008-10-01
We apply a four-term polynomial-interpolation extension of the van der Pauw Hall measurement technique to a 330 nm Mg-Pd bilayer during both absorption and desorption of hydrogen at room temperature. We show that standard versions of the van der Pauw DC Hall measurement technique produce an error of over 100% due to a drifting offset signal and can lead to unphysical interpretations of the physical processes occurring in this film. The four-term technique effectively removes this source of error, even when the offset signal is drifting by an amount larger than the Hall signal in the time interval between successive measurements. This technique can be used to increase the resolution of transport studies of any material in which the resistivity is rapidly changing, particularly when the material is changing from metallic to insulating behavior.
Yang, Gaobo; Zhang, ZhaoYang; Chen, Jie
2003-09-01
Object-based segmentation of image sequences is one of the issues often arise in the world of video processing and communications. In this paper, a robust semiautomatic video object segmentation scheme is proposed. To facilitate users defining the initial object contour efficiently and accurately, an improved intelligent scissors is proposed by trading off the accuracy of original intelligent scissors and the simplicity of bounding box. To avoid the accumulated errors during object tracking, video sequence is firstly decomposed into video clips according to the rigidity of video object and the motion complexity. Then a snake-based bi-directional tracking is utilized to interpolate the video object planes (VOPs) of successive frames. Experimental results demonstrate that it can achieve better spatial accuracy and temporal coherency than COST211 AM, with about 10-22% improvement of spatial accuracy and almost the same temporal coherency.
SPATIAL INTERPOLATION AS A TOOL FOR SPECTRAL UNMIXING OF REMOTELY SENSED IMAGES
Directory of Open Access Journals (Sweden)
L. Xi
2012-07-01
Full Text Available Super resolution-based spectral unmixing (SRSU is a recently developed method for spectral unmixing of remotely sensed imagery, but it is too complex to implement for common users who are interested in land cover mapping. This study makes use of spatial interpolation as an alternative approach to achieve super resolution reconstruction in SRSU. An ASTER image with three spectral bands was used as the test data. The algorithm is evaluated using root mean square error (RMSE compared with linear spectral unmixing and hard classification. The result shows that the proposed algorithm has higher unmixing accuracy than those of the other comparative algorithms, and it is proved as an efficient and convenient spectral unmixing tool of remotely sensed imagery.
International Nuclear Information System (INIS)
Yamada, Yoshifumi; Liu, Na; Ito, Satoshi
2006-01-01
The signal in the Fresnel transform technique corresponds to a blurred one of the spin density image. Because the amplitudes of adjacent sampled signals have a high interrelation, the signal amplitude at a point between sampled points can be estimated with a high degree of accuracy even if the sampling is so coarse as to generate aliasing in the reconstructed images. In this report, we describe a new aliasless image reconstruction technique in the phase scrambling Fourier transform (PSFT) imaging technique in which the PSFT signals are converted to Fresnel transform signals by multiplying them by a quadratic phase term and are then interpolated using polynomial expressions to generate fully encoded signals. Numerical simulation using MR images showed that almost completely aliasless images are reconstructed by this technique. Experiments using ultra-low-field PSFT MRI were conducted, and aliasless images were reconstructed from coarsely sampled PSFT signals. (author)
Digital elevation modeling via curvature interpolation for LiDAR data
Directory of Open Access Journals (Sweden)
Hwamog Kim
2016-03-01
Full Text Available Digital elevation model (DEM is a three-dimensional (3D representation of a terrain's surface - for a planet (including Earth, moon, or asteroid - created from point cloud data which measure terrain elevation. Its modeling requires surface reconstruction for the scattered data, which is an ill-posed problem and most computational algorithms become overly expensive as the number of sample points increases. This article studies an effective partial differential equation (PDE-based algorithm, called the curvature interpolation method (CIM. The new method iteratively utilizes curvature information, estimated from an intermediate surface, to construct a reliable image surface that contains all of the data points. The CIM is applied for DEM for point cloud data acquired by light detection and ranging (LiDAR technology. It converges to a piecewise smooth image, requiring O(N operations independently of the number of sample points, where $N$ is the number of grid points.
Directory of Open Access Journals (Sweden)
Kalle Remm
2011-08-01
Full Text Available Maps of the long-term mean precipitation involving local landscape variables were generated for the Baltic countries, and the effectiveness of seven modelling methods was compared. The precipitation data were recorded in 245 meteorological stations in 1966–2005, and 51 location-related explanatory variables were used. The similarity-based reasoning in the Constud software system outperformed other methods according to the validation fit, except for spring. The multivariate adaptive regression splines (MARS was another effective method on average. The inclusion of landscape variables, compared to reverse distance-weighted interpolation, highlights the effect of uplands, larger water bodies and forested areas. The long-term mean amount of precipitation, calculated as the station average, probably underestimates the real value for Estonia and overestimates it for Lithuania due to the uneven distribution of observation stations.
Pearlstine, Leonard; Higer, Aaron; Palaseanu, Monica; Fujisaki, Ikuko; Mazzotti, Frank
2007-01-01
The Everglades Depth Estimation Network (EDEN) is an integrated network of real-time water-level monitoring, ground-elevation modeling, and water-surface modeling that provides scientists and managers with current (2000-present), online water-stage and water-depth information for the entire freshwater portion of the Greater Everglades. Continuous daily spatial interpolations of the EDEN network stage data are presented on a 400-square-meter grid spacing. EDEN offers a consistent and documented dataset that can be used by scientists and managers to (1) guide large-scale field operations, (2) integrate hydrologic and ecological responses, and (3) support biological and ecological assessments that measure ecosystem responses to the implementation of the Comprehensive Everglades Restoration Plan (CERP) The target users are biologists and ecologists examining trophic level responses to hydrodynamic changes in the Everglades.
Denisenko, M. V.; Klenov, N. V.; Satanin, A. M.
2018-01-01
In this article the dynamics of the qubits states based on solution of the time-dependent Schrödinger equation is investigated. Using the Magnus method we obtain an explicit interpolation representation for the propagator, which allows to find wave function at an arbitrary time. To illustrate the effectiveness of the approach, the population of the levels a single and two coupled qubits have been calculated by applying the Magnus propagator and the result have been compared with the numerical solution of the Schrödinger equation. As a measure of the approximation of the wave function, we calculate fidelity, which indicates proximity when the exact and approximate evolution operator acts on the initial state. We discuss the possibility of extending the developed methods to multi-qubits system, when high-speed calculation methods of the operators of evolution is particularly relevant.
International Nuclear Information System (INIS)
Mol, Antonio Carlos A.; Pereira, Claudio Marcio N.A.; Freitas, Victor Goncalves G.; Jorge, Carlos Alexandre F.
2011-01-01
This paper reports the most recent development results of a simulation tool for assessment of radiation dose exposition by nuclear plant's personnel, using artificial intelligence and virtual reality technologies. The main purpose of this tool is to support training of nuclear plants' personnel, to optimize working tasks for minimisation of received dose. A finer grid of measurement points was considered within the nuclear plant's room, for different power operating conditions. Further, an intelligent system was developed, based on neural networks, to interpolate dose rate values among measured points. The intelligent dose prediction system is thus able to improve the simulation of dose received by personnel. This work describes the improvements implemented in this simulation tool.
Interpolate with DIVA and view the products in OceanBrowser : what's up ?
Watelet, Sylvain; Barth, Alexander; Beckers, Jean-Marie; Troupin, Charles
2017-04-01
The Data-Interpolating Variational Analysis (DIVA) software is a statistical tool designed to reconstruct a continuous field from discrete measurements. This method is based on the numerical implementation of the Variational Inverse Model (VIM), which consists of a minimization of a cost function, allowing the choice of the analyzed field fitting at best the data sets without presenting unrealistic strong variations. The problem is solved efficiently using a finite-element method. This method, equivalent to the Optimal Interpolation, is particularly suited to deal with irregularly-spaced observations and produces outputs on a regular grid (2D, 3D or 4D). The results are stored in NetCDF files, the most widespread format in the earth sciences community. OceanBrowser is a web-service that allows one to visualize gridded fields on-line. Within the SeaDataNet and EMODNET (Chemical lot) projects, several national ocean data centers have created gridded climatologies of different ocean properties using the data analysis software DIVA. In order to give a common viewing service to those interpolated products, the GHER has developed OceanBrowser which is based on open standards from the Open Geospatial Consortium (OGC), in particular Web Map Service (WMS) and Web Feature Service (WFS). These standards define a protocol for describing, requesting and querying two-dimensional maps at a given depth and time. DIVA and OceanBrowser are both softwares tools which are continuously upgraded and distributed for free through frequent version releases. The development is funded by the EMODnet and SeaDataNet projects and include many discussions and feedback from the users community. Here, we present two recent major upgrades. First, we have implemented a "customization" of DIVA analyses following the sea bottom, using the bottom depth gradient as a new source of information. The weaker the slope of the bottom ocean, the higher the correlation length. This correlation length being
Directory of Open Access Journals (Sweden)
Peter Congdon
2013-10-01
Full Text Available This paper considers estimation of disease prevalence for small areas (neighbourhoods when the available observations on prevalence are for an alternative partition of a region, such as service areas. Interpolation to neighbourhoods uses a kernel method extended to take account of two types of collateral information. The first is morbidity and service use data, such as hospital admissions, observed for neighbourhoods. Variations in morbidity and service use are expected to reflect prevalence. The second type of collateral information is ecological risk factors (e.g., pollution indices that are expected to explain variability in prevalence in service areas, but are typically observed only for neighbourhoods. An application involves estimating neighbourhood asthma prevalence in a London health region involving 562 neighbourhoods and 189 service (primary care areas.
Calibration of Pyrometers by Using Extrapolation and Interpolation Methods at NIM
Lu, X.; Yuan, Z.; Wang, J.; Bai, C.; Wang, T.; Dong, W.
2018-01-01
High-temperature fixed points (HTFPs) have been thoroughly investigated, and the performance of variable temperature blackbodies (VTBB) has also improved rapidly. These two are beginning to be used in the calibration of pyrometers; however, tungsten strip lamps (STSL) still play a role in the dissemination of the high-temperature scale in China. International Temperature Scale of 1990 values of HTFPs and the lamps were assigned on a primary standard pyrometer (PSP) and were traced to the primary standard of the high-temperature scale at the National Institute of Metrology. In this paper, two pyrometers calibrated by using extrapolation and interpolation methods are reported. The values of the calibration were compared against the STSL values and the PSP values on HTBB, and their uncertainties are calculated as well. Because the stability of the HTFPs was better than that of the lamps, the calibration chains based on the lamps are starting to be replaced by HTFPs and VTBBs in China.
GM-PHD Filter Combined with Track-Estimate Association and Numerical Interpolation
Directory of Open Access Journals (Sweden)
Jinguang Chen
2015-01-01
Full Text Available For the standard Gaussian mixture probability hypothesis density (GM-PHD filter, the number of targets can be overestimated if the clutter rate is too high or underestimated if the detection rate is too low. These problems seriously affect the accuracy of multitarget tracking for the number and the value of measurements and clutters cannot be distinguished and recognized. Therefore, we proposed an improved GM-PHD filter to tackle these problems. Firstly, a track-estimate association was implemented in the filtering process to detect and remove false-alarm targets. Secondly, a numerical interpolation technique was used to compensate the missing targets caused by low detection rate. At the end of this paper, simulation results were presented to demonstrate the proposed GM-PHD algorithm is more effective in estimating the number and state of targets than the previous ones.
Energy Technology Data Exchange (ETDEWEB)
Ng, H.P. [NUS Graduate School for Integrative Sciences and Engineering (Singapore); Biomedical Imaging Lab., Agency for Science Technology and Research (Singapore); Foong, K.W.C. [NUS Graduate School for Integrative Sciences and Engineering (Singapore); Dept. of Preventive Dentistry, National Univ. of Singapore (Singapore); Ong, S.H. [Dept. of Electrical and Computer Engineering, National Univ. of Singapore (Singapore); Div. of Bioengineering, National Univ. of Singapore (Singapore); Liu, J.; Nowinski, W.L. [Biomedical Imaging Lab., Agency for Science Technology and Research (Singapore); Goh, P.S. [Dept. of Diagnostic Radiology, National Univ. of Singapore (Singapore)
2007-06-15
The masseter plays a critical role in the mastication system. A hybrid method to shape-based interpolation is used to build the masseter model from magnetic resonance (MR) data sets. The main contribution here is the localizing of determinative slices in the data sets where clinicians are required to perform manual segmentations in order for an accurate model to be built. Shape-based criteria were used to locate the candidates for determinative slices and fuzzy-c-means (FCM) clustering technique was used to establish the determinative slices. Five masseter models were built in our work and the average overlap indices ({kappa}) achieved is 85.2%. This indicates that there is good agreement between the models and the manual contour tracings. In addition, the time taken, as compared to manually segmenting all the slices, is significantly lesser. (orig.)
Congdon, Peter
2013-10-14
This paper considers estimation of disease prevalence for small areas (neighbourhoods) when the available observations on prevalence are for an alternative partition of a region, such as service areas. Interpolation to neighbourhoods uses a kernel method extended to take account of two types of collateral information. The first is morbidity and service use data, such as hospital admissions, observed for neighbourhoods. Variations in morbidity and service use are expected to reflect prevalence. The second type of collateral information is ecological risk factors (e.g., pollution indices) that are expected to explain variability in prevalence in service areas, but are typically observed only for neighbourhoods. An application involves estimating neighbourhood asthma prevalence in a London health region involving 562 neighbourhoods and 189 service (primary care) areas.
Evaluating the perceived voice quality on VoIP network using interpolated FIR filter algorithm
Pal Singh, Harjit; Singh, Sarabjeet; Sarin, R. K.; Singh, Jasvir
2012-10-01
Voice over Internet Protocol (VoIP) is a popular communication service nowadays. VoIP reduces the cost of call transmission by passing voice and video packets through the available bandwidth for data packets through Internet protocol. The quality of the VoIP signal is degraded due to the various network impairments. The proposed scheme, interpolated finite impulse response, is implemented as post-processor after decoding the signal in VoIP system. The performance of the proposed scheme is evaluated for various network conditions. The results of the proposed scheme are measured with the objective measurement methods for signal quality evaluation. The performance of the proposed system is compared with the existing techniques for quality improvement in VoIP system. The results show much improvement in speech quality with the proposed scheme in comparison to other similar schemes.
Gülşen, Esra; Kurtulus, Bedri; Necati Yaylim, Tolga; Avsar, Ozgur
2017-04-01
In groundwater studies, quantification and detection of fluid flows in borehole is an important part of assessment aquifer characteristic at different depths. Monitoring wells disturbs the natural flow field and this disturbance creates different flow paths to an aquifer. Vertical flow fluid analyses are one of the important techniques to deal with the detection and quantification of these vertical flows in borehole/monitoring wells. Liwa region is located about 146 km to the south west of Abu Dhabi city and about 36 km southwest of Madinat Zayed. SWSR (Strategic Water Storage & Recovery Project) comprises three Schemes (A, B and C) and each scheme contains an infiltration basin in the center, 105 recovery wells, 10 clusters and each cluster comprises 3 monitoring wells with different depths; shallow ( 50 m), intermediate ( 75 m) and deep ( 100 m). The scope of this study is to calculate the transmissivity values at different depth and evaluate the Fluid Flow Log (FFL) data for Scheme A (105 recovery wells) in order to understand the aquifer characteristic at different depths. The transmissivity values at different depth levels are calculated using Razack and Huntley (1991) equation for vertical flow rates of 30 m3 /h, 60 m3 /h, 90 m3 /h, 120 m3 /h and then Empirical Bayesian Kriging is used for interpolation in Scheme A using ArcGIS 10.2 software. FFL are drawn by GeODin software. Derivative analysis of fluid flow data are done by Microsoft Office: Excel software. All statistical analyses are calculated by IBMSPSS software. The interpolation results show that the transmissivity values are higher at the top of the aquifer. In other word, the aquifer is found more productive at the upper part of the Liwa aquifer. We are very grateful for financial support and providing us the data to ZETAS Dubai Inc.
Received Signal Strength Database Interpolation by Kriging for a Wi-Fi Indoor Positioning System.
Jan, Shau-Shiun; Yeh, Shuo-Ju; Liu, Ya-Wen
2015-08-28
The main approach for a Wi-Fi indoor positioning system is based on the received signal strength (RSS) measurements, and the fingerprinting method is utilized to determine the user position by matching the RSS values with the pre-surveyed RSS database. To build a RSS fingerprint database is essential for an RSS based indoor positioning system, and building such a RSS fingerprint database requires lots of time and effort. As the range of the indoor environment becomes larger, labor is increased. To provide better indoor positioning services and to reduce the labor required for the establishment of the positioning system at the same time, an indoor positioning system with an appropriate spatial interpolation method is needed. In addition, the advantage of the RSS approach is that the signal strength decays as the transmission distance increases, and this signal propagation characteristic is applied to an interpolated database with the Kriging algorithm in this paper. Using the distribution of reference points (RPs) at measured points, the signal propagation model of the Wi-Fi access point (AP) in the building can be built and expressed as a function. The function, as the spatial structure of the environment, can create the RSS database quickly in different indoor environments. Thus, in this paper, a Wi-Fi indoor positioning system based on the Kriging fingerprinting method is developed. As shown in the experiment results, with a 72.2% probability, the error of the extended RSS database with Kriging is less than 3 dBm compared to the surveyed RSS database. Importantly, the positioning error of the developed Wi-Fi indoor positioning system with Kriging is reduced by 17.9% in average than that without Kriging.
Yang, Qi; Zhang, Yanzhu; Zhao, Tiebiao; Chen, YangQuan
2017-04-04
Image super-resolution using self-optimizing mask via fractional-order gradient interpolation and reconstruction aims to recover detailed information from low-resolution images and reconstruct them into high-resolution images. Due to the limited amount of data and information retrieved from low-resolution images, it is difficult to restore clear, artifact-free images, while still preserving enough structure of the image such as the texture. This paper presents a new single image super-resolution method which is based on adaptive fractional-order gradient interpolation and reconstruction. The interpolated image gradient via optimal fractional-order gradient is first constructed according to the image similarity and afterwards the minimum energy function is employed to reconstruct the final high-resolution image. Fractional-order gradient based interpolation methods provide an additional degree of freedom which helps optimize the implementation quality due to the fact that an extra free parameter α-order is being used. The proposed method is able to produce a rich texture detail while still being able to maintain structural similarity even under large zoom conditions. Experimental results show that the proposed method performs better than current single image super-resolution techniques. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
Kumari, Madhuri; Singh, Chander Kumar; Bakimchandra, Oinam; Basistha, Ashoke
2017-10-01
In mountainous region with heterogeneous topography, the geostatistical modeling of the rainfall using global data set may not confirm to the intrinsic hypothesis of stationarity. This study was focused on improving the precision of the interpolated rainfall maps by spatial stratification in complex terrain. Predictions of the normal annual rainfall data were carried out by ordinary kriging, universal kriging, and co-kriging, using 80-point observations in the Indian Himalayas extending over an area of 53,484 km2. A two-step spatial clustering approach is proposed. In the first step, the study area was delineated into two regions namely lowland and upland based on the elevation derived from the digital elevation model. The delineation was based on the natural break classification method. In the next step, the rainfall data was clustered into two groups based on its spatial location in lowland or upland. The terrain ruggedness index (TRI) was incorporated as a co-variable in co-kriging interpolation algorithm. The precision of the kriged and co-kriged maps was assessed by two accuracy measures, root mean square error and Chatfield's percent better. It was observed that the stratification of rainfall data resulted in 5-20 % of increase in the performance efficiency of interpolation methods. Co-kriging outperformed the kriging models at annual and seasonal scale. The result illustrates that the stratification of the study area improves the stationarity characteristic of the point data, thus enhancing the precision of the interpolated rainfall maps derived using geostatistical methods.
Csébfalvi, Balázs
2010-01-01
In this paper, we demonstrate that quasi-interpolation of orders two and four can be efficiently implemented on the Body-Centered Cubic (BCC) lattice by using tensor-product B-splines combined with appropriate discrete prefilters. Unlike the nonseparable box-spline reconstruction previously proposed for the BCC lattice, the prefiltered B-spline reconstruction can utilize the fast trilinear texture-fetching capability of the recent graphics cards. Therefore, it can be applied for rendering BCC-sampled volumetric data interactively. Furthermore, we show that a separable B-spline filter can suppress the postaliasing effect much more isotropically than a nonseparable box-spline filter of the same approximation power. Although prefilters that make the B-splines interpolating on the BCC lattice do not exist, we demonstrate that quasi-interpolating prefiltered linear and cubic B-spline reconstructions can still provide similar or higher image quality than the interpolating linear box-spline and prefiltered quintic box-spline reconstructions, respectively.
Geer, F.C. van; Zuur, A.F.
1997-01-01
This paper advocates an approach to extend single-output Box-Jenkins transfer/noise models for several groundwater head series to a multiple-output transfer/noise model. The approach links several groundwater head series and enables a spatial interpolation in terms of time series analysis. Our
Hofstra, N.; New, M.
2009-01-01
Angular-distance weighting (ADW) is a common approach for interpolation of an irregular network of meteorological observations to a regular grid. A widely used version of ADW employs the correlation decay distance (CDD) to (1) select stations that should contribute to each grid-point estimate and
K.O. Dzhaparidze (Kacha)
1998-01-01
textabstractIn this paper a convergence class is characterized for special series associated with Gelfond's interpolation problem (a generalization of the Abel-Goncharov problem) when the interpolation nodes are equidistantly distributed within the interval $[0,1]$. As a result, an expansion is
Liu, Xuan; Liu, Bo; Zhang, Li-jia; Xin, Xiang-jun; Zhang, Qi; Wang, Yong-jun; Tian, Qing-hua; Tian, Feng; Mao, Ya-ya
2018-01-01
Traditional clock recovery scheme achieves timing adjustment by digital interpolation, thus recovering the sampling sequence. Based on this, an improved clock recovery architecture joint channel equalization for coherent optical communication system is presented in this paper. The loop is different from the traditional clock recovery. In order to reduce the interpolation error caused by the distortion in the frequency domain of the interpolator and to suppress the spectral mirroring generated by the sampling rate change, the proposed algorithm joint equalization, improves the original interpolator in the loop, along with adaptive filtering, and makes error compensation for the original signals according to the balanced pre-filtering signals. Then the signals are adaptive interpolated through the feedback loop. Furthermore, the phase splitting timing recovery algorithm is adopted in this paper. The time error is calculated according to the improved algorithm when there is no transition between the adjacent symbols, making calculated timing error more accurate. Meanwhile, Carrier coarse synchronization module is placed before the beginning of timing recovery to eliminate the larger frequency offset interference, which effectively adjust the sampling clock phase. In this paper, the simulation results show that the timing error is greatly reduced after the loop is changed. Based on the phase splitting algorithm, the BER and MSE are better than those in the unvaried architecture. In the fiber channel, using MQAM modulation format, after 100 km-transmission of single-mode fiber, especially when ROF(roll-off factor) values tends to 0, the algorithm shows a better clock performance under different ROFs. When SNR values are less than 8, the BER could achieve 10-2 to 10-1 magnitude. Furthermore, the proposed timing recovery is more suitable for the situation with low SNR values.
International Nuclear Information System (INIS)
T Sutton; T Trumbull
2005-01-01
Secondary neutron energy spectra used by Monte Carlo codes are often provided in tabular format. Examples are the spectra obtained from ENDF/B-VI File 5 when the LF parameter has the value 1. These secondary spectra are tabulated on an incident energy mesh, and in a Monte Carlo calculation the tabulated spectra are generally interpolated to the energy of the incident neutron. A common method of interpolation involves the use of the unit base transformation. The details of the implementation vary from code to code, so here we will simply focus on the mathematics of the method. Given an incident neutron with energy E, the bracketing points E i and E i+1 on the incident energy mesh are determined. The corresponding secondary energy spectra are transformed to a dimensionless energy coordinate system in which the secondary energies lie between zero and one. A dimensionless secondary energy is then sampled from a spectrum obtained by linearly interpolating the transformed spectra--often using the method of statistical interpolation. Finally, the sampled secondary energy is transformed back into the normal energy coordinate system. For this inverse transformation, the minimum and maximum energies are linearly interpolated from the values given in the non-transformed secondary spectra. The purpose of the unit base transformation is to preserve (as nearly as possible) the physics of the secondary distribution--in particular the minimum and maximum energies possible for the secondary neutron. This method is used by several codes including MCNP and the new MC21 code that is the subject of this paper. In comparing MC21 results to those of MCNP, it was discovered that the nuclear data supplied to MCNP is structured in such a way that the code may not be doing the best possible job of preserving the physics of certain nuclear interactions. In this paper, we describe the problem and explain how it may be avoided
ZZ POINT-2007, linearly interpolable ENDF/B-VII.0 data for 14 temperatures
International Nuclear Information System (INIS)
Cullen, Dermott E.
2007-01-01
A - Description or function: The ENDF/B data library, ENDF/B-VII.0 was processed into the form of temperature dependent cross sections. The original evaluated data include cross sections represented in the form of a combination of resonance parameters and/or tabulated energy dependent cross sections, nominally at 0 Kelvin temperature. For use in applications, these ENDF/B-VII.0 data were processed into the form of temperature dependent cross sections at eight temperatures: 0, 300, 600, 900, 1200, 1500, 1800 and 2100 Kelvin. It has also been processed to six astrophysics like temperatures: 0.1, 1, 10, 100 eV, 1 and 10 keV. At each temperature the cross sections are tabulated and linearly interpolable in energy with a tolerance of 0.1 %. POINT 2007 contains all of the evaluations in the ENDF/B-VII general purpose library, which contains 78 new evaluations + 315 old ones: total 393 nuclides. It also includes 16 new elemental evaluations replaced by isotopic evaluations + 19 old ones. No special purpose ENDF/B-VII libraries, such as fission products, thermal scattering, photon interaction data are included. These evaluations include all cross sections over the energy range 10 e-5 eV to at least 20 MeV. The list of nuclides is indicated. B - Methods: The PREPRO 2007 code system was used to process the ENDF/B data. Listed below are the steps, including the PREPRO2007 codes, which were used to process the data in the order in which the codes were run. 1) Linearly interpolable, tabulated cross sections (LINEAR) 2) Including the resonance contribution (RECENT) 3) Doppler broaden all cross sections to temperature (SIGMA1) 4) Check data, define redundant cross sections by summation (FIXUP) 5) Update evaluation dictionary in MF/MT=1/451 (DICTIN) C - Restrictions: Due to recent changes in ENDF-6 Formats and Procedures only the latest version of the ENDF/B Pre-processing codes, namely PREPRO 2007, can be used to accurately process all current ENDF/B-VII evaluations. The use of
International Nuclear Information System (INIS)
Williams, Christopher S; Burggraf, Larry W; Petrosky, James C; Oxley, Mark E; Adamson, Paul E
2010-01-01
A three-dimensional Positron Annihilation Spectrometry System (3D PASS) for determination of 3D electron-positron (e - -e + ) momentum densities by measuring coincident annihilation photons was designed, constructed and characterized. 3D PASS collects a single data set including correlated photon energies and coincident photon positions which are typically collected separately by two-dimensional angular correlation of annihilation radiation (2D ACAR) and two-detector coincident Doppler broadening of annihilation radiation (CDBAR) spectrometry. 3D PASS is composed of two position-sensitive, high-purity germanium (HPGe) double-sided strip detectors (DSSD(s)) linked together by a 32-channel, 50 MHz digital electronics suite. The DSSDs data were analyzed to determine location of photon detection events using an interpolation method to achieve a spatial resolution less than the 5-mm width of the DSSDs' charge collection strips. The interpolation method relies on measuring a figure-of-merit proportional to the area of the transient charges observed on both strips directly adjacent to the charge collection strip detecting the full charge deposited by the annihilation photon. This sub-pixel resolution, corresponding to the error associated with event location within a sub-pixel was measured for both DSSDs using the approach outlined in Williams et al [1] and was on the order of ± 0.20 mm (± one-standard deviation). As a result of the sub-pixel resolution, the distance between the DSSDs and material sample was reduced by a factor of five compared to what is typically required in 2D ACAR systems was necessary to achieve 0.5-mrad angular resolution. This reduction in the system's footprint decreases attenuation of the annihilation photons in the air between the material sample and the DSSDs and increases the solid angle between the sample and the DSSDs, ultimately resulting in higher system detection efficiency. 3D PASS was characterized in the same manner comparable to
Using dynamical interpolation to map high-resolution altimeter data in the Western Mediterranean Sea
Roge, M.; Morrow, R.; Gerald, D.
2016-12-01
The main oceanographic objective of the future SWOT mission is to characterize the ocean mesoscale and submesoscale circulation by observing the fine range of ocean dynamics (from 15-300 km). However it will not capture the time evolution of short mesoscale signals. Despite the very high spatial resolution of the future satellite, the temporal resolution is not sufficient to track the evolution of the small, rapid features (exact repeat cycle of 21 days, with near repeats around 5-10 days, depending on the latitude). High resolution SWOT sea surface height snapshots alone will not allow us to follow the dynamics of ocean variability at these scales, such as the formation and evolution of small eddies. Here, we investigate a means to reconstruct the missing SSH signal in time between two satellite revisits. We use a shallow water quasi-geostrophic model developed by Ubelmann et al (2015). Based on potential vorticity conservation, it dynamically advects the SSH field, assuming that the quasi-geostrophic dynamics are principally captured by the first baroclinic mode. This model has been tested in energetic open ocean regions such as the Gulf Stream and the Californian Current, and has given improved results. Here we test this model in the Western Mediterranean Sea, where the first radius of deformation of Rossby is small (5-15 km), where the dynamics have a strong topographic control and strong spatial and seasonal variability. In this region, the technique provides a small improvement over linear interpolation in the coastal boundary current systems. The simple dynamical model is missing some physical mechanisms, needed to correctly represent the mesoscale circulation in this region, including a significant barotropic mode. We investigate modifications to the 1.5 layer model in this regional study, to include a topographic-beta effect and small-scale dissipation and an extension to a two-layer model. The results show an improved performance compared to simple linear
A Collision-Free G2 Continuous Path-Smoothing Algorithm Using Quadratic Polynomial Interpolation
Directory of Open Access Journals (Sweden)
Seong-Ryong Chang
2014-12-01
Full Text Available Most path-planning algorithms are used to obtain a collision-free path without considering continuity. On the other hand, a continuous path is needed for stable movement. In this paper, the searched path was converted into a G2 continuous path using the modified quadratic polynomial and membership function interpolation algorithm. It is simple, unique and provides a good geometric interpretation. In addition, a collision-checking and improvement algorithm is proposed. The collision-checking algorithm can check the collisions of a smoothed path. If collisions are detected, the collision improvement algorithm modifies the collision path to a collision-free path. The collision improvement algorithm uses a geometric method. This method uses the perpendicular line between a collision position and the collision piecewise linear path. The sub-waypoint is added, and the QPMI algorithm is applied again. As a result, the collision-smoothed path is converted into a collision-free smooth path without changing the continuity.
Prediction of selected Indian stock using a partitioning–interpolation based ARIMA–GARCH model
Directory of Open Access Journals (Sweden)
C. Narendra Babu
2015-07-01
Full Text Available Accurate long-term prediction of time series data (TSD is a very useful research challenge in diversified fields. As financial TSD are highly volatile, multi-step prediction of financial TSD is a major research problem in TSD mining. The two challenges encountered are, maintaining high prediction accuracy and preserving the data trend across the forecast horizon. The linear traditional models such as autoregressive integrated moving average (ARIMA and generalized autoregressive conditional heteroscedastic (GARCH preserve data trend to some extent, at the cost of prediction accuracy. Non-linear models like ANN maintain prediction accuracy by sacrificing data trend. In this paper, a linear hybrid model, which maintains prediction accuracy while preserving data trend, is proposed. A quantitative reasoning analysis justifying the accuracy of proposed model is also presented. A moving-average (MA filter based pre-processing, partitioning and interpolation (PI technique are incorporated by the proposed model. Some existing models and the proposed model are applied on selected NSE India stock market data. Performance results show that for multi-step ahead prediction, the proposed model outperforms the others in terms of both prediction accuracy and preserving data trend.
Directory of Open Access Journals (Sweden)
Michel Castro Moreira
Full Text Available ABSTRACT Water erosion is the process of disaggregation and transport of sediments, and rainfall erosivity is a numerical value that expresses the erosive capacity of rain. The scarcity of information on rainfall erosivity makes it difficult or impossible to use to estimate losses occasioned by the erosive process. The objective of this study was to develop Artificial Neural Networks (ANNs for spatial interpolation of the monthly and annual values of rainfall erosivity at any location in the state of Rio Grande do Sul, and a software that enables the use of these networks in a simple and fast manner. This experiment used 103 rainfall stations in Rio Grande do Sul and their surrounding area to generate synthetic rainfall series on the software ClimaBR 2.0. Rainfall erosivity was determined by summing the values of the EI30 and KE >25 indexes, considering two methodologies for obtaining the kinetic energy of rainfall. With these values of rainfall erosivity and latitude, longitude, and altitude of the stations, the ANNs were trained and tested for spatializations of rainfall erosivity. To facilitate the use of the ANNs, a computer program was generated, called netErosividade RS, which makes feasible the use of ANNs to estimate the values of rainfall erosivity for any location in the state of Rio Grande do Sul.
Wu, Wei; Tang, Xiao-Ping; Ma, Xue-Qing; Liu, Hong-Bin
2016-08-01
Soil temperature variability data provide valuable information on understanding land-surface ecosystem processes and climate change. This study developed and analyzed a spatial dataset of monthly mean soil temperature at a depth of 10 cm over a complex topographical region in southwestern China. The records were measured at 83 stations during the period of 1961-2000. Nine approaches were compared for interpolating soil temperature. The accuracy indicators were root mean square error (RMSE), modelling efficiency (ME), and coefficient of residual mass (CRM). The results indicated that thin plate spline with latitude, longitude, and elevation gave the best performance with RMSE varying between 0.425 and 0.592 °C, ME between 0.895 and 0.947, and CRM between -0.007 and 0.001. A spatial database was developed based on the best model. The dataset showed that larger seasonal changes of soil temperature were from autumn to winter over the region. The northern and eastern areas with hilly and low-middle mountains experienced larger seasonal changes.
Derivative-free generation and interpolation of convex Pareto optimal IMRT plans
International Nuclear Information System (INIS)
Hoffmann, Aswin L; Siem, Alex Y D; Hertog, Dick den; Kaanders, Johannes H A M; Huizenga, Henk
2006-01-01
In inverse treatment planning for intensity-modulated radiation therapy (IMRT), beamlet intensity levels in fluence maps of high-energy photon beams are optimized. Treatment plan evaluation criteria are used as objective functions to steer the optimization process. Fluence map optimization can be considered a multi-objective optimization problem, for which a set of Pareto optimal solutions exists: the Pareto efficient frontier (PEF). In this paper, a constrained optimization method is pursued to iteratively estimate the PEF up to some predefined error. We use the property that the PEF is convex for a convex optimization problem to construct piecewise-linear upper and lower bounds to approximate the PEF from a small initial set of Pareto optimal plans. A derivative-free Sandwich algorithm is presented in which these bounds are used with three strategies to determine the location of the next Pareto optimal solution such that the uncertainty in the estimated PEF is maximally reduced. We show that an intelligent initial solution for a new Pareto optimal plan can be obtained by interpolation of fluence maps from neighbouring Pareto optimal plans. The method has been applied to a simplified clinical test case using two convex objective functions to map the trade-off between tumour dose heterogeneity and critical organ sparing. All three strategies produce representative estimates of the PEF. The new algorithm is particularly suitable for dynamic generation of Pareto optimal plans in interactive treatment planning
Varanka, Dalia; Jiang, Bin; Yao, Xiaobai
2010-01-01
Measures of population pressure, referring in general to the stress upon the environment by human consumption of resources, are imperative for environmental sustainability studies and management. Development based on resource consumption is the predominant factor of population pressure. This paper presents a spatial model of population pressure by linking consumption associated with regional urbanism and ecosystem services. Maps representing relative geographic degree and extent of natural resource consumption and degree and extent of impacts on surrounding areas are new, and this research represents the theoretical research toward this goal. With development, such maps offer a visualization tool for planners of various services, amenities for people, and conservation planning for ecologist. Urbanization is commonly generalized by census numbers or impervious surface area. The potential geographical extent of urbanism encompasses the environmental resources of the surrounding region that sustain cities. This extent is interpolated using kriging of a variable based on population wealth data from the U.S. Census Bureau. When overlayed with land-use/land-cover data, the results indicate that the greatest estimates of population pressure fall within mixed forest areas. Mixed forest areas result from the spread of cedar woods in previously disturbed areas where further disturbance is then suppressed. Low density areas, such as suburbanization and abandoned farmland are characteristic of mixed forest areas.
Dai, K. Y.; Liu, G. R.; Lim, K. M.; Han, X.; Du, S. Y.
A meshfree model is presented for the static and dynamic analyses of functionally graded material (FGM) plates based on the radial point interpolation method (PIM). In the present method, the mid-plane of an FGM plate is represented by a set of distributed nodes while the material properties in its thickness direction are computed analytically to take into account their continuous variations from one surface to another. Several examples are successfully analyzed for static deflections, natural frequencies and dynamic responses of FGM plates with different volume fraction exponents and boundary conditions. The convergence rate and accuracy are studied and compared with the finite element method (FEM). The effects of the constituent fraction exponent on static deflection as well as natural frequency are also investigated in detail using different FGM models. Based on the current material gradient, it is found that as the volume fraction exponent increases, the mechanical characteristics of the FGM plate approach those of the pure metal plate blended in the FGM.
Directory of Open Access Journals (Sweden)
Marcin Kiedrzyński
2014-07-01
Full Text Available Attempts to study biodiversity hotspots on a regional scale should combine compositional and functionalist criteria. The detection of hotspots in this study uses one ecologically similar group of high conservation value species as hotspot indicators, as well as focal habitat indicators, to detect the distribution of suitable environmental conditions. The method is assessed with reference to thermophilous forests in Poland – key habitats for many rare and relict species. Twenty-six high conservation priority species were used as hotspot indicators, and ten plant taxa characteristic of the Quercetalia pubescenti-petraeae phytosociological order were used as focal habitat indicators. Species distribution data was based on a 10 × 10 km grid. The number of species per grid square was interpolated by the ordinary kriging geostatistical method. Our analysis largely determined the distribution of areas with concentration of thermophilous forest flora, but also regional disjunctions and geographical barriers. Indicator species richness can be interpreted as a reflection of the actual state of habitat conditions. It can also be used to determine the location of potential species refugia and possible past and future migration routes.