On Network-Error Correcting Convolutional Codes under the BSC Edge Error Model
Prasad, K
2010-01-01
Convolutional network-error correcting codes (CNECCs) are known to provide error correcting capability in acyclic instantaneous networks within the network coding paradigm under small field size conditions. In this work, we investigate the performance of CNECCs under the error model of the network where the edges are assumed to be statistically independent binary symmetric channels, each with the same probability of error $p_e$($0\\leq p_e<0.5$). We obtain bounds on the performance of such CNECCs based on a modified generating function (the transfer function) of the CNECCs. For a given network, we derive a mathematical condition on how small $p_e$ should be so that only single edge network-errors need to be accounted for, thus reducing the complexity of evaluating the probability of error of any CNECC. Simulations indicate that convolutional codes are required to possess different properties to achieve good performance in low $p_e$ and high $p_e$ regimes. For the low $p_e$ regime, convolutional codes with g...
A Matroidal Framework for Network-Error Correcting Codes
Prasad, K
2012-01-01
Matroidal networks were introduced by Dougherty et al. and have been well studied in the recent past. It was shown that a network has a scalar linear network coding solution if and only if it is matroidal associated with a representable matroid. A particularly interesting feature of this development is the ability to construct (scalar and vector) linearly solvable networks using certain classes of matroids. The current work attempts to establish a connection between matroid theory and network-error correcting codes. In a similar vein to the theory connecting matroids and network coding, we abstract the essential aspects of network-error correcting codes to arrive at the definition of a matroidal error correcting network. An acyclic network (with arbitrary sink demands) is then shown to possess a scalar linear error correcting network code if and only if there it is a matroidal error correcting network associated with a representable matroid. Therefore, constructing such network-error correcting codes implies ...
Low-cost ultrasonic distance sensor arrays with networked error correction.
Dai, Hongjun; Zhao, Shulin; Jia, Zhiping; Chen, Tianzhou
2013-09-05
Distance has been one of the basic factors in manufacturing and control fields, and ultrasonic distance sensors have been widely used as a low-cost measuring tool. However, the propagation of ultrasonic waves is greatly affected by environmental factors such as temperature, humidity and atmospheric pressure. In order to solve the problem of inaccurate measurement, which is significant within industry, this paper presents a novel ultrasonic distance sensor model using networked error correction (NEC) trained on experimental data. This is more accurate than other existing approaches because it uses information from indirect association with neighboring sensors, which has not been considered before. The NEC technique, focusing on optimization of the relationship of the topological structure of sensor arrays, is implemented for the compensation of erroneous measurements caused by the environment. We apply the maximum likelihood method to determine the optimal fusion data set and use a neighbor discovery algorithm to identify neighbor nodes at the top speed. Furthermore, we adopt the NEC optimization algorithm, which takes full advantage of the correlation coefficients for neighbor sensors. The experimental results demonstrate that the ranging errors of the NEC system are within 2.20%; furthermore, the mean absolute percentage error is reduced to 0.01% after three iterations of this method, which means that the proposed method performs extremely well. The optimized method of distance measurement we propose, with the capability of NEC, would bring a significant advantage for intelligent industrial automation.
Low-Cost Ultrasonic Distance Sensor Arrays with Networked Error Correction
Directory of Open Access Journals (Sweden)
Tianzhou Chen
2013-09-01
Full Text Available Distance has been one of the basic factors in manufacturing and control fields, and ultrasonic distance sensors have been widely used as a low-cost measuring tool. However, the propagation of ultrasonic waves is greatly affected by environmental factors such as temperature, humidity and atmospheric pressure. In order to solve the problem of inaccurate measurement, which is significant within industry, this paper presents a novel ultrasonic distance sensor model using networked error correction (NEC trained on experimental data. This is more accurate than other existing approaches because it uses information from indirect association with neighboring sensors, which has not been considered before. The NEC technique, focusing on optimization of the relationship of the topological structure of sensor arrays, is implemented for the compensation of erroneous measurements caused by the environment. We apply the maximum likelihood method to determine the optimal fusion data set and use a neighbor discovery algorithm to identify neighbor nodes at the top speed. Furthermore, we adopt the NEC optimization algorithm, which takes full advantage of the correlation coefficients for neighbor sensors. The experimental results demonstrate that the ranging errors of the NEC system are within 2.20%; furthermore, the mean absolute percentage error is reduced to 0.01% after three iterations of this method, which means that the proposed method performs extremely well. The optimized method of distance measurement we propose, with the capability of NEC, would bring a significant advantage for intelligent industrial automation.
Forward Error Correction Convolutional Codes for RTAs' Networks: An Overview
Directory of Open Access Journals (Sweden)
Salehe I. Mrutu
2014-06-01
Full Text Available For more than half a century, Forward Error Correction Convolutional Codes (FEC-CC have been in use to provide reliable data communication over various communication networks. The recent high increase of mobile communication services that require both bandwidth intensive and interactive Real Time Applications (RTAs impose an increased demand for fast and reliable wireless communication networks. Transmission burst errors; data decoding complexity and jitter are identified as key factors influencing the quality of service of RTAs implementation over wireless transmission media. This paper reviews FEC-CC as one of the most commonly used algorithm in Forward Error Correction for the purpose of improving its operational performance. Under this category, we have analyzed various previous works for their strengths and weaknesses in decoding FEC-CC. A comparison of various decoding algorithms is made based on their decoding computational complexity.
An upper bound on the number of errors corrected by a convolutional code
DEFF Research Database (Denmark)
Justesen, Jørn
2000-01-01
The number of errors that a convolutional codes can correct in a segment of the encoded sequence is upper bounded by the number of distinct syndrome sequences of the relevant length.......The number of errors that a convolutional codes can correct in a segment of the encoded sequence is upper bounded by the number of distinct syndrome sequences of the relevant length....
Convolution kernel design and efficient algorithm for sampling density correction.
Johnson, Kenneth O; Pipe, James G
2009-02-01
Sampling density compensation is an important step in non-cartesian image reconstruction. One of the common techniques to determine weights that compensate for differences in sampling density involves a convolution. A new convolution kernel is designed for sampling density attempting to minimize the error in a fully reconstructed image. The resulting weights obtained using this new kernel are compared with various previous methods, showing a reduction in reconstruction error. A computationally efficient algorithm is also presented that facilitates the calculation of the convolution of finite kernels. Both the kernel and the algorithm are extended to 3D. Copyright 2009 Wiley-Liss, Inc.
Upper bounds on the number of errors corrected by a convolutional code
DEFF Research Database (Denmark)
Justesen, Jørn
2004-01-01
We derive upper bounds on the weights of error patterns that can be corrected by a convolutional code with given parameters, or equivalently we give bounds on the code rate for a given set of error patterns. The bounds parallel the Hamming bound for block codes by relating the number of error pat...
Off-resonance artifacts correction with convolution in k-space (ORACLE).
Lin, Wei; Huang, Feng; Simonotto, Enrico; Duensing, George R; Reykowski, Arne
2012-06-01
Off-resonance artifacts hinder the wider applicability of echo-planar imaging and non-Cartesian MRI methods such as radial and spiral. In this work, a general and rapid method is proposed for off-resonance artifacts correction based on data convolution in k-space. The acquired k-space is divided into multiple segments based on their acquisition times. Off-resonance-induced artifact within each segment is removed by applying a convolution kernel, which is the Fourier transform of an off-resonance correcting spatial phase modulation term. The field map is determined from the inverse Fourier transform of a basis kernel, which is calibrated from data fitting in k-space. The technique was demonstrated in phantom and in vivo studies for radial, spiral and echo-planar imaging datasets. For radial acquisitions, the proposed method allows the self-calibration of the field map from the imaging data, when an alternating view-angle ordering scheme is used. An additional advantage for off-resonance artifacts correction based on data convolution in k-space is the reusability of convolution kernels to images acquired with the same sequence but different contrasts. Copyright © 2011 Wiley-Liss, Inc.
Convolution effect on TCR log response curve and the correction method for it
Chen, Q.; Liu, L. J.; Gao, J.
2016-09-01
Through-casing resistivity (TCR) logging has been successfully used in production wells for the dynamic monitoring of oil pools and the distribution of the residual oil, but its vertical resolution has limited its efficiency in identification of thin beds. The vertical resolution is limited by the distortion phenomenon of vertical response of TCR logging. The distortion phenomenon was studied in this work. It was found that the vertical response curve of TCR logging is the convolution of the true formation resistivity and the convolution function of TCR logging tool. Due to the effect of convolution, the measurement error at thin beds can reach 30% or even bigger. Thus the information of thin bed might be covered up very likely. The convolution function of TCR logging tool was obtained in both continuous and discrete way in this work. Through modified Lyle-Kalman deconvolution method, the true formation resistivity can be optimally estimated, so this inverse algorithm can correct the error caused by the convolution effect. Thus it can improve the vertical resolution of TCR logging tool for identification of thin beds.
Some physical factors influencing the accuracy of convolution scatter correction in SPECT.
Msaki, P; Axelsson, B; Larsson, S A
1989-03-01
Some important physical factors influencing the accuracy of convolution scatter correction techniques in SPECT are presented. In these techniques scatter correction in the projection relies on filter functions, QF, evaluated by Fourier transforms, from measured scatter functions, Qp, obtained from point spread functions. The spatial resolution has a marginal effect on Qp. Thus a single QF can be used in the scatter correction of SPECT measurements acquired with the low energy high resolution or the low energy general purpose collimators and over a wide range of patient-collimator distances. However, it is necessary to examine the details of the shape of point spread functions during evaluation of Qp. QF is completely described by scatter amplitude AF, slope BF and filter sum SF. SF is obtained by summation of the values of QF occupying a 31 x 31 pixels matrix. Regardless of differences in amplitude and slope, two filter functions are shown to be equivalent in terms of scatter correction ability, whenever their sums are equal. On the basis of filter sum, the observed small influence of ellipticity on QF implies that an average function can be used in scatter correcting SPECT measurements conducted with elliptic objects. SF is shown to increase with a decrease in photon energy and with an increase in window size. Thus, scatter correction by convolution may be severely hampered by photon statistics when SPECT imaging is done with low-energy photons. It is pointless to use unnecessarily large discriminator windows, in the hope of improving photon statistics, since most of the extra events acquired will eventually be subtracted during scatter correction. Regardless of the observed moderate reduction in SF when a lung-equivalent material replaces a portion of a water phantom, further studies are needed to develop a technique that is capable of handling attenuation and scatter corrections simultaneously. Whenever superficial and inner radioactive distributions coexist the
Scattering correction based on regularization de-convolution for Cone-Beam CT
Xie, Shi-peng
2016-01-01
In Cone-Beam CT (CBCT) imaging systems, the scattering phenomenon has a significant impact on the reconstructed image and is a long-lasting research topic on CBCT. In this paper, we propose a simple, novel and fast approach for mitigating scatter artifacts and increasing the image contrast in CBCT, belonging to the category of convolution-based method in which the projected data is de-convolved with a convolution kernel. A key step in this method is how to determine the convolution kernel. Compared with existing methods, the estimation of convolution kernel is based on bi-l1-l2-norm regularization imposed on both the intermediate the known scatter contaminated projection images and the convolution kernel. Our approach can reduce the scatter artifacts from 12.930 to 2.133.
Using GPU convolutions to correct optical distortion in closed-loop real-time missile simulations
Fronckowiak, Thomas, Jr.
2009-05-01
U.S. Army Aviation and Missile Research, Development and Engineering Center (AMRDEC) has long been a leader in in-band high fidelity scientific scene generation. Recent efforts to harness and exploit the parallel power of the Graphics Processor Unit (GPU), for both graphics and general purpose processing, have been paramount. The emergence of sophisticated image generation software packages, such as the Common Scene Generator (CSG) and the Joint Signature Image Generator (JSIG), have lead to a sharp increase in the performance of digital simulations and signal injection and projection systems in both tactical and strategic programs. One area of missile simulations that benefits from this technology is real-time modeling of optical effects, such as seeker dome distortion, glint, blurring effects, and correcting for facility misalignment and distortion. This paper discusses the on-going research of applying convolution filters to the GPU multi-pass rendering process to compensate for spatial distortion in the optical projection path for synthetic environments.
Scattering correction based on regularization de-convolution for Cone-Beam CT
Xie, Shi-peng; Yan, Rui-ju
2016-01-01
In Cone-Beam CT (CBCT) imaging systems, the scattering phenomenon has a significant impact on the reconstructed image and is a long-lasting research topic on CBCT. In this paper, we propose a simple, novel and fast approach for mitigating scatter artifacts and increasing the image contrast in CBCT, belonging to the category of convolution-based method in which the projected data is de-convolved with a convolution kernel. A key step in this method is how to determine the convolution kernel. Co...
DEFF Research Database (Denmark)
Silver, Jeremy D; Ritchie, Matthew E; Smyth, Gordon K
2009-01-01
is developed for exact maximum likelihood estimation (MLE) using high-quality optimization software and using the saddle-point estimates as starting values. "MLE" is shown to outperform heuristic estimators proposed by other authors, both in terms of estimation accuracy and in terms of performance on real data....... The saddle-point approximation is an adequate replacement in most practical situations. The performance of normexp for assessing differential expression is improved by adding a small offset to the corrected intensities....
Energy Technology Data Exchange (ETDEWEB)
Larsson, Anne [Department of Radiation Sciences, Radiation Physics, Umeaa University, SE-901 87 Umeaa (Sweden); Ljungberg, Michael [Medical Radiation Physics, Department of Clinical Sciences, Lund, Lund University, SE-221 85 Lund (Sweden); Mo, Susanna Jakobson [Department of Radiation Sciences, Diagnostic Radiology, Umeaa University, SE-901 87 Umeaa (Sweden); Riklund, Katrine [Department of Radiation Sciences, Diagnostic Radiology, Umeaa University, SE-901 87 Umeaa (Sweden); Johansson, Lennart [Department of Radiation Sciences, Radiation Physics, Umeaa University, SE-901 87 Umeaa (Sweden)
2006-11-21
Scatter and septal penetration deteriorate contrast and quantitative accuracy in single photon emission computed tomography (SPECT). In this study four different correction techniques for scatter and septal penetration are evaluated for {sup 123}I brain SPECT. One of the methods is a form of model-based compensation which uses the effective source scatter estimation (ESSE) for modelling scatter, and collimator-detector response (CDR) including both geometric and penetration components. The other methods, which operate on the 2D projection images, are convolution scatter subtraction (CSS) and two versions of transmission dependent convolution subtraction (TDCS), one of them proposed by us. This method uses CSS for correction for septal penetration, with a separate kernel, and TDCS for scatter correction. The corrections are evaluated for a dopamine transporter (DAT) study and a study of the regional cerebral blood flow (rCBF), performed with {sup 123}I. The images are produced using a recently developed Monte Carlo collimator routine added to the program SIMIND which can include interactions in the collimator. The results show that the method included in the iterative reconstruction is preferable to the other methods and that the new TDCS version gives better results compared with the other 2D methods.
Larsson, Anne; Ljungberg, Michael; Mo, Susanna Jakobson; Riklund, Katrine; Johansson, Lennart
2006-11-21
Scatter and septal penetration deteriorate contrast and quantitative accuracy in single photon emission computed tomography (SPECT). In this study four different correction techniques for scatter and septal penetration are evaluated for 123I brain SPECT. One of the methods is a form of model-based compensation which uses the effective source scatter estimation (ESSE) for modelling scatter, and collimator-detector response (CDR) including both geometric and penetration components. The other methods, which operate on the 2D projection images, are convolution scatter subtraction (CSS) and two versions of transmission dependent convolution subtraction (TDCS), one of them proposed by us. This method uses CSS for correction for septal penetration, with a separate kernel, and TDCS for scatter correction. The corrections are evaluated for a dopamine transporter (DAT) study and a study of the regional cerebral blood flow (rCBF), performed with 123I. The images are produced using a recently developed Monte Carlo collimator routine added to the program SIMIND which can include interactions in the collimator. The results show that the method included in the iterative reconstruction is preferable to the other methods and that the new TDCS version gives better results compared with the other 2D methods.
Larsson, Anne; Ljungberg, Michael; Jakobson Mo, Susanna; Riklund, Katrine; Johansson, Lennart
2006-11-01
Scatter and septal penetration deteriorate contrast and quantitative accuracy in single photon emission computed tomography (SPECT). In this study four different correction techniques for scatter and septal penetration are evaluated for 123I brain SPECT. One of the methods is a form of model-based compensation which uses the effective source scatter estimation (ESSE) for modelling scatter, and collimator-detector response (CDR) including both geometric and penetration components. The other methods, which operate on the 2D projection images, are convolution scatter subtraction (CSS) and two versions of transmission dependent convolution subtraction (TDCS), one of them proposed by us. This method uses CSS for correction for septal penetration, with a separate kernel, and TDCS for scatter correction. The corrections are evaluated for a dopamine transporter (DAT) study and a study of the regional cerebral blood flow (rCBF), performed with 123I. The images are produced using a recently developed Monte Carlo collimator routine added to the program SIMIND which can include interactions in the collimator. The results show that the method included in the iterative reconstruction is preferable to the other methods and that the new TDCS version gives better results compared with the other 2D methods.
Fundamentals of convolutional coding
Johannesson, Rolf
2015-01-01
Fundamentals of Convolutional Coding, Second Edition, regarded as a bible of convolutional coding brings you a clear and comprehensive discussion of the basic principles of this field * Two new chapters on low-density parity-check (LDPC) convolutional codes and iterative coding * Viterbi, BCJR, BEAST, list, and sequential decoding of convolutional codes * Distance properties of convolutional codes * Includes a downloadable solutions manual
Convolutive ICA for Spatio-Temporal Analysis of EEG
DEFF Research Database (Denmark)
Dyrholm, Mads; Makeig, Scott; Hansen, Lars Kai
2007-01-01
in the convolutive model can be correctly detected using Bayesian model selection. We demonstrate a framework for deconvolving an EEG ICA subspace. Initial results suggest that in some cases convolutive mixing may be a more realistic model for EEG signals than the instantaneous ICA model....
Compressing Convolutional Neural Networks
Chen, Wenlin; Wilson, James T.; Tyree, Stephen; Weinberger, Kilian Q.; Chen, Yixin
2015-01-01
Convolutional neural networks (CNN) are increasingly used in many areas of computer vision. They are particularly attractive because of their ability to "absorb" great quantities of labeled data through millions of parameters. However, as model sizes increase, so do the storage and memory requirements of the classifiers. We present a novel network architecture, Frequency-Sensitive Hashed Nets (FreshNets), which exploits inherent redundancy in both convolutional layers and fully-connected laye...
Convolution copula econometrics
Cherubini, Umberto; Mulinacci, Sabrina
2016-01-01
This book presents a novel approach to time series econometrics, which studies the behavior of nonlinear stochastic processes. This approach allows for an arbitrary dependence structure in the increments and provides a generalization with respect to the standard linear independent increments assumption of classical time series models. The book offers a solution to the problem of a general semiparametric approach, which is given by a concept called C-convolution (convolution of dependent variables), and the corresponding theory of convolution-based copulas. Intended for econometrics and statistics scholars with a special interest in time series analysis and copula functions (or other nonparametric approaches), the book is also useful for doctoral students with a basic knowledge of copula functions wanting to learn about the latest research developments in the field.
Efficient convolutional sparse coding
Energy Technology Data Exchange (ETDEWEB)
Wohlberg, Brendt
2017-06-20
Computationally efficient algorithms may be applied for fast dictionary learning solving the convolutional sparse coding problem in the Fourier domain. More specifically, efficient convolutional sparse coding may be derived within an alternating direction method of multipliers (ADMM) framework that utilizes fast Fourier transforms (FFT) to solve the main linear system in the frequency domain. Such algorithms may enable a significant reduction in computational cost over conventional approaches by implementing a linear solver for the most critical and computationally expensive component of the conventional iterative algorithm. The theoretical computational cost of the algorithm may be reduced from O(M.sup.3N) to O(MN log N), where N is the dimensionality of the data and M is the number of elements in the dictionary. This significant improvement in efficiency may greatly increase the range of problems that can practically be addressed via convolutional sparse representations.
DEFF Research Database (Denmark)
Pinkevych, Mykola; Cromer, Deborah; Tolstrup, Martin
2016-01-01
[This corrects the article DOI: 10.1371/journal.ppat.1005000.][This corrects the article DOI: 10.1371/journal.ppat.1005740.][This corrects the article DOI: 10.1371/journal.ppat.1005679.].......[This corrects the article DOI: 10.1371/journal.ppat.1005000.][This corrects the article DOI: 10.1371/journal.ppat.1005740.][This corrects the article DOI: 10.1371/journal.ppat.1005679.]....
Convolution Operators on Groups
Derighetti, Antoine
2011-01-01
This volume is devoted to a systematic study of the Banach algebra of the convolution operators of a locally compact group. Inspired by classical Fourier analysis we consider operators on Lp spaces, arriving at a description of these operators and Lp versions of the theorems of Wiener and Kaplansky-Helson.
Directory of Open Access Journals (Sweden)
Herb Silverman
1996-01-01
Full Text Available The radius of univalence is found for the convolution f∗g of functions f∈S (normalized univalent functions and g∈C (close-to-convex functions. A lower bound for the radius of univalence is also determined when f and g range over all of S. Finally, a characterization of C provides an inclusion relationship.
Invariant Scattering Convolution Networks
Bruna, Joan
2012-01-01
A wavelet scattering network computes a translation invariant image representation, which is stable to deformations and preserves high frequency information for classification. It cascades wavelet transform convolutions with non-linear modulus and averaging operators. The first network layer outputs SIFT-type descriptors whereas the next layers provide complementary invariant information which improves classification. The mathematical analysis of wavelet scattering networks explains important properties of deep convolution networks for classification. A scattering representation of stationary processes incorporates higher order moments and can thus discriminate textures having the same Fourier power spectrum. State of the art classification results are obtained for handwritten digits and texture discrimination, using a Gaussian kernel SVM and a generative PCA classifier.
Model Convolution: A Computational Approach to Digital Image Interpretation
Gardner, Melissa K.; Sprague, Brian L.; Pearson, Chad G.; Cosgrove, Benjamin D.; Bicek, Andrew D.; Bloom, Kerry; Salmon, E. D.
2010-01-01
Digital fluorescence microscopy is commonly used to track individual proteins and their dynamics in living cells. However, extracting molecule-specific information from fluorescence images is often limited by the noise and blur intrinsic to the cell and the imaging system. Here we discuss a method called “model-convolution,” which uses experimentally measured noise and blur to simulate the process of imaging fluorescent proteins whose spatial distribution cannot be resolved. We then compare model-convolution to the more standard approach of experimental deconvolution. In some circumstances, standard experimental deconvolution approaches fail to yield the correct underlying fluorophore distribution. In these situations, model-convolution removes the uncertainty associated with deconvolution and therefore allows direct statistical comparison of experimental and theoretical data. Thus, if there are structural constraints on molecular organization, the model-convolution method better utilizes information gathered via fluorescence microscopy, and naturally integrates experiment and theory. PMID:20461132
Hirschman, Isidore Isaac
2005-01-01
In studies of general operators of the same nature, general convolution transforms are immediately encountered as the objects of inversion. The relation between differential operators and integral transforms is the basic theme of this work, which is geared toward upper-level undergraduates and graduate students. It may be read easily by anyone with a working knowledge of real and complex variable theory. Topics include the finite and non-finite kernels, variation diminishing transforms, asymptotic behavior of kernels, real inversion theory, representation theory, the Weierstrass transform, and
Convolutional Goppa codes defined on fibrations
Curto, J I Iglesias; Martín, F J Plaza; Sotelo, G Serrano
2010-01-01
We define a new class of Convolutional Codes in terms of fibrations of algebraic varieties generalizaing our previous constructions of Convolutional Goppa Codes. Using this general construction we can give several examples of Maximum Distance Separable (MDS) Convolutional Codes.
Convolutional coding techniques for data protection
Massey, J. L.
1975-01-01
Results of research on the use of convolutional codes in data communications are presented. Convolutional coding fundamentals are discussed along with modulation and coding interaction. Concatenated coding systems and data compression with convolutional codes are described.
Consensus Convolutional Sparse Coding
Choudhury, Biswarup
2017-04-11
Convolutional sparse coding (CSC) is a promising direction for unsupervised learning in computer vision. In contrast to recent supervised methods, CSC allows for convolutional image representations to be learned that are equally useful for high-level vision tasks and low-level image reconstruction and can be applied to a wide range of tasks without problem-specific retraining. Due to their extreme memory requirements, however, existing CSC solvers have so far been limited to low-dimensional problems and datasets using a handful of low-resolution example images at a time. In this paper, we propose a new approach to solving CSC as a consensus optimization problem, which lifts these limitations. By learning CSC features from large-scale image datasets for the first time, we achieve significant quality improvements in a number of imaging tasks. Moreover, the proposed method enables new applications in high dimensional feature learning that has been intractable using existing CSC methods. This is demonstrated for a variety of reconstruction problems across diverse problem domains, including 3D multispectral demosaickingand 4D light field view synthesis.
Strongly-MDS convolutional codes
Gluesing-Luerssen, H; Rosenthal, J; Smarandache, R
2006-01-01
Maximum-distance separable (MDS) convolutional codes have the property that their free distance is maximal among all codes of the same rate and the same degree. In this paper, a class of MDS convolutional codes is introduced whose column distances reach the generalized Singleton bound at the earlies
Separating Underdetermined Convolutive Speech Mixtures
DEFF Research Database (Denmark)
Pedersen, Michael Syskind; Wang, DeLiang; Larsen, Jan
2006-01-01
a method for underdetermined blind source separation of convolutive mixtures. The proposed framework is applicable for separation of instantaneous as well as convolutive speech mixtures. It is possible to iteratively extract each speech signal from the mixture by combining blind source separation...
Topological convolution algebras
Alpay, Daniel
2012-01-01
In this paper we introduce a new family of topological convolution algebras of the form $\\bigcup_{p\\in\\mathbb N} L_2(S,\\mu_p)$, where $S$ is a Borel semi-group in a locally compact group $G$, which carries an inequality of the type $\\|f*g\\|_p\\le A_{p,q}\\|f\\|_q\\|g\\|_p$ for $p > q+d$ where $d$ pre-assigned, and $A_{p,q}$ is a constant. We give a sufficient condition on the measures $\\mu_p$ for such an inequality to hold. We study the functional calculus and the spectrum of the elements of these algebras, and present two examples, one in the setting of non commutative stochastic distributions, and the other related to Dirichlet series.
2002-01-01
Tile Calorimeter modules stored at CERN. The larger modules belong to the Barrel, whereas the smaller ones are for the two Extended Barrels. (The article was about the completion of the 64 modules for one of the latter.) The photo on the first page of the Bulletin n°26/2002, from 24 July 2002, illustrating the article «The ATLAS Tile Calorimeter gets into shape» was published with a wrong caption. We would like to apologise for this mistake and so publish it again with the correct caption.
2002-01-01
The photo on the second page of the Bulletin n°48/2002, from 25 November 2002, illustrating the article «Spanish Visit to CERN» was published with a wrong caption. We would like to apologise for this mistake and so publish it again with the correct caption. The Spanish delegation, accompanied by Spanish scientists at CERN, also visited the LHC superconducting magnet test hall (photo). From left to right: Felix Rodriguez Mateos of CERN LHC Division, Josep Piqué i Camps, Spanish Minister of Science and Technology, César Dopazo, Director-General of CIEMAT (Spanish Research Centre for Energy, Environment and Technology), Juan Antonio Rubio, ETT Division Leader at CERN, Manuel Aguilar-Benitez, Spanish Delegate to Council, Manuel Delfino, IT Division Leader at CERN, and Gonzalo León, Secretary-General of Scientific Policy to the Minister.
Directory of Open Access Journals (Sweden)
2012-01-01
Full Text Available Regarding Gorelik, G., & Shackelford, T.K. (2011. Human sexual conflict from molecules to culture. Evolutionary Psychology, 9, 564–587: The authors wish to correct an omission in citation to the existing literature. In the final paragraph on p. 570, we neglected to cite Burch and Gallup (2006 [Burch, R. L., & Gallup, G. G., Jr. (2006. The psychobiology of human semen. In S. M. Platek & T. K. Shackelford (Eds., Female infidelity and paternal uncertainty (pp. 141–172. New York: Cambridge University Press.]. Burch and Gallup (2006 reviewed the relevant literature on FSH and LH discussed in this paragraph, and should have been cited accordingly. In addition, Burch and Gallup (2006 should have been cited as the originators of the hypothesis regarding the role of FSH and LH in the semen of rapists. The authors apologize for this oversight.
Directory of Open Access Journals (Sweden)
2014-01-01
Full Text Available Regarding Tagler, M. J., and Jeffers, H. M. (2013. Sex differences in attitudes toward partner infidelity. Evolutionary Psychology, 11, 821–832: The authors wish to correct values in the originally published manuscript. Specifically, incorrect 95% confidence intervals around the Cohen's d values were reported on page 826 of the manuscript where we reported the within-sex simple effects for the significant Participant Sex × Infidelity Type interaction (first paragraph, and for attitudes toward partner infidelity (second paragraph. Corrected values are presented in bold below. The authors would like to thank Dr. Bernard Beins at Ithaca College for bringing these errors to our attention. Men rated sexual infidelity significantly more distressing (M = 4.69, SD = 0.74 than they rated emotional infidelity (M = 4.32, SD = 0.92, F(1, 322 = 23.96, p < .001, d = 0.44, 95% CI [0.23, 0.65], but there was little difference between women's ratings of sexual (M = 4.80, SD = 0.48 and emotional infidelity (M = 4.76, SD = 0.57, F(1, 322 = 0.48, p = .29, d = 0.08, 95% CI [−0.10, 0.26]. As expected, men rated sexual infidelity (M = 1.44, SD = 0.70 more negatively than they rated emotional infidelity (M = 2.66, SD = 1.37, F(1, 322 = 120.00, p < .001, d = 1.12, 95% CI [0.85, 1.39]. Although women also rated sexual infidelity (M = 1.40, SD = 0.62 more negatively than they rated emotional infidelity (M = 2.09, SD = 1.10, this difference was not as large and thus in the evolutionary theory supportive direction, F(1, 322 = 72.03, p < .001, d = 0.77, 95% CI [0.60, 0.94].
Blind recognition of punctured convolutional codes
Institute of Scientific and Technical Information of China (English)
LU Peizhong; LI Shen; ZOU Yan; LUO Xiangyang
2005-01-01
This paper presents an algorithm for blind recognition of punctured convolutional codes which is an important problem in adaptive modulation and coding. For a given finite sequence of convolutional code, the parity check matrix of the convolutional code is first computed by solving a linear system with adequate error tolerance. Then a minimal basic encoding matrix of the original convolutional code and its puncturing pattern are determined according to the known parity check matrix of the punctured convolutional code.
2015-10-01
In the article by Quintavalle et al (Quintavalle C, Anselmi CV, De Micco F, Roscigno G, Visconti G, Golia B, Focaccio A, Ricciardelli B, Perna E, Papa L, Donnarumma E, Condorelli G, Briguori C. Neutrophil gelatinase–associated lipocalin and contrast-induced acute kidney injury. Circ Cardiovasc Interv. 2015;8:e002673. DOI: 10.1161/CIRCINTERVENTIONS.115.002673.), which published online September 2, 2015, and appears in the September 2015 issue of the journal, a correction was needed. On page 1, the institutional affiliation for Elvira Donnarumma, PhD, “SDN Foundation,” has been changed to read, “IRCCS SDN, Naples, Italy.” The institutional affiliation for Laura Papa, PhD, “Institute for Endocrinology and Experimental Oncology, National Research Council, Naples, Italy,” has been changed to read, “Institute of Genetics and Biomedical Research, Milan Unit, Milan, Italy” and “Humanitas Research Hospital, Rozzano, Italy.” The authors regret this error.
Matrix convolution operators on groups
Chu, Cho-Ho
2008-01-01
In the last decade, convolution operators of matrix functions have received unusual attention due to their diverse applications. This monograph presents some new developments in the spectral theory of these operators. The setting is the Lp spaces of matrix-valued functions on locally compact groups. The focus is on the spectra and eigenspaces of convolution operators on these spaces, defined by matrix-valued measures. Among various spectral results, the L2-spectrum of such an operator is completely determined and as an application, the spectrum of a discrete Laplacian on a homogeneous graph is computed using this result. The contractivity properties of matrix convolution semigroups are studied and applications to harmonic functions on Lie groups and Riemannian symmetric spaces are discussed. An interesting feature is the presence of Jordan algebraic structures in matrix-harmonic functions.
General logarithmic image processing convolution.
Palomares, Jose M; González, Jesús; Ros, Eduardo; Prieto, Alberto
2006-11-01
The logarithmic image processing model (LIP) is a robust mathematical framework, which, among other benefits, behaves invariantly to illumination changes. This paper presents, for the first time, two general formulations of the 2-D convolution of separable kernels under the LIP paradigm. Although both formulations are mathematically equivalent, one of them has been designed avoiding the operations which are computationally expensive in current computers. Therefore, this fast LIP convolution method allows to obtain significant speedups and is more adequate for real-time processing. In order to support these statements, some experimental results are shown in Section V.
A REMARK ON CERTAIN CONVOLUTION OPERATOR
Institute of Scientific and Technical Information of China (English)
刘金林
1993-01-01
A certain operator D(a+p-1) defined by convolutions (or Hadamard products) is introduced. The object of this paper is to give an application of the convolution operator D(a+p-1) to the differential inequalities.
Fast Algorithms for Convolutional Neural Networks
Lavin, Andrew; Gray, Scott
2015-01-01
Deep convolutional neural networks take GPU days of compute time to train on large data sets. Pedestrian detection for self driving cars requires very low latency. Image recognition for mobile phones is constrained by limited processing resources. The success of convolutional neural networks in these situations is limited by how fast we can compute them. Conventional FFT based convolution is fast for large filters, but state of the art convolutional neural networks use small, 3x3 filters. We ...
Engineering Multirate Convolutions for Radar Imaging
Bierens, L.H.J.; Deprettere, E.F.
1996-01-01
We present a schematic design methodology for multirate convolution systems, based on combined algorithmic development and architecture design. It allows us to map the algebraic specification of a long convolution algorithm directly onto efficient fast convolution hardware based on short FFT process
Reed-Solomon convolutional codes
Gluesing-Luerssen, H; Schmale, W
2005-01-01
In this paper we will introduce a specific class of cyclic convolutional codes. The construction is based on Reed-Solomon block codes. The algebraic parameters as well as the distance of these codes are determined. This shows that some of these codes are optimal or near optimal.
Quasi-Convolution Pyramidal Blurring
Kraus, Martin
2008-01-01
Efficient image blurring techniques based on the pyramid algorithm can be implemented on modern graphics hardware; thus, image blurring with arbitrary blur width is possible in real time even for large images. However, pyramidal blurring methods do not achieve the image quality provided by convolution filters; in particular, the shape of the corresponding filter kernel varies locally, which potentially results in objectionable rendering artifacts. In this work, a new analysis filter is design...
Convolutive Blind Source Separation Methods
DEFF Research Database (Denmark)
Pedersen, Michael Syskind; Larsen, Jan; Kjems, Ulrik
2008-01-01
During the past decades, much attention has been given to the separation of mixed sources, in particular for the blind case where both the sources and the mixing process are unknown and only recordings of the mixtures are available. In several situations it is desirable to recover all sources from....... This may help practitioners and researchers new to the area of convolutive source separation obtain a complete overview of the field. Hopefully those with more experience in the field can identify useful tools, or find inspiration for new algorithms....
Incomplete convolutions in production and inventory models
Houtum, van G.J.; Zijm, W.H.M.
1997-01-01
In this paper, we study incomplete convolutions of continuous distribution functions, as they appear in the analysis of (multi-stage) production and inventory systems. Three example systems are discussed where these incomplete convolutions naturally arise. We derive explicit, nonrecursive formulae f
Independent Component Analysis in a convoluted world
DEFF Research Database (Denmark)
Dyrholm, Mads
2006-01-01
instantaneousICA, then select a physiologically interesting subspace, then remove the delayed temporal dependencies among the instantaneous ICA components by using convolutive ICA. By Bayesian model selection, in a real world EEG data set, it is shown that convolutive ICA is a better model for EEG than...
Model structure selection in convolutive mixtures
DEFF Research Database (Denmark)
Dyrholm, Mads; Makeig, Scott; Hansen, Lars Kai
2006-01-01
The CICAAR algorithm (convolutive independent component analysis with an auto-regressive inverse model) allows separation of white (i.i.d) source signals from convolutive mixtures. We introduce a source color model as a simple extension to the CICAAR which allows for a more parsimoneous represent......The CICAAR algorithm (convolutive independent component analysis with an auto-regressive inverse model) allows separation of white (i.i.d) source signals from convolutive mixtures. We introduce a source color model as a simple extension to the CICAAR which allows for a more parsimoneous...... representation in many practical mixtures. The new filter-CICAAR allows Bayesian model selection and can help answer questions like: 'Are we actually dealing with a convolutive mixture?'. We try to answer this question for EEG data....
Model structure selection in convolutive mixtures
DEFF Research Database (Denmark)
Dyrholm, Mads; Makeig, S.; Hansen, Lars Kai
2006-01-01
The CICAAR algorithm (convolutive independent component analysis with an auto-regressive inverse model) allows separation of white (i.i.d) source signals from convolutive mixtures. We introduce a source color model as a simple extension to the CICAAR which allows for a more parsimonious represent......The CICAAR algorithm (convolutive independent component analysis with an auto-regressive inverse model) allows separation of white (i.i.d) source signals from convolutive mixtures. We introduce a source color model as a simple extension to the CICAAR which allows for a more parsimonious...... representation in many practical mixtures. The new filter-CICAAR allows Bayesian model selection and can help answer questions like: ’Are we actually dealing with a convolutive mixture?’. We try to answer this question for EEG data....
Directory of Open Access Journals (Sweden)
Cătălin LUPU
2014-12-01
Full Text Available This article presents the development of optimal filters through covolution methods, necessary for restoring, correcting and improving fingerprints acquired from a sensor, able to provide the most ideal image in the output. After the image was binarized and equalized, Canny filter is applied in order to: eliminate the noise (filtering the image with a Gaussian filter, non-maxima suppression, module gradient adaptive binarization and extension edge points edges by hysteresis. The resulting image after applying Canny filter is not ideal. It is possible that the result will be an image with very fragmented edges and many pores in ridge. For the resulting image, a bank of convolution filters are applied one after another (Kirsch, Laplace, Roberts, Prewitt, Sobel, Frei-Chen, averaging convolution filter, circular convolution filter, lapacian convolution filter, gaussian convolution filter, LoG convolution filter, DoG, inverted filters, Wiener, the filter of ”equalization of the power spectrum” (intermediary filter between the Wiener filter and the inverted filter, the geometrical average filter , etc. with different features.
Two-dimensional cubic convolution.
Reichenbach, Stephen E; Geng, Frank
2003-01-01
The paper develops two-dimensional (2D), nonseparable, piecewise cubic convolution (PCC) for image interpolation. Traditionally, PCC has been implemented based on a one-dimensional (1D) derivation with a separable generalization to two dimensions. However, typical scenes and imaging systems are not separable, so the traditional approach is suboptimal. We develop a closed-form derivation for a two-parameter, 2D PCC kernel with support [-2,2] x [-2,2] that is constrained for continuity, smoothness, symmetry, and flat-field response. Our analyses, using several image models, including Markov random fields, demonstrate that the 2D PCC yields small improvements in interpolation fidelity over the traditional, separable approach. The constraints on the derivation can be relaxed to provide greater flexibility and performance.
Minimal memory requirements for pearl necklace encoders of quantum convolutional codes
Houshmand, Monireh; Wilde, Mark M
2010-01-01
One of the major goals in quantum computer science is to reduce the overhead associated with the implementation of quantum computers, and inevitably, routines for quantum error correction will account for most of this overhead. A particular technique for quantum error correction that may be useful in the outer layers of a concatenated scheme for fault tolerance is quantum convolutional coding. The encoder for a quantum convolutional code has a representation as a convolutional encoder or as a "pearl necklace" encoder. In the pearl necklace representation, it has not been particularly clear in the research literature how much quantum memory such an encoder would require for implementation. Here, we offer an algorithm that answers this question. The algorithm first constructs a weighted, directed acyclic graph where each vertex of the graph corresponds to a gate string in the pearl necklace encoder, and each path through the graph represents a non-commutative path through gates in the encoder. We show that the ...
Implementation of Convolution Encoder and Viterbi Decoder for Constraint Length 7 and Bit Rate 1/2
Directory of Open Access Journals (Sweden)
Mr. Sandesh Y.M
2013-11-01
Full Text Available Convolutional codes are non blocking codes that can be designed to either error detecting or correcting. Convolution coding has been used in communication systems including deep space communication and wireless communication. At the receiver end the original message sequence is obtained from the received data using Viterbi decoder. It implements Viterbi Algorithm which is a maximum likelihood algorithm, based on the minimum cumulative hamming distance it decides the optimal trellis path that is most likely followed at the encoder. In this paper I present the convolution encoder and Viterbi decoder for constraint length 7 and bit rate 1/2.
Convolutional Neural Network for Image Recognition
Seifnashri, Sahand
2015-01-01
The aim of this project is to use machine learning techniques especially Convolutional Neural Networks for image processing. These techniques can be used for Quark-Gluon discrimination using calorimeters data, but unfortunately I didn’t manage to get the calorimeters data and I just used the Jet data fromminiaodsim(ak4 chs). The Jet data was not good enough for Convolutional Neural Network which is designed for ’image’ recognition. This report is made of twomain part, part one is mainly about implementing Convolutional Neural Network on unphysical data such as MNIST digits and CIFAR-10 dataset and part 2 is about the Jet data.
PROPERTIES OF THE CONVOLUTION WITH PRESTARLIKE FUNCTIONS
Institute of Scientific and Technical Information of China (English)
Jacek DZIOK
2013-01-01
In the paper we investigate convolution properties related to the prestarlike functions and various inclusion relationships between defined classes of functions. Interest-ing applications involving the well-known classes of functions defined by linear operators are also considered.
Inf-convolution of G-expectations
Institute of Scientific and Technical Information of China (English)
BUCKDAHN; Rainer
2010-01-01
In this paper we will discuss the optimal risk transfer problems when risk measures are generated by G-expectations,and we present the relationship between inf-convolution of G-expectations and the infconvolution of drivers G.
Convolution kernels for multi-wavelength imaging
National Research Council Canada - National Science Library
Boucaud, Alexandre; Bocchio, Marco; Abergel, Alain; Orieux, François; Dole, Hervé; Hadj-Youcef, Mohamed Amine
2016-01-01
.... Given the knowledge of the PSF in each band, a straightforward way of processing images is to homogenise them all to a target PSF using convolution kernels, so that they appear as if they had been...
Gradient Flow Convolutive Blind Source Separation
DEFF Research Database (Denmark)
Pedersen, Michael Syskind; Nielsen, Chinton Møller
2004-01-01
Experiments have shown that the performance of instantaneous gradient flow beamforming by Cauwenberghs et al. is reduced significantly in reverberant conditions. By expanding the gradient flow principle to convolutive mixtures, separation in a reverberant environment is possible. By use of a circ......Experiments have shown that the performance of instantaneous gradient flow beamforming by Cauwenberghs et al. is reduced significantly in reverberant conditions. By expanding the gradient flow principle to convolutive mixtures, separation in a reverberant environment is possible. By use...
A guide to convolution arithmetic for deep learning
Dumoulin, Vincent; Visin, Francesco
2016-01-01
We introduce a guide to help deep learning practitioners understand and manipulate convolutional neural network architectures. The guide clarifies the relationship between various properties (input shape, kernel shape, zero padding, strides and output shape) of convolutional, pooling and transposed convolutional layers, as well as the relationship between convolutional and transposed convolutional layers. Relationships are derived for various cases, and are illustrated in order to make them i...
Convolutional Encoder and Viterbi Decoder Using SOPC For Variable Constraint Length
DEFF Research Database (Denmark)
Kulkarni, Anuradha; Dnyaneshwar, Mantri; Prasad, Neeli R.;
2013-01-01
Convolution encoder and Viterbi decoder are the basic and important blocks in any Code Division Multiple Accesses (CDMA). They are widely used in communication system due to their error correcting capability But the performance degrades with variable constraint length. In this context to have...... detailed analysis, this paper deals with the implementation of convolution encoder and Viterbi decoder using system on programming chip (SOPC). It uses variable constraint length of 7, 8 and 9 bits for 1/2 and 1/3 code rates. By analyzing the Viterbi algorithm it is seen that our algorithm has a better...
Fluence-convolution broad-beam (FCBB) dose calculation.
Lu, Weiguo; Chen, Mingli
2010-12-07
IMRT optimization requires a fast yet relatively accurate algorithm to calculate the iteration dose with small memory demand. In this paper, we present a dose calculation algorithm that approaches these goals. By decomposing the infinitesimal pencil beam (IPB) kernel into the central axis (CAX) component and lateral spread function (LSF) and taking the beam's eye view (BEV), we established a non-voxel and non-beamlet-based dose calculation formula. Both LSF and CAX are determined by a commissioning procedure using the collapsed-cone convolution/superposition (CCCS) method as the standard dose engine. The proposed dose calculation involves a 2D convolution of a fluence map with LSF followed by ray tracing based on the CAX lookup table with radiological distance and divergence correction, resulting in complexity of O(N(3)) both spatially and temporally. This simple algorithm is orders of magnitude faster than the CCCS method. Without pre-calculation of beamlets, its implementation is also orders of magnitude smaller than the conventional voxel-based beamlet-superposition (VBS) approach. We compared the presented algorithm with the CCCS method using simulated and clinical cases. The agreement was generally within 3% for a homogeneous phantom and 5% for heterogeneous and clinical cases. Combined with the 'adaptive full dose correction', the algorithm is well suitable for calculating the iteration dose during IMRT optimization.
Photon beam convolution using polyenergetic energy deposition kernels
Energy Technology Data Exchange (ETDEWEB)
Hoban, P.W.; Murray, D.C.; Round, W.H. (Waikato Univ., Hamilton (New Zealand). Dept. of Physics)
1994-04-01
In photon beam convolution calculations where polyenergetic energy deposition kernels (EDKs) are used, the primary photon energy spectrum should be correctly accounted for in Monte Carlo generation of EDKs. This requires the probability of interaction, determined by the linear attenuation coefficient, [mu], to be taken into account when primary photon interactions are forced to occur at the EDK origin. The use of primary and scattered EDKs generated with a fixed photon spectrum can give rise to an error in the dose calculation due to neglecting the effects of beam hardening with depth. The proportion of primary photon energy that is transferred to secondary electrons increases with depth of interaction, due to the increase in the ratio [mu][sub ab]/[mu] as the beam hardens. Convolution depth-dose curves calculated using polyenergetic EDKs generated for the primary photon spectra which exist at depths of 0, 20 and 40 cm in water, show a fall-off which is too steep when compared with EGS4 Monte Carlo results. A beam hardening correction factor applied to primary and scattered 0 cm EDKs, based on the ratio of kerma to terma at each depth, gives primary, scattered and total dose in good agreement with Monte Carlo results. (Author).
Kudekar, Shrinivas; Urbanke, Ruediger
2010-01-01
Convolutional LDPC ensembles, introduced by Felstrom and Zigangirov, have excellent thresholds and these thresholds are rapidly increasing as a function of the average degree. Several variations on the basic theme have been proposed to date, all of which share the good performance characteristics of convolutional LDPC ensembles. We describe the fundamental mechanism which explains why "convolutional-like" or "spatially coupled" codes perform so well. In essence, the spatial coupling of the individual code structure has the effect of increasing the belief-propagation (BP) threshold of the new ensemble to its maximum possible value, namely the maximum-a-posteriori (MAP) threshold of the underlying ensemble. For this reason we call this phenomenon "threshold saturation." This gives an entirely new way of approaching capacity. One significant advantage of such a construction is that one can create capacity-approaching ensembles with an error correcting radius which is increasing in the blocklength. Our proof make...
Interpolation by two-dimensional cubic convolution
Shi, Jiazheng; Reichenbach, Stephen E.
2003-08-01
This paper presents results of image interpolation with an improved method for two-dimensional cubic convolution. Convolution with a piecewise cubic is one of the most popular methods for image reconstruction, but the traditional approach uses a separable two-dimensional convolution kernel that is based on a one-dimensional derivation. The traditional, separable method is sub-optimal for the usual case of non-separable images. The improved method in this paper implements the most general non-separable, two-dimensional, piecewise-cubic interpolator with constraints for symmetry, continuity, and smoothness. The improved method of two-dimensional cubic convolution has three parameters that can be tuned to yield maximal fidelity for specific scene ensembles characterized by autocorrelation or power-spectrum. This paper illustrates examples for several scene models (a circular disk of parametric size, a square pulse with parametric rotation, and a Markov random field with parametric spatial detail) and actual images -- presenting the optimal parameters and the resulting fidelity for each model. In these examples, improved two-dimensional cubic convolution is superior to several other popular small-kernel interpolation methods.
ESTIMATING LOSS SEVERITY DISTRIBUTION: CONVOLUTION APPROACH
Directory of Open Access Journals (Sweden)
Ro J. Pak
2014-01-01
Full Text Available Financial loss can be classified into two types such as expected loss and unexpected loss. A current definition seeks to separate two losses from a total loss. In this article, however, we redefine a total loss as the sum of expected and unexpended losses; then the distribution of loss can be considered as the convolution of the distributions of both expected and unexpended losses. We propose to use a convolution of normal and exponential distribution for modelling a loss distribution. Subsequently, we compare its performance with other commonly used loss distributions. The examples of property insurance claim data are analyzed to show the applicability of this normal-exponential convolution model. Overall, we claim that the proposed model provides further useful information with regard to losses compared to existing models. We are able to provide new statistical quantities which are very critical and useful.
Uncertainty estimation by convolution using spatial statistics.
Sanchez-Brea, Luis Miguel; Bernabeu, Eusebio
2006-10-01
Kriging has proven to be a useful tool in image processing since it behaves, under regular sampling, as a convolution. Convolution kernels obtained with kriging allow noise filtering and include the effects of the random fluctuations of the experimental data and the resolution of the measuring devices. The uncertainty at each location of the image can also be determined using kriging. However, this procedure is slow since, currently, only matrix methods are available. In this work, we compare the way kriging performs the uncertainty estimation with the standard statistical technique for magnitudes without spatial dependence. As a result, we propose a much faster technique, based on the variogram, to determine the uncertainty using a convolutional procedure. We check the validity of this approach by applying it to one-dimensional images obtained in diffractometry and two-dimensional images obtained by shadow moire.
Astronomical Image Subtraction by Cross-Convolution
Yuan, Fang; Akerlof, Carl W.
2008-04-01
In recent years, there has been a proliferation of wide-field sky surveys to search for a variety of transient objects. Using relatively short focal lengths, the optics of these systems produce undersampled stellar images often marred by a variety of aberrations. As participants in such activities, we have developed a new algorithm for image subtraction that no longer requires high-quality reference images for comparison. The computational efficiency is comparable with similar procedures currently in use. The general technique is cross-convolution: two convolution kernels are generated to make a test image and a reference image separately transform to match as closely as possible. In analogy to the optimization technique for generating smoothing splines, the inclusion of an rms width penalty term constrains the diffusion of stellar images. In addition, by evaluating the convolution kernels on uniformly spaced subimages across the total area, these routines can accommodate point-spread functions that vary considerably across the focal plane.
Indian Academy of Sciences (India)
Barbara Jasiulis-Gołdyn; Anna Kula
2012-08-01
The paper deals with the notions of weak stability and weak generalized convolution with respect to a generalized convolution, introduced by Kucharczak and Urbanik. We study properties of such objects and give examples of weakly stable measures with respect to the Kendall convolution. Moreover, we show that in the context of non-commutative probability, two operations: the -convolution and the (,1)-convolution satisfy the Urbanik’s conditions for a generalized convolution, interpreted on the set of moment sequences. The weak stability reveals the relation between two operations.
Human Face Recognition Using Convolutional Neural Networks
Directory of Open Access Journals (Sweden)
Răzvan-Daniel Albu
2009-10-01
Full Text Available In this paper, I present a novel hybrid face recognition approach based on a convolutional neural architecture, designed to robustly detect highly variable face patterns. The convolutional network extracts successively larger features in a hierarchical set of layers. With the weights of the trained neural networks there are created kernel windows used for feature extraction in a 3-stage algorithm. I present experimental results illustrating the efficiency of the proposed approach. I use a database of 796 images of 159 individuals from Reims University which contains quite a high degree of variability in expression, pose, and facial details.
Spectral classification using convolutional neural networks
Hála, Pavel
2014-01-01
There is a great need for accurate and autonomous spectral classification methods in astrophysics. This thesis is about training a convolutional neural network (ConvNet) to recognize an object class (quasar, star or galaxy) from one-dimension spectra only. Author developed several scripts and C programs for datasets preparation, preprocessing and postprocessing of the data. EBLearn library (developed by Pierre Sermanet and Yann LeCun) was used to create ConvNets. Application on dataset of more than 60000 spectra yielded success rate of nearly 95%. This thesis conclusively proved great potential of convolutional neural networks and deep learning methods in astrophysics.
SAR ATR Based on Convolutional Neural Network
Directory of Open Access Journals (Sweden)
Tian Zhuangzhuang
2016-06-01
Full Text Available This study presents a new method of Synthetic Aperture Radar (SAR image target recognition based on a convolutional neural network. First, we introduce a class separability measure into the cost function to improve this network’s ability to distinguish between categories. Then, we extract SAR image features using the improved convolutional neural network and classify these features using a support vector machine. Experimental results using moving and stationary target acquisition and recognition SAR datasets prove the validity of this method.
Rojas-Villabona, Alvaro; Kitchen, Neil; Paddick, Ian
2016-11-01
Since its inception, doses applied using Gamma Knife Radiosurgery (GKR) have been calculated using a simple TMR algorithm, which assumes the patient's head is of even density, the same as water. This results in a significant approximation of the dose delivered by the Gamma Knife. We investigated how GKR dose calculations varied when using a new convolution algorithm clinically available for GKR planning that takes into account density variations in the head compared with the established calculation algorithm. Fifty-five patients undergoing GKR and harboring 85 lesions were voluntarily and prospectively enrolled into the study. Their clinical treatment plans were created and delivered using TMR 10, but were then recalculated using the density correction algorithm. Dosimetric differences between the planning algorithms were noted. Beam on time (BOT), which is directly proportional to dose, was the main value investigated. Changes of mean and maximum dose to organs at risk (OAR) were also assessed. Phantom studies were performed to investigate the effect of frame and pin materials on dose calculation using the convolution algorithm. Convolution yielded a mean increase in BOT of 7.4% (3.6%-11.6%). However, approximately 1.5% of this amount was due to the head contour being derived from the CT scans, as opposed to measurements using the Skull Scaling Instrument with TMR. Dose to the cochlea calculated with the convolution algorithm was approximately 7% lower than with the TMR 10 algorithm. No significant difference in relative dose distribution was noted and CT artifact typically caused by the stereotactic frame, glue embolization material or different fixation pin materials did not systematically affect convolution isodoses. Nonetheless, substantial error was introduced to the convolution calculation in one target located exactly in the area of major CT artifact caused by a fixation pin. Inhomogeneity correction using the convolution algorithm results in a considerable
On a Generalized Hankel Type Convolution of Generalized Functions
Indian Academy of Sciences (India)
S P Malgonde; G S Gaikawad
2001-11-01
The classical generalized Hankel type convolution are defined and extended to a class of generalized functions. Algebraic properties of the convolution are explained and the existence and significance of an identity element are discussed.
Semantic segmentation of bioimages using convolutional neural networks
CSIR Research Space (South Africa)
Wiehman, S
2016-07-01
Full Text Available Convolutional neural networks have shown great promise in both general image segmentation problems as well as bioimage segmentation. In this paper, the application of different convolutional network architectures is explored on the C. elegans live...
A note on maximal estimates for stochastic convolutions
Veraar, M.; Weis, L.
2011-01-01
In stochastic partial differential equations it is important to have pathwise regularity properties of stochastic convolutions. In this note we present a new sufficient condition for the pathwise continuity of stochastic convolutions in Banach spaces.
Parallel Multi Channel Convolution using General Matrix Multiplication
VASUDEVAN, ARAVIND; Anderson, Andrew; Gregg, David
2017-01-01
Convolutional neural networks (CNNs) have emerged as one of the most successful machine learning technologies for image and video processing. The most computationally intensive parts of CNNs are the convolutional layers, which convolve multi-channel images with multiple kernels. A common approach to implementing convolutional layers is to expand the image into a column matrix (im2col) and perform Multiple Channel Multiple Kernel (MCMK) convolution using an existing parallel General Matrix Mul...
Convolutional Neural Networks for SAR Image Segmentation
DEFF Research Database (Denmark)
Malmgren-Hansen, David; Nobel-Jørgensen, Morten
2015-01-01
Segmentation of Synthetic Aperture Radar (SAR) images has several uses, but it is a difficult task due to a number of properties related to SAR images. In this article we show how Convolutional Neural Networks (CNNs) can easily be trained for SAR image segmentation with good results. Besides...
Discrete Fresnel Transform and Its Circular Convolution
Ouyang, Xing; Gunning, Fatima; Zhang, Hongyu; Guan, Yong Liang
2015-01-01
Discrete trigonometric transformations, such as the discrete Fourier and cosine/sine transforms, are important in a variety of applications due to their useful properties. For example, one well-known property is the convolution theorem for Fourier transform. In this letter, we derive a discrete Fresnel transform (DFnT) from the infinitely periodic optical gratings, as a linear trigonometric transform. Compared to the previous formulations of DFnT, the DFnT in this letter has no degeneracy, which hinders its mathematic applications, due to destructive interferences. The circular convolution property of the DFnT is studied for the first time. It is proved that the DFnT of a circular convolution of two sequences equals either one circularly convolving with the DFnT of the other. As circular convolution is a fundamental process in discrete systems, the DFnT not only gives the coefficients of the Talbot image, but can also be useful for optical and digital signal processing and numerical evaluation of the Fresnel ...
Properties of derivations on some convolution algebras
DEFF Research Database (Denmark)
Pedersen, Thomas Vils
2014-01-01
For all convolution algebras L1[0; 1); L1 loc and A(!) = T n L1(!n), the derivations are of the form Dμf = Xf μ for suitable measures μ, where (Xf)(t) = tf(t). We describe the (weakly) compact as well as the (weakly) Montel derivations on these algebras in terms of properties of the measure μ...
Epileptiform spike detection via convolutional neural networks
DEFF Research Database (Denmark)
Johansen, Alexander Rosenberg; Jin, Jing; Maszczyk, Tomasz
2016-01-01
The EEG of epileptic patients often contains sharp waveforms called "spikes", occurring between seizures. Detecting such spikes is crucial for diagnosing epilepsy. In this paper, we develop a convolutional neural network (CNN) for detecting spikes in EEG of epileptic patients in an automated fash...
Quasi-cyclic unit memory convolutional codes
DEFF Research Database (Denmark)
Justesen, Jørn; Paaske, Erik; Ballan, Mark
1990-01-01
Unit memory convolutional codes with generator matrices, which are composed of circulant submatrices, are introduced. This structure facilitates the analysis of efficient search for good codes. Equivalences among such codes and some of the basic structural properties are discussed. In particular...
Convolutions with the Continuous Primitive Integral
Directory of Open Access Journals (Sweden)
Erik Talvila
2009-01-01
I⊂ℝ. When g∈L1, the estimate is ‖f∗g‖≤‖f‖‖g‖1. There are results on differentiation and integration of convolutions. A type of Fubini theorem is proved for the continuous primitive integral.
Online multipath convolutional coding for real-time transmission
Thai, Tuan Tran; Lacan, Jerome
2012-01-01
Most of multipath multimedia streaming proposals use Forward Error Correction (FEC) approach to protect from packet losses. However, FEC does not sustain well burst of losses even when packets from a given FEC block are spread over multiple paths. In this article, we propose an online multipath convolutional coding for real-time multipath streaming based on an on-the-fly coding scheme called Tetrys. We evaluate the benefits brought out by this coding scheme inside an existing FEC multipath load splitting proposal known as Encoded Multipath Streaming (EMS). We demonstrate that Tetrys consistently outperforms FEC in both uniform and burst losses with EMS scheme. We also propose a modification of the standard EMS algorithm that greatly improves the performance in terms of packet recovery. Finally, we analyze different spreading policies of the Tetrys redundancy traffic between available paths and observe that the longer propagation delay path should be preferably used to carry repair packets.
Convolution theorems: partitioning the space of integral transforms
Lindsey, Alan R.; Suter, Bruce W.
1999-03-01
Investigating a number of different integral transforms uncovers distinct patterns in the type of translation convolution theorems afforded by each. It is shown that transforms based on separable kernels (aka Fourier, Laplace and their relatives) have a form of the convolution theorem providing for a transform domain product of the convolved functions. However, transforms based on kernels not separable in the function and transform variables mandate a convolution theorem of a different type; namely in the transform domain the convolution becomes another convolution--one function with the transform of the other.
Decoding of Convolutional Codes over the Erasure Channel
Tomás, Virtudes; Smarandache, Roxana
2010-01-01
In this paper we study the decoding capabilities of convolutional codes over the erasure channel. Of special interest will be maximum distance profile (MDP) convolutional codes. These are codes which have a maximum possible column distance increase. We show how this strong minimum distance condition of MDP convolutional codes help us to solve error situations that maximum distance separable (MDS) block codes fail to solve. Towards this goal, we define two subclasses of MDP codes: reverse-MDP convolutional codes and complete-MDP convolutional codes. Reverse-MDP codes have the capability to recover a maximum number of erasures using an algorithm which runs backward in time. Complete-MDP convolutional codes are both MDP and reverse-MDP codes. They are capable to recover the state of the decoder under the mildest condition. We show that complete-MDP convolutional codes perform in certain sense better than MDS block codes of the same rate over the erasure channel.
A convolutional neural network neutrino event classifier
Aurisano, A.; Radovic, A.; Rocco, D.; Himmel, A.; Messier, M. D.; Niner, E.; Pawloski, G.; Psihas, F.; Sousa, A.; Vahle, P.
2016-09-01
Convolutional neural networks (CNNs) have been widely applied in the computer vision community to solve complex problems in image recognition and analysis. We describe an application of the CNN technology to the problem of identifying particle interactions in sampling calorimeters used commonly in high energy physics and high energy neutrino physics in particular. Following a discussion of the core concepts of CNNs and recent innovations in CNN architectures related to the field of deep learning, we outline a specific application to the NOvA neutrino detector. This algorithm, CVN (Convolutional Visual Network) identifies neutrino interactions based on their topology without the need for detailed reconstruction and outperforms algorithms currently in use by the NOvA collaboration.
Transition Mean Values of Shifted Convolution Sums
Petrow, Ian
2011-01-01
Let f be a classical holomorphic cusp form for SL_2(Z) of weight k which is a normalized eigenfunction for the Hecke algebra, and let \\lambda(n) be its eigenvalues. In this paper we study "shifted convolution sums" of the eigenvalues \\lambda(n) after averaging over many shifts h and obtain asymptotic estimates. The result is somewhat surprising: one encounters a transition region depending on the ratio of the square of the length of the average over h to the length of the shifted convolution sum. The phenomenon is similar to that encountered by Conrey, Farmer and Soundararajan in their 2000 paper Transition Mean Values of Real Characters, and the connection of both results to Eisenstein series and multiple Dirichlet series is discussed.
A Convolutional Neural Network Neutrino Event Classifier
Aurisano, A; Rocco, D; Himmel, A; Messier, M D; Niner, E; Pawloski, G; Psihas, F; Sousa, A; Vahle, P
2016-01-01
Convolutional neural networks (CNNs) have been widely applied in the computer vision community to solve complex problems in image recognition and analysis. We describe an application of the CNN technology to the problem of identifying particle interactions in sampling calorimeters used commonly in high energy physics and high energy neutrino physics in particular. Following a discussion of the core concepts of CNNs and recent innovations in CNN architectures related to the field of deep learning, we outline a specific application to the NOvA neutrino detector. This algorithm, CVN (Convolutional Visual Network) identifies neutrino interactions based on their topology without the need for detailed reconstruction and outperforms algorithms currently in use by the NOvA collaboration.
Rational Convolution Roots of Isobaric Polynomials
Conci, Aura; Li, Huilan; MacHenry, Trueman
2014-01-01
In this paper, we exhibit two matrix representations of the rational roots of generalized Fibonacci polynomials (GFPs) under convolution product, in terms of determinants and permanents, respectively. The underlying root formulas for GFPs and for weighted isobaric polynomials (WIPs), which appeared in an earlier paper by MacHenry and Tudose, make use of two types of operators. These operators are derived from the generating functions for Stirling numbers of the first kind and second kind. Hen...
A Generative Model for Deep Convolutional Learning
Pu, Yunchen; Yuan, Xin; Carin, Lawrence
2015-01-01
A generative model is developed for deep (multi-layered) convolutional dictionary learning. A novel probabilistic pooling operation is integrated into the deep model, yielding efficient bottom-up (pretraining) and top-down (refinement) probabilistic learning. Experimental results demonstrate powerful capabilities of the model to learn multi-layer features from images, and excellent classification results are obtained on the MNIST and Caltech 101 datasets.
Convolutional Neural Network Based dem Super Resolution
Chen, Zixuan; Wang, Xuewen; Xu, Zekai; Hou, Wenguang
2016-06-01
DEM super resolution is proposed in our previous publication to improve the resolution for a DEM on basis of some learning examples. Meanwhile, the nonlocal algorithm is introduced to deal with it and lots of experiments show that the strategy is feasible. In our publication, the learning examples are defined as the partial original DEM and their related high measurements due to this way can avoid the incompatibility between the data to be processed and the learning examples. To further extent the applications of this new strategy, the learning examples should be diverse and easy to obtain. Yet, it may cause the problem of incompatibility and unrobustness. To overcome it, we intend to investigate a convolutional neural network based method. The input of the convolutional neural network is a low resolution DEM and the output is expected to be its high resolution one. A three layers model will be adopted. The first layer is used to detect some features from the input, the second integrates the detected features to some compressed ones and the final step transforms the compressed features as a new DEM. According to this designed structure, some learning DEMs will be taken to train it. Specifically, the designed network will be optimized by minimizing the error of the output and its expected high resolution DEM. In practical applications, a testing DEM will be input to the convolutional neural network and a super resolution will be obtained. Many experiments show that the CNN based method can obtain better reconstructions than many classic interpolation methods.
Binary Error Correcting Network Codes
Wang, Qiwen; Li, Shuo-Yen Robert
2011-01-01
We consider network coding for networks experiencing worst-case bit-flip errors, and argue that this is a reasonable model for highly dynamic wireless network transmissions. We demonstrate that in this setup prior network error-correcting schemes can be arbitrarily far from achieving the optimal network throughput. We propose a new metric for errors under this model. Using this metric, we prove a new Hamming-type upper bound on the network capacity. We also show a commensurate lower bound based on GV-type codes that can be used for error-correction. The codes used to attain the lower bound are non-coherent (do not require prior knowledge of network topology). The end-to-end nature of our design enables our codes to be overlaid on classical distributed random linear network codes. Further, we free internal nodes from having to implement potentially computationally intensive link-by-link error-correction.
The convolution theorem for two-dimensional continuous wavelet transform
Institute of Scientific and Technical Information of China (English)
ZHANG CHI
2013-01-01
In this paper , application of two -dimensional continuous wavelet transform to image processes is studied. We first show that the convolution and correlation of two continuous wavelets satisfy the required admissibility and regularity conditions ,and then we derive the convolution and correlation theorem for two-dimensional continuous wavelet transform. Finally, we present numerical example showing the usefulness of applying the convolution theorem for two -dimensional continuous wavelet transform to perform image restoration in the presence of additive noise.
An Algorithm for the Convolution of Legendre Series
Hale, Nicholas
2014-01-01
An O(N2) algorithm for the convolution of compactly supported Legendre series is described. The algorithm is derived from the convolution theorem for Legendre polynomials and the recurrence relation satisfied by spherical Bessel functions. Combining with previous work yields an O(N 2) algorithm for the convolution of Chebyshev series. Numerical results are presented to demonstrate the improved efficiency over the existing algorithm. © 2014 Society for Industrial and Applied Mathematics.
BERNOULLI CONVOLUTIONS ASSOCIATED WITH CERTAIN NON-PISOT NUMBERS
Institute of Scientific and Technical Information of China (English)
Feng Dejun; Wang Yang
2003-01-01
The Bernoulli convolution vλ measure is shown to be absolutely continuous with L2 density for almost all 1/2<λ<1,and singular if λ-1 is a Pisot number.It is an open question whether the Pisot type Bernoulli convolutions are the only singular ones.In this paper,we construct a family of non-Pisot type Bernoulli convolutions vλ such that their density functions,if they excist,are not L2.We also construct other Bernolulli convolutions whose density functions,if they exist,behave rather badly.
Convolutions Induced Discrete Probability Distributions and a New Fibonacci Constant
Rajan, Arulalan; Rao, Vittal; Rao, Ashok
2010-01-01
This paper proposes another constant that can be associated with Fibonacci sequence. In this work, we look at the probability distributions generated by the linear convolution of Fibonacci sequence with itself, and the linear convolution of symmetrized Fibonacci sequence with itself. We observe that for a distribution generated by the linear convolution of the standard Fibonacci sequence with itself, the variance converges to 8.4721359... . Also, for a distribution generated by the linear convolution of symmetrized Fibonacci sequences, the variance converges in an average sense to 17.1942 ..., which is approximately twice that we get with common Fibonacci sequence.
Applications of convolution voltammetry in electroanalytical chemistry.
Bentley, Cameron L; Bond, Alan M; Hollenkamp, Anthony F; Mahon, Peter J; Zhang, Jie
2014-02-18
The robustness of convolution voltammetry for determining accurate values of the diffusivity (D), bulk concentration (C(b)), and stoichiometric number of electrons (n) has been demonstrated by applying the technique to a series of electrode reactions in molecular solvents and room temperature ionic liquids (RTILs). In acetonitrile, the relatively minor contribution of nonfaradaic current facilitates analysis with macrodisk electrodes, thus moderate scan rates can be used without the need to perform background subtraction to quantify the diffusivity of iodide [D = 1.75 (±0.02) × 10(-5) cm(2) s(-1)] in this solvent. In the RTIL 1-ethyl-3-methylimidazolium bis(trifluoromethanesulfonyl)imide, background subtraction is necessary at a macrodisk electrode but can be avoided at a microdisk electrode, thereby simplifying the analytical procedure and allowing the diffusivity of iodide [D = 2.70 (±0.03) × 10(-7) cm(2) s(-1)] to be quantified. Use of a convolutive procedure which simultaneously allows D and nC(b) values to be determined is also demonstrated. Three conditions under which a technique of this kind may be applied are explored and are related to electroactive species which display slow dissolution kinetics, undergo a single multielectron transfer step, or contain multiple noninteracting redox centers using ferrocene in an RTIL, 1,4-dinitro-2,3,5,6-tetramethylbenzene, and an alkynylruthenium trimer, respectively, as examples. The results highlight the advantages of convolution voltammetry over steady-state techniques such as rotating disk electrode voltammetry and microdisk electrode voltammetry, as it is not restricted by the mode of diffusion (planar or radial), hence removing limitations on solvent viscosity, electrode geometry, and voltammetric scan rate.
Convolution neural networks for ship type recognition
Rainey, Katie; Reeder, John D.; Corelli, Alexander G.
2016-05-01
Algorithms to automatically recognize ship type from satellite imagery are desired for numerous maritime applications. This task is difficult, and example imagery accurately labeled with ship type is hard to obtain. Convolutional neural networks (CNNs) have shown promise in image recognition settings, but many of these applications rely on the availability of thousands of example images for training. This work attempts to under- stand for which types of ship recognition tasks CNNs might be well suited. We report the results of baseline experiments applying a CNN to several ship type classification tasks, and discuss many of the considerations that must be made in approaching this problem.
Fourier transforms and convolutions for the experimentalist
Jennison, RC
1961-01-01
Fourier Transforms and Convolutions for the Experimentalist provides the experimentalist with a guide to the principles and practical uses of the Fourier transformation. It aims to bridge the gap between the more abstract account of a purely mathematical approach and the rule of thumb calculation and intuition of the practical worker. The monograph springs from a lecture course which the author has given in recent years and for which he has drawn upon a number of sources, including a set of notes compiled by the late Dr. I. C. Browne from a series of lectures given by Mr. J . A. Ratcliffe of t
Zebrafish tracking using convolutional neural networks
XU, Zhiping; Cheng, Xi En
2017-01-01
Keeping identity for a long term after occlusion is still an open problem in the video tracking of zebrafish-like model animals, and accurate animal trajectories are the foundation of behaviour analysis. We utilize the highly accurate object recognition capability of a convolutional neural network (CNN) to distinguish fish of the same congener, even though these animals are indistinguishable to the human eye. We used data augmentation and an iterative CNN training method to optimize the accuracy for our classification task, achieving surprisingly accurate trajectories of zebrafish of different size and age zebrafish groups over different time spans. This work will make further behaviour analysis more reliable. PMID:28211462
Zebrafish tracking using convolutional neural networks
Xu, Zhiping; Cheng, Xi En
2017-02-01
Keeping identity for a long term after occlusion is still an open problem in the video tracking of zebrafish-like model animals, and accurate animal trajectories are the foundation of behaviour analysis. We utilize the highly accurate object recognition capability of a convolutional neural network (CNN) to distinguish fish of the same congener, even though these animals are indistinguishable to the human eye. We used data augmentation and an iterative CNN training method to optimize the accuracy for our classification task, achieving surprisingly accurate trajectories of zebrafish of different size and age zebrafish groups over different time spans. This work will make further behaviour analysis more reliable.
Chen, Liang-Chieh; Papandreou, George; Kokkinos, Iasonas; Murphy, Kevin; Yuille, Alan L
2017-04-27
In this work we address the task of semantic image segmentation with Deep Learning and make three main contributions that are experimentally shown to have substantial practical merit. First, we highlight convolution with upsampled filters, or 'atrous convolution', as a powerful tool in dense prediction tasks. Atrous convolution allows us to explicitly control the resolution at which feature responses are computed within Deep Convolutional Neural Networks. It also allows us to effectively enlarge the field of view of filters to incorporate larger context without increasing the number of parameters or the amount of computation. Second, we propose atrous spatial pyramid pooling (ASPP) to robustly segment objects at multiple scales. ASPP probes an incoming convolutional feature layer with filters at multiple sampling rates and effective fields-of-views, thus capturing objects as well as image context at multiple scales. Third, we improve the localization of object boundaries by combining methods from DCNNs and probabilistic graphical models. The commonly deployed combination of max-pooling and downsampling in DCNNs achieves invariance but has a toll on localization accuracy. We overcome this by combining the responses at the final DCNN layer with a fully connected Conditional Random Field (CRF), which is shown both qualitatively and quantitatively to improve localization performance. Our proposed "DeepLab" system sets the new state-of-art at the PASCAL VOC-2012 semantic image segmentation task, reaching 79.7% mIOU in the test set, and advances the results on three other datasets: PASCAL-Context, PASCAL-Person-Part, and Cityscapes. All of our code is made publicly available online.
One dimensional Convolutional Goppa Codes over the projective line
Pérez, J A Domínguez; Sotelo, G Serrano
2011-01-01
We give a general method to construct MDS one-dimensional convolutional codes. Our method generalizes previous constructions of H. Gluesing-Luerssen and B. Langfeld. Moreover we give a classification of one-dimensional Convolutional Goppa Codes and propose a characterization of MDS codes of this type.
Explicit solutions of fractional diffusion equations via Generalized Gamma Convolution
D'Ovidio, Mirko
2010-01-01
In this paper we deal with Mellin convolution of generalized Gamma densities which brings to integrals of modified Bessel functions of the second kind. Such convolutions allow us to write explicitly the solutions of the time-fractional diffusion equations involving the adjoint operators of a square Bessel process and a Bessel process.
Convolution of Lorentz Invariant Ultradistributions and Field Theory
Bollini, C G
2003-01-01
In this work, a general definition of convolution between two arbitrary four dimensional Lorentz invariant (fdLi) Tempered Ultradistributions is given, in both: Minkowskian and Euclidean Space (Spherically symmetric tempered ultradistributions). The product of two arbitrary fdLi distributions of exponential type is defined via the convolution of its corresponding Fourier Transforms. Several examples of convolution of two fdLi Tempered Ultradistributions are given. In particular we calculate exactly the convolution of two Feynman's massless propagators. An expression for the Fourier Transform of a Lorentz invariant Tempered Ultradistribution in terms of modified Bessel distributions is obtained in this work (Generalization of Bochner's formula to Minkowskian space). At the same time, and in a previous step used for the deduction of the convolution formula, we obtain the generalization to the Minkowskian space, of the dimensional regularization of the perturbation theory of Green Functions in the Euclidean conf...
Vehicle Detection Based on Visual Saliency and Deep Sparse Convolution Hierarchical Model
Institute of Scientific and Technical Information of China (English)
CAI Yingfeng; WANG Hai; CHEN Xiaobo; GAO Li; CHEN Long
2016-01-01
Traditional vehicle detection algorithms use traverse search based vehicle candidate generation and hand crafted based classifier training for vehicle candidate verification. These types of methods generally have high processing times and low vehicle detection performance. To address this issue, a visual saliency and deep sparse convolution hierarchical model based vehicle detection algorithm is proposed. A visual saliency calculation is firstly used to generate a small vehicle candidate area. The vehicle candidate sub images are then loaded into a sparse deep convolution hierarchical model with an SVM-based classifier to perform the final detection. The experimental results demonstrate that the proposed method is with 94.81% correct rate and 0.78% false detection rate on the existing datasets and the real road pictures captured by our group, which outperforms the existing state-of-the-art algorithms. More importantly, high discriminative multi-scale features are generated by deep sparse convolution network which has broad application prospects in target recognition in the field of intelligent vehicle.
Vehicle detection based on visual saliency and deep sparse convolution hierarchical model
Cai, Yingfeng; Wang, Hai; Chen, Xiaobo; Gao, Li; Chen, Long
2016-07-01
Traditional vehicle detection algorithms use traverse search based vehicle candidate generation and hand crafted based classifier training for vehicle candidate verification. These types of methods generally have high processing times and low vehicle detection performance. To address this issue, a visual saliency and deep sparse convolution hierarchical model based vehicle detection algorithm is proposed. A visual saliency calculation is firstly used to generate a small vehicle candidate area. The vehicle candidate sub images are then loaded into a sparse deep convolution hierarchical model with an SVM-based classifier to perform the final detection. The experimental results demonstrate that the proposed method is with 94.81% correct rate and 0.78% false detection rate on the existing datasets and the real road pictures captured by our group, which outperforms the existing state-of-the-art algorithms. More importantly, high discriminative multi-scale features are generated by deep sparse convolution network which has broad application prospects in target recognition in the field of intelligent vehicle.
Very Deep Convolutional Neural Networks for Morphologic Classification of Erythrocytes.
Durant, Thomas J S; Olson, Eben M; Schulz, Wade L; Torres, Richard
2017-09-06
Morphologic profiling of the erythrocyte population is a widely used and clinically valuable diagnostic modality, but one that relies on a slow manual process associated with significant labor cost and limited reproducibility. Automated profiling of erythrocytes from digital images by capable machine learning approaches would augment the throughput and value of morphologic analysis. To this end, we sought to evaluate the performance of leading implementation strategies for convolutional neural networks (CNNs) when applied to classification of erythrocytes based on morphology. Erythrocytes were manually classified into 1 of 10 classes using a custom-developed Web application. Using recent literature to guide architectural considerations for neural network design, we implemented a "very deep" CNN, consisting of >150 layers, with dense shortcut connections. The final database comprised 3737 labeled cells. Ensemble model predictions on unseen data demonstrated a harmonic mean of recall and precision metrics of 92.70% and 89.39%, respectively. Of the 748 cells in the test set, 23 misclassification errors were made, with a correct classification frequency of 90.60%, represented as a harmonic mean across the 10 morphologic classes. These findings indicate that erythrocyte morphology profiles could be measured with a high degree of accuracy with "very deep" CNNs. Further, these data support future efforts to expand classes and optimize practical performance in a clinical environment as a prelude to full implementation as a clinical tool. © 2017 American Association for Clinical Chemistry.
Compressed imaging by sparse random convolution.
Marcos, Diego; Lasser, Theo; López, Antonio; Bourquard, Aurélien
2016-01-25
The theory of compressed sensing (CS) shows that signals can be acquired at sub-Nyquist rates if they are sufficiently sparse or compressible. Since many images bear this property, several acquisition models have been proposed for optical CS. An interesting approach is random convolution (RC). In contrast with single-pixel CS approaches, RC allows for the parallel capture of visual information on a sensor array as in conventional imaging approaches. Unfortunately, the RC strategy is difficult to implement as is in practical settings due to important contrast-to-noise-ratio (CNR) limitations. In this paper, we introduce a modified RC model circumventing such difficulties by considering measurement matrices involving sparse non-negative entries. We then implement this model based on a slightly modified microscopy setup using incoherent light. Our experiments demonstrate the suitability of this approach for dealing with distinct CS scenarii, including 1-bit CS.
Relationships among transforms, convolutions, and first variations
Directory of Open Access Journals (Sweden)
Jeong Gyoo Kim
1999-01-01
Full Text Available In this paper, we establish several interesting relationships involving the Fourier-Feynman transform, the convolution product, and the first variation for functionals F on Wiener space of the form F(x=f(〈α1,x〉,…,〈αn,x〉, (* where 〈αj,x〉 denotes the Paley-Wiener-Zygmund stochastic integral ∫0Tαj(tdx(t.
Robust Convolutional Neural Networks for Image Recognition
Directory of Open Access Journals (Sweden)
Hayder M. Albeahdili
2015-11-01
Full Text Available Recently image recognition becomes vital task using several methods. One of the most interesting used methods is using Convolutional Neural Network (CNN. It is widely used for this purpose. However, since there are some tasks that have small features that are considered an essential part of a task, then classification using CNN is not efficient because most of those features diminish before reaching the final stage of classification. In this work, analyzing and exploring essential parameters that can influence model performance. Furthermore different elegant prior contemporary models are recruited to introduce new leveraging model. Finally, a new CNN architecture is proposed which achieves state-of-the-art classification results on the different challenge benchmarks. The experimented are conducted on MNIST, CIFAR-10, and CIFAR-100 datasets. Experimental results showed that the results outperform and achieve superior results comparing to the most contemporary approaches.
An exactly solvable self-convolutive recurrence
Martin, Richard J
2011-01-01
We consider a self-convolutive recurrence whose solution is the sequence of coefficients in the asymptotic expansion of the logarithmic derivative of the confluent hypergeometic function $U(a,b,z)$. By application of the Hilbert transform we convert this expression into an explicit, non-recursive solution in which the $n$th coefficient is expressed as the $(n-1)$th moment of a measure, and also as the trace of the $(n-1)$th iterate of a linear operator. Applications of these sequences, and hence of the explicit solution provided, are found in quantum field theory as the number of Feynman diagrams of a certain type and order, in Brownian motion theory, and in combinatorics.
Robust smile detection using convolutional neural networks
Bianco, Simone; Celona, Luigi; Schettini, Raimondo
2016-11-01
We present a fully automated approach for smile detection. Faces are detected using a multiview face detector and aligned and scaled using automatically detected eye locations. Then, we use a convolutional neural network (CNN) to determine whether it is a smiling face or not. To this end, we investigate different shallow CNN architectures that can be trained even when the amount of learning data is limited. We evaluate our complete processing pipeline on the largest publicly available image database for smile detection in an uncontrolled scenario. We investigate the robustness of the method to different kinds of geometric transformations (rotation, translation, and scaling) due to imprecise face localization, and to several kinds of distortions (compression, noise, and blur). To the best of our knowledge, this is the first time that this type of investigation has been performed for smile detection. Experimental results show that our proposal outperforms state-of-the-art methods on both high- and low-quality images.
Image quality of mixed convolution kernel in thoracic computed tomography.
Neubauer, Jakob; Spira, Eva Maria; Strube, Juliane; Langer, Mathias; Voss, Christian; Kotter, Elmar
2016-11-01
The mixed convolution kernel alters his properties geographically according to the depicted organ structure, especially for the lung. Therefore, we compared the image quality of the mixed convolution kernel to standard soft and hard kernel reconstructions for different organ structures in thoracic computed tomography (CT) images.Our Ethics Committee approved this prospective study. In total, 31 patients who underwent contrast-enhanced thoracic CT studies were included after informed consent. Axial reconstructions were performed with hard, soft, and mixed convolution kernel. Three independent and blinded observers rated the image quality according to the European Guidelines for Quality Criteria of Thoracic CT for 13 organ structures. The observers rated the depiction of the structures in all reconstructions on a 5-point Likert scale. Statistical analysis was performed with the Friedman Test and post hoc analysis with the Wilcoxon rank-sum test.Compared to the soft convolution kernel, the mixed convolution kernel was rated with a higher image quality for lung parenchyma, segmental bronchi, and the border between the pleura and the thoracic wall (P kernel, the mixed convolution kernel was rated with a higher image quality for aorta, anterior mediastinal structures, paratracheal soft tissue, hilar lymph nodes, esophagus, pleuromediastinal border, large and medium sized pulmonary vessels and abdomen (P kernel cannot fully substitute the standard CT reconstructions. Hard and soft convolution kernel reconstructions still seem to be mandatory for thoracic CT.
Terminated LDPC Convolutional Codes over GF(2^p)
Uchikawa, Hironori; Sakaniwa, Kohichi
2010-01-01
In this paper, we present a construction method of terminated non-binary low-density parity-check (LDPC) convolutional codes. Our construction method is an expansion of Felstrom and Zigangirov construction for non-binary LDPC convolutional codes. The rate-compatibility of the non-binary LDPC convolutional codes is also discussed. The proposed rate-compatible code is designed from one single mother (2,4)-regular non-binary LDPC convolutional code of rate 1/2. Higher-rate codes are produced by puncturing the mother code and lower-rate codes are produced by multiplicatively repeating the mother code. For moderate values of the syndrome former memory, simulation results show that mother non-binary LDPC convolutional code outperform binary LDPC convolutional codes with comparable constraint bit length. And the derived low-rate and high-rate non-binary LDPC convolutional codes exhibit good decoding performance without loss of large gap to the Shannon limits.
Yao, Kun
2015-01-01
We demonstrate a convolutional neural network trained to reproduce the Kohn-Sham kinetic energy of hydrocarbons from electron density. The output of the network is used as a non-local correction to the conventional local and semi-local kinetic functionals. We show that this approximation qualitatively reproduces Kohn-Sham potential energy surfaces when used with conventional exchange correlation functionals. Numerical noise inherited from the non-linearity of the neural network is identified as the major challenge for the model. Finally we examine the features in the density learned by the neural network to anticipate the prospects of generalizing these models.
Larsson, Anne; Johansson, Lennart
2003-11-21
In single photon emission computed tomography (SPECT), transmission-dependent convolution subtraction has been shown to be useful when correcting for scattered events. The method is based on convolution subtraction, but includes a matrix of scatter fractions instead of a global scatter fraction. The method can be extended to iteratively improve the scatter estimate, but in this note we show that this requires a modification of the theory to use scatter-to-total scatter fractions for the first iteration only and scatter-to-primary fractions thereafter. To demonstrate this, scatter correction is performed on a Monte Carlo simulated image of a point source of activity in water. The modification of the theory is compared to corrections where the scatter fractions are based on the scatter-to-total ratio, using one and ten iterations. The resulting ratios of subtracted to original counts are compared to the true scatter-to-total ratio of the simulation and the most accurate result is found for our modification of the theory.
The Law of Large Numbers for the Free Multiplicative Convolution
DEFF Research Database (Denmark)
Haagerup, Uffe; Möller, Sören
2013-01-01
In classical probability the law of large numbers for the multiplicative convolution follows directly from the law for the additive convolution. In free probability this is not the case. The free additive law was proved by D. Voiculescu in 1986 for probability measures with bounded support...... for the case of bounded support. In contrast to the classical multiplicative convolution case, the limit measure for the free multiplicative law of large numbers is not a Dirac measure, unless the original measure is a Dirac measure. We also show that the mean value of lnx is additive with respect to the free...
Interpolating and filtering decoding algorithm for convolution codes
Directory of Open Access Journals (Sweden)
O. O. Shpylka
2010-01-01
Full Text Available There has been synthesized interpolating and filtering decoding algorithm for convolution codes on maximum of a posteriori probability criterion, in which combined filtering coder state and interpolation of information signs on sliding interval are processed
FPGA-based digital convolution for wireless applications
Guan, Lei
2017-01-01
This book presents essential perspectives on digital convolutions in wireless communications systems and illustrates their corresponding efficient real-time field-programmable gate array (FPGA) implementations. Covering these digital convolutions from basic concept to vivid simulation/illustration, the book is also supplemented with MS PowerPoint presentations to aid in comprehension. FPGAs or generic all programmable devices will soon become widespread, serving as the “brains” of all types of real-time smart signal processing systems, like smart networks, smart homes and smart cities. The book examines digital convolution by bringing together the following main elements: the fundamental theory behind the mathematical formulae together with corresponding physical phenomena; virtualized algorithm simulation together with benchmark real-time FPGA implementations; and detailed, state-of-the-art case studies on wireless applications, including popular linear convolution in digital front ends (DFEs); nonlinear...
Multipath Convolutional-Recursive Neural Networks for Object Recognition
2014-01-01
Part 8: Pattern Recognition; International audience; Extracting good representations from images is essential for many computer vision tasks. While progress in deep learning shows the importance of learning hierarchical features, it is also important to learn features through multiple paths. This paper presents Multipath Convolutional-Recursive Neural Networks(M-CRNNs), a novel scheme which aims to learn image features from multiple paths using models based on combination of convolutional and...
Approximation of integral operators using product-convolution expansions
Escande, Paul; Weiss, Pierre
2016-01-01
We consider a class of linear integral operators with impulse responses varying regularly in time or space. These operators appear in a large number of applications ranging from signal/image processing to biology. Evaluating their action on functions is a computationally intensive problem necessary for many practical problems. We analyze a technique called product-convolution expansion: the operator is locally approximated by a convolution, allowing to design fast numerical algorithms ba...
Approximation of integral operators using convolution-product expansions
Escande, Paul; Weiss, Pierre
2016-01-01
We consider a class of linear integral operators with impulse responses varying regularly in time or space. These operators appear in a large number of applications ranging from signal/image processing to biology. Evaluating their action on functions is a computation-ally intensive problem necessary for many practical problems. We analyze a technique called convolution-product expansion: the operator is locally approximated by a convolution, allowing to design fast numerical algorithms based ...
Two-dimensional Block of Spatial Convolution Algorithm and Simulation
Mussa Mohamed Ahmed
2012-01-01
This paper proposes an algorithm based on sub image-segmentation strategy. The proposed scheme divides a grayscale image into overlapped 6×6 blocks each of which is segmented into four small 3x3 non-overlapped sub-images. A new spatial approach for efficiently computing 2-dimensional linear convolution or cross-correlation between suitable flipped and fixed filter coefficients (sub image for cross-correlation) and corresponding input sub image is presented. Computation of convolution is itera...
Traffic sign recognition with deep convolutional neural networks
KARAMATIĆ, BORIS
2016-01-01
The problem of detection and recognition of traffic signs is becoming an important problem when it comes to the development of self driving cars and advanced driver assistance systems. In this thesis we will develop a system for detection and recognition of traffic signs. For the problem of detection we will use aggregate channel features and for the problem of recognition we will use a deep convolutional neural network. We will describe how convolutional neural networks work, how they are co...
Multiscale Convolutional Neural Networks for Hand Detection
Directory of Open Access Journals (Sweden)
Shiyang Yan
2017-01-01
Full Text Available Unconstrained hand detection in still images plays an important role in many hand-related vision problems, for example, hand tracking, gesture analysis, human action recognition and human-machine interaction, and sign language recognition. Although hand detection has been extensively studied for decades, it is still a challenging task with many problems to be tackled. The contributing factors for this complexity include heavy occlusion, low resolution, varying illumination conditions, different hand gestures, and the complex interactions between hands and objects or other hands. In this paper, we propose a multiscale deep learning model for unconstrained hand detection in still images. Deep learning models, and deep convolutional neural networks (CNNs in particular, have achieved state-of-the-art performances in many vision benchmarks. Developed from the region-based CNN (R-CNN model, we propose a hand detection scheme based on candidate regions generated by a generic region proposal algorithm, followed by multiscale information fusion from the popular VGG16 model. Two benchmark datasets were applied to validate the proposed method, namely, the Oxford Hand Detection Dataset and the VIVA Hand Detection Challenge. We achieved state-of-the-art results on the Oxford Hand Detection Dataset and had satisfactory performance in the VIVA Hand Detection Challenge.
Metaheuristic Algorithms for Convolution Neural Network
Fanany, Mohamad Ivan; Arymurthy, Aniati Murni
2016-01-01
A typical modern optimization technique is usually either heuristic or metaheuristic. This technique has managed to solve some optimization problems in the research area of science, engineering, and industry. However, implementation strategy of metaheuristic for accuracy improvement on convolution neural networks (CNN), a famous deep learning method, is still rarely investigated. Deep learning relates to a type of machine learning technique, where its aim is to move closer to the goal of artificial intelligence of creating a machine that could successfully perform any intellectual tasks that can be carried out by a human. In this paper, we propose the implementation strategy of three popular metaheuristic approaches, that is, simulated annealing, differential evolution, and harmony search, to optimize CNN. The performances of these metaheuristic methods in optimizing CNN on classifying MNIST and CIFAR dataset were evaluated and compared. Furthermore, the proposed methods are also compared with the original CNN. Although the proposed methods show an increase in the computation time, their accuracy has also been improved (up to 7.14 percent). PMID:27375738
Convolution kernels for multi-wavelength imaging
Boucaud, Alexandre; Abergel, Alain; Orieux, François; Dole, Hervé; Hadj-Youcef, Mohamed Amine
2016-01-01
Astrophysical images issued from different instruments and/or spectral bands often require to be processed together, either for fitting or comparison purposes. However each image is affected by an instrumental response, also known as PSF, that depends on the characteristics of the instrument as well as the wavelength and the observing strategy. Given the knowledge of the PSF in each band, a straightforward way of processing images is to homogenise them all to a target PSF using convolution kernels, so that they appear as if they had been acquired by the same instrument. We propose an algorithm that generates such PSF-matching kernels, based on Wiener filtering with a tunable regularisation parameter. This method ensures all anisotropic features in the PSFs to be taken into account. We compare our method to existing procedures using measured Herschel/PACS and SPIRE PSFs and simulated JWST/MIRI PSFs. Significant gains up to two orders of magnitude are obtained with respect to the use of kernels computed assumin...
Event Discrimination using Convolutional Neural Networks
Menon, Hareesh; Hughes, Richard; Daling, Alec; Winer, Brian
2017-01-01
Convolutional Neural Networks (CNNs) are computational models that have been shown to be effective at classifying different types of images. We present a method to use CNNs to distinguish events involving the production of a top quark pair and a Higgs boson from events involving the production of a top quark pair and several quark and gluon jets. To do this, we generate and simulate data using MADGRAPH and DELPHES for a general purpose LHC detector at 13 TeV. We produce images using a particle flow algorithm by binning the particles geometrically based on their position in the detector and weighting the bins by the energy of each particle within each bin, and by defining channels based on particle types (charged track, neutral hadronic, neutral EM, lepton, heavy flavor). Our classification results are competitive with standard machine learning techniques. We have also looked into the classification of the substructure of the events, in a process known as scene labeling. In this context, we look for the presence of boosted objects (such as top quarks) with substructure encompassed within single jets. Preliminary results on substructure classification will be presented.
Convolution kernels for multi-wavelength imaging
Boucaud, A.; Bocchio, M.; Abergel, A.; Orieux, F.; Dole, H.; Hadj-Youcef, M. A.
2016-12-01
Astrophysical images issued from different instruments and/or spectral bands often require to be processed together, either for fitting or comparison purposes. However each image is affected by an instrumental response, also known as point-spread function (PSF), that depends on the characteristics of the instrument as well as the wavelength and the observing strategy. Given the knowledge of the PSF in each band, a straightforward way of processing images is to homogenise them all to a target PSF using convolution kernels, so that they appear as if they had been acquired by the same instrument. We propose an algorithm that generates such PSF-matching kernels, based on Wiener filtering with a tunable regularisation parameter. This method ensures all anisotropic features in the PSFs to be taken into account. We compare our method to existing procedures using measured Herschel/PACS and SPIRE PSFs and simulated JWST/MIRI PSFs. Significant gains up to two orders of magnitude are obtained with respect to the use of kernels computed assuming Gaussian or circularised PSFs. A software to compute these kernels is available at https://github.com/aboucaud/pypher
Do Convolutional Neural Networks Learn Class Hierarchy?
Alsallakh, Bilal; Jourabloo, Amin; Ye, Mao; Liu, Xiaoming; Ren, Liu
2017-08-29
Convolutional Neural Networks (CNNs) currently achieve state-of-the-art accuracy in image classification. With a growing number of classes, the accuracy usually drops as the possibilities of confusion increase. Interestingly, the class confusion patterns follow a hierarchical structure over the classes. We present visual-analytics methods to reveal and analyze this hierarchy of similar classes in relation with CNN-internal data. We found that this hierarchy not only dictates the confusion patterns between the classes, it furthermore dictates the learning behavior of CNNs. In particular, the early layers in these networks develop feature detectors that can separate high-level groups of classes quite well, even after a few training epochs. In contrast, the latter layers require substantially more epochs to develop specialized feature detectors that can separate individual classes. We demonstrate how these insights are key to significant improvement in accuracy by designing hierarchy-aware CNNs that accelerate model convergence and alleviate overfitting. We further demonstrate how our methods help in identifying various quality issues in the training data.
Colonoscopic polyp detection using convolutional neural networks
Park, Sun Young; Sargent, Dusty
2016-03-01
Computer aided diagnosis (CAD) systems for medical image analysis rely on accurate and efficient feature extraction methods. Regardless of which type of classifier is used, the results will be limited if the input features are not diagnostically relevant and do not properly discriminate between the different classes of images. Thus, a large amount of research has been dedicated to creating feature sets that capture the salient features that physicians are able to observe in the images. Successful feature extraction reduces the semantic gap between the physician's interpretation and the computer representation of images, and helps to reduce the variability in diagnosis between physicians. Due to the complexity of many medical image classification tasks, feature extraction for each problem often requires domainspecific knowledge and a carefully constructed feature set for the specific type of images being classified. In this paper, we describe a method for automatic diagnostic feature extraction from colonoscopy images that may have general application and require a lower level of domain-specific knowledge. The work in this paper expands on our previous CAD algorithm for detecting polyps in colonoscopy video. In that work, we applied an eigenimage model to extract features representing polyps, normal tissue, diverticula, etc. from colonoscopy videos taken from various viewing angles and imaging conditions. Classification was performed using a conditional random field (CRF) model that accounted for the spatial and temporal adjacency relationships present in colonoscopy video. In this paper, we replace the eigenimage feature descriptor with features extracted from a convolutional neural network (CNN) trained to recognize the same image types in colonoscopy video. The CNN-derived features show greater invariance to viewing angles and image quality factors when compared to the eigenimage model. The CNN features are used as input to the CRF classifier as before. We report
Noise-enhanced convolutional neural networks.
Audhkhasi, Kartik; Osoba, Osonde; Kosko, Bart
2016-06-01
Injecting carefully chosen noise can speed convergence in the backpropagation training of a convolutional neural network (CNN). The Noisy CNN algorithm speeds training on average because the backpropagation algorithm is a special case of the generalized expectation-maximization (EM) algorithm and because such carefully chosen noise always speeds up the EM algorithm on average. The CNN framework gives a practical way to learn and recognize images because backpropagation scales with training data. It has only linear time complexity in the number of training samples. The Noisy CNN algorithm finds a special separating hyperplane in the network's noise space. The hyperplane arises from the likelihood-based positivity condition that noise-boosts the EM algorithm. The hyperplane cuts through a uniform-noise hypercube or Gaussian ball in the noise space depending on the type of noise used. Noise chosen from above the hyperplane speeds training on average. Noise chosen from below slows it on average. The algorithm can inject noise anywhere in the multilayered network. Adding noise to the output neurons reduced the average per-iteration training-set cross entropy by 39% on a standard MNIST image test set of handwritten digits. It also reduced the average per-iteration training-set classification error by 47%. Adding noise to the hidden layers can also reduce these performance measures. The noise benefit is most pronounced for smaller data sets because the largest EM hill-climbing gains tend to occur in the first few iterations. This noise effect can assist random sampling from large data sets because it allows a smaller random sample to give the same or better performance than a noiseless sample gives.
Voltage measurements at the vacuum post-hole convolute of the Z pulsed-power accelerator
Directory of Open Access Journals (Sweden)
E. M. Waisman
2014-12-01
Full Text Available Presented are voltage measurements taken near the load region on the Z pulsed-power accelerator using an inductive voltage monitor (IVM. Specifically, the IVM was connected to, and thus monitored the voltage at, the bottom level of the accelerator’s vacuum double post-hole convolute. Additional voltage and current measurements were taken at the accelerator’s vacuum-insulator stack (at a radius of 1.6 m by using standard D-dot and B-dot probes, respectively. During postprocessing, the measurements taken at the stack were translated to the location of the IVM measurements by using a lossless propagation model of the Z accelerator’s magnetically insulated transmission lines (MITLs and a lumped inductor model of the vacuum post-hole convolute. Across a wide variety of experiments conducted on the Z accelerator, the voltage histories obtained from the IVM and the lossless propagation technique agree well in overall shape and magnitude. However, large-amplitude, high-frequency oscillations are more pronounced in the IVM records. It is unclear whether these larger oscillations represent true voltage oscillations at the convolute or if they are due to noise pickup and/or transit-time effects and other resonant modes in the IVM. Results using a transit-time-correction technique and Fourier analysis support the latter. Regardless of which interpretation is correct, both true voltage oscillations and the excitement of resonant modes could be the result of transient electrical breakdowns in the post-hole convolute, though more information is required to determine definitively if such breakdowns occurred. Despite the larger oscillations in the IVM records, the general agreement found between the lossless propagation results and the results of the IVM shows that large voltages are transmitted efficiently through the MITLs on Z. These results are complementary to previous studies [R. D. McBride et al., Phys. Rev. ST Accel. Beams 13, 120401 (2010
Evaluation of convolutional neural networks for visual recognition.
Nebauer, C
1998-01-01
Convolutional neural networks provide an efficient method to constrain the complexity of feedforward neural networks by weight sharing and restriction to local connections. This network topology has been applied in particular to image classification when sophisticated preprocessing is to be avoided and raw images are to be classified directly. In this paper two variations of convolutional networks--neocognitron and a modification of neocognitron--are compared with classifiers based on fully connected feedforward layers (i.e., multilayer perceptron, nearest neighbor classifier, auto-encoding network) with respect to their visual recognition performance. Beside the original neocognitron a modification of the neocognitron is proposed which combines neurons from perceptron with the localized network structure of neocognitron. Instead of training convolutional networks by time-consuming error backpropagation, in this work a modular procedure is applied whereby layers are trained sequentially from the input to the output layer in order to recognize features of increasing complexity. For a quantitative experimental comparison with standard classifiers two very different recognition tasks have been chosen: handwritten digit recognition and face recognition. In the first example on handwritten digit recognition the generalization of convolutional networks is compared to fully connected networks. In several experiments the influence of variations of position, size, and orientation of digits is determined and the relation between training sample size and validation error is observed. In the second example recognition of human faces is investigated under constrained and variable conditions with respect to face orientation and illumination and the limitations of convolutional networks are discussed.
Brain and art: illustrations of the cerebral convolutions. A review.
Lazić, D; Marinković, S; Tomić, I; Mitrović, D; Starčević, A; Milić, I; Grujičić, M; Marković, B
2014-08-01
Aesthetics and functional significance of the cerebral cortical relief gave us the idea to find out how often the convolutions are presented in fine art, and in which techniques, conceptual meaning and pathophysiological aspect. We examined 27,614 art works created by 2,856 authors and presented in art literature, and in Google images search. The cerebral gyri were shown in 0.85% of the art works created by 2.35% of the authors. The concept of the brain was first mentioned in ancient Egypt some 3,700 years ago. The first artistic drawing of the convolutions was made by Leonardo da Vinci, and the first colour picture by an unknown Italian author. Rembrandt van Rijn was the first to paint the gyri. Dozens of modern authors, who are professional artists, medical experts or designers, presented the cerebralc onvolutions in drawings, paintings, digital works or sculptures, with various aesthetic, symbolic and metaphorical connotation. Some artistic compositions and natural forms show a gyral pattern. The convolutions, whose cortical layers enable the cognitive functions, can be affected by various disorders. Some artists suffered from those disorders, and some others presented them in their artworks. The cerebral convolutions or gyri, thanks to their extensive cortical mantle, are the specific morphological basis for the human mind, but also the structures with their own aesthetics. Contemporary authors relatively often depictor model the cerebral convolutions, either from the aesthetic or conceptual aspect. In this way, they make a connection between the neuroscience and fineart.
Cheng, Phillip M; Malhi, Harshawn S
2016-11-28
The purpose of this study is to evaluate transfer learning with deep convolutional neural networks for the classification of abdominal ultrasound images. Grayscale images from 185 consecutive clinical abdominal ultrasound studies were categorized into 11 categories based on the text annotation specified by the technologist for the image. Cropped images were rescaled to 256 × 256 resolution and randomized, with 4094 images from 136 studies constituting the training set, and 1423 images from 49 studies constituting the test set. The fully connected layers of two convolutional neural networks based on CaffeNet and VGGNet, previously trained on the 2012 Large Scale Visual Recognition Challenge data set, were retrained on the training set. Weights in the convolutional layers of each network were frozen to serve as fixed feature extractors. Accuracy on the test set was evaluated for each network. A radiologist experienced in abdominal ultrasound also independently classified the images in the test set into the same 11 categories. The CaffeNet network classified 77.3% of the test set images accurately (1100/1423 images), with a top-2 accuracy of 90.4% (1287/1423 images). The larger VGGNet network classified 77.9% of the test set accurately (1109/1423 images), with a top-2 accuracy of VGGNet was 89.7% (1276/1423 images). The radiologist classified 71.7% of the test set images correctly (1020/1423 images). The differences in classification accuracies between both neural networks and the radiologist were statistically significant (p neural networks may be used to construct effective classifiers for abdominal ultrasound images.
Implementation of FFT convolution and multigrid superposition models in the FOCUS RTP system
Miften, Moyed; Wiesmeyer, Mark; Monthofer, Suzanne; Krippner, Ken
2000-04-01
In radiotherapy treatment planning, convolution/superposition algorithms currently represent the best practical approach for accurate photon dose calculation in heterogeneous tissues. In this work, the implementation, accuracy and performance of the FFT convolution (FFTC) and multigrid superposition (MGS) algorithms are presented. The FFTC and MGS models use the same `TERMA' calculation and are commissioned using the same parameters. Both models use the same spectra, incorporate the same off-axis softening and base incident lateral fluence on the same measurements. In addition, corrections are explicitly applied to the polyenergetic and parallel kernel approximations, and electron contamination is modelled. Spectra generated by Monte Carlo (MC) modelling of treatment heads are used. Calculations using the MC spectra were in excellent agreement with measurements for many linear accelerator types. To speed up the calculations, a number of calculation techniques were implemented, including separate primary and scatter dose calculation, the FFT technique which assumes kernel invariance for the convolution calculation and a multigrid (MG) acceleration technique for the superposition calculation. Timing results show that the FFTC model is faster than MGS by a factor of 4 and 8 for small and large field sizes, respectively. Comparisons with measured data and BEAM MC results for a wide range of clinical beam setups show that (a) FFTC and MGS doses match measurements to better than 2% or 2 mm in homogeneous media; (b) MGS is more accurate than FFTC in lung phantoms where MGS doses are within 3% or 3 mm of BEAM results and (c) FFTC overestimates the dose in lung by a maximum of 9% compared to BEAM.
Glaucoma detection based on deep convolutional neural network.
Xiangyu Chen; Yanwu Xu; Damon Wing Kee Wong; Tien Yin Wong; Jiang Liu
2015-08-01
Glaucoma is a chronic and irreversible eye disease, which leads to deterioration in vision and quality of life. In this paper, we develop a deep learning (DL) architecture with convolutional neural network for automated glaucoma diagnosis. Deep learning systems, such as convolutional neural networks (CNNs), can infer a hierarchical representation of images to discriminate between glaucoma and non-glaucoma patterns for diagnostic decisions. The proposed DL architecture contains six learned layers: four convolutional layers and two fully-connected layers. Dropout and data augmentation strategies are adopted to further boost the performance of glaucoma diagnosis. Extensive experiments are performed on the ORIGA and SCES datasets. The results show area under curve (AUC) of the receiver operating characteristic curve in glaucoma detection at 0.831 and 0.887 in the two databases, much better than state-of-the-art algorithms. The method could be used for glaucoma detection.
Two dimensional convolute integers for machine vision and image recognition
Edwards, Thomas R.
1988-01-01
Machine vision and image recognition require sophisticated image processing prior to the application of Artificial Intelligence. Two Dimensional Convolute Integer Technology is an innovative mathematical approach for addressing machine vision and image recognition. This new technology generates a family of digital operators for addressing optical images and related two dimensional data sets. The operators are regression generated, integer valued, zero phase shifting, convoluting, frequency sensitive, two dimensional low pass, high pass and band pass filters that are mathematically equivalent to surface fitted partial derivatives. These operators are applied non-recursively either as classical convolutions (replacement point values), interstitial point generators (bandwidth broadening or resolution enhancement), or as missing value calculators (compensation for dead array element values). These operators show frequency sensitive feature selection scale invariant properties. Such tasks as boundary/edge enhancement and noise or small size pixel disturbance removal can readily be accomplished. For feature selection tight band pass operators are essential. Results from test cases are given.
Improving displayed resolution in convolution reconstruction of digital holograms
Institute of Scientific and Technical Information of China (English)
FAN Qi; ZHAO Jian-lin; ZHANG Yan-cao; WANG Jun; DI Jiang-lei
2006-01-01
In digital holographic microscopy,when the object is placed near the CCD,the Fresnel approximation is no longer valid and the convolution approach has to be applied.With this approach,the sampling spacing of the reconstructed image plane is equal to the pixel size of the CCD.If the lateral resolution of the reconstructed image is higher than that of the CCD,Nyquist sampling criterion is violated and aliasing errors will be introduced.In this Letter,a new method is proposed to solve this problem by investigating convolution reconstruction of digital holograms.By appending enough zeros to the angular spectrum between the two FFT's in convolution reconstruction of digital holograms,the displayed resolution of the reconstructed image can be improved.Experimental results show a good agreement with theoretical analysis.
Throughput Scaling Of Convolution For Error-Tolerant Multimedia Applications
Anam, Mohammad Ashraful
2012-01-01
Convolution and cross-correlation are the basis of filtering and pattern or template matching in multimedia signal processing. We propose two throughput scaling options for any one-dimensional convolution kernel in programmable processors by adjusting the imprecision (distortion) of computation. Our approach is based on scalar quantization, followed by two forms of tight packing in floating-point (one of which is proposed in this paper) that allow for concurrent calculation of multiple results. We illustrate how our approach can operate as an optional pre- and post-processing layer for off-the-shelf optimized convolution routines. This is useful for multimedia applications that are tolerant to processing imprecision and for cases where the input signals are inherently noisy (error tolerant multimedia applications). Indicative experimental results with a digital music matching system and an MPEG-7 audio descriptor system demonstrate that the proposed approach offers up to 175% increase in processing throughput...
Spatially variant convolution with scaled B-splines.
Muñoz-Barrutia, Arrate; Artaechevarria, Xabier; Ortiz-de-Solorzano, Carlos
2010-01-01
We present an efficient algorithm to compute multidimensional spatially variant convolutions--or inner products--between N-dimensional signals and B-splines--or their derivatives--of any order and arbitrary sizes. The multidimensional B-splines are computed as tensor products of 1-D B-splines, and the input signal is expressed in a B-spline basis. The convolution is then computed by using an adequate combination of integration and scaled finite differences as to have, for moderate and large scale values, a computational complexity that does not depend on the scaling factor. To show in practice the benefit of using our spatially variant convolution approach, we present an adaptive noise filter that adjusts the kernel size to the local image characteristics and a high sensitivity local ridge detector.
The Existence of Strongly-MDS Convolutional Codes
Hutchinson, Ryan
2008-01-01
It is known that maximum distance separable and maximum distance profile convolutional codes exist over large enough finite fields of any characteristic for all parameters $(n,k,\\delta)$. It has been conjectured that the same is true for convolutional codes that are strongly maximum distance separable. Using methods from linear systems theory, we resolve this conjecture by showing that, over a large enough finite field of any characteristic, codes which are simultaneously maximum distance profile and strongly maximum distance separable exist for all parameters $(n,k,\\delta)$.
Convolutional cylinder-type block-circulant cycle codes
Directory of Open Access Journals (Sweden)
Mohammad Gholami
2013-06-01
Full Text Available In this paper, we consider a class of column-weight two quasi-cyclic low-density paritycheck codes in which the girth can be large enough, as an arbitrary multiple of 8. Then we devote a convolutional form to these codes, such that their generator matrix can be obtained by elementary row and column operations on the parity-check matrix. Finally, we show that the free distance of the convolutional codes is equal to the minimum distance of their block counterparts.
Inferring low-dimensional microstructure representations using convolutional neural networks
Lubbers, Nicholas; Barros, Kipton
2016-01-01
We apply recent advances in machine learning and computer vision to a central problem in materials informatics: The statistical representation of microstructural images. We use activations in a pre-trained convolutional neural network to provide a high-dimensional characterization of a set of synthetic microstructural images. Next, we use manifold learning to obtain a low-dimensional embedding of this statistical characterization. We show that the low-dimensional embedding extracts the parameters used to generate the images. According to a variety of metrics, the convolutional neural network method yields dramatically better embeddings than the analogous method derived from two-point correlations alone.
Detection of phase transition via convolutional neural network
Tanaka, Akinori
2016-01-01
We design a Convolutional Neural Network (CNN) which studies correlation between discretized inverse temperature and spin configuration of 2D Ising model and show that it can find a feature of the phase transition without teaching any a priori information for it. We also define a new order parameter via the CNN and show that it provides well approximated critical inverse temperature. In addition, we compare the activation functions for convolution layer and find that the Rectified Linear Unit (ReLU) is important to detect the phase transition of 2D Ising model.
Directory of Open Access Journals (Sweden)
Kılıç Emrah
2016-12-01
Full Text Available In this paper, we consider Gauthier’s generalized convolution and then define its binomial analogue as well as alternating binomial analogue. We formulate these convolutions and give some applications of them.
Directory of Open Access Journals (Sweden)
Oliver Serang
Full Text Available Exact Bayesian inference can sometimes be performed efficiently for special cases where a function has commutative and associative symmetry of its inputs (called "causal independence". For this reason, it is desirable to exploit such symmetry on big data sets. Here we present a method to exploit a general form of this symmetry on probabilistic adder nodes by transforming those probabilistic adder nodes into a probabilistic convolution tree with which dynamic programming computes exact probabilities. A substantial speedup is demonstrated using an illustration example that can arise when identifying splice forms with bottom-up mass spectrometry-based proteomics. On this example, even state-of-the-art exact inference algorithms require a runtime more than exponential in the number of splice forms considered. By using the probabilistic convolution tree, we reduce the runtime to O(k log(k2 and the space to O(k log(k where k is the number of variables joined by an additive or cardinal operator. This approach, which can also be used with junction tree inference, is applicable to graphs with arbitrary dependency on counting variables or cardinalities and can be used on diverse problems and fields like forward error correcting codes, elemental decomposition, and spectral demixing. The approach also trivially generalizes to multiple dimensions.
Radio Signal Augmentation for Improved Training of a Convolutional Neural Network
2016-09-01
parameters of the network. Examples of these parameters include: • Input data dimensions and channels (e.g., image size and colors) • Size of convolutional ...filters • Number of convolutional filters • Pooling/downsampling size and method (e.g., max-pool or average) • Number of convolution and pooling...TECHNICAL REPORT 3055 September 2016 Radio Signal Augmentation for Improved Training of a Convolutional Neural Network Daniel
Efficient Partitioning of Algorithms for Long Convolutions and their Mapping onto Architectures
Bierens, L.; Deprettere, E.
1998-01-01
We present an efficient approach for the partitioning of algorithms implementing long convolutions. The dependence graph (DG) of a convolution algorithm is locally sequential globally parallel (LSGP) partitioned into smaller, less complex convolution algorithms. The LSGP partitioned DG is mapped ont
General Purpose Convolution Algorithm in S4 Classes by Means of FFT
Directory of Open Access Journals (Sweden)
Peter Ruckdeschel
2014-08-01
By means of object orientation this default algorithm is overloaded by more specific algorithms where possible, in particular where explicit convolution formulae are available. Our focus is on R package distr which implements this approach, overloading operator + for convolution; based on this convolution, we define a whole arithmetics of mathematical operations acting on distribution objects, comprising operators +, -, *, /, and ^.
Robust Fusion of Irregularly Sampled Data Using Adaptive Normalized Convolution
Pham, T.Q.; Van Vliet, L.J.; Schutte, K.
2006-01-01
We present a novel algorithm for image fusion from irregularly sampled data. The method is based on the framework of normalized convolution (NC), in which the local signal is approximated through a projection onto a subspace. The use of polynomial basis functions in this paper makes NC equivalent to
A single Chip Implementation for Fast Convolution of Long Sequences
Zwartenkot, H.T.J.; Boerrigter, M.J.G.; Bierens, L.H.J.; Smit, J.
1996-01-01
Usually, long convolutions are computed by programmable DSP boards using long FFTs. Typical operational requirements such as minimum power dissipation, minimum volume and high dynamic range/accuracy, make this solution often inefficient and even unacceptable. In this paper we present a single chip f
Unsupervised pre-training for fully convolutional neural networks
Wiehman, Stiaan; Kroon, Steve; Villiers, De Hendrik
2017-01-01
Unsupervised pre-training of neural networks has been shown to act as a regularization technique, improving performance and reducing model variance. Recently, fully convolutional networks (FCNs) have shown state-of-the-art results on various semantic segmentation tasks. Unfortunately, there is no ef
Face recognition: a convolutional neural-network approach.
Lawrence, S; Giles, C L; Tsoi, A C; Back, A D
1997-01-01
We present a hybrid neural-network for human face recognition which compares favourably with other methods. The system combines local image sampling, a self-organizing map (SOM) neural network, and a convolutional neural network. The SOM provides a quantization of the image samples into a topological space where inputs that are nearby in the original space are also nearby in the output space, thereby providing dimensionality reduction and invariance to minor changes in the image sample, and the convolutional neural network provides partial invariance to translation, rotation, scale, and deformation. The convolutional network extracts successively larger features in a hierarchical set of layers. We present results using the Karhunen-Loeve transform in place of the SOM, and a multilayer perceptron (MLP) in place of the convolutional network for comparison. We use a database of 400 images of 40 individuals which contains quite a high degree of variability in expression, pose, and facial details. We analyze the computational complexity and discuss how new classes could be added to the trained recognizer.
CICAAR - Convolutive ICA with an Auto-Regressive Inverse Model
DEFF Research Database (Denmark)
Dyrholm, Mads; Hansen, Lars Kai
2004-01-01
We invoke an auto-regressive IIR inverse model for convolutive ICA and derive expressions for the likelihood and its gradient. We argue that optimization will give a stable inverse. When there are more sensors than sources the mixing model parameters are estimated in a second step by least squares...
Real-time rendering of optical effects using spatial convolution
Rokita, Przemyslaw
1998-03-01
Simulation of special effects such as: defocus effect, depth-of-field effect, raindrops or water film falling on the windshield, may be very useful in visual simulators and in all computer graphics applications that need realistic images of outdoor scenery. Those effects are especially important in rendering poor visibility conditions in flight and driving simulators, but can also be applied, for example, in composing computer graphics and video sequences- -i.e. in Augmented Reality systems. This paper proposes a new approach to the rendering of those optical effects by iterative adaptive filtering using spatial convolution. The advantage of this solution is that the adaptive convolution can be done in real-time by existing hardware. Optical effects mentioned above can be introduced into the image computed using conventional camera model by applying to the intensity of each pixel the convolution filter having an appropriate point spread function. The algorithms described in this paper can be easily implemented int the visualization pipeline--the final effect may be obtained by iterative filtering using a single hardware convolution filter or with the pipeline composed of identical 3 X 3 filters placed as the stages of this pipeline. Another advantage of the proposed solution is that the extension based on proposed algorithm can be added to the existing rendering systems as a final stage of the visualization pipeline.
Convolution operators defined by singular measures on the motion group
Brandolini, Luca; Thangavelu, Sundaram; Travaglini, Giancarlo
2010-01-01
This paper contains an $L^{p}$ improving result for convolution operators defined by singular measures associated to hypersurfaces on the motion group. This needs only mild geometric properties of the surfaces, and it extends earlier results on Radon type transforms on $\\mathbb{R}^{n}$. The proof relies on the harmonic analysis on the motion group.
Review of the convolution algorithm for evaluating service integrated systems
DEFF Research Database (Denmark)
Iversen, Villy Bæk
1997-01-01
In this paper we give a review of the applicability of the convolution algorithm. By this we are able to evaluate communication networks end--to--end with e.g. BPP multi-ratetraffic models insensitive to the holding time distribution. Rearrangement, minimum allocation, and maximum allocation are ...
Behaviour at infinity of solutions of twisted convolution equations
Energy Technology Data Exchange (ETDEWEB)
Volchkov, Valerii V; Volchkov, Vitaly V [Donetsk National University, Donetsk (Ukraine)
2012-02-28
We obtain a precise characterization of the minimal rate of growth at infinity of non-trivial solutions of twisted convolution equations in unbounded domains of C{sup n}. As an application, we obtain definitive versions of the two-radii theorem for twisted spherical means.
Two-level convolution formula for nuclear structure function
Ma, Boqiang
1990-05-01
A two-level convolution formula for the nuclear structure function is derived in considering the nucleus as a composite system of baryon-mesons which are also composite systems of quark-gluons again. The results show that the European Muon Colaboration effect can not be explained by the nuclear effects as nucleon Fermi motion and nuclear binding contributions.
Two-Dimensional Tail-Biting Convolutional Codes
Alfandary, Liam
2011-01-01
The multidimensional convolutional codes are an extension of the notion of convolutional codes (CCs) to several dimensions of time. This paper explores the class of two-dimensional convolutional codes (2D CCs) and 2D tail-biting convolutional codes (2D TBCCs), in particular, from several aspects. First, we derive several basic algebraic properties of these codes, applying algebraic methods in order to find bijective encoders, create parity check matrices and to inverse encoders. Next, we discuss the minimum distance and weight distribution properties of these codes. Extending an existing tree-search algorithm to two dimensions, we apply it to find codes with high minimum distance. Word-error probability asymptotes for sample codes are given and compared with other codes. The results of this approach suggest that 2D TBCCs can perform better than comparable 1D TBCCs or other codes. We then present several novel iterative suboptimal algorithms for soft decoding 2D CCs, which are based on belief propagation. Two ...
Yetter－Drinfel‘d Module and Convolution Module
Institute of Scientific and Technical Information of China (English)
张良云; 王栓宏
2002-01-01
In this paper,we first give a sufficent and necessary condition for a Hopf algebra to be a Yetter-Drinfel'd module,and prove that the finite dual of a Yetter-Drinfel'd module is still a Yetter-Drinfel'd module,Finally,we introduce a concept of convolution module.
On the generalized Hamming weights of convolutional codes
Rosenthal, J.; York, E.V.
1995-01-01
Motivated by applications in cryptology K. Wei introduced in 1991 the concept of a generalized Hamming weight for a linear block code. In this paper we define generalized Hamming weights for the class of convolutional codes and we derive several of their basic properties. By restricting to convoluti
Maximum-likelihood estimation of circle parameters via convolution.
Zelniker, Emanuel E; Clarkson, I Vaughan L
2006-04-01
The accurate fitting of a circle to noisy measurements of circumferential points is a much studied problem in the literature. In this paper, we present an interpretation of the maximum-likelihood estimator (MLE) and the Delogne-Kåsa estimator (DKE) for circle-center and radius estimation in terms of convolution on an image which is ideal in a certain sense. We use our convolution-based MLE approach to find good estimates for the parameters of a circle in digital images. In digital images, it is then possible to treat these estimates as preliminary estimates into various other numerical techniques which further refine them to achieve subpixel accuracy. We also investigate the relationship between the convolution of an ideal image with a "phase-coded kernel" (PCK) and the MLE. This is related to the "phase-coded annulus" which was introduced by Atherton and Kerbyson who proposed it as one of a number of new convolution kernels for estimating circle center and radius. We show that the PCK is an approximate MLE (AMLE). We compare our AMLE method to the MLE and the DKE as well as the Cramér-Rao Lower Bound in ideal images and in both real and synthetic digital images.
Robust Fusion of Irregularly Sampled Data Using Adaptive Normalized Convolution
Pham, T.Q.; Vliet, L.J. van; Schutte, K.
2006-01-01
We present a novel algorithm for image fusion from irregularly sampled data. The method is based on the framework of normalized convolution (NC), in which the local signal is approximated through a projection onto a subspace. The use of polynomial basis functions in this paper makes NC equivalent to
A single Chip Implementation for Fast Convolution of Long Sequences
Zwartenkot, H.T.J.; Boerrigter, M.J.G.; Bierens, L.H.J.; Smit, J.
1996-01-01
Usually, long convolutions are computed by programmable DSP boards using long FFTs. Typical operational requirements such as minimum power dissipation, minimum volume and high dynamic range/accuracy, make this solution often inefficient and even unacceptable. In this paper we present a single chip
Comments on "A New Random-Error-Correction Code"
DEFF Research Database (Denmark)
Paaske, Erik
1979-01-01
This correspondence investigates the error propagation properties of six different systems using a (12, 6) systematic double-error-correcting convolutional encoder and a one-step majority-logic feedback decoder. For the generally accepted assumption that channel errors are much more likely to occur...
Testing Error Correcting Codes by Multicanonical Sampling of Rare Events
Iba, Yukito; Hukushima, Koji
2007-01-01
The idea of rare event sampling is applied to the estimation of the performance of error-correcting codes. The essence of the idea is importance sampling of the pattern of noises in the channel by Multicanonical Monte Carlo, which enables efficient estimation of tails of the distribution of bit error rate. The idea is successfully tested with a convolutional code.
Kinetic Energy of Hydrocarbons as a Function of Electron Density and Convolutional Neural Networks.
Yao, Kun; Parkhill, John
2016-03-01
We demonstrate a convolutional neural network trained to reproduce the Kohn-Sham kinetic energy of hydrocarbons from an input electron density. The output of the network is used as a nonlocal correction to conventional local and semilocal kinetic functionals. We show that this approximation qualitatively reproduces Kohn-Sham potential energy surfaces when used with conventional exchange correlation functionals. The density which minimizes the total energy given by the functional is examined in detail. We identify several avenues to improve on this exploratory work, by reducing numerical noise and changing the structure of our functional. Finally we examine the features in the density learned by the neural network to anticipate the prospects of generalizing these models.
FAST-PT: a novel algorithm to calculate convolution integrals in cosmological perturbation theory
McEwen, Joseph E.; Fang, Xiao; Hirata, Christopher M.; Blazek, Jonathan A.
2016-09-01
We present a novel algorithm, FAST-PT, for performing convolution or mode-coupling integrals that appear in nonlinear cosmological perturbation theory. The algorithm uses several properties of gravitational structure formation—the locality of the dark matter equations and the scale invariance of the problem—as well as Fast Fourier Transforms to describe the input power spectrum as a superposition of power laws. This yields extremely fast performance, enabling mode-coupling integral computations fast enough to embed in Monte Carlo Markov Chain parameter estimation. We describe the algorithm and demonstrate its application to calculating nonlinear corrections to the matter power spectrum, including one-loop standard perturbation theory and the renormalization group approach. We also describe our public code (in Python) to implement this algorithm. The code, along with a user manual and example implementations, is available at https://github.com/JoeMcEwen/FAST-PT.
Low-Power Maximum a Posteriori (MAP Algorithm for WiMAX Convolutional Turbo Decoder
Directory of Open Access Journals (Sweden)
Chitralekha Ngangbam
2013-05-01
Full Text Available We propose to design a Low-Power Memory-Reduced Traceback MAP iterative decoding of convolutional turbo code (CTC which has large data access with large memories consumption and verify the functionality by using simulation tool. The traceback maximum a posteriori algorithm (MAP decoding provides the best performance in terms of bit error rate (BER and reduce the power consumption of the state metric cache (SMC without losing the correction performance. The computation and accessing of different metrics reduce the size of the SMC with no requires complicated reversion checker, path selection, and reversion flag cache. Radix-2*2 and radix-4 traceback structures provide a tradeoff between power consumption and operating frequency for double-binary (DB MAP decoding. These two traceback structures achieve an around 25% power reduction of the SMC, and around 12% power reduction of the DB MAP decoders for WiMAX standard
FAST-PT: a novel algorithm to calculate convolution integrals in cosmological perturbation theory
McEwen, Joseph E; Hirata, Christopher M; Blazek, Jonathan A
2016-01-01
We present a novel algorithm, FAST-PT, for performing convolution or mode-coupling integrals that appear in nonlinear cosmological perturbation theory. The algorithm uses several properties of gravitational structure formation -- the locality of the dark matter equations and the scale invariance of the problem -- as well as Fast Fourier Transforms to describe the input power spectrum as a superposition of power laws. This yields extremely fast performance, enabling mode-coupling integral computations fast enough to embed in Monte Carlo Markov Chain parameter estimation. We describe the algorithm and demonstrate its application to calculating nonlinear corrections to the matter power spectrum, including one-loop standard perturbation theory and the renormalization group approach. We also describe our public code (in Python) to implement this algorithm, including the applications described here.
Larsson, Anne; Johansson, Lennart
2003-11-01
In single photon emission computed tomography (SPECT), transmission-dependent convolution subtraction has been shown to be useful when correcting for scattered events. The method is based on convolution subtraction, but includes a matrix of scatter fractions instead of a global scatter fraction. The method can be extended to iteratively improve the scatter estimate, but in this note we show that this requires a modification of the theory to use scatter-to-total scatter fractions for the first iteration only and scatter-to-primary fractions thereafter. To demonstrate this, scatter correction is performed on a Monte Carlo simulated image of a point source of activity in water. The modification of the theory is compared to corrections where the scatter fractions are based on the scatter-to-total ratio, using one and ten iterations. The resulting ratios of subtracted to original counts are compared to the true scatter-to-total ratio of the simulation and the most accurate result is found for our modification of the theory.
An Implementation of Error Minimization Data Transmission in OFDM using Modified Convolutional Code
Directory of Open Access Journals (Sweden)
Hendy Briantoro
2016-04-01
Full Text Available This paper presents about error minimization in OFDM system. In conventional system, usually using channel coding such as BCH Code or Convolutional Code. But, performance BCH Code or Convolutional Code is not good in implementation of OFDM System. Error bits of OFDM system without channel coding is 5.77%. Then, we used convolutional code with code rate 1/2, it can reduce error bitsonly up to 3.85%. So, we proposed OFDM system with Modified Convolutional Code. In this implementation, we used Software Define Radio (SDR, namely Universal Software Radio Peripheral (USRP NI 2920 as the transmitter and receiver. The result of OFDM system using Modified Convolutional Code with code rate is able recover all character received so can decrease until 0% error bit. Increasing performance of Modified Convolutional Code is about 1 dB in BER of 10-4 from BCH Code and Convolutional Code. So, performance of Modified Convolutional better than BCH Code or Convolutional Code. Keywords: OFDM, BCH Code, Convolutional Code, Modified Convolutional Code, SDR, USRP
Energy Technology Data Exchange (ETDEWEB)
Gonzalez-Castano, D. M.; Gonzalez, L. Brualla; Gago-Arias, M. A.; Pardo-Montero, J.; Gomez, F.; Luna-Vega, V.; Sanchez, M.; Lobato, R. [Radiation Physics Laboratory, Universidad de Santiago de Compostela, 15782 (Spain) and Dpto de Fisica de Particulas, Universidad de Santiago de Compostela, 15782 (Spain); Servicio de Radiofisica ERESA, Consorcio Hospital General Universitario de Valencia, 46014 (Spain); Dpto de Fisica de Particulas, Universidad de Santiago de Compostela, 15782 (Spain); Radiation Physics Laboratory, Universidad de Santiago de Compostela, 15782 Spain and Dpto de Fisica de Particulas, Universidad de Santiago de Compostela, 15782 (Spain); Servicio de Radiofisica y Proteccion Radiologica, Hospital Clinico Universitario de Santiago, Santiago de Compostela, 15782 (Spain)
2012-01-15
Purpose: This work contains an alternative methodology for obtaining correction factors for ionization chamber (IC) dosimetry of small fields and composite fields such as IMRT. The method is based on the convolution/superposition (C/S) of an IC response function (RF) with the dose distribution in a certain plane which includes chamber position. This method is an alternative to the full Monte Carlo (MC) approach that has been used previously by many authors for the same objective. Methods: The readout of an IC at a point inside a phantom irradiated by a certain beam can be obtained as the convolution of the dose spatial distribution caused by the beam and the IC two-dimensional RF. The proposed methodology has been applied successfully to predict the response of a PTW 30013 IC when measuring different nonreference fields, namely: output factors of 6 MV small fields, beam profiles of cobalt 60 narrow fields and 6 MV radiosurgery segments. The two-dimensional RF of a PTW 30013 IC was obtained by MC simulation of the absorbed dose to cavity air when the IC was scanned by a 0.6 x 0.6 mm{sup 2} cross section parallel pencil beam at low depth in a water phantom. For each of the cases studied, the results of the IC direct measurement were compared with the corresponding obtained by the C/S method. Results: For all of the cases studied, the agreement between the IC direct measurement and the IC calculated response was excellent (better than 1.5%). Conclusions: This method could be implemented in TPS in order to calculate dosimetry correction factors when an experimental IMRT treatment verification with in-phantom ionization chamber is performed. The miss-response of the IC due to the nonreference conditions could be quickly corrected by this method rather than employing MC derived correction factors. This method can be considered as an alternative to the plan-class associated correction factors proposed recently as part of an IAEA work group on nonstandard field dosimetry.
Image Super-Resolution Using Deep Convolutional Networks.
Dong, Chao; Loy, Chen Change; He, Kaiming; Tang, Xiaoou
2016-02-01
We propose a deep learning method for single image super-resolution (SR). Our method directly learns an end-to-end mapping between the low/high-resolution images. The mapping is represented as a deep convolutional neural network (CNN) that takes the low-resolution image as the input and outputs the high-resolution one. We further show that traditional sparse-coding-based SR methods can also be viewed as a deep convolutional network. But unlike traditional methods that handle each component separately, our method jointly optimizes all layers. Our deep CNN has a lightweight structure, yet demonstrates state-of-the-art restoration quality, and achieves fast speed for practical on-line usage. We explore different network structures and parameter settings to achieve trade-offs between performance and speed. Moreover, we extend our network to cope with three color channels simultaneously, and show better overall reconstruction quality.
The analysis of VERITAS muon images using convolutional neural networks
Feng, Qi
2016-01-01
Imaging atmospheric Cherenkov telescopes (IACTs) are sensitive to rare gamma-ray photons, buried in the background of charged cosmic-ray (CR) particles, the flux of which is several orders of magnitude greater. The ability to separate gamma rays from CR particles is important, as it is directly related to the sensitivity of the instrument. This gamma-ray/CR-particle classification problem in IACT data analysis can be treated with the rapidly-advancing machine learning algorithms, which have the potential to outperform the traditional box-cut methods on image parameters. We present preliminary results of a precise classification of a small set of muon events using a convolutional neural networks model with the raw images as input features. We also show the possibility of using the convolutional neural networks model for regression problems, such as the radius and brightness measurement of muon events, which can be used to calibrate the throughput efficiency of IACTs.
Convolution theorems for the linear canonical transform and their applications
Institute of Scientific and Technical Information of China (English)
DENG Bing; TAO Ran; WANG Yue
2006-01-01
As generalization of the fractional Fourier transform (FRFT), the linear canonical transform (LCT) has been used in several areas, including optics and signal processing. Many properties for this transform are already known, but the convolution theorems, similar to the version of the Fourier transform, are still to be determined. In this paper, the authors derive the convolution theorems for the LCT, and explore the sampling theorem and multiplicative filter for the band limited signal in the linear canonical domain. Finally, the sampling and reconstruction formulas are deduced, together with the construction methodology for the above mentioned multiplicative filter in the time domain based on fast Fourier transform (FFT), which has much lower computational load than the construction method in the linear canonical domain.
Star-galaxy Classification Using Deep Convolutional Neural Networks
Kim, Edward J
2016-01-01
Most existing star-galaxy classifiers use the reduced summary information from catalogs, requiring careful feature extraction and selection. The latest advances in machine learning that use deep convolutional neural networks allow a machine to automatically learn the features directly from data, minimizing the need for input from human experts. We present a star-galaxy classification framework that uses deep convolutional neural networks (ConvNets) directly on the reduced, calibrated pixel values. Using data from the Sloan Digital Sky Survey (SDSS) and the Canada-France-Hawaii Telescope Lensing Survey (CFHTLenS), we demonstrate that ConvNets are able to produce accurate and well-calibrated probabilistic classifications that are competitive with conventional machine learning techniques. Future advances in deep learning may bring more success with current and forthcoming photometric surveys, such as the Dark Energy Survey (DES) and the Large Synoptic Survey Telescope (LSST), because deep neural networks require...
The analysis of VERITAS muon images using convolutional neural networks
Feng, Qi; Lin, Tony T. Y.; VERITAS Collaboration
2017-06-01
Imaging atmospheric Cherenkov telescopes (IACTs) are sensitive to rare gamma-ray photons, buried in the background of charged cosmic-ray (CR) particles, the flux of which is several orders of magnitude greater. The ability to separate gamma rays from CR particles is important, as it is directly related to the sensitivity of the instrument. This gamma-ray/CR-particle classification problem in IACT data analysis can be treated with the rapidly-advancing machine learning algorithms, which have the potential to outperform the traditional box-cut methods on image parameters. We present preliminary results of a precise classification of a small set of muon events using a convolutional neural networks model with the raw images as input features. We also show the possibility of using the convolutional neural networks model for regression problems, such as the radius and brightness measurement of muon events, which can be used to calibrate the throughput efficiency of IACTs.
a Convolutional Network for Semantic Facade Segmentation and Interpretation
Schmitz, Matthias; Mayer, Helmut
2016-06-01
In this paper we present an approach for semantic interpretation of facade images based on a Convolutional Network. Our network processes the input images in a fully convolutional way and generates pixel-wise predictions. We show that there is no need for large datasets to train the network when transfer learning is employed, i. e., a part of an already existing network is used and fine-tuned, and when the available data is augmented by using deformed patches of the images for training. The network is trained end-to-end with patches of the images and each patch is augmented independently. To undo the downsampling for the classification, we add deconvolutional layers to the network. Outputs of different layers of the network are combined to achieve more precise pixel-wise predictions. We demonstrate the potential of our network based on results for the eTRIMS (Korč and Förstner, 2009) dataset reduced to facades.
Self-Taught convolutional neural networks for short text clustering.
Xu, Jiaming; Xu, Bo; Wang, Peng; Zheng, Suncong; Tian, Guanhua; Zhao, Jun; Xu, Bo
2017-01-12
Short text clustering is a challenging problem due to its sparseness of text representation. Here we propose a flexible Self-Taught Convolutional neural network framework for Short Text Clustering (dubbed STC(2)), which can flexibly and successfully incorporate more useful semantic features and learn non-biased deep text representation in an unsupervised manner. In our framework, the original raw text features are firstly embedded into compact binary codes by using one existing unsupervised dimensionality reduction method. Then, word embeddings are explored and fed into convolutional neural networks to learn deep feature representations, meanwhile the output units are used to fit the pre-trained binary codes in the training process. Finally, we get the optimal clusters by employing K-means to cluster the learned representations. Extensive experimental results demonstrate that the proposed framework is effective, flexible and outperform several popular clustering methods when tested on three public short text datasets.
Trajectory Generation Method with Convolution Operation on Velocity Profile
Energy Technology Data Exchange (ETDEWEB)
Lee, Geon [Hanyang Univ., Seoul (Korea, Republic of); Kim, Doik [Korea Institute of Science and Technology, Daejeon (Korea, Republic of)
2014-03-15
The use of robots is no longer limited to the field of industrial robots and is now expanding into the fields of service and medical robots. In this light, a trajectory generation method that can respond instantaneously to the external environment is strongly required. Toward this end, this study proposes a method that enables a robot to change its trajectory in real-time using a convolution operation. The proposed method generates a trajectory in real time and satisfies the physical limits of the robot system such as acceleration and velocity limit. Moreover, a new way to improve the previous method, which generates inefficient trajectories in some cases owing to the characteristics of the trapezoidal shape of trajectories, is proposed by introducing a triangle shape. The validity and effectiveness of the proposed method is shown through a numerical simulation and a comparison with the previous convolution method.
On New Bijective Convolution Operator Act for Analytic Functions
Directory of Open Access Journals (Sweden)
Oqlah Al-Refai
2009-01-01
Full Text Available Problem statement: We introduced a new bijective convolution linear operator defined on the class of normalized analytic functions. This operator was motivated by many researchers namely Srivastava, Owa, Ruscheweyh and many others. The operator was essential to obtain new classes of analytic functions. Approach: Simple technique of Ruscheweyh was used in our preliminary approach to create new bijective convolution linear operator. The preliminary concept of Hadamard products was mentioned and the concept of subordination was given to give sharp proofs for certain sufficient conditions of the linear operator aforementioned. In fact, the subordinating factor sequence was used to derive different types of subordination results. Results: Having the linear operator, subordination theorems were established by using standard concept of subordination. The results reduced to well-known results studied by various researchers. Coefficient bounds and inclusion properties, growth and closure theorems for some subclasses were also obtained. Conclusion: Therefore, many interesting results could be obtained and some applications could be gathered.
Fibonacci Sequence, Recurrence Relations, Discrete Probability Distributions and Linear Convolution
Rajan, Arulalan; Rao, Ashok; Jamadagni, H S
2012-01-01
The classical Fibonacci sequence is known to exhibit many fascinating properties. In this paper, we explore the Fibonacci sequence and integer sequences generated by second order linear recurrence relations with positive integer coe?cients from the point of view of probability distributions that they induce. We obtain the generalizations of some of the known limiting properties of these probability distributions and present certain optimal properties of the classical Fibonacci sequence in this context. In addition, we also look at the self linear convolution of linear recurrence relations with positive integer coefficients. Analysis of self linear convolution is focused towards locating the maximum in the resulting sequence. This analysis, also highlights the influence that the largest positive real root, of the "characteristic equation" of the linear recurrence relations with positive integer coefficients, has on the location of the maximum. In particular, when the largest positive real root is 2,the locatio...
3D Medical Image Interpolation Based on Parametric Cubic Convolution
Institute of Scientific and Technical Information of China (English)
无
2007-01-01
In the process of display, manipulation and analysis of biomedical image data, they usually need to be converted to data of isotropic discretization through the process of interpolation, while the cubic convolution interpolation is widely used due to its good tradeoff between computational cost and accuracy. In this paper, we present a whole concept for the 3D medical image interpolation based on cubic convolution, and the six methods, with the different sharp control parameter, which are formulated in details. Furthermore, we also give an objective comparison for these methods using data sets with the different slice spacing. Each slice in these data sets is estimated by each interpolation method and compared with the original slice using three measures: mean-squared difference, number of sites of disagreement, and largest difference. According to the experimental results, we present a recommendation for 3D medical images under the different situations in the end.
Infimal Convolution Regularisation Functionals of BV and Lp Spaces
Burger, Martin
2016-02-03
We study a general class of infimal convolution type regularisation functionals suitable for applications in image processing. These functionals incorporate a combination of the total variation seminorm and Lp norms. A unified well-posedness analysis is presented and a detailed study of the one-dimensional model is performed, by computing exact solutions for the corresponding denoising problem and the case p=2. Furthermore, the dependency of the regularisation properties of this infimal convolution approach to the choice of p is studied. It turns out that in the case p=2 this regulariser is equivalent to the Huber-type variant of total variation regularisation. We provide numerical examples for image decomposition as well as for image denoising. We show that our model is capable of eliminating the staircasing effect, a well-known disadvantage of total variation regularisation. Moreover as p increases we obtain almost piecewise affine reconstructions, leading also to a better preservation of hat-like structures.
Protein secondary structure prediction using deep convolutional neural fields
Sheng Wang; Jian Peng; Jianzhu Ma; Jinbo Xu
2015-01-01
Protein secondary structure (SS) prediction is important for studying protein structure and function. When only the sequence (profile) information is used as input feature, currently the best predictors can obtain ~80% Q3 accuracy, which has not been improved in the past decade. Here we present DeepCNF (Deep Convolutional Neural Fields) for protein SS prediction. DeepCNF is a Deep Learning extension of Conditional Neural Fields (CNF), which is an integration of Conditional Random Fields (CRF)...
Interleaved Convolutional Code and Its Viterbi Decoder Architecture
2003-01-01
We propose an area-efficient high-speed interleaved Viterbi decoder architecture, which is based on the state-parallel architecture with register exchange path memory structure, for interleaved convolutional code. The state-parallel architecture uses as many add-compare-select (ACS) units as the number of trellis states. By replacing each delay (or storage) element in state metrics memory (or path metrics memory) and path memory (or survival memory) with delays, interleaved Viterbi decoder ...
Design of SVD/SGK Convolution Filters for Image Processing
1980-01-01
of filters by transforming one-dimensional linear phase filters * into two-dimensional linear phase filters . By assuming that the prototype filter is a...linear phase filter , his algorithm transforms a one-dimensional filter h(u) into a two-dimensional filter W (u,v) by means of transformation given by...significance of their implementation of the designed filter is that a large two-dimensional convolution *A linear phase filter implies symmetry of the filter. 13
Efficient Convolutional Neural Network with Binary Quantization Layer
Ravanbakhsh, Mahdyar; Mousavi, Hossein; Nabi, Moin; Marcenaro, Lucio; Regazzoni, Carlo
2016-01-01
In this paper we introduce a novel method for segmentation that can benefit from general semantics of Convolutional Neural Network (CNN). Our segmentation proposes visually and semantically coherent image segments. We use binary encoding of CNN features to overcome the difficulty of the clustering on the high-dimensional CNN feature space. These binary encoding can be embedded into the CNN as an extra layer at the end of the network. This results in real-time segmentation. To the best of our ...
Fusing Deep Convolutional Networks for Large Scale Visual Concept Classification
Ergun, Hilal; Sert, Mustafa
2016-01-01
Deep learning architectures are showing great promise in various computer vision domains including image classification, object detection, event detection and action recognition. In this study, we investigate various aspects of convolutional neural networks (CNNs) from the big data perspective. We analyze recent studies and different network architectures both in terms of running time and accuracy. We present extensive empirical information along with best practices for big data practitioners...
An obstruction for q-deformation of the convolution product
Van Leeuwen, H; van Leeuwen, Hans; Maassen, Hans
1995-01-01
We consider two independent q-Gaussian random variables X and Y and a function f chosen in such a way that f(X) and X have the same distribution. For 0 < q < 1 we find that at least the fourth moments of X + Y and f(X) + Y are different. We conclude that no q-deformed convolution product can exist for functions of independent q-Gaussian random variables.
Contour Detection Using Cost-Sensitive Convolutional Neural Networks
Hwang, Jyh-Jing; Liu, Tyng-Luh
2014-01-01
We address the problem of contour detection via per-pixel classifications of edge point. To facilitate the process, the proposed approach leverages with DenseNet, an efficient implementation of multiscale convolutional neural networks (CNNs), to extract an informative feature vector for each pixel and uses an SVM classifier to accomplish contour detection. The main challenge lies in adapting a pre-trained per-image CNN model for yielding per-pixel image features. We propose to base on the Den...
Image interpolation by two-dimensional parametric cubic convolution.
Shi, Jiazheng; Reichenbach, Stephen E
2006-07-01
Cubic convolution is a popular method for image interpolation. Traditionally, the piecewise-cubic kernel has been derived in one dimension with one parameter and applied to two-dimensional (2-D) images in a separable fashion. However, images typically are statistically nonseparable, which motivates this investigation of nonseparable cubic convolution. This paper derives two new nonseparable, 2-D cubic-convolution kernels. The first kernel, with three parameters (designated 2D-3PCC), is the most general 2-D, piecewise-cubic interpolator defined on [-2, 2] x [-2, 2] with constraints for biaxial symmetry, diagonal (or 90 degrees rotational) symmetry, continuity, and smoothness. The second kernel, with five parameters (designated 2D-5PCC), relaxes the constraint of diagonal symmetry, based on the observation that many images have rotationally asymmetric statistical properties. This paper also develops a closed-form solution for determining the optimal parameter values for parametric cubic-convolution kernels with respect to ensembles of scenes characterized by autocorrelation (or power spectrum). This solution establishes a practical foundation for adaptive interpolation based on local autocorrelation estimates. Quantitative fidelity analyses and visual experiments indicate that these new methods can outperform several popular interpolation methods. An analysis of the error budgets for reconstruction error associated with blurring and aliasing illustrates that the methods improve interpolation fidelity for images with aliased components. For images with little or no aliasing, the methods yield results similar to other popular methods. Both 2D-3PCC and 2D-5PCC are low-order polynomials with small spatial support and so are easy to implement and efficient to apply.
GPU Acceleration of Image Convolution using Spatially-varying Kernel
Hartung, Steven; Shukla, Hemant; Miller, J. Patrick; Pennypacker, Carlton
2012-01-01
Image subtraction in astronomy is a tool for transient object discovery such as asteroids, extra-solar planets and supernovae. To match point spread functions (PSFs) between images of the same field taken at different times a convolution technique is used. Particularly suitable for large-scale images is a computationally intensive spatially-varying kernel. The underlying algorithm is inherently massively parallel due to unique kernel generation at every pixel location. The spatially-varying k...
Nuclear norm regularized convolutional Max Pos@Top machine
Li, Qinfeng
2016-11-18
In this paper, we propose a novel classification model for the multiple instance data, which aims to maximize the number of positive instances ranked before the top-ranked negative instances. This method belongs to a recently emerged performance, named as Pos@Top. Our proposed classification model has a convolutional structure that is composed by four layers, i.e., the convolutional layer, the activation layer, the max-pooling layer and the full connection layer. In this paper, we propose an algorithm to learn the convolutional filters and the full connection weights to maximize the Pos@Top measure over the training set. Also, we try to minimize the rank of the filter matrix to explore the low-dimensional space of the instances in conjunction with the classification results. The rank minimization is conducted by the nuclear norm minimization of the filter matrix. In addition, we develop an iterative algorithm to solve the corresponding problem. We test our method on several benchmark datasets. The experimental results show the superiority of our method compared with other state-of-the-art Pos@Top maximization methods.
Fine-grained representation learning in convolutional autoencoders
Luo, Chang; Wang, Jie
2016-03-01
Convolutional autoencoders (CAEs) have been widely used as unsupervised feature extractors for high-resolution images. As a key component in CAEs, pooling is a biologically inspired operation to achieve scale and shift invariances, and the pooled representation directly affects the CAEs' performance. Fine-grained pooling, which uses small and dense pooling regions, encodes fine-grained visual cues and enhances local characteristics. However, it tends to be sensitive to spatial rearrangements. In most previous works, pooled features were obtained by empirically modulating parameters in CAEs. We see the CAE as a whole and propose a fine-grained representation learning law to extract better fine-grained features. This representation learning law suggests two directions for improvement. First, we probabilistically evaluate the discrimination-invariance tradeoff with fine-grained granularity in the pooled feature maps, and suggest the proper filter scale in the convolutional layer and appropriate whitening parameters in preprocessing step. Second, pooling approaches are combined with the sparsity degree in pooling regions, and we propose the preferable pooling approach. Experimental results on two independent benchmark datasets demonstrate that our representation learning law could guide CAEs to extract better fine-grained features and performs better in multiclass classification task. This paper also provides guidance for selecting appropriate parameters to obtain better fine-grained representation in other convolutional neural networks.
Deep Convolutional Neural Network for Inverse Problems in Imaging
Jin, Kyong Hwan; McCann, Michael T.; Froustey, Emmanuel; Unser, Michael
2017-09-01
In this paper, we propose a novel deep convolutional neural network (CNN)-based algorithm for solving ill-posed inverse problems. Regularized iterative algorithms have emerged as the standard approach to ill-posed inverse problems in the past few decades. These methods produce excellent results, but can be challenging to deploy in practice due to factors including the high computational cost of the forward and adjoint operators and the difficulty of hyper parameter selection. The starting point of our work is the observation that unrolled iterative methods have the form of a CNN (filtering followed by point-wise non-linearity) when the normal operator (H*H, the adjoint of H times H) of the forward model is a convolution. Based on this observation, we propose using direct inversion followed by a CNN to solve normal-convolutional inverse problems. The direct inversion encapsulates the physical model of the system, but leads to artifacts when the problem is ill-posed; the CNN combines multiresolution decomposition and residual learning in order to learn to remove these artifacts while preserving image structure. We demonstrate the performance of the proposed network in sparse-view reconstruction (down to 50 views) on parallel beam X-ray computed tomography in synthetic phantoms as well as in real experimental sinograms. The proposed network outperforms total variation-regularized iterative reconstruction for the more realistic phantoms and requires less than a second to reconstruct a 512 x 512 image on GPU.
A model of traffic signs recognition with convolutional neural network
Hu, Haihe; Li, Yujian; Zhang, Ting; Huo, Yi; Kuang, Wenqing
2016-10-01
In real traffic scenes, the quality of captured images are generally low due to some factors such as lighting conditions, and occlusion on. All of these factors are challengeable for automated recognition algorithms of traffic signs. Deep learning has provided a new way to solve this kind of problems recently. The deep network can automatically learn features from a large number of data samples and obtain an excellent recognition performance. We therefore approach this task of recognition of traffic signs as a general vision problem, with few assumptions related to road signs. We propose a model of Convolutional Neural Network (CNN) and apply the model to the task of traffic signs recognition. The proposed model adopts deep CNN as the supervised learning model, directly takes the collected traffic signs image as the input, alternates the convolutional layer and subsampling layer, and automatically extracts the features for the recognition of the traffic signs images. The proposed model includes an input layer, three convolutional layers, three subsampling layers, a fully-connected layer, and an output layer. To validate the proposed model, the experiments are implemented using the public dataset of China competition of fuzzy image processing. Experimental results show that the proposed model produces a recognition accuracy of 99.01 % on the training dataset, and yield a record of 92% on the preliminary contest within the fourth best.
Super-resolution reconstruction algorithm based on adaptive convolution kernel size selection
Gao, Hang; Chen, Qian; Sui, Xiubao; Zeng, Junjie; Zhao, Yao
2016-09-01
Restricted by the detector technology and optical diffraction limit, the spatial resolution of infrared imaging system is difficult to achieve significant improvement. Super-Resolution (SR) reconstruction algorithm is an effective way to solve this problem. Among them, the SR algorithm based on multichannel blind deconvolution (MBD) estimates the convolution kernel only by low resolution observation images, according to the appropriate regularization constraints introduced by a priori assumption, to realize the high resolution image restoration. The algorithm has been shown effective when each channel is prime. In this paper, we use the significant edges to estimate the convolution kernel and introduce an adaptive convolution kernel size selection mechanism, according to the uncertainty of the convolution kernel size in MBD processing. To reduce the interference of noise, we amend the convolution kernel in an iterative process, and finally restore a clear image. Experimental results show that the algorithm can meet the convergence requirement of the convolution kernel estimation.
The double Mellin-Barnes type integrals and their applications to convolution theory
Hai, Nguyen Thanh
1992-01-01
This book presents new results in the theory of the double Mellin-Barnes integrals popularly known as the general H-function of two variables.A general integral convolution is constructed by the authors and it contains Laplace convolution as a particular case and possesses a factorization property for one-dimensional H-transform. Many examples of convolutions for classical integral transforms are obtained and they can be applied for the evaluation of series and integrals.
An area-efficient 2-D convolution implementation on FPGA for space applications
Gambardella, Giulio; Tiotto, Gabriele; Prinetto, Paolo Ernesto; Di Carlo, Stefano; Indaco, Marco; Rolfo, Daniele
2011-01-01
The 2-D Convolution is an algorithm widely used in image and video processing. Although its computation is simple, its implementation requires a high computational power and an intensive use of memory. Field Programmable Gate Arrays (FPGA) architectures were proposed to accelerate calculations of 2-D Convolution and the use of buffers implemented on FPGAs are used to avoid direct memory access. In this paper we present an implementation of the 2-D Convolution algorithm on a FPGA architecture ...
IMAGE DE-BLURRING USING WIENER DE-CONVOLUTION AND WAVELET FOR DIFFERENT BLURRING KERNEL
M.Tech Research Scholar Shuchi Singh*, Asst Professor Vipul Awasthi, Asst Professor NitinSahu
2016-01-01
Image de-convolution is an active research area of recovering a sharp image after blurring by a convolution. One of the problems in image de-convolution is how to preserve the texture structures while removing blur in presence of noise. Various methods have been used for such as gradient based methods, sparsity based methods, and nonlocal self-similarity methods. In this thesis, we have used the conventional non-blind method of Wiener de-convolution. Further Wavelet denoising has been used to...
Finding strong lenses in CFHTLS using convolutional neural networks
Jacobs, C.; Glazebrook, K.; Collett, T.; More, A.; McCarthy, C.
2017-10-01
We train and apply convolutional neural networks, a machine learning technique developed to learn from and classify image data, to Canada-France-Hawaii Telescope Legacy Survey (CFHTLS) imaging for the identification of potential strong lensing systems. An ensemble of four convolutional neural networks was trained on images of simulated galaxy-galaxy lenses. The training sets consisted of a total of 62 406 simulated lenses and 64 673 non-lens negative examples generated with two different methodologies. An ensemble of trained networks was applied to all of the 171 deg2 of the CFHTLS wide field image data, identifying 18 861 candidates including 63 known and 139 other potential lens candidates. A second search of 1.4 million early-type galaxies selected from the survey catalogue as potential deflectors, identified 2465 candidates including 117 previously known lens candidates, 29 confirmed lenses/high-quality lens candidates, 266 novel probable or potential lenses and 2097 candidates we classify as false positives. For the catalogue-based search we estimate a completeness of 21-28 per cent with respect to detectable lenses and a purity of 15 per cent, with a false-positive rate of 1 in 671 images tested. We predict a human astronomer reviewing candidates produced by the system would identify 20 probable lenses and 100 possible lenses per hour in a sample selected by the robot. Convolutional neural networks are therefore a promising tool for use in the search for lenses in current and forthcoming surveys such as the Dark Energy Survey and the Large Synoptic Survey Telescope.
Learning Contextual Dependence With Convolutional Hierarchical Recurrent Neural Networks
Zuo, Zhen; Shuai, Bing; Wang, Gang; Liu, Xiao; Wang, Xingxing; Wang, Bing; Chen, Yushi
2016-07-01
Existing deep convolutional neural networks (CNNs) have shown their great success on image classification. CNNs mainly consist of convolutional and pooling layers, both of which are performed on local image areas without considering the dependencies among different image regions. However, such dependencies are very important for generating explicit image representation. In contrast, recurrent neural networks (RNNs) are well known for their ability of encoding contextual information among sequential data, and they only require a limited number of network parameters. General RNNs can hardly be directly applied on non-sequential data. Thus, we proposed the hierarchical RNNs (HRNNs). In HRNNs, each RNN layer focuses on modeling spatial dependencies among image regions from the same scale but different locations. While the cross RNN scale connections target on modeling scale dependencies among regions from the same location but different scales. Specifically, we propose two recurrent neural network models: 1) hierarchical simple recurrent network (HSRN), which is fast and has low computational cost; and 2) hierarchical long-short term memory recurrent network (HLSTM), which performs better than HSRN with the price of more computational cost. In this manuscript, we integrate CNNs with HRNNs, and develop end-to-end convolutional hierarchical recurrent neural networks (C-HRNNs). C-HRNNs not only make use of the representation power of CNNs, but also efficiently encodes spatial and scale dependencies among different image regions. On four of the most challenging object/scene image classification benchmarks, our C-HRNNs achieve state-of-the-art results on Places 205, SUN 397, MIT indoor, and competitive results on ILSVRC 2012.
Polynomial theory of error correcting codes
Cancellieri, Giovanni
2015-01-01
The book offers an original view on channel coding, based on a unitary approach to block and convolutional codes for error correction. It presents both new concepts and new families of codes. For example, lengthened and modified lengthened cyclic codes are introduced as a bridge towards time-invariant convolutional codes and their extension to time-varying versions. The novel families of codes include turbo codes and low-density parity check (LDPC) codes, the features of which are justified from the structural properties of the component codes. Design procedures for regular LDPC codes are proposed, supported by the presented theory. Quasi-cyclic LDPC codes, in block or convolutional form, represent one of the most original contributions of the book. The use of more than 100 examples allows the reader gradually to gain an understanding of the theory, and the provision of a list of more than 150 definitions, indexed at the end of the book, permits rapid location of sought information.
Continuous speech recognition based on convolutional neural network
Zhang, Qing-qing; Liu, Yong; Pan, Jie-lin; Yan, Yong-hong
2015-07-01
Convolutional Neural Networks (CNNs), which showed success in achieving translation invariance for many image processing tasks, are investigated for continuous speech recognitions in the paper. Compared to Deep Neural Networks (DNNs), which have been proven to be successful in many speech recognition tasks nowadays, CNNs can reduce the NN model sizes significantly, and at the same time achieve even better recognition accuracies. Experiments on standard speech corpus TIMIT showed that CNNs outperformed DNNs in the term of the accuracy when CNNs had even smaller model size.
Training Convolutional Neural Networks for Translational Invariance on SAR ATR
DEFF Research Database (Denmark)
Malmgren-Hansen, David; Engholm, Rasmus; Østergaard Pedersen, Morten
2016-01-01
In this paper we present a comparison of the robustness of Convolutional Neural Networks (CNN) to other classifiers in the presence of uncertainty of the objects localization in SAR image. We present a framework for simulating simple SAR images, translating the object of interest systematically...... and testing the classification performance. Our results show that where other classification methods are very sensitive to even small translations, CNN is quite robust to translational variance, making it much more useful in relation to Automatic Target Recognition (ATR) in a real life context....
Convolution Algebra for Fluid Modes with Finite Energy
1992-04-01
PHILLIPS LABORATORY AIR FORCE SYSTEMS COMMAND UNITED STATES AIR FORCE HANSCOM AIR FORCE BASE, MASSACHIUSETTS 01731-5000 94-22604 "This technical report ’-as...with finite spatial and temporal extents. At Boston University, we have developed a full form of wavelet expansion which has the advantage over more...distribution: 00 bX =00 0l if, TZ< VPf (X) = V •a,,,’(x) = E bnb (x) where b, =otherwise (34) V=o ,i=o a._, otherwise 7 The convolution of two
Plant species classification using deep convolutional neural network
DEFF Research Database (Denmark)
Dyrmann, Mads; Karstoft, Henrik; Midtiby, Henrik Skov
2016-01-01
Information on which weed species are present within agricultural fields is important for site specific weed management. This paper presents a method that is capable of recognising plant species in colour images by using a convolutional neural network. The network is built from scratch trained...... stabilisation and illumination, and images shot with hand-held mobile phones in fields with changing lighting conditions and different soil types. For these 22 species, the network is able to achieve a classification accuracy of 86.2%....
Convolutional neural networks for synthetic aperture radar classification
Profeta, Andrew; Rodriguez, Andres; Clouse, H. Scott
2016-05-01
For electro-optical object recognition, convolutional neural networks (CNNs) are the state-of-the-art. For large datasets, CNNs are able to learn meaningful features used for classification. However, their application to synthetic aperture radar (SAR) has been limited. In this work we experimented with various CNN architectures on the MSTAR SAR dataset. As the input to the CNN we used the magnitude and phase (2 channels) of the SAR imagery. We used the deep learning toolboxes CAFFE and Torch7. Our results show that we can achieve 93% accuracy on the MSTAR dataset using CNNs.
A Fortran 90 code for magnetohydrodynamics. Part 1, Banded convolution
Energy Technology Data Exchange (ETDEWEB)
Walker, D.W.
1992-03-01
This report describes progress in developing a Fortran 90 version of the KITE code for studying plasma instabilities in Tokamaks. In particular, the evaluation of convolution terms appearing in the numerical solution is discussed, and timing results are presented for runs performed on an 8k processor Connection Machine (CM-2). Estimates of the performance on a full-size 64k CM-2 are given, and range between 100 and 200 Mflops. The advantages of having a Fortran 90 version of the KITE code are stressed, and the future use of such a code on the newly announced CM5 and Paragon computers, from Thinking Machines Corporation and Intel, is considered.
Image reconstruction of simulated specimens using convolution back projection
Directory of Open Access Journals (Sweden)
Mohd. Farhan Manzoor
2001-04-01
Full Text Available This paper reports about the reconstruction of cross-sections of composite structures. The convolution back projection (CBP algorithm has been used to capture the attenuation field over the specimen. Five different test cases have been taken up for evaluation. These cases represent varying degrees of complexity. In addition, the role of filters on the nature of the reconstruction errors has also been discussed. Numerical results obtained in the study reveal that CBP algorithm is a useful tool for qualitative as well as quantitative assessment of composite regions encountered in engineering applications.
Faster GPU-based convolutional gridding via thread coarsening
Merry, Bruce
2016-01-01
Convolutional gridding is a processor-intensive step in interferometric imaging. While it is possible to use graphics processing units (GPUs) to accelerate this operation, existing methods use only a fraction of the available flops. We apply thread coarsening to improve the efficiency of an existing algorithm, and observe performance gains of up to $3.2\\times$ for single-polarization gridding and $1.9\\times$ for quad-polarization gridding on a GeForce GTX 980, and smaller but still significant gains on a Radeon R9 290X.
Convolution seal for transition duct in turbine system
Energy Technology Data Exchange (ETDEWEB)
Flanagan, James Scott; LeBegue, Jeffrey Scott; McMahan, Kevin Weston; Dillard, Daniel Jackson; Pentecost, Ronnie Ray
2015-03-10
A turbine system is disclosed. In one embodiment, the turbine system includes a transition duct. The transition duct includes an inlet, an outlet, and a passage extending between the inlet and the outlet and defining a longitudinal axis, a radial axis, and a tangential axis. The outlet of the transition duct is offset from the inlet along the longitudinal axis and the tangential axis. The transition duct further includes an interface member for interfacing with a turbine section. The turbine system further includes a convolution seal contacting the interface member to provide a seal between the interface member and the turbine section.
Convolution seal for transition duct in turbine system
Energy Technology Data Exchange (ETDEWEB)
Flanagan, James Scott; LeBegue, Jeffrey Scott; McMahan, Kevin Weston; Dillard, Daniel Jackson; Pentecost, Ronnie Ray
2015-05-26
A turbine system is disclosed. In one embodiment, the turbine system includes a transition duct. The transition duct includes an inlet, an outlet, and a passage extending between the inlet and the outlet and defining a longitudinal axis, a radial axis, and a tangential axis. The outlet of the transition duct is offset from the inlet along the longitudinal axis and the tangential axis. The transition duct further includes an interface feature for interfacing with an adjacent transition duct. The turbine system further includes a convolution seal contacting the interface feature to provide a seal between the interface feature and the adjacent transition duct.
Tandem mass spectrometry data quality assessment by self-convolution
Directory of Open Access Journals (Sweden)
Tham Wai
2007-09-01
Full Text Available Abstract Background Many algorithms have been developed for deciphering the tandem mass spectrometry (MS data sets. They can be essentially clustered into two classes. The first performs searches on theoretical mass spectrum database, while the second based itself on de novo sequencing from raw mass spectrometry data. It was noted that the quality of mass spectra affects significantly the protein identification processes in both instances. This prompted the authors to explore ways to measure the quality of MS data sets before subjecting them to the protein identification algorithms, thus allowing for more meaningful searches and increased confidence level of proteins identified. Results The proposed method measures the qualities of MS data sets based on the symmetric property of b- and y-ion peaks present in a MS spectrum. Self-convolution on MS data and its time-reversal copy was employed. Due to the symmetric nature of b-ions and y-ions peaks, the self-convolution result of a good spectrum would produce a highest mid point intensity peak. To reduce processing time, self-convolution was achieved using Fast Fourier Transform and its inverse transform, followed by the removal of the "DC" (Direct Current component and the normalisation of the data set. The quality score was defined as the ratio of the intensity at the mid point to the remaining peaks of the convolution result. The method was validated using both theoretical mass spectra, with various permutations, and several real MS data sets. The results were encouraging, revealing a high percentage of positive prediction rates for spectra with good quality scores. Conclusion We have demonstrated in this work a method for determining the quality of tandem MS data set. By pre-determining the quality of tandem MS data before subjecting them to protein identification algorithms, spurious protein predictions due to poor tandem MS data are avoided, giving scientists greater confidence in the
Visualizing Vector Fields Using Line Integral Convolution and Dye Advection
Shen, Han-Wei; Johnson, Christopher R.; Ma, Kwan-Liu
1996-01-01
We present local and global techniques to visualize three-dimensional vector field data. Using the Line Integral Convolution (LIC) method to image the global vector field, our new algorithm allows the user to introduce colored 'dye' into the vector field to highlight local flow features. A fast algorithm is proposed that quickly recomputes the dyed LIC images. In addition, we introduce volume rendering methods that can map the LIC texture on any contour surface and/or translucent region defined by additional scalar quantities, and can follow the advection of colored dye throughout the volume.
Medical image fusion using the convolution of Meridian distributions.
Agrawal, Mayank; Tsakalides, Panagiotis; Achim, Alin
2010-01-01
The aim of this paper is to introduce a novel non-Gaussian statistical model-based approach for medical image fusion based on the Meridian distribution. The paper also includes a new approach to estimate the parameters of generalized Cauchy distribution. The input images are first decomposed using the Dual-Tree Complex Wavelet Transform (DT-CWT) with the subband coefficients modelled as Meridian random variables. Then, the convolution of Meridian distributions is applied as a probabilistic prior to model the fused coefficients, and the weights used to combine the source images are optimised via Maximum Likelihood (ML) estimation. The superior performance of the proposed method is demonstrated using medical images.
Faster GPU-based convolutional gridding via thread coarsening
Merry, B.
2016-07-01
Convolutional gridding is a processor-intensive step in interferometric imaging. While it is possible to use graphics processing units (GPUs) to accelerate this operation, existing methods use only a fraction of the available flops. We apply thread coarsening to improve the efficiency of an existing algorithm, and observe performance gains of up to 3.2 × for single-polarization gridding and 1.9 × for quad-polarization gridding on a GeForce GTX 980, and smaller but still significant gains on a Radeon R9 290X.
Low-dose CT denoising with convolutional neural network
Chen, Hu; Zhang, Weihua; Liao, Peixi; Li, Ke; Zhou, Jiliu; Wang, Ge
2016-01-01
To reduce the potential radiation risk, low-dose CT has attracted much attention. However, simply lowering the radiation dose will lead to significant deterioration of the image quality. In this paper, we propose a noise reduction method for low-dose CT via deep neural network without accessing original projection data. A deep convolutional neural network is trained to transform low-dose CT images towards normal-dose CT images, patch by patch. Visual and quantitative evaluation demonstrates a competing performance of the proposed method.
Digital image correlation based on a fast convolution strategy
Yuan, Yuan; Zhan, Qin; Xiong, Chunyang; Huang, Jianyong
2017-10-01
In recent years, the efficiency of digital image correlation (DIC) methods has attracted increasing attention because of its increasing importance for many engineering applications. Based on the classical affine optical flow (AOF) algorithm and the well-established inverse compositional Gauss-Newton algorithm, which is essentially a natural extension of the AOF algorithm under a nonlinear iterative framework, this paper develops a set of fast convolution-based DIC algorithms for high-efficiency subpixel image registration. Using a well-developed fast convolution technique, the set of algorithms establishes a series of global data tables (GDTs) over the digital images, which allows the reduction of the computational complexity of DIC significantly. Using the pre-calculated GDTs, the subpixel registration calculations can be implemented efficiently in a look-up-table fashion. Both numerical simulation and experimental verification indicate that the set of algorithms significantly enhances the computational efficiency of DIC, especially in the case of a dense data sampling for the digital images. Because the GDTs need to be computed only once, the algorithms are also suitable for efficiently coping with image sequences that record the time-varying dynamics of specimen deformations.
Convolutional neural network architectures for predicting DNA–protein binding
Zeng, Haoyang; Edwards, Matthew D.; Liu, Ge; Gifford, David K.
2016-01-01
Motivation: Convolutional neural networks (CNN) have outperformed conventional methods in modeling the sequence specificity of DNA–protein binding. Yet inappropriate CNN architectures can yield poorer performance than simpler models. Thus an in-depth understanding of how to match CNN architecture to a given task is needed to fully harness the power of CNNs for computational biology applications. Results: We present a systematic exploration of CNN architectures for predicting DNA sequence binding using a large compendium of transcription factor datasets. We identify the best-performing architectures by varying CNN width, depth and pooling designs. We find that adding convolutional kernels to a network is important for motif-based tasks. We show the benefits of CNNs in learning rich higher-order sequence features, such as secondary motifs and local sequence context, by comparing network performance on multiple modeling tasks ranging in difficulty. We also demonstrate how careful construction of sequence benchmark datasets, using approaches that control potentially confounding effects like positional or motif strength bias, is critical in making fair comparisons between competing methods. We explore how to establish the sufficiency of training data for these learning tasks, and we have created a flexible cloud-based framework that permits the rapid exploration of alternative neural network architectures for problems in computational biology. Availability and Implementation: All the models analyzed are available at http://cnn.csail.mit.edu. Contact: gifford@mit.edu Supplementary information: Supplementary data are available at Bioinformatics online. PMID:27307608
Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks.
Dosovitskiy, Alexey; Fischer, Philipp; Springenberg, Jost Tobias; Riedmiller, Martin; Brox, Thomas
2016-09-01
Deep convolutional networks have proven to be very successful in learning task specific features that allow for unprecedented performance on various computer vision tasks. Training of such networks follows mostly the supervised learning paradigm, where sufficiently many input-output pairs are required for training. Acquisition of large training sets is one of the key challenges, when approaching a new task. In this paper, we aim for generic feature learning and present an approach for training a convolutional network using only unlabeled data. To this end, we train the network to discriminate between a set of surrogate classes. Each surrogate class is formed by applying a variety of transformations to a randomly sampled 'seed' image patch. In contrast to supervised network training, the resulting feature representation is not class specific. It rather provides robustness to the transformations that have been applied during training. This generic feature representation allows for classification results that outperform the state of the art for unsupervised learning on several popular datasets (STL-10, CIFAR-10, Caltech-101, Caltech-256). While features learned with our approach cannot compete with class specific features from supervised training on a classification task, we show that they are advantageous on geometric matching problems, where they also outperform the SIFT descriptor.
Multichannel Convolutional Neural Network for Biological Relation Extraction
Quan, Chanqin; Sun, Xiao; Bai, Wenjun
2016-01-01
The plethora of biomedical relations which are embedded in medical logs (records) demands researchers' attention. Previous theoretical and practical focuses were restricted on traditional machine learning techniques. However, these methods are susceptible to the issues of “vocabulary gap” and data sparseness and the unattainable automation process in feature extraction. To address aforementioned issues, in this work, we propose a multichannel convolutional neural network (MCCNN) for automated biomedical relation extraction. The proposed model has the following two contributions: (1) it enables the fusion of multiple (e.g., five) versions in word embeddings; (2) the need for manual feature engineering can be obviated by automated feature learning with convolutional neural network (CNN). We evaluated our model on two biomedical relation extraction tasks: drug-drug interaction (DDI) extraction and protein-protein interaction (PPI) extraction. For DDI task, our system achieved an overall f-score of 70.2% compared to the standard linear SVM based system (e.g., 67.0%) on DDIExtraction 2013 challenge dataset. And for PPI task, we evaluated our system on Aimed and BioInfer PPI corpus; our system exceeded the state-of-art ensemble SVM system by 2.7% and 5.6% on f-scores. PMID:28053977
Robust Visual Tracking via Convolutional Networks Without Training.
Kaihua Zhang; Qingshan Liu; Yi Wu; Ming-Hsuan Yang
2016-04-01
Deep networks have been successfully applied to visual tracking by learning a generic representation offline from numerous training images. However, the offline training is time-consuming and the learned generic representation may be less discriminative for tracking specific objects. In this paper, we present that, even without offline training with a large amount of auxiliary data, simple two-layer convolutional networks can be powerful enough to learn robust representations for visual tracking. In the first frame, we extract a set of normalized patches from the target region as fixed filters, which integrate a series of adaptive contextual filters surrounding the target to define a set of feature maps in the subsequent frames. These maps measure similarities between each filter and useful local intensity patterns across the target, thereby encoding its local structural information. Furthermore, all the maps together form a global representation, via which the inner geometric layout of the target is also preserved. A simple soft shrinkage method that suppresses noisy values below an adaptive threshold is employed to de-noise the global representation. Our convolutional networks have a lightweight structure and perform favorably against several state-of-the-art methods on the recent tracking benchmark data set with 50 challenging videos.
Convolutional neural network features based change detection in satellite images
Mohammed El Amin, Arabi; Liu, Qingjie; Wang, Yunhong
2016-07-01
With the popular use of high resolution remote sensing (HRRS) satellite images, a huge research efforts have been placed on change detection (CD) problem. An effective feature selection method can significantly boost the final result. While hand-designed features have proven difficulties to design features that effectively capture high and mid-level representations, the recent developments in machine learning (Deep Learning) omit this problem by learning hierarchical representation in an unsupervised manner directly from data without human intervention. In this letter, we propose approaching the change detection problem from a feature learning perspective. A novel deep Convolutional Neural Networks (CNN) features based HR satellite images change detection method is proposed. The main guideline is to produce a change detection map directly from two images using a pretrained CNN. This method can omit the limited performance of hand-crafted features. Firstly, CNN features are extracted through different convolutional layers. Then, a concatenation step is evaluated after an normalization step, resulting in a unique higher dimensional feature map. Finally, a change map was computed using pixel-wise Euclidean distance. Our method has been validated on real bitemporal HRRS satellite images according to qualitative and quantitative analyses. The results obtained confirm the interest of the proposed method.
Convolution approach to the piNN system
Blankleider, B
1994-01-01
The unitary NN-piNN model contains a serious theoretical flaw: unitarity is obtained at the price of having to use an effective piNN coupling constant that is smaller than the experimental one. This is but one aspect of a more general renormalization problem whose origin lies in the truncation of Hilbert space used to derive the equations. Here we present a new theoretical approach to the piNN problem where unitary equations are obtained without having to truncate Hilbert space. Indeed, the only approximation made is the neglect of connected three-body forces. As all possible dressings of one-particle propagators and vertices are retained in our model, we overcome the renormalization problems inherent in previous piNN theories. The key element of our derivation is the use of convolution integrals that have enabled us to sum all the possible disconnected time-ordered graphs. We also discuss how the convolution method can be extended to sum all the time orderings of a connected graph. This has enabled us to cal...
Real-Time Video Convolutional Face Finder on Embedded Platforms
Directory of Open Access Journals (Sweden)
Franck Mamalet
2007-03-01
Full Text Available A high-level optimization methodology is applied for implementing the well-known convolutional face finder (CFF algorithm for real-time applications on mobile phones, such as teleconferencing, advanced user interfaces, image indexing, and security access control. CFF is based on a feature extraction and classification technique which consists of a pipeline of convolutions and subsampling operations. The design of embedded systems requires a good trade-off between performance and code size due to the limited amount of available resources. The followed methodology copes with the main drawbacks of the original implementation of CFF such as floating-point computation and memory allocation, in order to allow parallelism exploitation and perform algorithm optimizations. Experimental results show that our embedded face detection system can accurately locate faces with less computational load and memory cost. It runs on a 275 MHz Starcore DSP at 35 QCIF images/s with state-of-the-art detection rates and very low false alarm rates.
A Mathematical Motivation for Complex-Valued Convolutional Networks.
Tygert, Mark; Bruna, Joan; Chintala, Soumith; LeCun, Yann; Piantino, Serkan; Szlam, Arthur
2016-05-01
A complex-valued convolutional network (convnet) implements the repeated application of the following composition of three operations, recursively applying the composition to an input vector of nonnegative real numbers: (1) convolution with complex-valued vectors, followed by (2) taking the absolute value of every entry of the resulting vectors, followed by (3) local averaging. For processing real-valued random vectors, complex-valued convnets can be viewed as data-driven multiscale windowed power spectra, data-driven multiscale windowed absolute spectra, data-driven multiwavelet absolute values, or (in their most general configuration) data-driven nonlinear multiwavelet packets. Indeed, complex-valued convnets can calculate multiscale windowed spectra when the convnet filters are windowed complex-valued exponentials. Standard real-valued convnets, using rectified linear units (ReLUs), sigmoidal (e.g., logistic or tanh) nonlinearities, or max pooling, for example, do not obviously exhibit the same exact correspondence with data-driven wavelets (whereas for complex-valued convnets, the correspondence is much more than just a vague analogy). Courtesy of the exact correspondence, the remarkably rich and rigorous body of mathematical analysis for wavelets applies directly to (complex-valued) convnets.
Convolutional Network Coding Based on Matrix Power Series Representation
Guo, Wangmei; Sun, Qifu Tyler
2011-01-01
In this paper, convolutional network coding is formulated by means of matrix power series representation of the local encoding kernel (LEK) matrices and global encoding kernel (GEK) matrices to establish its theoretical fundamentals for practical implementations. From the encoding perspective, the GEKs of a convolutional network code (CNC) are shown to be uniquely determined by its LEK matrix $K(z)$ if $K_0$, the constant coefficient matrix of $K(z)$, is nilpotent. This will simplify the CNC design because a nilpotent $K_0$ suffices to guarantee a unique set of GEKs. Besides, the relation between coding topology and $K(z)$ is also discussed. From the decoding perspective, the main theme is to justify that the first $L+1$ terms of the GEK matrix $F(z)$ at a sink $r$ suffice to check whether the code is decodable at $r$ with delay $L$ and to start decoding if so. The concomitant decoding scheme avoids dealing with $F(z)$, which may contain infinite terms, as a whole and hence reduces the complexity of decodabil...
Classification of Histology Sections via Multispectral Convolutional Sparse Coding.
Zhou, Yin; Chang, Hang; Barner, Kenneth; Spellman, Paul; Parvin, Bahram
2014-06-01
Image-based classification of histology sections plays an important role in predicting clinical outcomes. However this task is very challenging due to the presence of large technical variations (e.g., fixation, staining) and biological heterogeneities (e.g., cell type, cell state). In the field of biomedical imaging, for the purposes of visualization and/or quantification, different stains are typically used for different targets of interest (e.g., cellular/subcellular events), which generates multi-spectrum data (images) through various types of microscopes and, as a result, provides the possibility of learning biological-component-specific features by exploiting multispectral information. We propose a multispectral feature learning model that automatically learns a set of convolution filter banks from separate spectra to efficiently discover the intrinsic tissue morphometric signatures, based on convolutional sparse coding (CSC). The learned feature representations are then aggregated through the spatial pyramid matching framework (SPM) and finally classified using a linear SVM. The proposed system has been evaluated using two large-scale tumor cohorts, collected from The Cancer Genome Atlas (TCGA). Experimental results show that the proposed model 1) outperforms systems utilizing sparse coding for unsupervised feature learning (e.g., PSD-SPM [5]); 2) is competitive with systems built upon features with biological prior knowledge (e.g., SMLSPM [4]).
Convolutional Neural Network Based Fault Detection for Rotating Machinery
Janssens, Olivier; Slavkovikj, Viktor; Vervisch, Bram; Stockman, Kurt; Loccufier, Mia; Verstockt, Steven; Van de Walle, Rik; Van Hoecke, Sofie
2016-09-01
Vibration analysis is a well-established technique for condition monitoring of rotating machines as the vibration patterns differ depending on the fault or machine condition. Currently, mainly manually-engineered features, such as the ball pass frequencies of the raceway, RMS, kurtosis an crest, are used for automatic fault detection. Unfortunately, engineering and interpreting such features requires a significant level of human expertise. To enable non-experts in vibration analysis to perform condition monitoring, the overhead of feature engineering for specific faults needs to be reduced as much as possible. Therefore, in this article we propose a feature learning model for condition monitoring based on convolutional neural networks. The goal of this approach is to autonomously learn useful features for bearing fault detection from the data itself. Several types of bearing faults such as outer-raceway faults and lubrication degradation are considered, but also healthy bearings and rotor imbalance are included. For each condition, several bearings are tested to ensure generalization of the fault-detection system. Furthermore, the feature-learning based approach is compared to a feature-engineering based approach using the same data to objectively quantify their performance. The results indicate that the feature-learning system, based on convolutional neural networks, significantly outperforms the classical feature-engineering based approach which uses manually engineered features and a random forest classifier. The former achieves an accuracy of 93.61 percent and the latter an accuracy of 87.25 percent.
Asymptotic formula for the moments of Bernoulli convolutions
Directory of Open Access Journals (Sweden)
E. A. Timofeev
2016-01-01
Full Text Available Abstract. Asymptotic Formula for the Moments of Bernoulli Convolutions Timofeev E. A. Received February 8, 2016 For each λ, 0 < λ < 1, we define a random variable ∞ Yλ =(1−λξnλn, n=0 where ξn are independent random variables with P{ξn =0}=P{ξn =1}= 1. 2 The distribution of Yλ is called a symmetric Bernoulli convolution. The main result of this paper is Mn =EYλn =nlogλ22logλ(1−λ+0.5logλ2−0.5eτ(−logλn1+O(n−0.99, where is a 1-periodic function, 1k2πikx τ(x= kα −lnλ e k̸=0 1 (1 − λ2πit(1 − 22πitπ−2πit2−2πitζ(2πit, 2i sh(π2t α(t = − and ζ(z is the Riemann zeta function. The article is published in the author’s wording.
Transforming Musical Signals through a Genre Classifying Convolutional Neural Network
Geng, S.; Ren, G.; Ogihara, M.
2017-05-01
Convolutional neural networks (CNNs) have been successfully applied on both discriminative and generative modeling for music-related tasks. For a particular task, the trained CNN contains information representing the decision making or the abstracting process. One can hope to manipulate existing music based on this 'informed' network and create music with new features corresponding to the knowledge obtained by the network. In this paper, we propose a method to utilize the stored information from a CNN trained on musical genre classification task. The network was composed of three convolutional layers, and was trained to classify five-second song clips into five different genres. After training, randomly selected clips were modified by maximizing the sum of outputs from the network layers. In addition to the potential of such CNNs to produce interesting audio transformation, more information about the network and the original music could be obtained from the analysis of the generated features since these features indicate how the network 'understands' the music.
Single-Cell Phenotype Classification Using Deep Convolutional Neural Networks.
Dürr, Oliver; Sick, Beate
2016-10-01
Deep learning methods are currently outperforming traditional state-of-the-art computer vision algorithms in diverse applications and recently even surpassed human performance in object recognition. Here we demonstrate the potential of deep learning methods to high-content screening-based phenotype classification. We trained a deep learning classifier in the form of convolutional neural networks with approximately 40,000 publicly available single-cell images from samples treated with compounds from four classes known to lead to different phenotypes. The input data consisted of multichannel images. The construction of appropriate feature definitions was part of the training and carried out by the convolutional network, without the need for expert knowledge or handcrafted features. We compare our results against the recent state-of-the-art pipeline in which predefined features are extracted from each cell using specialized software and then fed into various machine learning algorithms (support vector machine, Fisher linear discriminant, random forest) for classification. The performance of all classification approaches is evaluated on an untouched test image set with known phenotype classes. Compared to the best reference machine learning algorithm, the misclassification rate is reduced from 8.9% to 6.6%.
Enhancing Neutron Beam Production with a Convoluted Moderator
Energy Technology Data Exchange (ETDEWEB)
Iverson, Erik B [ORNL; Baxter, David V [Center for the Exploration of Energy and Matter, Indiana University; Muhrer, Guenter [Los Alamos National Laboratory (LANL); Ansell, Stuart [ISIS Facility, Rutherford Appleton Laboratory (ISIS); Gallmeier, Franz X [ORNL; Dalgliesh, Robert [ISIS Facility, Rutherford Appleton Laboratory (ISIS); Lu, Wei [ORNL; Kaiser, Helmut [Center for the Exploration of Energy and Matter, Indiana University
2014-10-01
We describe a new concept for a neutron moderating assembly resulting in the more efficient production of slow neutron beams. The Convoluted Moderator, a heterogeneous stack of interleaved moderating material and nearly transparent single-crystal spacers, is a directionally-enhanced neutron beam source, improving beam effectiveness over an angular range comparable to the range accepted by neutron beam lines and guides. We have demonstrated gains of 50% in slow neutron intensity for a given fast neutron production rate while simultaneously reducing the wavelength-dependent emission time dispersion by 25%, both coming from a geometric effect in which the neutron beam lines view a large surface area of moderating material in a relatively small volume. Additionally, we have confirmed a Bragg-enhancement effect arising from coherent scattering within the single-crystal spacers. We have not observed hypothesized refractive effects leading to additional gains at long wavelength. In addition to confirmation of the validity of the Convoluted Moderator concept, our measurements provide a series of benchmark experiments suitable for developing simulation and analysis techniques for practical optimization and eventual implementation at slow neutron source facilities.
Multiple deep convolutional neural networks averaging for face alignment
Zhang, Shaohua; Yang, Hua; Yin, Zhouping
2015-05-01
Face alignment is critical for face recognition, and the deep learning-based method shows promise for solving such issues, given that competitive results are achieved on benchmarks with additional benefits, such as dispensing with handcrafted features and initial shape. However, most existing deep learning-based approaches are complicated and quite time-consuming during training. We propose a compact face alignment method for fast training without decreasing its accuracy. Rectified linear unit is employed, which allows all networks approximately five times faster convergence than a tanh neuron. An eight learnable layer deep convolutional neural network (DCNN) based on local response normalization and a padding convolutional layer (PCL) is designed to provide reliable initial values during prediction. A model combination scheme is presented to further reduce errors, while showing that only two network architectures and hyperparameter selection procedures are required in our approach. A three-level cascaded system is ultimately built based on the DCNNs and model combination mode. Extensive experiments validate the effectiveness of our method and demonstrate comparable accuracy with state-of-the-art methods on BioID, labeled face parts in the wild, and Helen datasets.
Classifications of multispectral colorectal cancer tissues using convolution neural network
Directory of Open Access Journals (Sweden)
Hawraa Haj-Hassan
2017-01-01
Full Text Available Background: Colorectal cancer (CRC is the third most common cancer among men and women. Its diagnosis in early stages, typically done through the analysis of colon biopsy images, can greatly improve the chances of a successful treatment. This paper proposes to use convolution neural networks (CNNs to predict three tissue types related to the progression of CRC: benign hyperplasia (BH, intraepithelial neoplasia (IN, and carcinoma (Ca. Methods: Multispectral biopsy images of thirty CRC patients were retrospectively analyzed. Images of tissue samples were divided into three groups, based on their type (10 BH, 10 IN, and 10 Ca. An active contour model was used to segment image regions containing pathological tissues. Tissue samples were classified using a CNN containing convolution, max-pooling, and fully-connected layers. Available tissue samples were split into a training set, for learning the CNN parameters, and test set, for evaluating its performance. Results: An accuracy of 99.17% was obtained from segmented image regions, outperforming existing approaches based on traditional feature extraction, and classification techniques. Conclusions: Experimental results demonstrate the effectiveness of CNN for the classification of CRC tissue types, in particular when using presegmented regions of interest.
An optimal nonorthogonal separation of the anisotropic Gaussian convolution filter.
Lampert, Christoph H; Wirjadi, Oliver
2006-11-01
We give an analytical and geometrical treatment of what it means to separate a Gaussian kernel along arbitrary axes in R(n), and we present a separation scheme that allows us to efficiently implement anisotropic Gaussian convolution filters for data of arbitrary dimensionality. Based on our previous analysis we show that this scheme is optimal with regard to the number of memory accesses and interpolation operations needed. The proposed method relies on nonorthogonal convolution axes and works completely in image space. Thus, it avoids the need for a fast Fourier transform (FFT)-subroutine. Depending on the accuracy and speed requirements, different interpolation schemes and methods to implement the one-dimensional Gaussian (finite impulse response and infinite impulse response) can be integrated. Special emphasis is put on analyzing the performance and accuracy of the new method. In particular, we show that without any special optimization of the source code, it can perform anisotropic Gaussian filtering faster than methods relying on the FFT.
Thermalnet: a Deep Convolutional Network for Synthetic Thermal Image Generation
Kniaz, V. V.; Gorbatsevich, V. S.; Mizginov, V. A.
2017-05-01
Deep convolutional neural networks have dramatically changed the landscape of the modern computer vision. Nowadays methods based on deep neural networks show the best performance among image recognition and object detection algorithms. While polishing of network architectures received a lot of scholar attention, from the practical point of view the preparation of a large image dataset for a successful training of a neural network became one of major challenges. This challenge is particularly profound for image recognition in wavelengths lying outside the visible spectrum. For example no infrared or radar image datasets large enough for successful training of a deep neural network are available to date in public domain. Recent advances of deep neural networks prove that they are also capable to do arbitrary image transformations such as super-resolution image generation, grayscale image colorisation and imitation of style of a given artist. Thus a natural question arise: how could be deep neural networks used for augmentation of existing large image datasets? This paper is focused on the development of the Thermalnet deep convolutional neural network for augmentation of existing large visible image datasets with synthetic thermal images. The Thermalnet network architecture is inspired by colorisation deep neural networks.
Trainable Convolution Filters and Their Application to Face Recognition.
Kumar, Ritwik; Banerjee, Arunava; Vemuri, Baba C; Pfister, Hanspeter
2012-07-01
In this paper, we present a novel image classification system that is built around a core of trainable filter ensembles that we call Volterra kernel classifiers. Our system treats images as a collection of possibly overlapping patches and is composed of three components: (1) A scheme for a single patch classification that seeks a smooth, possibly nonlinear, functional mapping of the patches into a range space, where patches of the same class are close to one another, while patches from different classes are far apart-in the L_2 sense. This mapping is accomplished using trainable convolution filters (or Volterra kernels) where the convolution kernel can be of any shape or order. (2) Given a corpus of Volterra classifiers with various kernel orders and shapes for each patch, a boosting scheme for automatically selecting the best weighted combination of the classifiers to achieve higher per-patch classification rate. (3) A scheme for aggregating the classification information obtained for each patch via voting for the parent image classification. We demonstrate the effectiveness of the proposed technique using face recognition as an application area and provide extensive experiments on the Yale, CMU PIE, Extended Yale B, Multi-PIE, and MERL Dome benchmark face data sets. We call the Volterra kernel classifiers applied to face recognition Volterrafaces. We show that our technique, which falls into the broad class of embedding-based face image discrimination methods, consistently outperforms various state-of-the-art methods in the same category.
Video-based face recognition via convolutional neural networks
Bao, Tianlong; Ding, Chunhui; Karmoshi, Saleem; Zhu, Ming
2017-06-01
Face recognition has been widely studied recently while video-based face recognition still remains a challenging task because of the low quality and large intra-class variation of video captured face images. In this paper, we focus on two scenarios of video-based face recognition: 1)Still-to-Video(S2V) face recognition, i.e., querying a still face image against a gallery of video sequences; 2)Video-to-Still(V2S) face recognition, in contrast to S2V scenario. A novel method was proposed in this paper to transfer still and video face images to an Euclidean space by a carefully designed convolutional neural network, then Euclidean metrics are used to measure the distance between still and video images. Identities of still and video images that group as pairs are used as supervision. In the training stage, a joint loss function that measures the Euclidean distance between the predicted features of training pairs and expanding vectors of still images is optimized to minimize the intra-class variation while the inter-class variation is guaranteed due to the large margin of still images. Transferred features are finally learned via the designed convolutional neural network. Experiments are performed on COX face dataset. Experimental results show that our method achieves reliable performance compared with other state-of-the-art methods.
Real-Time Video Convolutional Face Finder on Embedded Platforms
Directory of Open Access Journals (Sweden)
Mamalet Franck
2007-01-01
Full Text Available A high-level optimization methodology is applied for implementing the well-known convolutional face finder (CFF algorithm for real-time applications on mobile phones, such as teleconferencing, advanced user interfaces, image indexing, and security access control. CFF is based on a feature extraction and classification technique which consists of a pipeline of convolutions and subsampling operations. The design of embedded systems requires a good trade-off between performance and code size due to the limited amount of available resources. The followed methodology copes with the main drawbacks of the original implementation of CFF such as floating-point computation and memory allocation, in order to allow parallelism exploitation and perform algorithm optimizations. Experimental results show that our embedded face detection system can accurately locate faces with less computational load and memory cost. It runs on a 275 MHz Starcore DSP at 35 QCIF images/s with state-of-the-art detection rates and very low false alarm rates.
Deep Convolutional Neural Networks for large-scale speech tasks.
Sainath, Tara N; Kingsbury, Brian; Saon, George; Soltau, Hagen; Mohamed, Abdel-rahman; Dahl, George; Ramabhadran, Bhuvana
2015-04-01
Convolutional Neural Networks (CNNs) are an alternative type of neural network that can be used to reduce spectral variations and model spectral correlations which exist in signals. Since speech signals exhibit both of these properties, we hypothesize that CNNs are a more effective model for speech compared to Deep Neural Networks (DNNs). In this paper, we explore applying CNNs to large vocabulary continuous speech recognition (LVCSR) tasks. First, we determine the appropriate architecture to make CNNs effective compared to DNNs for LVCSR tasks. Specifically, we focus on how many convolutional layers are needed, what is an appropriate number of hidden units, what is the best pooling strategy. Second, investigate how to incorporate speaker-adapted features, which cannot directly be modeled by CNNs as they do not obey locality in frequency, into the CNN framework. Third, given the importance of sequence training for speech tasks, we introduce a strategy to use ReLU+dropout during Hessian-free sequence training of CNNs. Experiments on 3 LVCSR tasks indicate that a CNN with the proposed speaker-adapted and ReLU+dropout ideas allow for a 12%-14% relative improvement in WER over a strong DNN system, achieving state-of-the art results in these 3 tasks.
Coronary artery calcification (CAC) classification with deep convolutional neural networks
Liu, Xiuming; Wang, Shice; Deng, Yufeng; Chen, Kuan
2017-03-01
Coronary artery calcification (CAC) is a typical marker of the coronary artery disease, which is one of the biggest causes of mortality in the U.S. This study evaluates the feasibility of using a deep convolutional neural network (DCNN) to automatically detect CAC in X-ray images. 1768 posteroanterior (PA) view chest X-Ray images from Sichuan Province Peoples Hospital, China were collected retrospectively. Each image is associated with a corresponding diagnostic report written by a trained radiologist (907 normal, 861 diagnosed with CAC). Onequarter of the images were randomly selected as test samples; the rest were used as training samples. DCNN models consisting of 2,4,6 and 8 convolutional layers were designed using blocks of pre-designed CNN layers. Each block was implemented in Theano with Graphics Processing Units (GPU). Human-in-the-loop learning was also performed on a subset of 165 images with framed arteries by trained physicians. The results from the DCNN models were compared to the diagnostic reports. The average diagnostic accuracies for models with 2,4,6,8 layers were 0.85, 0.87, 0.88, and 0.89 respectively. The areas under the curve (AUC) were 0.92, 0.95, 0.95, and 0.96. As the model grows deeper, the AUC or diagnostic accuracies did not have statistically significant changes. The results of this study indicate that DCNN models have promising potential in the field of intelligent medical image diagnosis practice.
Using convolutional decoding to improve time delay and phase estimation in digital communications
Energy Technology Data Exchange (ETDEWEB)
Ormesher, Richard C. (Albuquerque, NM); Mason, John J. (Albuquerque, NM)
2010-01-26
The time delay and/or phase of a communication signal received by a digital communication receiver can be estimated based on a convolutional decoding operation that the communication receiver performs on the received communication signal. If the original transmitted communication signal has been spread according to a spreading operation, a corresponding despreading operation can be integrated into the convolutional decoding operation.
moco: Fast Motion Correction for Calcium Imaging
Directory of Open Access Journals (Sweden)
Alexander eDubbs
2016-02-01
Full Text Available Motion correction is the first step in a pipeline of algorithms to analyze calcium imaging videos and extract biologically relevant information, for example the network structure of the neurons therein. Fast motion correction is especially critical for closed-loop activity triggered stimulation experiments, where accurate detection and targeting of specific cells in necessary. We introduce a novel motion-correction algorithm that uses a Fourier-transform approach, and a combination of judicious downsampling and the accelerated computation of many $L_2$ norms using dynamic programming and two-dimensional, fft-accelerated convolutions, to enhance its efficiency. Its accuracy is comparable to that of established community-used algorithms, and it is more stable to large translational motions. It is programmed in Java and is compatible with ImageJ.
moco: Fast Motion Correction for Calcium Imaging.
Dubbs, Alexander; Guevara, James; Yuste, Rafael
2016-01-01
Motion correction is the first step in a pipeline of algorithms to analyze calcium imaging videos and extract biologically relevant information, for example the network structure of the neurons therein. Fast motion correction is especially critical for closed-loop activity triggered stimulation experiments, where accurate detection and targeting of specific cells in necessary. We introduce a novel motion-correction algorithm which uses a Fourier-transform approach, and a combination of judicious downsampling and the accelerated computation of many L 2 norms using dynamic programming and two-dimensional, fft-accelerated convolutions, to enhance its efficiency. Its accuracy is comparable to that of established community-used algorithms, and it is more stable to large translational motions. It is programmed in Java and is compatible with ImageJ.
Directory of Open Access Journals (Sweden)
Congyi Lyu
2016-12-01
Full Text Available Vision-based object tracking has lots of applications in robotics, like surveillance, navigation, motion capturing, and so on. However, the existing object tracking systems still suffer from the challenging problem of high computation consumption in the image processing algorithms. The problem can prevent current systems from being used in many robotic applications which have limitations of payload and power, for example, micro air vehicles. In these applications, the central processing unit- or graphics processing unit-based computers are not good choices due to the high weight and power consumption. To address the problem, this article proposed a real-time object tracking system based on field-programmable gate array, convolution neural network, and visual servo technology. The time-consuming image processing algorithms, such as distortion correction, color space convertor, and Sobel edge, Harris corner features detector, and convolution neural network were redesigned using the programmable gates in field-programmable gate array. Based on the field-programmable gate array-based image processing, an image-based visual servo controller was designed to drive a two degree of freedom manipulator to track the target in real time. Finally, experiments on the proposed system were performed to illustrate the effectiveness of the real-time object tracking system.
An investigation of error correcting techniques for OMV and AXAF
Ingels, Frank; Fryer, John
1991-01-01
The original objectives of this project were to build a test system for the NASA 255/223 Reed/Solomon encoding/decoding chip set and circuit board. This test system was then to be interfaced with a convolutional system at MSFC to examine the performance of the concantinated codes. After considerable work, it was discovered that the convolutional system could not function as needed. This report documents the design, construction, and testing of the test apparatus for the R/S chip set. The approach taken was to verify the error correcting behavior of the chip set by injecting known error patterns onto data and observing the results. Error sequences were generated using pseudo-random number generator programs, with Poisson time distribution between errors and Gaussian burst lengths. Sample means, variances, and number of un-correctable errors were calculated for each data set before testing.
The effect of whitening transformation on pooling operations in convolutional autoencoders
Li, Zuhe; Fan, Yangyu; Liu, Weihua
2015-12-01
Convolutional autoencoders (CAEs) are unsupervised feature extractors for high-resolution images. In the pre-processing step, whitening transformation has widely been adopted to remove redundancy by making adjacent pixels less correlated. Pooling is a biologically inspired operation to reduce the resolution of feature maps and achieve spatial invariance in convolutional neural networks. Conventionally, pooling methods are mainly determined empirically in most previous work. Therefore, our main purpose is to study the relationship between whitening processing and pooling operations in convolutional autoencoders for image classification. We propose an adaptive pooling approach based on the concepts of information entropy to test the effect of whitening on pooling in different conditions. Experimental results on benchmark datasets indicate that the performance of pooling strategies is associated with the distribution of feature activations, which can be affected by whitening processing. This provides guidance for the selection of pooling methods in convolutional autoencoders and other convolutional neural networks.
Dura-Bernal, Salvador; Garreau, Guillaume; Georgiou, Julius; Andreou, Andreas G; Denham, Susan L; Wennekers, Thomas
2013-10-01
The ability to recognize the behavior of individuals is of great interest in the general field of safety (e.g. building security, crowd control, transport analysis, independent living for the elderly). Here we report a new real-time acoustic system for human action and behavior recognition that integrates passive audio and active micro-Doppler sonar signatures over multiple time scales. The system architecture is based on a six-layer convolutional neural network, trained and evaluated using a dataset of 10 subjects performing seven different behaviors. Probabilistic combination of system output through time for each modality separately yields 94% (passive audio) and 91% (micro-Doppler sonar) correct behavior classification; probabilistic multimodal integration increases classification performance to 98%. This study supports the efficacy of micro-Doppler sonar systems in characterizing human actions, which can then be efficiently classified using ConvNets. It also demonstrates that the integration of multiple sources of acoustic information can significantly improve the system's performance.
Directory of Open Access Journals (Sweden)
Yi Zhang
2017-04-01
Full Text Available Ear detection is an important step in ear recognition approaches. Most existing ear detection techniques are based on manually designing features or shallow learning algorithms. However, researchers found that the pose variation, occlusion, and imaging conditions provide a great challenge to the traditional ear detection methods under uncontrolled conditions. This paper proposes an efficient technique involving Multiple Scale Faster Region-based Convolutional Neural Networks (Faster R-CNN to detect ears from 2D profile images in natural images automatically. Firstly, three regions of different scales are detected to infer the information about the ear location context within the image. Then an ear region filtering approach is proposed to extract the correct ear region and eliminate the false positives automatically. In an experiment with a test set of 200 web images (with variable photographic conditions, 98% of ears were accurately detected. Experiments were likewise conducted on the Collection J2 of University of Notre Dame Biometrics Database (UND-J2 and University of Beira Interior Ear dataset (UBEAR, which contain large occlusion, scale, and pose variations. Detection rates of 100% and 98.22%, respectively, demonstrate the effectiveness of the proposed approach.
Directory of Open Access Journals (Sweden)
C. Yao
2017-09-01
Full Text Available With the rapid development of Precision Agriculture (PA promoted by high-resolution remote sensing, it makes significant sense in management and estimation of agriculture through crop classification of high-resolution remote sensing image. Due to the complex and fragmentation of the features and the surroundings in the circumstance of high-resolution, the accuracy of the traditional classification methods has not been able to meet the standard of agricultural problems. In this case, this paper proposed a classification method for high-resolution agricultural remote sensing images based on convolution neural networks(CNN. For training, a large number of training samples were produced by panchromatic images of GF-1 high-resolution satellite of China. In the experiment, through training and testing on the CNN under the toolbox of deep learning by MATLAB, the crop classification finally got the correct rate of 99.66 % after the gradual optimization of adjusting parameter during training. Through improving the accuracy of image classification and image recognition, the applications of CNN provide a reference value for the field of remote sensing in PA.
Yao, C.; Zhang, Y.; Zhang, Y.; Liu, H.
2017-09-01
With the rapid development of Precision Agriculture (PA) promoted by high-resolution remote sensing, it makes significant sense in management and estimation of agriculture through crop classification of high-resolution remote sensing image. Due to the complex and fragmentation of the features and the surroundings in the circumstance of high-resolution, the accuracy of the traditional classification methods has not been able to meet the standard of agricultural problems. In this case, this paper proposed a classification method for high-resolution agricultural remote sensing images based on convolution neural networks(CNN). For training, a large number of training samples were produced by panchromatic images of GF-1 high-resolution satellite of China. In the experiment, through training and testing on the CNN under the toolbox of deep learning by MATLAB, the crop classification finally got the correct rate of 99.66 % after the gradual optimization of adjusting parameter during training. Through improving the accuracy of image classification and image recognition, the applications of CNN provide a reference value for the field of remote sensing in PA.
A convolution model of rock bed thermal storage units
Sowell, E. F.; Curry, R. L.
1980-01-01
A method is presented whereby a packed-bed thermal storage unit is dynamically modeled for bi-directional flow and arbitrary input flow stream temperature variations. The method is based on the principle of calculating the output temperature as the sum of earlier input temperatures, each multiplied by a predetermined 'response factor', i.e., discrete convolution. A computer implementation of the scheme, in the form of a subroutine for a widely used solar simulation program (TRNSYS) is described and numerical results compared with other models. Also, a method for efficient computation of the required response factors is described; this solution is for a triangular input pulse, previously unreported, although the solution method is also applicable for other input functions. This solution requires a single integration of a known function which is easily carried out numerically to the required precision.
Drug-Drug Interaction Extraction via Convolutional Neural Networks
Directory of Open Access Journals (Sweden)
Shengyu Liu
2016-01-01
Full Text Available Drug-drug interaction (DDI extraction as a typical relation extraction task in natural language processing (NLP has always attracted great attention. Most state-of-the-art DDI extraction systems are based on support vector machines (SVM with a large number of manually defined features. Recently, convolutional neural networks (CNN, a robust machine learning method which almost does not need manually defined features, has exhibited great potential for many NLP tasks. It is worth employing CNN for DDI extraction, which has never been investigated. We proposed a CNN-based method for DDI extraction. Experiments conducted on the 2013 DDIExtraction challenge corpus demonstrate that CNN is a good choice for DDI extraction. The CNN-based DDI extraction method achieves an F-score of 69.75%, which outperforms the existing best performing method by 2.75%.
Drug-Drug Interaction Extraction via Convolutional Neural Networks.
Liu, Shengyu; Tang, Buzhou; Chen, Qingcai; Wang, Xiaolong
2016-01-01
Drug-drug interaction (DDI) extraction as a typical relation extraction task in natural language processing (NLP) has always attracted great attention. Most state-of-the-art DDI extraction systems are based on support vector machines (SVM) with a large number of manually defined features. Recently, convolutional neural networks (CNN), a robust machine learning method which almost does not need manually defined features, has exhibited great potential for many NLP tasks. It is worth employing CNN for DDI extraction, which has never been investigated. We proposed a CNN-based method for DDI extraction. Experiments conducted on the 2013 DDIExtraction challenge corpus demonstrate that CNN is a good choice for DDI extraction. The CNN-based DDI extraction method achieves an F-score of 69.75%, which outperforms the existing best performing method by 2.75%.
Convolutional Neural Networks Applied to House Numbers Digit Classification
Sermanet, Pierre; LeCun, Yann
2012-01-01
We classify digits of real-world house numbers using convolutional neural networks (ConvNets). ConvNets are hierarchical feature learning neural networks whose structure is biologically inspired. Unlike many popular vision approaches that are hand-designed, ConvNets can automatically learn a unique set of features optimized for a given task. We augmented the traditional ConvNet architecture by learning multi-stage features and by using Lp pooling and establish a new state-of-the-art of 94.85% accuracy on the SVHN dataset (45.2% error improvement). Furthermore, we analyze the benefits of different pooling methods and multi-stage features in ConvNets. The source code and a tutorial are available at eblearn.sf.net.
INVARIANT DESCRIPTOR LEARNING USING A SIAMESE CONVOLUTIONAL NEURAL NETWORK
Directory of Open Access Journals (Sweden)
L. Chen
2016-06-01
Full Text Available In this paper we describe learning of a descriptor based on the Siamese Convolutional Neural Network (CNN architecture and evaluate our results on a standard patch comparison dataset. The descriptor learning architecture is composed of an input module, a Siamese CNN descriptor module and a cost computation module that is based on the L2 Norm. The cost function we use pulls the descriptors of matching patches close to each other in feature space while pushing the descriptors for non-matching pairs away from each other. Compared to related work, we optimize the training parameters by combining a moving average strategy for gradients and Nesterov's Accelerated Gradient. Experiments show that our learned descriptor reaches a good performance and achieves state-of-art results in terms of the false positive rate at a 95 % recall rate on standard benchmark datasets.
Hybrid Algorithm for the Optimization of Training Convolutional Neural Network
Directory of Open Access Journals (Sweden)
Hayder M. Albeahdili
2015-10-01
Full Text Available The training optimization processes and efficient fast classification are vital elements in the development of a convolution neural network (CNN. Although stochastic gradient descend (SGD is a Prevalence algorithm used by many researchers for the optimization of training CNNs, it has vast limitations. In this paper, it is endeavor to diminish and tackle drawbacks inherited from SGD by proposing an alternate algorithm for CNN training optimization. A hybrid of genetic algorithm (GA and particle swarm optimization (PSO is deployed in this work. In addition to SGD, PSO and genetic algorithm (PSO-GA are also incorporated as a combined and efficient mechanism in achieving non trivial solutions. The proposed unified method achieves state-of-the-art classification results on the different challenge benchmark datasets such as MNIST, CIFAR-10, and SVHN. Experimental results showed that the results outperform and achieve superior results to most contemporary approaches.
Facial Expression Recognition Using 3D Convolutional Neural Network
Directory of Open Access Journals (Sweden)
Young-Hyen Byeon
2014-12-01
Full Text Available This paper is concerned with video-based facial expression recognition frequently used in conjunction with HRI (Human-Robot Interaction that can naturally interact between human and robot. For this purpose, we design a 3D-CNN(3D Convolutional Neural Networks by augmenting dimensionality reduction methods such as PCA(Principal Component Analysis and TMPCA(Tensor-based Multilinear Principal Component Analysis to recognize simultaneously the successive frames with facial expression images obtained through video camera. The 3D-CNN can achieve some degree of shift and deformation invariance using local receptive fields and spatial subsampling through dimensionality reduction of redundant CNN’s output. The experimental results on video-based facial expression database reveal that the presented method shows a good performance in comparison to the conventional methods such as PCA and TMPCA.
FPGA Prototyping of RNN Decoder for Convolutional Codes
Directory of Open Access Journals (Sweden)
Salcic Zoran
2006-01-01
Full Text Available This paper presents prototyping of a recurrent type neural network (RNN convolutional decoder using system-level design specification and design flow that enables easy mapping to the target FPGA architecture. Implementation and the performance measurement results have shown that an RNN decoder for hard-decision decoding coupled with a simple hard-limiting neuron activation function results in a very low complexity, which easily fits into standard Altera FPGA. Moreover, the design methodology allowed modeling of complete testbed for prototyping RNN decoders in simulation and real-time environment (same FPGA, thus enabling evaluation of BER performance characteristics of the decoder for various conditions of communication channel in real time.
Fast convolution with free-space Green's functions
Vico, Felipe; Greengard, Leslie; Ferrando, Miguel
2016-10-01
We introduce a fast algorithm for computing volume potentials - that is, the convolution of a translation invariant, free-space Green's function with a compactly supported source distribution defined on a uniform grid. The algorithm relies on regularizing the Fourier transform of the Green's function by cutting off the interaction in physical space beyond the domain of interest. This permits the straightforward application of trapezoidal quadrature and the standard FFT, with superalgebraic convergence for smooth data. Moreover, the method can be interpreted as employing a Nystrom discretization of the corresponding integral operator, with matrix entries which can be obtained explicitly and rapidly. This is of use in the design of preconditioners or fast direct solvers for a variety of volume integral equations. The method proposed permits the computation of any derivative of the potential, at the cost of an additional FFT.
Structured learning via convolutional neural networks for vehicle detection
Maqueda, Ana I.; del Blanco, Carlos R.; Jaureguizar, Fernando; García, Narciso
2017-05-01
One of the main tasks in a vision-based traffic monitoring system is the detection of vehicles. Recently, deep neural networks have been successfully applied to this end, outperforming previous approaches. However, most of these works generally rely on complex and high-computational region proposal networks. Others employ deep neural networks as a segmentation strategy to achieve a semantic representation of the object of interest, which has to be up-sampled later. In this paper, a new design for a convolutional neural network is applied to vehicle detection in highways for traffic monitoring. This network generates a spatially structured output that encodes the vehicle locations. Promising results have been obtained in the GRAM-RTM dataset.
SOME ASYMPTOTIC PROPERTIES OF THE CONVOLUTION TRANSFORMS OF FRACTAL MEASURES
Institute of Scientific and Technical Information of China (English)
Cao Li
2012-01-01
We study the asymptotic behavior near the boundary of u(x,y) =Ky * μ (x),defined on the half-space R+ × RN by the convolution of an approximate identity Ky (.) (y ＞0) and a measure μ on RN.The Poisson and the heat kernel are unified as special cases in our setting.We are mainly interested in the relationship between the rate of growth at boundary of u and the s-density of a singular measure μ.Then a boundary limit theorem of Fatou's type for singular measures is proved.Meanwhile,the asymptotic behavior of a quotient of Kμ and Kv is also studied,then the corresponding Fatou-Doob's boundary relative limit is obtained.In particular,some results about the singular boundary behavior of harmonic and heat functions can be deduced simultaneously from ours.At the end,an application in fractal geometry is given.
Plane-wave decomposition by spherical-convolution microphone array
Rafaely, Boaz; Park, Munhum
2001-05-01
Reverberant sound fields are widely studied, as they have a significant influence on the acoustic performance of enclosures in a variety of applications. For example, the intelligibility of speech in lecture rooms, the quality of music in auditoria, the noise level in offices, and the production of 3D sound in living rooms are all affected by the enclosed sound field. These sound fields are typically studied through frequency response measurements or statistical measures such as reverberation time, which do not provide detailed spatial information. The aim of the work presented in this seminar is the detailed analysis of reverberant sound fields. A measurement and analysis system based on acoustic theory and signal processing, designed around a spherical microphone array, is presented. Detailed analysis is achieved by decomposition of the sound field into waves, using spherical Fourier transform and spherical convolution. The presentation will include theoretical review, simulation studies, and initial experimental results.
Convolution Equivalent L\\'evy Processes and First Passage Times
Griffin, Philip S
2012-01-01
We investigate the behaviour of L\\'{e}vy processes with convolution equivalent L\\'evy measures, up to the time of first passage over a high level $u$. Such problems arise naturally in the context of insurance risk where $u$ is the initial reserve. We obtain a precise asymptotic estimate on the probability of first passage occurring by time $T$. This result is then used to study the process conditioned on first passage by time $T$. The existence of a limiting process as $u\\to \\infty$ is demonstrated, which leads to precise estimates for the probability of other events relating to first passage, such as the overshoot. A discussion of these results, as they relate to insurance risk, is also given.
Star-galaxy classification using deep convolutional neural networks
Kim, Edward J.; Brunner, Robert J.
2017-02-01
Most existing star-galaxy classifiers use the reduced summary information from catalogues, requiring careful feature extraction and selection. The latest advances in machine learning that use deep convolutional neural networks (ConvNets) allow a machine to automatically learn the features directly from the data, minimizing the need for input from human experts. We present a star-galaxy classification framework that uses deep ConvNets directly on the reduced, calibrated pixel values. Using data from the Sloan Digital Sky Survey and the Canada-France-Hawaii Telescope Lensing Survey, we demonstrate that ConvNets are able to produce accurate and well-calibrated probabilistic classifications that are competitive with conventional machine learning techniques. Future advances in deep learning may bring more success with current and forthcoming photometric surveys, such as the Dark Energy Survey and the Large Synoptic Survey Telescope, because deep neural networks require very little, manual feature engineering.
Invariant Descriptor Learning Using a Siamese Convolutional Neural Network
Chen, L.; Rottensteiner, F.; Heipke, C.
2016-06-01
In this paper we describe learning of a descriptor based on the Siamese Convolutional Neural Network (CNN) architecture and evaluate our results on a standard patch comparison dataset. The descriptor learning architecture is composed of an input module, a Siamese CNN descriptor module and a cost computation module that is based on the L2 Norm. The cost function we use pulls the descriptors of matching patches close to each other in feature space while pushing the descriptors for non-matching pairs away from each other. Compared to related work, we optimize the training parameters by combining a moving average strategy for gradients and Nesterov's Accelerated Gradient. Experiments show that our learned descriptor reaches a good performance and achieves state-of-art results in terms of the false positive rate at a 95 % recall rate on standard benchmark datasets.
Exploiting Narrowband Efficiency for Broadband Convolutive Blind Source Separation
Directory of Open Access Journals (Sweden)
Aichner Robert
2007-01-01
Full Text Available Based on a recently presented generic broadband blind source separation (BSS algorithm for convolutive mixtures, we propose in this paper a novel algorithm combining advantages of broadband algorithms with the computational efficiency of narrowband techniques. By selective application of the Szegö theorem which relates properties of Toeplitz and circulant matrices, a new normalization is derived as a special case of the generic broadband algorithm. This results in a computationally efficient and fast converging algorithm without introducing typical narrowband problems such as the internal permutation problem or circularity effects. Moreover, a novel regularization method for the generic broadband algorithm is proposed and subsequently also derived for the proposed algorithm. Experimental results in realistic acoustic environments show improved performance of the novel algorithm compared to previous approximations.
Learning Building Extraction in Aerial Scenes with Convolutional Networks.
Yuan, Jiangye
2017-09-11
Extracting buildings from aerial scene images is an important task with many applications. However, this task is highly difficult to automate due to extremely large variations of building appearances, and still heavily relies on manual work. To attack this problem, we design a deep convolutional network with a simple structure that integrates activation from multiple layers for pixel-wise prediction, and introduce the signed distance function of building boundaries as the output representation, which has an enhanced representation power. To train the network, we leverage abundant building footprint data from geographic information systems (GIS) to generate large amounts of labeled data. The trained model achieves a superior performance on datasets that are significantly larger and more complex than those used in prior work, demonstrating that the proposed method provides a promising and scalable solution for automating this labor-intensive task.
Training strategy for convolutional neural networks in pedestrian gender classification
Ng, Choon-Boon; Tay, Yong-Haur; Goi, Bok-Min
2017-06-01
In this work, we studied a strategy for training a convolutional neural network in pedestrian gender classification with limited amount of labeled training data. Unsupervised learning by k-means clustering on pedestrian images was used to learn the filters to initialize the first layer of the network. As a form of pre-training, supervised learning for the related task of pedestrian classification was performed. Finally, the network was fine-tuned for gender classification. We found that this strategy improved the network's generalization ability in gender classification, achieving better test results when compared to random weights initialization and slightly more beneficial than merely initializing the first layer filters by unsupervised learning. This shows that unsupervised learning followed by pre-training with pedestrian images is an effective strategy to learn useful features for pedestrian gender classification.
XOR-FREE Implementation of Convolutional Encoder for Reconfigurable Hardware
Directory of Open Access Journals (Sweden)
Gaurav Purohit
2016-01-01
Full Text Available This paper presents a novel XOR-FREE algorithm to implement the convolutional encoder using reconfigurable hardware. The approach completely removes the XOR processing of a chosen nonsystematic, feedforward generator polynomial of larger constraint length. The hardware (HW implementation of new architecture uses Lookup Table (LUT for storing the parity bits. The design implements architectural reconfigurability by modifying the generator polynomial of the same constraint length and code rate to reduce the design complexity. The proposed architecture reduces the dynamic power up to 30% and improves the hardware cost and propagation delay up to 20% and 32%, respectively. The performance of the proposed architecture is validated in MATLAB Simulink and tested on Zynq-7 series FPGA.
Radio Frequency Interference mitigation using deep convolutional neural networks
Akeret, Joel; Lucchi, Aurelien; Refregier, Alexandre
2016-01-01
We propose a novel approach for mitigating radio frequency interference (RFI) signals in radio data using the latest advances in deep learning. We employ a special type of Convolutional Neural Network, the U-Net, that enables the classification of clean signal and RFI signatures in 2D time-ordered data acquired from a radio telescope. We train and assess the performance of this network using the HIDE & SEEK radio data simulation and processing packages, as well as data collected at the Bleien Observatory. We find that our U-Net implementation can outperform classical RFI mitigation algorithms such as SEEK's SumThreshold implementation. We publish our U-Net software package on GitHub under GPLv3 license.
Radio frequency interference mitigation using deep convolutional neural networks
Akeret, J.; Chang, C.; Lucchi, A.; Refregier, A.
2017-01-01
We propose a novel approach for mitigating radio frequency interference (RFI) signals in radio data using the latest advances in deep learning. We employ a special type of Convolutional Neural Network, the U-Net, that enables the classification of clean signal and RFI signatures in 2D time-ordered data acquired from a radio telescope. We train and assess the performance of this network using the HIDE &SEEK radio data simulation and processing packages, as well as early Science Verification data acquired with the 7m single-dish telescope at the Bleien Observatory. We find that our U-Net implementation is showing competitive accuracy to classical RFI mitigation algorithms such as SEEK's SUMTHRESHOLD implementation. We publish our U-Net software package on GitHub under GPLv3 license.
Interleaved Convolutional Code and Its Viterbi Decoder Architecture
Directory of Open Access Journals (Sweden)
Jun Jin Kong
2003-12-01
Full Text Available We propose an area-efficient high-speed interleaved Viterbi decoder architecture, which is based on the state-parallel architecture with register exchange path memory structure, for interleaved convolutional code. The state-parallel architecture uses as many add-compare-select (ACS units as the number of trellis states. By replacing each delay (or storage element in state metrics memory (or path metrics memory and path memory (or survival memory with I delays, interleaved Viterbi decoder is obtained where I is the interleaving degree. The decoding speed of this decoder architecture is as fast as the operating clock speed. The latency of proposed interleaved Viterbi decoder is Ã‚Â“decoding depth (DD ÃƒÂ— interleaving degree (I+Ã¢Â€Â‰extra delaysÃ¢Â€Â‰(A,Ã‚Â” which increases linearly with the interleaving degree I.
Rapid Exact Signal Scanning With Deep Convolutional Neural Networks
Thom, Markus; Gritschneder, Franz
2017-03-01
A rigorous formulation of the dynamics of a signal processing scheme aimed at dense signal scanning without any loss in accuracy is introduced and analyzed. Related methods proposed in the recent past lack a satisfactory analysis of whether they actually fulfill any exactness constraints. This is improved through an exact characterization of the requirements for a sound sliding window approach. The tools developed in this paper are especially beneficial if Convolutional Neural Networks are employed, but can also be used as a more general framework to validate related approaches to signal scanning. The proposed theory helps to eliminate redundant computations and renders special case treatment unnecessary, resulting in a dramatic boost in efficiency particularly on massively parallel processors. This is demonstrated both theoretically in a computational complexity analysis and empirically on modern parallel processors.
Convolutional Neural Networks for patient-specific ECG classification.
Kiranyaz, Serkan; Ince, Turker; Hamila, Ridha; Gabbouj, Moncef
2015-01-01
We propose a fast and accurate patient-specific electrocardiogram (ECG) classification and monitoring system using an adaptive implementation of 1D Convolutional Neural Networks (CNNs) that can fuse feature extraction and classification into a unified learner. In this way, a dedicated CNN will be trained for each patient by using relatively small common and patient-specific training data and thus it can also be used to classify long ECG records such as Holter registers in a fast and accurate manner. Alternatively, such a solution can conveniently be used for real-time ECG monitoring and early alert system on a light-weight wearable device. The experimental results demonstrate that the proposed system achieves a superior classification performance for the detection of ventricular ectopic beats (VEB) and supraventricular ectopic beats (SVEB).
Plane-wave decomposition by spherical-convolution microphone array
Rafaely, Boaz; Park, Munhum
2004-05-01
Reverberant sound fields are widely studied, as they have a significant influence on the acoustic performance of enclosures in a variety of applications. For example, the intelligibility of speech in lecture rooms, the quality of music in auditoria, the noise level in offices, and the production of 3D sound in living rooms are all affected by the enclosed sound field. These sound fields are typically studied through frequency response measurements or statistical measures such as reverberation time, which do not provide detailed spatial information. The aim of the work presented in this seminar is the detailed analysis of reverberant sound fields. A measurement and analysis system based on acoustic theory and signal processing, designed around a spherical microphone array, is presented. Detailed analysis is achieved by decomposition of the sound field into waves, using spherical Fourier transform and spherical convolution. The presentation will include theoretical review, simulation studies, and initial experimental results.
Enhanced Line Integral Convolution with Flow Feature Detection
Lane, David; Okada, Arthur
1996-01-01
The Line Integral Convolution (LIC) method, which blurs white noise textures along a vector field, is an effective way to visualize overall flow patterns in a 2D domain. The method produces a flow texture image based on the input velocity field defined in the domain. Because of the nature of the algorithm, the texture image tends to be blurry. This sometimes makes it difficult to identify boundaries where flow separation and reattachments occur. We present techniques to enhance LIC texture images and use colored texture images to highlight flow separation and reattachment boundaries. Our techniques have been applied to several flow fields defined in 3D curvilinear multi-block grids and scientists have found the results to be very useful.
Plant species classification using deep convolutional neural network
DEFF Research Database (Denmark)
Dyrmann, Mads; Karstoft, Henrik; Midtiby, Henrik Skov
2016-01-01
Information on which weed species are present within agricultural fields is important for site specific weed management. This paper presents a method that is capable of recognising plant species in colour images by using a convolutional neural network. The network is built from scratch trained...... and tested on a total of 10,413 images containing 22 weed and crop species at early growth stages. These images originate from six different data sets, which have variations with respect to lighting, resolution, and soil type. This includes images taken under controlled conditions with regard to camera...... stabilisation and illumination, and images shot with hand-held mobile phones in fields with changing lighting conditions and different soil types. For these 22 species, the network is able to achieve a classification accuracy of 86.2%....
Robust Fusion of Irregularly Sampled Data Using Adaptive Normalized Convolution
Directory of Open Access Journals (Sweden)
Schutte Klamer
2006-01-01
Full Text Available We present a novel algorithm for image fusion from irregularly sampled data. The method is based on the framework of normalized convolution (NC, in which the local signal is approximated through a projection onto a subspace. The use of polynomial basis functions in this paper makes NC equivalent to a local Taylor series expansion. Unlike the traditional framework, however, the window function of adaptive NC is adapted to local linear structures. This leads to more samples of the same modality being gathered for the analysis, which in turn improves signal-to-noise ratio and reduces diffusion across discontinuities. A robust signal certainty is also adapted to the sample intensities to minimize the influence of outliers. Excellent fusion capability of adaptive NC is demonstrated through an application of super-resolution image reconstruction.
Modeling Task fMRI Data via Deep Convolutional Autoencoder.
Huang, Heng; Hu, Xintao; Zhao, Yu; Makkie, Milad; Dong, Qinglin; Zhao, Shijie; Guo, Lei; Liu, Tianming
2017-06-15
Task-based fMRI (tfMRI) has been widely used to study functional brain networks under task performance. Modeling tfMRI data is challenging due to at least two problems: the lack of the ground truth of underlying neural activity and the highly complex intrinsic structure of tfMRI data. To better understand brain networks based on fMRI data, data-driven approaches have been proposed, for instance, Independent Component Analysis (ICA) and Sparse Dictionary Learning (SDL). However, both ICA and SDL only build shallow models, and they are under the strong assumption that original fMRI signal could be linearly decomposed into time series components with their corresponding spatial maps. As growing evidence shows that human brain function is hierarchically organized, new approaches that can infer and model the hierarchical structure of brain networks are widely called for. Recently, deep convolutional neural network (CNN) has drawn much attention, in that deep CNN has proven to be a powerful method for learning high-level and mid-level abstractions from low-level raw data. Inspired by the power of deep CNN, in this study, we developed a new neural network structure based on CNN, called Deep Convolutional Auto-Encoder (DCAE), in order to take the advantages of both data-driven approach and CNN's hierarchical feature abstraction ability for the purpose of learning mid-level and high-level features from complex, large-scale tfMRI time series in an unsupervised manner. The DCAE has been applied and tested on the publicly available human connectome project (HCP) tfMRI datasets, and promising results are achieved.
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.
He, Kaiming; Zhang, Xiangyu; Ren, Shaoqing; Sun, Jian
2015-09-01
Existing deep convolutional neural networks (CNNs) require a fixed-size (e.g., 224 × 224) input image. This requirement is "artificial" and may reduce the recognition accuracy for the images or sub-images of an arbitrary size/scale. In this work, we equip the networks with another pooling strategy, "spatial pyramid pooling", to eliminate the above requirement. The new network structure, called SPP-net, can generate a fixed-length representation regardless of image size/scale. Pyramid pooling is also robust to object deformations. With these advantages, SPP-net should in general improve all CNN-based image classification methods. On the ImageNet 2012 dataset, we demonstrate that SPP-net boosts the accuracy of a variety of CNN architectures despite their different designs. On the Pascal VOC 2007 and Caltech101 datasets, SPP-net achieves state-of-the-art classification results using a single full-image representation and no fine-tuning. The power of SPP-net is also significant in object detection. Using SPP-net, we compute the feature maps from the entire image only once, and then pool features in arbitrary regions (sub-images) to generate fixed-length representations for training the detectors. This method avoids repeatedly computing the convolutional features. In processing test images, our method is 24-102 × faster than the R-CNN method, while achieving better or comparable accuracy on Pascal VOC 2007. In ImageNet Large Scale Visual Recognition Challenge (ILSVRC) 2014, our methods rank #2 in object detection and #3 in image classification among all 38 teams. This manuscript also introduces the improvement made for this competition.
Cantilever tilt causing amplitude related convolution in dynamic mode atomic force microscopy.
Wang, Chunmei; Sun, Jielin; Itoh, Hiroshi; Shen, Dianhong; Hu, Jun
2011-01-01
It is well known that the topography in atomic force microscopy (AFM) is a convolution of the tip's shape and the sample's geometry. The classical convolution model was established in contact mode assuming a static probe, but it is no longer valid in dynamic mode AFM. It is still not well understood whether or how the vibration of the probe in dynamic mode affects the convolution. Such ignorance complicates the interpretation of the topography. Here we propose a convolution model for dynamic mode by taking into account the typical design of the cantilever tilt in AFMs, which leads to a different convolution from that in contact mode. Our model indicates that the cantilever tilt results in a dynamic convolution affected by the absolute value of the amplitude, especially in the case that corresponding contact convolution has sharp edges beyond certain angle. The effect was experimentally demonstrated by a perpendicular SiO(2)/Si super-lattice structure. Our model is useful for quantitative characterizations in dynamic mode, especially in probe characterization and critical dimension measurements.
Convoluted cells as a marker for maternal cell contamination in CVS cultures
DEFF Research Database (Denmark)
Hertz, Jens Michael; Jensen, P K; Therkelsen, A J
1987-01-01
In order to identify cells of maternal origin in CVS cultures, tissue from 1st trimester abortions were cultivated and the cultures stained in situ for X-chromatin. Convoluted cells and maternal fibroblasts were found to be positive. By chromosome analysis of cultures from 105 diagnostic placenta...... biopsies, obtained by the transabdominal route, metaphases of maternal origin were found in nine cases. In eight of these cases colonies of convoluted cells were observed. We conclude that convoluted cells are of maternal origin and are a reliable marker for maternal cell contamination in CVS cultures....
SU-E-T-536: Inhomogeneity Correction in Planning of Gamma Knife Treatments for Acoustic Schwannoma
Energy Technology Data Exchange (ETDEWEB)
Lu, L [Ohio State Univ, Columbus, OH (United States); Gupta, N [Ohio State University, Columbus, OH (United States); Hessler, J [University of Cincinnati, Cincinnati, OH (United States); Liu, A [Upper Arlington High School, Columbus, OH (United States); Weldon, M [Ohio State University, Columbus, OH (United States); McGregor, J [Ohio State University, Columbus, OH (United States); Ammirati, M [Ohio State University, Columbus, OH (United States); Guiou, M [Ohio State University, Columbus, OH (United States); Xia, F [Ohio State University, Columbus, OH (United States); Grecula, J [Ohio State University, Columbus, OH (United States)
2014-06-01
Purpose: To find out the dose difference on targets and organs at risk for the treatment of acoustic schwannoma if the inhomogeneity correction (Convolution algorithm) is applied. Methods: Images of patients treated for acoustic schwannoma with Gamma Knife using TMR 10 algorithm were retrieved from database and replanned with Convolution and TMR 10 algorithm respectively. These patients were treated using a preplan scheme in following: (1) Before the actual treatment day, using the MRI image that was taken without a head frame on the patient's skull, a pre-treatment plan was made based on the default skull coordinates in the Gamma Knife treatment planning system (LGP); (2) then on treatment day, a head frame was placed on the patient's skull, and a CT image was taken. The CT image with head frame was registered and fused with the completed preplan; (3) the treatment plan was finalized and the treatment was delivered. To find out the dosimetry impact of inhomogeneity correction, we used the retrieved CT images to replan the treatment using Convolution algorithm in LGP software version 10.1.1. The dose distributions and the dose volume histograms for targets and OARs were compared for these two dose calculation algorithms. Results: The dose calculated with the Convolution algorithm in general is slightly lower than the one from TMR 10 around the boney area. The effect from the inhomogeneity correction is observable but not significant, and varies with the location of the tumor. Conclusion: Inhomogeneity correction slightly improve the dose accuracy for acoustic schwannoma Gamma Knife treatments although the correction may not be very significant. Our Result provides evidence for dose prescription adjustment to treat acoustic schwannoma. The actual clinical outcome of switching from using TMR10 to using Convolution needs to be further investigated.
Suzuki, S; Arai, H
1990-04-01
In single-photon emission computed tomography (SPECT) and X-ray CT one-dimensional (1-D) convolution method is used for their image reconstruction from projections. The method makes a 1-D convolution filtering on projection data with a 1-D filter in the space domain, and back projects the filtered data for reconstruction. Images can also be reconstructed by first forming the 2-D backprojection images from projections and then convoluting them with a 2-D space-domain filter. This is the reconstruction by the 2-D convolution method, and it has the opposite reconstruction process to the 1-D convolution method. Since the 2-D convolution method is inferior to the 1-D convolution method in speed in reconstruction, it has no practical use. In the actual reconstruction by the 2-D convolution method, convolution is made on a finite plane which is called convolution window. A convolution window of size N X N needs a 2-D discrete filter of the same size. If better reconstructions are achieved with small convolution windows, the reconstruction time for the 2-D convolution method can be reduced. For this purpose, 2-D filters of a simple function form are proposed which can give good reconstructions with small convolution windows. They are here defined on a finite plane, depending on the window size used, although a filter function is usually defined on the infinite plane. They are however set so that they better approximate the property of a 2-D filter function defined on the infinite plane. Filters of size N X N are thus determined. Their value varies with window size. The filters are applied to image reconstructions of SPECT.(ABSTRACT TRUNCATED AT 250 WORDS)
The statistical nature of the second order corrections to the thermal SZE
2004-01-01
This paper shows that the accepted expressions for the second order corrections in the parameter $z$ to the thermal Sunyaev-Zel'dovich effect can be accurately reproduced by a simple convolution integral approach. This representation allows to separate the second order SZE corrections into two type of components. One associated to a single line broadening, directly related to the even derivative terms present in the distortion intensity curve, while the other is related to a frequency shift, ...
A new convolution algorithm for loss probablity analysis in multiservice networks
DEFF Research Database (Denmark)
Huang, Qian; Ko, King-Tim; Iversen, Villy Bæk
2011-01-01
Performance analysis in multiservice loss systems generally focuses on accurate and efficient calculation methods for traffic loss probability. Convolution algorithm is one of the existing efficient numerical methods. Exact loss probabilities are obtainable from the convolution algorithm in systems...... where the bandwidth is fully shared by all traffic classes; but not available for systems with trunk reservation, i.e. part of the bandwidth is reserved for a special class of traffic. A proposal known as asymmetric convolution algorithm (ACA) has been made to overcome the deficiency of the convolution...... algorithm. It obtains an approximation of the channel occupancy distribution in multiservice systems with trunk reservation. However, the ACA approximation is only accurate with two traffic flows; increased approximation errors are observed for systems with three or more traffic flows. In this paper, we...
Time-Domain Convolutive Blind Source Separation Employing Selective-Tap Adaptive Algorithms
Directory of Open Access Journals (Sweden)
Pan Qiongfeng
2007-01-01
Full Text Available We investigate novel algorithms to improve the convergence and reduce the complexity of time-domain convolutive blind source separation (BSS algorithms. First, we propose MMax partial update time-domain convolutive BSS (MMax BSS algorithm. We demonstrate that the partial update scheme applied in the MMax LMS algorithm for single channel can be extended to multichannel time-domain convolutive BSS with little deterioration in performance and possible computational complexity saving. Next, we propose an exclusive maximum selective-tap time-domain convolutive BSS algorithm (XM BSS that reduces the interchannel coherence of the tap-input vectors and improves the conditioning of the autocorrelation matrix resulting in improved convergence rate and reduced misalignment. Moreover, the computational complexity is reduced since only half of the tap inputs are selected for updating. Simulation results have shown a significant improvement in convergence rate compared to existing techniques.
Mishchenko, Michael I.
2014-01-01
This Essay traces the centuries-long history of the phenomenological disciplines of directional radiometry and radiative transfer in turbid media, discusses their fundamental weaknesses, and outlines the convoluted process of their conversion into legitimate branches of physical optics.
ICA if fMRI based on a convolutive mixture model
DEFF Research Database (Denmark)
Hansen, Lars Kai
2003-01-01
mixing relevant for spatial ICA. Convolutive ICA has many computational problems and no standard solution is available. In this study a new predictive estimation method is used for finding the mixing coefficients and the source signals of a convolutive mixture and it is applied in temporal mode...... challenge with previous independent component analyses is the convolutive nature of the mixing process in fMRI. In temporal ICA we assume that the measured fMRI response is an instantaneous, spatially varying, mixture of independent time functions. However, the convolutive structure of the hemodynamic....... The mixing is represented by “mixture coefficient images” quantifying the local response to a given source at a certain time lag. This is the first communication to address this important issue in the context of fMRI ICA. Data: A single slice holding 128x128 pixels and passing through primary visual cortex...
Numerical simulation of seismic wave propagation in complex media by convolutional differentiator
Institute of Scientific and Technical Information of China (English)
LI Xin-fu; LI Xiao-fan
2008-01-01
We apply the forward modeling algorithm constituted by the convolutional Forsyte polynomial differentiator pro- posed by former worker to seismic wave simulation of complex heterogeneous media and compare the efficiency and accuracy between this method and other seismic simulation methods such as finite difference and pseudospec- tral method. Numerical experiments demonstrate that the algorithm constituted by convolutional Forsyte polyno- mial differentiator has high efficiency and accuracy and needs less computational resources, so it is a numerical modeling method with much potential.
Learning text representation using recurrent convolutional neural network with highway layers
Wen, Ying; Zhang, Weinan; Luo, Rui; Wang, Jun
2016-01-01
Recently, the rapid development of word embedding and neural networks has brought new inspiration to various NLP and IR tasks. In this paper, we describe a staged hybrid model combining Recurrent Convolutional Neural Networks (RCNN) with highway layers. The highway network module is incorporated in the middle takes the output of the bi-directional Recurrent Neural Network (Bi-RNN) module in the first stage and provides the Convolutional Neural Network (CNN) module in the last stage with the i...
Relative n-widths of periodic convolution classes with NCVD-kernel and B-kernel
Institute of Scientific and Technical Information of China (English)
无
2010-01-01
In this paper,we consider the relative n-widths of two kinds of periodic convolution classes,Kp(K) and Bp(G),whose convolution kernels are NCVD-kernel K and B-kernel G. The asymptotic estimations of Kn(Kp(K),Kp(K))q and Kn(Bp(G),Bp(G))q are obtained for p=1 and ∞,1≤ q≤∞.
Witoonchart, Peerajak; Chongstitvatana, Prabhas
2017-08-01
In this study, for the first time, we show how to formulate a structured support vector machine (SSVM) as two layers in a convolutional neural network, where the top layer is a loss augmented inference layer and the bottom layer is the normal convolutional layer. We show that a deformable part model can be learned with the proposed structured SVM neural network by backpropagating the error of the deformable part model to the convolutional neural network. The forward propagation calculates the loss augmented inference and the backpropagation calculates the gradient from the loss augmented inference layer to the convolutional layer. Thus, we obtain a new type of convolutional neural network called an Structured SVM convolutional neural network, which we applied to the human pose estimation problem. This new neural network can be used as the final layers in deep learning. Our method jointly learns the structural model parameters and the appearance model parameters. We implemented our method as a new layer in the existing Caffe library. Copyright © 2017 Elsevier Ltd. All rights reserved.
Examples of minimal-memory, non-catastrophic quantum convolutional encoders
Wilde, Mark M; Hosseini-Khayat, Saied
2010-01-01
One of the most important open questions in the theory of quantum convolutional coding is to determine a minimal-memory, non-catastrophic, polynomial-depth convolutional encoder for an arbitrary quantum convolutional code. Here, we present a technique that finds quantum convolutional encoders with such desirable properties for several example quantum convolutional codes (an exposition of our technique in full generality will appear elsewhere). We first show how to encode the well-studied Forney-Grassl-Guha (FGG) code with an encoder that exploits just one memory qubit (the former Grassl-Roetteler encoder requires 15 memory qubits). We then show how our technique can find an online decoder corresponding to this encoder, and we also detail the operation of our technique on a different example of a quantum convolutional code. Finally, the reduction in memory for the FGG encoder makes it feasible to simulate the performance of a quantum turbo code employing it, and we present the results of such simulations.
An efficient method for correcting the edge artifact due to smoothing.
Maisog, J M; Chmielowska, J
1998-01-01
Spatial smoothing is a common pre-processing step in the analysis of functional brain imaging data. It can increase sensitivity to signals of specific shapes and sizes (Rosenfeld and Kak [1982]: Digital Picture Processing, vol. 2. Orlando, Fla.: Academic; Worsley et al. [1996]: Hum Brain Mapping 4:74-90). Also, some amount of spatial smoothness is required if methods from the theory of Gaussian random fields are to be used (Holmes [1994]: Statistical Issues in Functional Brain Mapping. PhD thesis, University of Glasgow). Smoothing is most often implemented as a convolution of the imaging data with a smoothing kernel, and convolution is most efficiently performed using the Convolution Theorem and the Fast Fourier Transform (Cooley and Tukey [1965]: Math Comput 19:297-301; Priestly [1981]: Spectral Analysis and Time Series. San Diego: Academic; Press et al. [1992]: Numerical Recipes in C: The Art of Scientific Computing, 2nd ed. Cambridge: Cambridge University Press). An undesirable side effect of smoothing is an artifact along the edges of the brain, where brain voxels become smoothed with non-brain voxels. This results in a dark rim which might be mistaken for hypoactivity. In this short methodological paper, we present a method for correcting functional brain images for the edge artifact due to smoothing, while retaining the use of the Convolution Theorem and the Fast Fourier Transform for efficient calculation of convolutions.
A deep convolutional neural network for recognizing foods
Jahani Heravi, Elnaz; Habibi Aghdam, Hamed; Puig, Domenec
2015-12-01
Controlling the food intake is an efficient way that each person can undertake to tackle the obesity problem in countries worldwide. This is achievable by developing a smartphone application that is able to recognize foods and compute their calories. State-of-art methods are chiefly based on hand-crafted feature extraction methods such as HOG and Gabor. Recent advances in large-scale object recognition datasets such as ImageNet have revealed that deep Convolutional Neural Networks (CNN) possess more representation power than the hand-crafted features. The main challenge with CNNs is to find the appropriate architecture for each problem. In this paper, we propose a deep CNN which consists of 769; 988 parameters. Our experiments show that the proposed CNN outperforms the state-of-art methods and improves the best result of traditional methods 17%. Moreover, using an ensemble of two CNNs that have been trained two different times, we are able to improve the classification performance 21:5%.
A quantum algorithm for Viterbi decoding of classical convolutional codes
Grice, Jon R.; Meyer, David A.
2015-07-01
We present a quantum Viterbi algorithm (QVA) with better than classical performance under certain conditions. In this paper, the proposed algorithm is applied to decoding classical convolutional codes, for instance, large constraint length and short decode frames . Other applications of the classical Viterbi algorithm where is large (e.g., speech processing) could experience significant speedup with the QVA. The QVA exploits the fact that the decoding trellis is similar to the butterfly diagram of the fast Fourier transform, with its corresponding fast quantum algorithm. The tensor-product structure of the butterfly diagram corresponds to a quantum superposition that we show can be efficiently prepared. The quantum speedup is possible because the performance of the QVA depends on the fanout (number of possible transitions from any given state in the hidden Markov model) which is in general much less than . The QVA constructs a superposition of states which correspond to all legal paths through the decoding lattice, with phase as a function of the probability of the path being taken given received data. A specialized amplitude amplification procedure is applied one or more times to recover a superposition where the most probable path has a high probability of being measured.
Predicting Semantic Descriptions from Medical Images with Convolutional Neural Networks.
Schlegl, Thomas; Waldstein, Sebastian M; Vogl, Wolf-Dieter; Schmidt-Erfurth, Ursula; Langs, Georg
2015-01-01
Learning representative computational models from medical imaging data requires large training data sets. Often, voxel-level annotation is unfeasible for sufficient amounts of data. An alternative to manual annotation, is to use the enormous amount of knowledge encoded in imaging data and corresponding reports generated during clinical routine. Weakly supervised learning approaches can link volume-level labels to image content but suffer from the typical label distributions in medical imaging data where only a small part consists of clinically relevant abnormal structures. In this paper we propose to use a semantic representation of clinical reports as a learning target that is predicted from imaging data by a convolutional neural network. We demonstrate how we can learn accurate voxel-level classifiers based on weak volume-level semantic descriptions on a set of 157 optical coherence tomography (OCT) volumes. We specifically show how semantic information increases classification accuracy for intraretinal cystoid fluid (IRC), subretinal fluid (SRF) and normal retinal tissue, and how the learning algorithm links semantic concepts to image content and geometry.
Classification and Segmentation of Satellite Orthoimagery Using Convolutional Neural Networks
Directory of Open Access Journals (Sweden)
Martin Längkvist
2016-04-01
Full Text Available The availability of high-resolution remote sensing (HRRS data has opened up the possibility for new interesting applications, such as per-pixel classification of individual objects in greater detail. This paper shows how a convolutional neural network (CNN can be applied to multispectral orthoimagery and a digital surface model (DSM of a small city for a full, fast and accurate per-pixel classification. The predicted low-level pixel classes are then used to improve the high-level segmentation. Various design choices of the CNN architecture are evaluated and analyzed. The investigated land area is fully manually labeled into five categories (vegetation, ground, roads, buildings and water, and the classification accuracy is compared to other per-pixel classification works on other land areas that have a similar choice of categories. The results of the full classification and segmentation on selected segments of the map show that CNNs are a viable tool for solving both the segmentation and object recognition task for remote sensing data.
Protein Secondary Structure Prediction Using Deep Convolutional Neural Fields
Wang, Sheng; Peng, Jian; Ma, Jianzhu; Xu, Jinbo
2016-01-01
Protein secondary structure (SS) prediction is important for studying protein structure and function. When only the sequence (profile) information is used as input feature, currently the best predictors can obtain ~80% Q3 accuracy, which has not been improved in the past decade. Here we present DeepCNF (Deep Convolutional Neural Fields) for protein SS prediction. DeepCNF is a Deep Learning extension of Conditional Neural Fields (CNF), which is an integration of Conditional Random Fields (CRF) and shallow neural networks. DeepCNF can model not only complex sequence-structure relationship by a deep hierarchical architecture, but also interdependency between adjacent SS labels, so it is much more powerful than CNF. Experimental results show that DeepCNF can obtain ~84% Q3 accuracy, ~85% SOV score, and ~72% Q8 accuracy, respectively, on the CASP and CAMEO test proteins, greatly outperforming currently popular predictors. As a general framework, DeepCNF can be used to predict other protein structure properties such as contact number, disorder regions, and solvent accessibility.
Building Extraction from Remote Sensing Data Using Fully Convolutional Networks
Bittner, K.; Cui, S.; Reinartz, P.
2017-05-01
Building detection and footprint extraction are highly demanded for many remote sensing applications. Though most previous works have shown promising results, the automatic extraction of building footprints still remains a nontrivial topic, especially in complex urban areas. Recently developed extensions of the CNN framework made it possible to perform dense pixel-wise classification of input images. Based on these abilities we propose a methodology, which automatically generates a full resolution binary building mask out of a Digital Surface Model (DSM) using a Fully Convolution Network (FCN) architecture. The advantage of using the depth information is that it provides geometrical silhouettes and allows a better separation of buildings from background as well as through its invariance to illumination and color variations. The proposed framework has mainly two steps. Firstly, the FCN is trained on a large set of patches consisting of normalized DSM (nDSM) as inputs and available ground truth building mask as target outputs. Secondly, the generated predictions from FCN are viewed as unary terms for a Fully connected Conditional Random Fields (FCRF), which enables us to create a final binary building mask. A series of experiments demonstrate that our methodology is able to extract accurate building footprints which are close to the buildings original shapes to a high degree. The quantitative and qualitative analysis show the significant improvements of the results in contrast to the multy-layer fully connected network from our previous work.
BUILDING EXTRACTION FROM REMOTE SENSING DATA USING FULLY CONVOLUTIONAL NETWORKS
Directory of Open Access Journals (Sweden)
K. Bittner
2017-05-01
Full Text Available Building detection and footprint extraction are highly demanded for many remote sensing applications. Though most previous works have shown promising results, the automatic extraction of building footprints still remains a nontrivial topic, especially in complex urban areas. Recently developed extensions of the CNN framework made it possible to perform dense pixel-wise classification of input images. Based on these abilities we propose a methodology, which automatically generates a full resolution binary building mask out of a Digital Surface Model (DSM using a Fully Convolution Network (FCN architecture. The advantage of using the depth information is that it provides geometrical silhouettes and allows a better separation of buildings from background as well as through its invariance to illumination and color variations. The proposed framework has mainly two steps. Firstly, the FCN is trained on a large set of patches consisting of normalized DSM (nDSM as inputs and available ground truth building mask as target outputs. Secondly, the generated predictions from FCN are viewed as unary terms for a Fully connected Conditional Random Fields (FCRF, which enables us to create a final binary building mask. A series of experiments demonstrate that our methodology is able to extract accurate building footprints which are close to the buildings original shapes to a high degree. The quantitative and qualitative analysis show the significant improvements of the results in contrast to the multy-layer fully connected network from our previous work.
On the Relationship between Visual Attributes and Convolutional Networks
Castillo, Victor
2015-06-02
One of the cornerstone principles of deep models is their abstraction capacity, i.e. their ability to learn abstract concepts from ‘simpler’ ones. Through extensive experiments, we characterize the nature of the relationship between abstract concepts (specifically objects in images) learned by popular and high performing convolutional networks (conv-nets) and established mid-level representations used in computer vision (specifically semantic visual attributes). We focus on attributes due to their impact on several applications, such as object description, retrieval and mining, and active (and zero-shot) learning. Among the findings we uncover, we show empirical evidence of the existence of Attribute Centric Nodes (ACNs) within a conv-net, which is trained to recognize objects (not attributes) in images. These special conv-net nodes (1) collectively encode information pertinent to visual attribute representation and discrimination, (2) are unevenly and sparsely distribution across all layers of the conv-net, and (3) play an important role in conv-net based object recognition.
Innervation of the renal proximal convoluted tubule of the rat
Energy Technology Data Exchange (ETDEWEB)
Barajas, L.; Powers, K. (Harbor-UCLA Medical Center, Torrance (USA))
1989-12-01
Experimental data suggest the proximal tubule as a major site of neurogenic influence on tubular function. The functional and anatomical axial heterogeneity of the proximal tubule prompted this study of the distribution of innervation sites along the early, mid, and late proximal convoluted tubule (PCT) of the rat. Serial section autoradiograms, with tritiated norepinephrine serving as a marker for monoaminergic nerves, were used in this study. Freehand clay models and graphic reconstructions of proximal tubules permitted a rough estimation of the location of the innervation sites along the PCT. In the subcapsular nephrons, the early PCT (first third) was devoid of innervation sites with most of the innervation occurring in the mid (middle third) and in the late (last third) PCT. Innervation sites were found in the early PCT in nephrons located deeper in the cortex. In juxtamedullary nephrons, innervation sites could be observed on the PCT as it left the glomerulus. This gradient of PCT innervation can be explained by the different tubulovascular relationships of nephrons at different levels of the cortex. The absence of innervation sites in the early PCT of subcapsular nephrons suggests that any influence of the renal nerves on the early PCT might be due to an effect of neurotransmitter released from renal nerves reaching the early PCT via the interstitium and/or capillaries.
Image reconstruction from incomplete convolution data via total variation regularization
Directory of Open Access Journals (Sweden)
Zhida Shen
2015-02-01
Full Text Available Variational models with Total Variation (TV regularization have long been known to preserve image edges and produce high quality reconstruction. On the other hand, recent theory on compressive sensing has shown that it is feasible to accurately reconstruct images from a few linear measurements via TV regularization. However, in general TV models are difficult to solve due to the nondifferentiability and the universal coupling of variables. In this paper, we propose the use of alternating direction method for image reconstruction from highly incomplete convolution data, where an image is reconstructed as a minimizer of an energy function that sums a TV term for image regularity and a least squares term for data fitting. Our algorithm, called RecPK, takes advantage of problem structures and has an extremely low per-iteration cost. To demonstrate the efficiency of RecPK, we compare it with TwIST, a state-of-the-art algorithm for minimizing TV models. Moreover, we also demonstrate the usefulness of RecPK in image zooming.
Convolutional networks for fast, energy-efficient neuromorphic computing.
Esser, Steven K; Merolla, Paul A; Arthur, John V; Cassidy, Andrew S; Appuswamy, Rathinakumar; Andreopoulos, Alexander; Berg, David J; McKinstry, Jeffrey L; Melano, Timothy; Barch, Davis R; di Nolfo, Carmelo; Datta, Pallab; Amir, Arnon; Taba, Brian; Flickner, Myron D; Modha, Dharmendra S
2016-10-11
Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that (i) approach state-of-the-art classification accuracy across eight standard datasets encompassing vision and speech, (ii) perform inference while preserving the hardware's underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1,200 and 2,600 frames/s and using between 25 and 275 mW (effectively >6,000 frames/s per Watt), and (iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. This approach allows the algorithmic power of deep learning to be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer.
Method for Veterbi decoding of large constraint length convolutional codes
Hsu, In-Shek; Truong, Trieu-Kie; Reed, Irving S.; Jing, Sun
1988-05-01
A new method of Viterbi decoding of convolutional codes lends itself to a pipline VLSI architecture using a single sequential processor to compute the path metrics in the Viterbi trellis. An array method is used to store the path information for NK intervals where N is a number, and K is constraint length. The selected path at the end of each NK interval is then selected from the last entry in the array. A trace-back method is used for returning to the beginning of the selected path back, i.e., to the first time unit of the interval NK to read out the stored branch metrics of the selected path which correspond to the message bits. The decoding decision made in this way is no longer maximum likelihood, but can be almost as good, provided that constraint length K in not too small. The advantage is that for a long message, it is not necessary to provide a large memory to store the trellis derived information until the end of the message to select the path that is to be decoded; the selection is made at the end of every NK time unit, thus decoding a long message in successive blocks.
Gearbox Fault Identification and Classification with Convolutional Neural Networks
Directory of Open Access Journals (Sweden)
ZhiQiang Chen
2015-01-01
Full Text Available Vibration signals of gearbox are sensitive to the existence of the fault. Based on vibration signals, this paper presents an implementation of deep learning algorithm convolutional neural network (CNN used for fault identification and classification in gearboxes. Different combinations of condition patterns based on some basic fault conditions are considered. 20 test cases with different combinations of condition patterns are used, where each test case includes 12 combinations of different basic condition patterns. Vibration signals are preprocessed using statistical measures from the time domain signal such as standard deviation, skewness, and kurtosis. In the frequency domain, the spectrum obtained with FFT is divided into multiple bands, and the root mean square (RMS value is calculated for each one so the energy maintains its shape at the spectrum peaks. The achieved accuracy indicates that the proposed approach is highly reliable and applicable in fault diagnosis of industrial reciprocating machinery. Comparing with peer algorithms, the present method exhibits the best performance in the gearbox fault diagnosis.
Classification of breast cancer cytological specimen using convolutional neural network
Żejmo, Michał; Kowal, Marek; Korbicz, Józef; Monczak, Roman
2017-01-01
The paper presents a deep learning approach for automatic classification of breast tumors based on fine needle cytology. The main aim of the system is to distinguish benign from malignant cases based on microscopic images. Experiment was carried out on cytological samples derived from 50 patients (25 benign cases + 25 malignant cases) diagnosed in Regional Hospital in Zielona Góra. To classify microscopic images, we used convolutional neural networks (CNN) of two types: GoogLeNet and AlexNet. Due to the very large size of images of cytological specimen (on average 200000 × 100000 pixels), they were divided into smaller patches of size 256 × 256 pixels. Breast cancer classification usually is based on morphometric features of nuclei. Therefore, training and validation patches were selected using Support Vector Machine (SVM) so that suitable amount of cell material was depicted. Neural classifiers were tuned using GPU accelerated implementation of gradient descent algorithm. Training error was defined as a cross-entropy classification loss. Classification accuracy was defined as the percentage ratio of successfully classified validation patches to the total number of validation patches. The best accuracy rate of 83% was obtained by GoogLeNet model. We observed that more misclassified patches belong to malignant cases.
Infimal convolution of total generalized variation functionals for dynamic MRI.
Schloegl, Matthias; Holler, Martin; Schwarzl, Andreas; Bredies, Kristian; Stollberger, Rudolf
2017-07-01
To accelerate dynamic MR applications using infimal convolution of total generalized variation functionals (ICTGV) as spatio-temporal regularization for image reconstruction. ICTGV comprises a new image prior tailored to dynamic data that achieves regularization via optimal local balancing between spatial and temporal regularity. Here it is applied for the first time to the reconstruction of dynamic MRI data. CINE and perfusion scans were investigated to study the influence of time dependent morphology and temporal contrast changes. ICTGV regularized reconstruction from subsampled MR data is formulated as a convex optimization problem. Global solutions are obtained by employing a duality based non-smooth optimization algorithm. The reconstruction error remains on a low level with acceleration factors up to 16 for both CINE and dynamic contrast-enhanced MRI data. The GPU implementation of the algorithm suites clinical demands by reducing reconstruction times of one dataset to less than 4 min. ICTGV based dynamic magnetic resonance imaging reconstruction allows for vast undersampling and therefore enables for very high spatial and temporal resolutions, spatial coverage and reduced scan time. With the proposed distinction of model and regularization parameters it offers a new and robust method of flexible decomposition into components with different degrees of temporal regularity. Magn Reson Med 78:142-155, 2017. © 2016 International Society for Magnetic Resonance in Medicine. © 2016 International Society for Magnetic Resonance in Medicine.
Multi-modal vertebrae recognition using Transformed Deep Convolution Network.
Cai, Yunliang; Landis, Mark; Laidley, David T; Kornecki, Anat; Lum, Andrea; Li, Shuo
2016-07-01
Automatic vertebra recognition, including the identification of vertebra locations and naming in multiple image modalities, are highly demanded in spinal clinical diagnoses where large amount of imaging data from various of modalities are frequently and interchangeably used. However, the recognition is challenging due to the variations of MR/CT appearances or shape/pose of the vertebrae. In this paper, we propose a method for multi-modal vertebra recognition using a novel deep learning architecture called Transformed Deep Convolution Network (TDCN). This new architecture can unsupervisely fuse image features from different modalities and automatically rectify the pose of vertebra. The fusion of MR and CT image features improves the discriminativity of feature representation and enhances the invariance of the vertebra pattern, which allows us to automatically process images from different contrast, resolution, protocols, even with different sizes and orientations. The feature fusion and pose rectification are naturally incorporated in a multi-layer deep learning network. Experiment results show that our method outperforms existing detection methods and provides a fully automatic location+naming+pose recognition for routine clinical practice. Copyright © 2016 Elsevier Ltd. All rights reserved.
Synthesising Primary Reflections by Marchenko Redatuming and Convolutional Interferometry
Curtis, A.
2015-12-01
Standard active-source seismic processing and imaging steps such as velocity analysis and reverse time migration usually provide best results when all reflected waves in the input data are primaries (waves that reflect only once). Multiples (recorded waves that reflect multiple times) represent a source of coherent noise in data that must be suppressed to avoid imaging artefacts. Consequently, multiple-removal methods have been a primcipal direction of active-source seismic research for decades. We describe a new method to estimate primaries directly, which obviates the need for multiple removal. Primaries are constructed within convolutional interferometry by combining first arriving events of up-going and direct wave down-going Green's functions to virtual receivers in the subsurface. The required up-going wavefields to virtual receivers along discrete subsurface boundaries can be constructed using Marchenko redatuming. Crucially, this is possible without detailed models of the Earth's subsurface velocity structure: similarly to most migration techniques, the method only requires surface reflection data and estimates of direct (non-reflected) arrivals between subsurface sources and the acquisition surface. The method is demonstrated on a stratified synclinal model. It is shown both to improve reverse time migration compared to standard methods, and to be particularly robust against errors in the reference velocity model used.
New Parallel Interference Cancellation for Convolutionally Coded CDMA Systems
Institute of Scientific and Technical Information of China (English)
Xu Guo-xiong; Gan Liang-cai; Huang Tian-xi
2004-01-01
Based on BCJR algorithm proposed by Bahl et al and linear soft decision feedback, a reduced-complexity parallel interference cancellation (simplified PIC) for convolutionally coded DS CDMA systems is proposed. By computer simulation, we compare the simplified PIC with the exact PIC. It shows that the simplified PIC can achieve the performance close to the exact PIC if the mean values of coded symbols are linearly computed in terms of the sum of initial a prior log-likelihood rate (LLR) and updated a prior LLR, while a significant performance loss will occur if the mean values of coded symbols are linearly computed in terms of the updated a prior LLR only. Meanwhile, we also compare the simplified PIC with MF receiver and conventional PICs. The simulation results show that the simplified PIC dominantly outperforms the MF receiver and conventional PICs, at signal-noise rate (SNR) of 7 dB, for example, the bit error rate is about 10-4 for the simplified PIC, which is far below that of matched-filter receiver and conventional PIC.
Toward an optimal convolutional neural network for traffic sign recognition
Habibi Aghdam, Hamed; Jahani Heravi, Elnaz; Puig, Domenec
2015-12-01
Convolutional Neural Networks (CNN) beat the human performance on German Traffic Sign Benchmark competition. Both the winner and the runner-up teams trained CNNs to recognize 43 traffic signs. However, both networks are not computationally efficient since they have many free parameters and they use highly computational activation functions. In this paper, we propose a new architecture that reduces the number of the parameters 27% and 22% compared with the two networks. Furthermore, our network uses Leaky Rectified Linear Units (ReLU) as the activation function that only needs a few operations to produce the result. Specifically, compared with the hyperbolic tangent and rectified sigmoid activation functions utilized in the two networks, Leaky ReLU needs only one multiplication operation which makes it computationally much more efficient than the two other functions. Our experiments on the Gertman Traffic Sign Benchmark dataset shows 0:6% improvement on the best reported classification accuracy while it reduces the overall number of parameters 85% compared with the winner network in the competition.
Thomaz, Ricardo L.; Carneiro, Pedro C.; Patrocinio, Ana C.
2017-03-01
Breast cancer is the leading cause of death for women in most countries. The high levels of mortality relate mostly to late diagnosis and to the direct proportionally relationship between breast density and breast cancer development. Therefore, the correct assessment of breast density is important to provide better screening for higher risk patients. However, in modern digital mammography the discrimination among breast densities is highly complex due to increased contrast and visual information for all densities. Thus, a computational system for classifying breast density might be a useful tool for aiding medical staff. Several machine-learning algorithms are already capable of classifying small number of classes with good accuracy. However, machinelearning algorithms main constraint relates to the set of features extracted and used for classification. Although well-known feature extraction techniques might provide a good set of features, it is a complex task to select an initial set during design of a classifier. Thus, we propose feature extraction using a Convolutional Neural Network (CNN) for classifying breast density by a usual machine-learning classifier. We used 307 mammographic images downsampled to 260x200 pixels to train a CNN and extract features from a deep layer. After training, the activation of 8 neurons from a deep fully connected layer are extracted and used as features. Then, these features are feedforward to a single hidden layer neural network that is cross-validated using 10-folds to classify among four classes of breast density. The global accuracy of this method is 98.4%, presenting only 1.6% of misclassification. However, the small set of samples and memory constraints required the reuse of data in both CNN and MLP-NN, therefore overfitting might have influenced the results even though we cross-validated the network. Thus, although we presented a promising method for extracting features and classifying breast density, a greater database is
QingJie, Wei; WenBin, Wang
2017-06-01
In this paper, the image retrieval using deep convolutional neural network combined with regularization and PRelu activation function is studied, and improves image retrieval accuracy. Deep convolutional neural network can not only simulate the process of human brain to receive and transmit information, but also contains a convolution operation, which is very suitable for processing images. Using deep convolutional neural network is better than direct extraction of image visual features for image retrieval. However, the structure of deep convolutional neural network is complex, and it is easy to over-fitting and reduces the accuracy of image retrieval. In this paper, we combine L1 regularization and PRelu activation function to construct a deep convolutional neural network to prevent over-fitting of the network and improve the accuracy of image retrieval
A convolution-superposition dose calculation engine for GPUs
Energy Technology Data Exchange (ETDEWEB)
Hissoiny, Sami; Ozell, Benoit; Despres, Philippe [Departement de genie informatique et genie logiciel, Ecole polytechnique de Montreal, 2500 Chemin de Polytechnique, Montreal, Quebec H3T 1J4 (Canada); Departement de radio-oncologie, CRCHUM-Centre hospitalier de l' Universite de Montreal, 1560 rue Sherbrooke Est, Montreal, Quebec H2L 4M1 (Canada)
2010-03-15
Purpose: Graphic processing units (GPUs) are increasingly used for scientific applications, where their parallel architecture and unprecedented computing power density can be exploited to accelerate calculations. In this paper, a new GPU implementation of a convolution/superposition (CS) algorithm is presented. Methods: This new GPU implementation has been designed from the ground-up to use the graphics card's strengths and to avoid its weaknesses. The CS GPU algorithm takes into account beam hardening, off-axis softening, kernel tilting, and relies heavily on raytracing through patient imaging data. Implementation details are reported as well as a multi-GPU solution. Results: An overall single-GPU acceleration factor of 908x was achieved when compared to a nonoptimized version of the CS algorithm implemented in PlanUNC in single threaded central processing unit (CPU) mode, resulting in approximatively 2.8 s per beam for a 3D dose computation on a 0.4 cm grid. A comparison to an established commercial system leads to an acceleration factor of approximately 29x or 0.58 versus 16.6 s per beam in single threaded mode. An acceleration factor of 46x has been obtained for the total energy released per mass (TERMA) calculation and a 943x acceleration factor for the CS calculation compared to PlanUNC. Dose distributions also have been obtained for a simple water-lung phantom to verify that the implementation gives accurate results. Conclusions: These results suggest that GPUs are an attractive solution for radiation therapy applications and that careful design, taking the GPU architecture into account, is critical in obtaining significant acceleration factors. These results potentially can have a significant impact on complex dose delivery techniques requiring intensive dose calculations such as intensity-modulated radiation therapy (IMRT) and arc therapy. They also are relevant for adaptive radiation therapy where dose results must be obtained rapidly.
A convolutional neural network approach for objective video quality assessment.
Le Callet, Patrick; Viard-Gaudin, Christian; Barba, Dominique
2006-09-01
This paper describes an application of neural networks in the field of objective measurement method designed to automatically assess the perceived quality of digital videos. This challenging issue aims to emulate human judgment and to replace very complex and time consuming subjective quality assessment. Several metrics have been proposed in literature to tackle this issue. They are based on a general framework that combines different stages, each of them addressing complex problems. The ambition of this paper is not to present a global perfect quality metric but rather to focus on an original way to use neural networks in such a framework in the context of reduced reference (RR) quality metric. Especially, we point out the interest of such a tool for combining features and pooling them in order to compute quality scores. The proposed approach solves some problems inherent to objective metrics that should predict subjective quality score obtained using the single stimulus continuous quality evaluation (SSCQE) method. This latter has been adopted by video quality expert group (VQEG) in its recently finalized reduced referenced and no reference (RRNR-TV) test plan. The originality of such approach compared to previous attempts to use neural networks for quality assessment, relies on the use of a convolutional neural network (CNN) that allows a continuous time scoring of the video. Objective features are extracted on a frame-by-frame basis on both the reference and the distorted sequences; they are derived from a perceptual-based representation and integrated along the temporal axis using a time-delay neural network (TDNN). Experiments conducted on different MPEG-2 videos, with bit rates ranging 2-6 Mb/s, show the effectiveness of the proposed approach to get a plausible model of temporal pooling from the human vision system (HVS) point of view. More specifically, a linear correlation criteria, between objective and subjective scoring, up to 0.92 has been obtained on
Text-Attentional Convolutional Neural Network for Scene Text Detection
He, Tong; Huang, Weilin; Qiao, Yu; Yao, Jian
2016-06-01
Recent deep learning models have demonstrated strong capabilities for classifying text and non-text components in natural images. They extract a high-level feature computed globally from a whole image component (patch), where the cluttered background information may dominate true text features in the deep representation. This leads to less discriminative power and poorer robustness. In this work, we present a new system for scene text detection by proposing a novel Text-Attentional Convolutional Neural Network (Text-CNN) that particularly focuses on extracting text-related regions and features from the image components. We develop a new learning mechanism to train the Text-CNN with multi-level and rich supervised information, including text region mask, character label, and binary text/nontext information. The rich supervision information enables the Text-CNN with a strong capability for discriminating ambiguous texts, and also increases its robustness against complicated background components. The training process is formulated as a multi-task learning problem, where low-level supervised information greatly facilitates main task of text/non-text classification. In addition, a powerful low-level detector called Contrast- Enhancement Maximally Stable Extremal Regions (CE-MSERs) is developed, which extends the widely-used MSERs by enhancing intensity contrast between text patterns and background. This allows it to detect highly challenging text patterns, resulting in a higher recall. Our approach achieved promising results on the ICDAR 2013 dataset, with a F-measure of 0.82, improving the state-of-the-art results substantially.
Deep convolutional networks for pancreas segmentation in CT imaging
Roth, Holger R.; Farag, Amal; Lu, Le; Turkbey, Evrim B.; Summers, Ronald M.
2015-03-01
Automatic organ segmentation is an important prerequisite for many computer-aided diagnosis systems. The high anatomical variability of organs in the abdomen, such as the pancreas, prevents many segmentation methods from achieving high accuracies when compared to state-of-the-art segmentation of organs like the liver, heart or kidneys. Recently, the availability of large annotated training sets and the accessibility of affordable parallel computing resources via GPUs have made it feasible for "deep learning" methods such as convolutional networks (ConvNets) to succeed in image classification tasks. These methods have the advantage that used classification features are trained directly from the imaging data. We present a fully-automated bottom-up method for pancreas segmentation in computed tomography (CT) images of the abdomen. The method is based on hierarchical coarse-to-fine classification of local image regions (superpixels). Superpixels are extracted from the abdominal region using Simple Linear Iterative Clustering (SLIC). An initial probability response map is generated, using patch-level confidences and a two-level cascade of random forest classifiers, from which superpixel regions with probabilities larger 0.5 are retained. These retained superpixels serve as a highly sensitive initial input of the pancreas and its surroundings to a ConvNet that samples a bounding box around each superpixel at different scales (and random non-rigid deformations at training time) in order to assign a more distinct probability of each superpixel region being pancreas or not. We evaluate our method on CT images of 82 patients (60 for training, 2 for validation, and 20 for testing). Using ConvNets we achieve maximum Dice scores of an average 68% +/- 10% (range, 43-80%) in testing. This shows promise for accurate pancreas segmentation, using a deep learning approach and compares favorably to state-of-the-art methods.
Convolutional neural networks for prostate cancer recurrence prediction
Kumar, Neeraj; Verma, Ruchika; Arora, Ashish; Kumar, Abhay; Gupta, Sanchit; Sethi, Amit; Gann, Peter H.
2017-03-01
Accurate prediction of the treatment outcome is important for cancer treatment planning. We present an approach to predict prostate cancer (PCa) recurrence after radical prostatectomy using tissue images. We used a cohort whose case vs. control (recurrent vs. non-recurrent) status had been determined using post-treatment follow up. Further, to aid the development of novel biomarkers of PCa recurrence, cases and controls were paired based on matching of other predictive clinical variables such as Gleason grade, stage, age, and race. For this cohort, tissue resection microarray with up to four cores per patient was available. The proposed approach is based on deep learning, and its novelty lies in the use of two separate convolutional neural networks (CNNs) - one to detect individual nuclei even in the crowded areas, and the other to classify them. To detect nuclear centers in an image, the first CNN predicts distance transform of the underlying (but unknown) multi-nuclear map from the input HE image. The second CNN classifies the patches centered at nuclear centers into those belonging to cases or controls. Voting across patches extracted from image(s) of a patient yields the probability of recurrence for the patient. The proposed approach gave 0.81 AUC for a sample of 30 recurrent cases and 30 non-recurrent controls, after being trained on an independent set of 80 case-controls pairs. If validated further, such an approach might help in choosing between a combination of treatment options such as active surveillance, radical prostatectomy, radiation, and hormone therapy. It can also generalize to the prediction of treatment outcomes in other cancers.
Classifying Radio Galaxies with the Convolutional Neural Network
Aniyan, A. K.; Thorat, K.
2017-06-01
We present the application of a deep machine learning technique to classify radio images of extended sources on a morphological basis using convolutional neural networks (CNN). In this study, we have taken the case of the Fanaroff-Riley (FR) class of radio galaxies as well as radio galaxies with bent-tailed morphology. We have used archival data from the Very Large Array (VLA)—Faint Images of the Radio Sky at Twenty Centimeters survey and existing visually classified samples available in the literature to train a neural network for morphological classification of these categories of radio sources. Our training sample size for each of these categories is ˜200 sources, which has been augmented by rotated versions of the same. Our study shows that CNNs can classify images of the FRI and FRII and bent-tailed radio galaxies with high accuracy (maximum precision at 95%) using well-defined samples and a “fusion classifier,” which combines the results of binary classifications, while allowing for a mechanism to find sources with unusual morphologies. The individual precision is highest for bent-tailed radio galaxies at 95% and is 91% and 75% for the FRI and FRII classes, respectively, whereas the recall is highest for FRI and FRIIs at 91% each, while the bent-tailed class has a recall of 79%. These results show that our results are comparable to that of manual classification, while being much faster. Finally, we discuss the computational and data-related challenges associated with the morphological classification of radio galaxies with CNNs.
Text-Attentional Convolutional Neural Network for Scene Text Detection.
He, Tong; Huang, Weilin; Qiao, Yu; Yao, Jian
2016-06-01
Recent deep learning models have demonstrated strong capabilities for classifying text and non-text components in natural images. They extract a high-level feature globally computed from a whole image component (patch), where the cluttered background information may dominate true text features in the deep representation. This leads to less discriminative power and poorer robustness. In this paper, we present a new system for scene text detection by proposing a novel text-attentional convolutional neural network (Text-CNN) that particularly focuses on extracting text-related regions and features from the image components. We develop a new learning mechanism to train the Text-CNN with multi-level and rich supervised information, including text region mask, character label, and binary text/non-text information. The rich supervision information enables the Text-CNN with a strong capability for discriminating ambiguous texts, and also increases its robustness against complicated background components. The training process is formulated as a multi-task learning problem, where low-level supervised information greatly facilitates the main task of text/non-text classification. In addition, a powerful low-level detector called contrast-enhancement maximally stable extremal regions (MSERs) is developed, which extends the widely used MSERs by enhancing intensity contrast between text patterns and background. This allows it to detect highly challenging text patterns, resulting in a higher recall. Our approach achieved promising results on the ICDAR 2013 data set, with an F-measure of 0.82, substantially improving the state-of-the-art results.
Text-Attentional Convolutional Neural Networks for Scene Text Detection.
He, Tong; Huang, Weilin; Qiao, Yu; Yao, Jian
2016-03-28
Recent deep learning models have demonstrated strong capabilities for classifying text and non-text components in natural images. They extract a high-level feature computed globally from a whole image component (patch), where the cluttered background information may dominate true text features in the deep representation. This leads to less discriminative power and poorer robustness. In this work, we present a new system for scene text detection by proposing a novel Text-Attentional Convolutional Neural Network (Text-CNN) that particularly focuses on extracting text-related regions and features from the image components. We develop a new learning mechanism to train the Text-CNN with multi-level and rich supervised information, including text region mask, character label, and binary text/nontext information. The rich supervision information enables the Text-CNN with a strong capability for discriminating ambiguous texts, and also increases its robustness against complicated background components. The training process is formulated as a multi-task learning problem, where low-level supervised information greatly facilitates main task of text/non-text classification. In addition, a powerful low-level detector called Contrast- Enhancement Maximally Stable Extremal Regions (CE-MSERs) is developed, which extends the widely-used MSERs by enhancing intensity contrast between text patterns and background. This allows it to detect highly challenging text patterns, resulting in a higher recall. Our approach achieved promising results on the ICDAR 2013 dataset, with a F-measure of 0.82, improving the state-of-the-art results substantially.
Single-trial EEG RSVP classification using convolutional neural networks
Shamwell, Jared; Lee, Hyungtae; Kwon, Heesung; Marathe, Amar R.; Lawhern, Vernon; Nothwang, William
2016-05-01
Traditionally, Brain-Computer Interfaces (BCI) have been explored as a means to return function to paralyzed or otherwise debilitated individuals. An emerging use for BCIs is in human-autonomy sensor fusion where physiological data from healthy subjects is combined with machine-generated information to enhance the capabilities of artificial systems. While human-autonomy fusion of physiological data and computer vision have been shown to improve classification during visual search tasks, to date these approaches have relied on separately trained classification models for each modality. We aim to improve human-autonomy classification performance by developing a single framework that builds codependent models of human electroencephalograph (EEG) and image data to generate fused target estimates. As a first step, we developed a novel convolutional neural network (CNN) architecture and applied it to EEG recordings of subjects classifying target and non-target image presentations during a rapid serial visual presentation (RSVP) image triage task. The low signal-to-noise ratio (SNR) of EEG inherently limits the accuracy of single-trial classification and when combined with the high dimensionality of EEG recordings, extremely large training sets are needed to prevent overfitting and achieve accurate classification from raw EEG data. This paper explores a new deep CNN architecture for generalized multi-class, single-trial EEG classification across subjects. We compare classification performance from the generalized CNN architecture trained across all subjects to the individualized XDAWN, HDCA, and CSP neural classifiers which are trained and tested on single subjects. Preliminary results show that our CNN meets and slightly exceeds the performance of the other classifiers despite being trained across subjects.
Output-sensitive 3D line integral convolution.
Falk, Martin; Weiskopf, Daniel
2008-01-01
We propose an output-sensitive visualization method for 3D line integral convolution (LIC) whose rendering speed is largely independent of the data set size and mostly governed by the complexity of the output on the image plane. Our approach of view-dependent visualization tightly links the LIC generation with the volume rendering of the LIC result in order to avoid the computation of unnecessary LIC points: early-ray termination and empty-space leaping techniques are used to skip the computation of the LIC integral in a lazy-evaluation approach; both ray casting and texture slicing can be used as volume-rendering techniques. The input noise is modeled in object space to allow for temporal coherence under object and camera motion. Different noise models are discussed, covering dense representations based on filtered white noise all the way to sparse representations similar to oriented LIC. Aliasing artifacts are avoided by frequency control over the 3D noise and by employing a 3D variant of MIPmapping. A range of illumination models is applied to the LIC streamlines: different codimension-2 lighting models and a novel gradient-based illumination model that relies on precomputed gradients and does not require any direct calculation of gradients after the LIC integral is evaluated. We discuss the issue of proper sampling of the LIC and volume-rendering integrals by employing a frequency-space analysis of the noise model and the precomputed gradients. Finally, we demonstrate that our visualization approach lends itself to a fast graphics processing unit (GPU) implementation that supports both steady and unsteady flow. Therefore, this 3D LIC method allows users to interactively explore 3D flow by means of high-quality, view-dependent, and adaptive LIC volume visualization. Applications to flow visualization in combination with feature extraction and focus-and-context visualization are described, a comparison to previous methods is provided, and a detailed performance
A fast convolution-based methodology to simulate 2-D/3-D cardiac ultrasound images.
Gao, Hang; Choi, Hon Fai; Claus, Piet; Boonen, Steven; Jaecques, Siegfried; Van Lenthe, G Harry; Van der Perre, Georges; Lauriks, Walter; D'hooge, Jan
2009-02-01
This paper describes a fast convolution-based methodology for simulating ultrasound images in a 2-D/3-D sector format as typically used in cardiac ultrasound. The conventional convolution model is based on the assumption of a space-invariant point spread function (PSF) and typically results in linear images. These characteristics are not representative for cardiac data sets. The spatial impulse response method (IRM) has excellent accuracy in the linear domain; however, calculation time can become an issue when scatterer numbers become significant and when 3-D volumetric data sets need to be computed. As a solution to these problems, the current manuscript proposes a new convolution-based methodology in which the data sets are produced by reducing the conventional 2-D/3-D convolution model to multiple 1-D convolutions (one for each image line). As an example, simulated 2-D/3-D phantom images are presented along with their gray scale histogram statistics. In addition, the computation time is recorded and contrasted to a commonly used implementation of IRM (Field II). It is shown that COLE can produce anatomically plausible images with local Rayleigh statistics but at improved calculation time (1200 times faster than the reference method).
Data convolution and combination operation (COCOA) for motion ghost artifacts reduction.
Huang, Feng; Lin, Wei; Börnert, Peter; Li, Yu; Reykowski, Arne
2010-07-01
A novel method, data convolution and combination operation, is introduced for the reduction of ghost artifacts due to motion or flow during data acquisition. Since neighboring k-space data points from different coil elements have strong correlations, a new "synthetic" k-space with dispersed motion artifacts can be generated through convolution for each coil. The corresponding convolution kernel can be self-calibrated using the acquired k-space data. The synthetic and the acquired data sets can be checked for consistency to identify k-space areas that are motion corrupted. Subsequently, these two data sets can be combined appropriately to produce a k-space data set showing a reduced level of motion induced error. If the acquired k-space contains isolated error, the error can be completely eliminated through data convolution and combination operation. If the acquired k-space data contain widespread errors, the application of the convolution also significantly reduces the overall error. Results with simulated and in vivo data demonstrate that this self-calibrated method robustly reduces ghost artifacts due to swallowing, breathing, or blood flow, with a minimum impact on the image signal-to-noise ratio. (c) 2010 Wiley-Liss, Inc.
Chacko, Nikhil; Liebling, Michael; Blu, Thierry
2013-10-01
Discretization of continuous (analog) convolution operators by direct sampling of the convolution kernel and use of fast Fourier transforms is highly efficient. However, it assumes the input and output signals are band-limited, a condition rarely met in practice, where signals have finite support or abrupt edges and sampling is nonideal. Here, we propose to approximate signals in analog, shift-invariant function spaces, which do not need to be band-limited, resulting in discrete coefficients for which we derive discrete convolution kernels that accurately model the analog convolution operator while taking into account nonideal sampling devices (such as finite fill-factor cameras). This approach retains the efficiency of direct sampling but not its limiting assumption. We propose fast forward and inverse algorithms that handle finite-length, periodic, and mirror-symmetric signals with rational sampling rates. We provide explicit convolution kernels for computing coherent wave propagation in the context of digital holography. When compared to band-limited methods in simulations, our method leads to fewer reconstruction artifacts when signals have sharp edges or when using nonideal sampling devices.
Iterative sinc-convolution method for solving planar D-bar equation with application to EIT.
Abbasi, Mahdi; Naghsh-Nilchi, Ahmad-Reza
2012-08-01
The numerical solution of D-bar integral equations is the key in inverse scattering solution of many complex problems in science and engineering including conductivity imaging. Recently, a couple of methodologies were considered for the numerical solution of D-bar integral equation, namely product integrals and multigrid. The first one involves high computational complexity and other one has low convergence rate disadvantages. In this paper, a new and efficient sinc-convolution algorithm is introduced to solve the two-dimensional D-bar integral equation to overcome both of these disadvantages and to resolve the singularity problem not tackled before effectively. The method of sinc-convolution is based on using collocation to replace multidimensional convolution-form integrals- including the two-dimensional D-bar integral equations - by a system of algebraic equations. Separation of variables in the proposed method allows elimination of the formulation of the huge full matrices and therefore reduces the computational complexity drastically. In addition, the sinc-convolution method converges exponentially with a convergence rate of O(e-cN). Simulation results on solving a test electrical impedance tomography problem confirm the efficiency of the proposed sinc-convolution-based algorithm. Copyright © 2012 John Wiley & Sons, Ltd.
1980-10-01
shortcut is available; note that on the right-hand side of Equation (26) the first term leads to Eular Convolution and the second to Mean Value...Convolution. Eular Convolution and Mean Value Convolution are just special cases of R-K(2,a) Convolution (see Table 2). TABLE 2. SPECIAL CASES OF R-K(2,a)C...Convolution Eular 0 Mean Value for 1/2 1/2 Trapezoidal I For a single real pole filter, F(s) - 1 (28) and any input, G(s), the approximation using R-K(2
Convolutional Sparse Coding for Static and Dynamic Images Analysis
Directory of Open Access Journals (Sweden)
B. A. Knyazev
2014-01-01
Full Text Available The objective of this work is to improve performance of static and dynamic objects recognition. For this purpose a new image representation model and a transformation algorithm are proposed. It is examined and illustrated that limitations of previous methods make it difficult to achieve this objective. Static images, specifically handwritten digits of the widely used MNIST dataset, are the primary focus of this work. Nevertheless, preliminary qualitative results of image sequences analysis based on the suggested model are presented.A general analytical form of the Gabor function, often employed to generate filters, is described and discussed. In this research, this description is required for computing parameters of responses returned by our algorithm. The recursive convolution operator is introduced, which allows extracting free shape features of visual objects. The developed parametric representation model is compared with sparse coding based on energy function minimization.In the experimental part of this work, errors of estimating the parameters of responses are determined. Also, parameters statistics and their correlation coefficients for more than 106 responses extracted from the MNIST dataset are calculated. It is demonstrated that these data correspond well with previous research studies on Gabor filters as well as with works on visual cortex primary cells of mammals, in which similar responses were observed. A comparative test of the developed model with three other approaches is conducted; speed and accuracy scores of handwritten digits classification are presented. A support vector machine with a linear or radial basic function is used for classification of images and their representations while principal component analysis is used in some cases to prepare data beforehand. High accuracy is not attained due to the specific difficulties of combining our model with a support vector machine (a 3.99% error rate. However, another method is
Deep Manifold Learning Combined With Convolutional Neural Networks for Action Recognition.
Chen, Xin; Weng, Jian; Lu, Wei; Xu, Jiaming; Weng, Jiasi
2017-09-15
Learning deep representations have been applied in action recognition widely. However, there have been a few investigations on how to utilize the structural manifold information among different action videos to enhance the recognition accuracy and efficiency. In this paper, we propose to incorporate the manifold of training samples into deep learning, which is defined as deep manifold learning (DML). The proposed DML framework can be adapted to most existing deep networks to learn more discriminative features for action recognition. When applied to a convolutional neural network, DML embeds the previous convolutional layer's manifold into the next convolutional layer; thus, the discriminative capacity of the next layer can be promoted. We also apply the DML on a restricted Boltzmann machine, which can alleviate the overfitting problem. Experimental results on four standard action databases (i.e., UCF101, HMDB51, KTH, and UCF sports) show that the proposed method outperforms the state-of-the-art methods.
Image retrieval method based on metric learning for convolutional neural network
Wang, Jieyuan; Qian, Ying; Ye, Qingqing; Wang, Biao
2017-09-01
At present, the research of content-based image retrieval (CBIR) focuses on learning effective feature for the representations of origin images and similarity measures. The retrieval accuracy and efficiency are crucial to a CBIR. With the rise of deep learning, convolutional network is applied in the domain of image retrieval and achieved remarkable results, but the image visual feature extraction of convolutional neural network exist high dimension problems, this problem makes the image retrieval and speed ineffective. This paper uses the metric learning for the image visual features extracted from the convolutional neural network, decreased the feature redundancy, improved the retrieval performance. The work in this paper is also a necessary part for further implementation of feature hashing to the approximate-nearest-neighbor (ANN) retrieval method.
Wu, Xuecheng; Wu, Yingchun; Yang, Jing; Wang, Zhihua; Zhou, Binwu; Gréhan, Gérard; Cen, Kefa
2013-05-20
Application of the modified convolution method to reconstruct digital inline holography of particle illuminated by an elliptical Gaussian beam is investigated. Based on the analysis on the formation of particle hologram using the Collins formula, the convolution method is modified to compensate the astigmatism by adding two scaling factors. Both simulated and experimental holograms of transparent droplets and opaque particles are used to test the algorithm, and the reconstructed images are compared with that using FRFT reconstruction. Results show that the modified convolution method can accurately reconstruct the particle image. This method has an advantage that the reconstructed images in different depth positions have the same size and resolution with the hologram. This work shows that digital inline holography has great potential in particle diagnostics in curvature containers.
Perez-Carrasco, Jose Antonio; Acha, Begona; Serrano, Carmen; Camunas-Mesa, Luis; Serrano-Gotarredona, Teresa; Linares-Barranco, Bernabe
2010-04-01
Address-event representation (AER) is an emergent hardware technology which shows a high potential for providing in the near future a solid technological substrate for emulating brain-like processing structures. When used for vision, AER sensors and processors are not restricted to capturing and processing still image frames, as in commercial frame-based video technology, but sense and process visual information in a pixel-level event-based frameless manner. As a result, vision processing is practically simultaneous to vision sensing, since there is no need to wait for sensing full frames. Also, only meaningful information is sensed, communicated, and processed. Of special interest for brain-like vision processing are some already reported AER convolutional chips, which have revealed a very high computational throughput as well as the possibility of assembling large convolutional neural networks in a modular fashion. It is expected that in a near future we may witness the appearance of large scale convolutional neural networks with hundreds or thousands of individual modules. In the meantime, some research is needed to investigate how to assemble and configure such large scale convolutional networks for specific applications. In this paper, we analyze AER spiking convolutional neural networks for texture recognition hardware applications. Based on the performance figures of already available individual AER convolution chips, we emulate large scale networks using a custom made event-based behavioral simulator. We have developed a new event-based processing architecture that emulates with AER hardware Manjunath's frame-based feature recognition software algorithm, and have analyzed its performance using our behavioral simulator. Recognition rate performance is not degraded. However, regarding speed, we show that recognition can be achieved before an equivalent frame is fully sensed and transmitted.
Change of Scale Formulas for Wiener Integrals Related to Fourier-Feynman Transform and Convolution
Directory of Open Access Journals (Sweden)
Bong Jin Kim
2014-01-01
Full Text Available Cameron and Storvick discovered change of scale formulas for Wiener integrals of functionals in Banach algebra S on classical Wiener space. Yoo and Skoug extended these results for functionals in the Fresnel class F(B and in a generalized Fresnel class FA1,A2 on abstract Wiener space. We express Fourier-Feynman transform and convolution product of functionals in S as limits of Wiener integrals. Moreover we obtain change of scale formulas for Wiener integrals related to Fourier-Feynman transform and convolution product of these functionals.
A convolutional learning system for object classification in 3-D Lidar data.
Prokhorov, Danil
2010-05-01
In this brief, a convolutional learning system for classification of segmented objects represented in 3-D as point clouds of laser reflections is proposed. Several novelties are discussed: (1) extension of the existing convolutional neural network (CNN) framework to direct processing of 3-D data in a multiview setting which may be helpful for rotation-invariant consideration, (2) improvement of CNN training effectiveness by employing a stochastic meta-descent (SMD) method, and (3) combination of unsupervised and supervised training for enhanced performance of CNN. CNN performance is illustrated on a two-class data set of objects in a segmented outdoor environment.
Pindza, Edson; Maré, Eben
2017-03-01
A modified discrete singular convolution method is proposed. The method is based on the single (SE) and double (DE) exponential transformation to speed up the convergence of the existing methods. Numerical computations are performed on a wide variety of singular boundary value and singular perturbed problems in one and two dimensions. The obtained results from discrete singular convolution methods based on single and double exponential transformations are compared with each other, and with the existing methods too. Numerical results confirm that these methods are considerably efficient and accurate in solving singular and regular problems. Moreover, the method can be applied to a wide class of nonlinear partial differential equations.
Convolutional neural networks applied to neutrino events in a liquid argon time projection chamber
Energy Technology Data Exchange (ETDEWEB)
Acciarri, R.; Adams, C.; An, R.; Asaadi, J.; Auger, M.; Bagby, L.; Baller, B.; Barr, G.; Bass, M.; Bay, F.; Bishai, M.; Blake, A.; Bolton, T.; Bugel, L.; Camilleri, L.; Caratelli, D.; Carls, B.; Fernandez, R. Castillo; Cavanna, F.; Chen, H.; Church, E.; Cianci, D.; Collin, G. H.; Conrad, J. M.; Convery, M.; Crespo-Anad?n, J. I.; Del Tutto, M.; Devitt, D.; Dytman, S.; Eberly, B.; Ereditato, A.; Sanchez, L. Escudero; Esquivel, J.; Fleming, B. T.; Foreman, W.; Furmanski, A. P.; Garvey, G. T.; Genty, V.; Goeldi, D.; Gollapinni, S.; Graf, N.; Gramellini, E.; Greenlee, H.; Grosso, R.; Guenette, R.; Hackenburg, A.; Hamilton, P.; Hen, O.; Hewes, J.; Hill, C.; Ho, J.; Horton-Smith, G.; James, C.; de Vries, J. Jan; Jen, C. -M.; Jiang, L.; Johnson, R. A.; Jones, B. J. P.; Joshi, J.; Jostlein, H.; Kaleko, D.; Karagiorgi, G.; Ketchum, W.; Kirby, B.; Kirby, M.; Kobilarcik, T.; Kreslo, I.; Laube, A.; Li, Y.; Lister, A.; Littlejohn, B. R.; Lockwitz, S.; Lorca, D.; Louis, W. C.; Luethi, M.; Lundberg, B.; Luo, X.; Marchionni, A.; Mariani, C.; Marshall, J.; Caicedo, D. A. Martinez; Meddage, V.; Miceli, T.; Mills, G. B.; Moon, J.; Mooney, M.; Moore, C. D.; Mousseau, J.; Murrells, R.; Naples, D.; Nienaber, P.; Nowak, J.; Palamara, O.; Paolone, V.; Papavassiliou, V.; Pate, S. F.; Pavlovic, Z.; Porzio, D.; Pulliam, G.; Qian, X.; Raaf, J. L.; Rafique, A.; Rochester, L.; von Rohr, C. Rudolf; Russell, B.; Schmitz, D. W.; Schukraft, A.; Seligman, W.; Shaevitz, M. H.; Sinclair, J.; Snider, E. L.; Soderberg, M.; S?ldner-Rembold, S.; Soleti, S. R.; Spentzouris, P.; Spitz, J.; St. John, J.; Strauss, T.; Szelc, A. M.; Tagg, N.; Terao, K.; Thomson, M.; Toups, M.; Tsai, Y. -T.; Tufanli, S.; Usher, T.; Van de Water, R. G.; Viren, B.; Weber, M.; Weston, J.; Wickremasinghe, D. A.; Wolbers, S.; Wongjirad, T.; Woodruff, K.; Yang, T.; Zeller, G. P.; Zennamo, J.; Zhang, C.
2017-03-01
We present several studies of convolutional neural networks applied to data coming from the MicroBooNE detector, a liquid argon time projection chamber (LArTPC). The algorithms studied include the classification of single particle images, the localization of single particle and neutrino interactions in an image, and the detection of a simulated neutrino event overlaid with cosmic ray backgrounds taken from real detector data. These studies demonstrate the potential of convolutional neural networks for particle identification or event detection on simulated neutrino interactions. We also address technical issues that arise when applying this technique to data from a large LArTPC at or near ground level.
Meda Inequality for Rearrangements of the Convolution on the Heisenberg Group and Some Applications
Directory of Open Access Journals (Sweden)
V. S. Guliyev
2009-01-01
Full Text Available The Meda inequality for rearrangements of the convolution operator on the Heisenberg group ℍn is proved. By using the Meda inequality, an O'Neil-type inequality for the convolution is obtained. As applications of these results, some sufficient and necessary conditions for the boundedness of the fractional maximal operator MΩ,α and fractional integral operator IΩ,α with rough kernels in the spaces Lp(ℍn are found. Finally, we give some comments on the extension of our results to the case of homogeneous groups.
Convolutional Neural Networks Applied to Neutrino Events in a Liquid Argon Time Projection Chamber
Acciarri, R; An, R; Asaadi, J; Auger, M; Bagby, L; Baller, B; Barr, G; Bass, M; Bay, F; Bishai, M; Blake, A; Bolton, T; Bugel, L; Camilleri, L; Caratelli, D; Carls, B; Fernandez, R Castillo; Cavanna, F; Chen, H; Church, E; Cianci, D; Collin, G H; Conrad, J M; Convery, M; Crespo-Anadón, J I; Del Tutto, M; Devitt, D; Dytman, S; Eberly, B; Ereditato, A; Sanchez, L Escudero; Esquivel, J; Fleming, B T; Foreman, W; Furmanski, A P; Garvey, G T; Genty, V; Goeldi, D; Gollapinni, S; Graf, N; Gramellini, E; Greenlee, H; Grosso, R; Guenette, R; Hackenburg, A; Hamilton, P; Hen, O; Hewes, J; Hill, C; Ho, J; Horton-Smith, G; James, C; de Vries, J Jan; Jen, C -M; Jiang, L; Johnson, R A; Jones, B J P; Joshi, J; Jostlein, H; Kaleko, D; Karagiorgi, G; Ketchum, W; Kirby, B; Kirby, M; Kobilarcik, T; Kreslo, I; Laube, A; Li, Y; Lister, A; Littlejohn, B R; Lockwitz, S; Lorca, D; Louis, W C; Luethi, M; Lundberg, B; Luo, X; Marchionni, A; Mariani, C; Marshall, J; Caicedo, D A Martinez; Meddage, V; Miceli, T; Mills, G B; Moon, J; Mooney, M; Moore, C D; Mousseau, J; Murrells, R; Naples, D; Nienaber, P; Nowak, J; Palamara, O; Paolone, V; Papavassiliou, V; Pate, S F; Pavlovic, Z; Porzio, D; Pulliam, G; Qian, X; Raaf, J L; Rafique, A; Rochester, L; von Rohr, C Rudolf; Russell, B; Schmitz, D W; Schukraft, A; Seligman, W; Shaevitz, M H; Sinclair, J; Snider, E L; Soderberg, M; Söldner-Rembold, S; Soleti, S R; Spentzouris, P; Spitz, J; John, J St; Strauss, T; Szelc, A M; Tagg, N; Terao, K; Thomson, M; Toups, M; Tsai, Y -T; Tufanli, S; Usher, T; Van de Water, R G; Viren, B; Weber, M; Weston, J; Wickremasinghe, D A; Wolbers, S; Wongjirad, T; Woodruff, K; Yang, T; Zeller, G P; Zennamo, J; Zhang, C
2016-01-01
We present several studies of convolutional neural networks applied to data coming from the MicroBooNE detector, a liquid argon time projection chamber (LArTPC). The algorithms studied include the classification of single particle images, the localization of single particle and neutrino interactions in an image, and the detection of a simulated neutrino event overlaid with cosmic ray backgrounds taken from real detector data. These studies demonstrate the potential of convolutional neural networks for particle identification or event detection on simulated neutrino interactions. We also address technical issues that arise when applying this technique to data from a large LArTPC at or near ground level.
Nascov, Victor; Logofătu, Petre Cătălin
2009-08-01
We describe a fast computational algorithm able to evaluate the Rayleigh-Sommerfeld diffraction formula, based on a special formulation of the convolution theorem and the fast Fourier transform. What is new in our approach compared to other algorithms is the use of a more general type of convolution with a scale parameter, which allows for independent sampling intervals in the input and output computation windows. Comparison between the calculations made using our algorithm and direct numeric integration show a very good agreement, while the computation speed is increased by orders of magnitude.
Implementation of large kernel 2-D convolution in limited FPGA resource
Zhong, Sheng; Li, Yang; Yan, Luxin; Zhang, Tianxu; Cao, Zhiguo
2007-12-01
2-D Convolution is a simple mathematical operation which is fundamental to many common image processing operators. Using FPGA to implement the convolver can greatly reduce the DSP's heavy burden in signal processing. But with the limit resource the FPGA can implement a convolver with small 2-D kernel. In this paper, An FIFO type line delayer is presented to serve as the data buffer for convolution to reduce the data fetching operation. A finite state machine is applied to control the reuse of multipliers and adders arrays. With these two techniques, a resource limited FPGA can be used to implement a larger kernel convolver which is commonly used in image process systems.
Transfer Function Bounds for Partial-unit-memory Convolutional Codes Based on Reduced State Diagram
Lee, P. J.
1984-01-01
The performance of a coding system consisting of a convolutional encoder and a Viterbi decoder is analytically found by the well-known transfer function bounding technique. For the partial-unit-memory byte-oriented convolutional encoder with m sub 0 binary memory cells and (k sub 0 m sub 0) inputs, a state diagram of 2(K) (sub 0) was for the transfer function bound. A reduced state diagram of (2 (m sub 0) +1) is used for easy evaluation of transfer function bounds for partial-unit-memory codes.
Convolutional neural networks applied to neutrino events in a liquid argon time projection chamber
Acciarri, R.; Adams, C.; An, R.; Asaadi, J.; Auger, M.; Bagby, L.; Baller, B.; Barr, G.; Bass, M.; Bay, F.; Bishai, M.; Blake, A.; Bolton, T.; Bugel, L.; Camilleri, L.; Caratelli, D.; Carls, B.; Castillo Fernandez, R.; Cavanna, F.; Chen, H.; Church, E.; Cianci, D.; Collin, G. H.; Conrad, J. M.; Convery, M.; Crespo-Anadón, J. I.; Del Tutto, M.; Devitt, D.; Dytman, S.; Eberly, B.; Ereditato, A.; Escudero Sanchez, L.; Esquivel, J.; Fleming, B. T.; Foreman, W.; Furmanski, A. P.; Garvey, G. T.; Genty, V.; Goeldi, D.; Gollapinni, S.; Graf, N.; Gramellini, E.; Greenlee, H.; Grosso, R.; Guenette, R.; Hackenburg, A.; Hamilton, P.; Hen, O.; Hewes, J.; Hill, C.; Ho, J.; Horton-Smith, G.; James, C.; de Vries, J. Jan; Jen, C.-M.; Jiang, L.; Johnson, R. A.; Jones, B. J. P.; Joshi, J.; Jostlein, H.; Kaleko, D.; Karagiorgi, G.; Ketchum, W.; Kirby, B.; Kirby, M.; Kobilarcik, T.; Kreslo, I.; Laube, A.; Li, Y.; Lister, A.; Littlejohn, B. R.; Lockwitz, S.; Lorca, D.; Louis, W. C.; Luethi, M.; Lundberg, B.; Luo, X.; Marchionni, A.; Mariani, C.; Marshall, J.; Martinez Caicedo, D. A.; Meddage, V.; Miceli, T.; Mills, G. B.; Moon, J.; Mooney, M.; Moore, C. D.; Mousseau, J.; Murrells, R.; Naples, D.; Nienaber, P.; Nowak, J.; Palamara, O.; Paolone, V.; Papavassiliou, V.; Pate, S. F.; Pavlovic, Z.; Porzio, D.; Pulliam, G.; Qian, X.; Raaf, J. L.; Rafique, A.; Rochester, L.; von Rohr, C. Rudolf; Russell, B.; Schmitz, D. W.; Schukraft, A.; Seligman, W.; Shaevitz, M. H.; Sinclair, J.; Snider, E. L.; Soderberg, M.; Söldner-Rembold, S.; Soleti, S. R.; Spentzouris, P.; Spitz, J.; St. John, J.; Strauss, T.; Szelc, A. M.; Tagg, N.; Terao, K.; Thomson, M.; Toups, M.; Tsai, Y.-T.; Tufanli, S.; Usher, T.; Van de Water, R. G.; Viren, B.; Weber, M.; Weston, J.; Wickremasinghe, D. A.; Wolbers, S.; Wongjirad, T.; Woodruff, K.; Yang, T.; Zeller, G. P.; Zennamo, J.; Zhang, C.
2017-03-01
We present several studies of convolutional neural networks applied to data coming from the MicroBooNE detector, a liquid argon time projection chamber (LArTPC). The algorithms studied include the classification of single particle images, the localization of single particle and neutrino interactions in an image, and the detection of a simulated neutrino event overlaid with cosmic ray backgrounds taken from real detector data. These studies demonstrate the potential of convolutional neural networks for particle identification or event detection on simulated neutrino interactions. We also address technical issues that arise when applying this technique to data from a large LArTPC at or near ground level.
Anatomically informed convolution kernels for the projection of fMRI data on the cortical surface.
Operto, Grégory; Bulot, Rémy; Anton, Jean-Luc; Coulon, Olivier
2006-01-01
We present here a method that aims at producing representations of functional brain data on the cortical surface from functional MRI volumes. Such representations are required for subsequent cortical-based functional analysis. We propose a projection technique based on the definition, around each node of the grey/white matter interface mesh, of convolution kernels whose shape and distribution rely on the geometry of the local anatomy. For one anatomy, a set of convolution kernels is computed that can be used to project any functional data registered with this anatomy. The method is presented together with experiments on synthetic data and real statistical t-maps.
Patient-specific dosimetry based on quantitative SPECT imaging and 3D-DFT convolution
Energy Technology Data Exchange (ETDEWEB)
Akabani, G.; Hawkins, W.G.; Eckblade, M.B.; Leichner, P.K. [Univ. of Nebraska Medical Center, Omaha, NE (United States)
1999-01-01
The objective of this study was to validate the use of a 3-D discrete Fourier Transform (3D-DFT) convolution method to carry out the dosimetry for I-131 for soft tissues in radioimmunotherapy procedures. To validate this convolution method, mathematical and physical phantoms were used as a basis of comparison with Monte Carlo transport (MCT) calculations which were carried out using the EGS4 system code. The mathematical phantom consisted of a sphere containing uniform and nonuniform activity distributions. The physical phantom consisted of a cylinder containing uniform and nonuniform activity distributions. Quantitative SPECT reconstruction was carried out using the Circular Harmonic Transform (CHT) algorithm.
Positive convolution structure for a class of Heckman-Opdam hypergeometric functions of type BC
Rösler, Margit
2009-01-01
In this paper, we derive explicit product formulas and positive convolution structures for three continuous classes of Heckman-Opdam hypergeometric functions of type $BC$. For specific discrete series of multiplicities these hypergeometric functions occur as the spherical functions of non-compact Grassmann manifolds $G/K$ over one of the (skew) fields $\\mathbb F= \\mathbb R, \\mathbb C, \\mathbb H.$ We write the product formula of these spherical functions in an explicit form which allows analytic continuation with respect to the parameters. In each of the three cases, we obtain a series of hypergroup algebras which include the commutative convolution algebras of $K$-biinvariant functions on $G$.
Serang, Oliver
2015-08-01
Observations depending on sums of random variables are common throughout many fields; however, no efficient solution is currently known for performing max-product inference on these sums of general discrete distributions (max-product inference can be used to obtain maximum a posteriori estimates). The limiting step to max-product inference is the max-convolution problem (sometimes presented in log-transformed form and denoted as "infimal convolution," "min-convolution," or "convolution on the tropical semiring"), for which no O(k log(k)) method is currently known. Presented here is an O(k log(k)) numerical method for estimating the max-convolution of two nonnegative vectors (e.g., two probability mass functions), where k is the length of the larger vector. This numerical max-convolution method is then demonstrated by performing fast max-product inference on a convolution tree, a data structure for performing fast inference given information on the sum of n discrete random variables in O(nk log(nk)log(n)) steps (where each random variable has an arbitrary prior distribution on k contiguous possible states). The numerical max-convolution method can be applied to specialized classes of hidden Markov models to reduce the runtime of computing the Viterbi path from nk(2) to nk log(k), and has potential application to the all-pairs shortest paths problem.
Comparative evaluation of scatter correction techniques in 3D positron emission tomography
Zaidi, H
2000-01-01
Much research and development has been concentrated on the scatter compensation required for quantitative 3D PET. Increasingly sophisticated scatter correction procedures are under investigation, particularly those based on accurate scatter models, and iterative reconstruction-based scatter compensation approaches. The main difference among the correction methods is the way in which the scatter component in the selected energy window is estimated. Monte Carlo methods give further insight and might in themselves offer a possible correction procedure. Methods: Five scatter correction methods are compared in this paper where applicable. The dual-energy window (DEW) technique, the convolution-subtraction (CVS) method, two variants of the Monte Carlo-based scatter correction technique (MCBSC1 and MCBSC2) and our newly developed statistical reconstruction-based scatter correction (SRBSC) method. These scatter correction techniques are evaluated using Monte Carlo simulation studies, experimental phantom measurements...
Classifying Multi-year Land Use and Land Cover using Deep Convolutional Neural Networks
Seo, B.
2015-12-01
Cultivated ecosystems constitute a particularly frequent form of human land use. Long-term management of a cultivated ecosystem requires us to know temporal change of land use and land cover (LULC) of the target system. Land use and land cover changes (LUCC) in agricultural ecosystem is often rapid and unexpectedly occurs. Thus, longitudinal LULC is particularly needed to examine trends of ecosystem functions and ecosystem services of the target system. Multi-temporal classification of land use and land cover (LULC) in complex heterogeneous landscape remains a challenge. Agricultural landscapes often made up of a mosaic of numerous LULC classes, thus spatial heterogeneity is large. Moreover, temporal and spatial variation within a LULC class is also large. Under such a circumstance, standard classifiers would fail to identify the LULC classes correctly due to the heterogeneity of the target LULC classes. Because most standard classifiers search for a specific pattern of features for a class, they fail to detect classes with noisy and/or transformed feature data sets. Recently, deep learning algorithms have emerged in the machine learning communities and shown superior performance on a variety of tasks, including image classification and object recognition. In this paper, we propose to use convolutional neural networks (CNN) to learn from multi-spectral data to classify agricultural LULC types. Based on multi-spectral satellite data, we attempted to classify agricultural LULC classes in Soyang watershed, South Korea for the three years' study period (2009-2011). The classification performance of support vector machine (SVM) and CNN classifiers were compared for different years. Preliminary results demonstrate that the proposed method can improve classification performance compared to the SVM classifier. The SVM classifier failed to identify classes when trained on a year to predict another year, whilst CNN could reconstruct LULC maps of the catchment over the study
2015-12-15
convolution, activation functions, and pooling. For a model trained on classes, the output from the classification layer comprises + 1...Keypoint Density-based Region Proposal for Fine-Grained Object Detection and Classification using Regions with Convolutional Neural Network...Convolutional Neural Networks (CNNs) enable them to outperform conventional techniques on standard object detection and classification tasks, their
Error Analysis of Padding Schemes for DFT’s of Convolutions and Derivatives
2012-01-31
spectral techniques for geoid computations over large regions. Journal of Geodesy , 70(6), 357-373. Tziavos IN, Sideris MG, Forsberg R, Schwarz KP...362- 378. Zhang C (1995) A general formula and its inverse formula for gravimetric transformations by use of convolution and deconvolution techniques. Journal of Geodesy , 70(1-2), 51-64. 24
Shifting and Variational Properties for Fourier-Feynman Transform and Convolution
Directory of Open Access Journals (Sweden)
Byoung Soo Kim
2015-01-01
Full Text Available Shifting, scaling, modulation, and variational properties for Fourier-Feynman transform of functionals in a Banach algebra S are given. Cameron and Storvick's translation theorem can be obtained as a corollary of our result. We also study shifting, scaling, and modulation properties for the convolution product of functionals in S.
Ferre, S.; Hoenderop, J.G.J.; Bindels, R.J.M.
2011-01-01
In healthy individuals, Mg(2+) homeostasis is tightly regulated by the concerted action of intestinal absorption, exchange with bone, and renal excretion. The kidney, more precisely the distal convoluted tubule (DCT), is the final determinant of plasma Mg(2+) concentrations. Positional cloning strat
John Michael Salgado Cebola
2016-01-01
Comparative study between the performance of Convolutional Networks using pretrained models and statistical generative models on tasks of image classification in semi-supervised enviroments.Study of multiple ensembles using these techniques and generated data from estimated pdfs.Pretrained Convents, LDA, pLSA, Fisher Vectors, Sparse-coded SPMs, TSVMs being the key models worked upon.
Long-term Recurrent Convolutional Networks for Visual Recognition and Description
2014-11-17
scription task only requires a single convolutional network since the input consists of a single image. A variety of deep and multi- modal models [8...architectures for video description (see Figure 4). For each architecture, we assume we have predictions of objects, subjects, and verbs present in the video from
Strahl, Stefan B; Ramekers, Dyan; Nagelkerke, Marjolijn M B; Schwarz, Konrad E; Spitzer, Philipp; Klis, Sjaak F L; Grolman, Wilko; Versnel, Huib
2016-01-01
The electrically evoked compound action potential (eCAP) is a routinely performed measure of the auditory nerve in cochlear implant users. Using a convolution model of the eCAP, additional information about the neural firing properties can be obtained, which may provide relevant information about th
Fast 2D Convolutions and Cross-Correlations Using Scalable Architectures.
Carranza, Cesar; Llamocca, Daniel; Pattichis, Marios
2017-05-01
The manuscript describes fast and scalable architectures and associated algorithms for computing convolutions and cross-correlations. The basic idea is to map 2D convolutions and cross-correlations to a collection of 1D convolutions and cross-correlations in the transform domain. This is accomplished through the use of the discrete periodic radon transform for general kernels and the use of singular value decomposition -LU decompositions for low-rank kernels. The approach uses scalable architectures that can be fitted into modern FPGA and Zynq-SOC devices. Based on different types of available resources, for P×P blocks, 2D convolutions and cross-correlations can be computed in just O(P) clock cycles up to O(P(2)) clock cycles. Thus, there is a trade-off between performance and required numbers and types of resources. We provide implementations of the proposed architectures using modern programmable devices (Virtex-7 and Zynq-SOC). Based on the amounts and types of required resources, we show that the proposed approaches significantly outperform current methods.
New molecular players facilitating Mg(2+) reabsorption in the distal convoluted tubule.
Glaudemans, B.; Knoers, N.V.A.M.; Hoenderop, J.G.J.; Bindels, R.J.M.
2010-01-01
The renal distal convoluted tubule (DCT) has an essential role in maintaining systemic magnesium (Mg(2+)) concentration. The DCT is the final determinant of plasma Mg(2+) levels, as the more distal nephron segments are largely impermeable to Mg(2+). In the past decade, positional candidate strategie
Pawara, Pornntiwa; Okafor, Emmanuel; Surinta, Olarik; Schomaker, Lambertus; Wiering, Marco
2017-01-01
The use of machine learning and computer vision methods for recognizing different plants from images has attracted lots of attention from the community. This paper aims at comparing local feature descriptors and bags of visual words with different classifiers to deep convolutional neural networks (C
Convolutional neural network based sensor fusion for forward looking ground penetrating radar
Sakaguchi, Rayn; Crosskey, Miles; Chen, David; Walenz, Brett; Morton, Kenneth
2016-05-01
Forward looking ground penetrating radar (FLGPR) is an alternative buried threat sensing technology designed to offer additional standoff compared to downward looking GPR systems. Due to additional flexibility in antenna configurations, FLGPR systems can accommodate multiple sensor modalities on the same platform that can provide complimentary information. The different sensor modalities present challenges in both developing informative feature extraction methods, and fusing sensor information in order to obtain the best discrimination performance. This work uses convolutional neural networks in order to jointly learn features across two sensor modalities and fuse the information in order to distinguish between target and non-target regions. This joint optimization is possible by modifying the traditional image-based convolutional neural network configuration to extract data from multiple sources. The filters generated by this process create a learned feature extraction method that is optimized to provide the best discrimination performance when fused. This paper presents the results of applying convolutional neural networks and compares these results to the use of fusion performed with a linear classifier. This paper also compares performance between convolutional neural networks architectures to show the benefit of fusing the sensor information in different ways.
Yetter-Drinfel'd模与卷积Hopf模%Yetter-Drinfel'd Module and Convolution Module
Institute of Scientific and Technical Information of China (English)
张良云; 王栓宏
2002-01-01
In this paper, we first give a sufficient and necessary condition for a Hopf algebra to be a Yetter-Drinfe]'d module, and prove that the finite dual of a YetterDrinfel'd module is still a Yetter-Drinfel'd module. Finally, we introduce a concept of convolution module.
A generalized recursive convolution method for time-domain propagation in porous media.
Dragna, Didier; Pineau, Pierre; Blanc-Benon, Philippe
2015-08-01
An efficient numerical method, referred to as the auxiliary differential equation (ADE) method, is proposed to compute convolutions between relaxation functions and acoustic variables arising in sound propagation equations in porous media. For this purpose, the relaxation functions are approximated in the frequency domain by rational functions. The time variation of the convolution is thus governed by first-order differential equations which can be straightforwardly solved. The accuracy of the method is first investigated and compared to that of recursive convolution methods. It is shown that, while recursive convolution methods are first or second-order accurate in time, the ADE method does not introduce any additional error. The ADE method is then applied for outdoor sound propagation using the equations proposed by Wilson et al. in the ground [(2007). Appl. Acoust. 68, 173-200]. A first one-dimensional case is performed showing that only five poles are necessary to accurately approximate the relaxation functions for typical applications. Finally, the ADE method is used to compute sound propagation in a three-dimensional geometry over an absorbing ground. Results obtained with Wilson's equations are compared to those obtained with Zwikker and Kosten's equations and with an impedance surface for different flow resistivities.
Institute of Scientific and Technical Information of China (English)
WUMianbin; XIALiming; 等
2002-01-01
Continuous hydrolysis of chitosan was performed in a convoluted fibrous bed bioreactor (CFBB) with immobilized T. reesei. At dilution rate of 0.4d-1 and substrate concentration of 2% (mass vs. volume), the average degree of polymerization of hydrolysate can be kept at 1.25-1.35, which can be easily regulated by changing dilution rate or inlet chitosan concentration.
The beta-binomial convolution model for 2 × 2 tables with missing cell counts
Eisinga, Rob
2009-01-01
This paper considers the beta-binomial convolution model for the analysis of 2×2 tables with missing cell counts.We discuss maximumlikelihood (ML) parameter estimation using the expectation–maximization algorithm and study information loss relative to complete data estimators. We also examine bias o
A new family of windows--convolution windows and their applications
Institute of Scientific and Technical Information of China (English)
ZHANG; Jieqiu; LIANG; Changhong; CHEN; Yanpu
2005-01-01
A new family of windows is constructed by convolutions via a few rectangular windows with same time width and is thus referred to as convolution windows. The expressions of the second-order up to the eighth-order convolution windows in both the time and frequency domains are derived. Their applications in high accuracy harmonic analysis of periodic signals are investigated. Comparisons between the proposed windows and some known windows with the same width shows that, when the synchronous deviation of data sampling is slight, the proposed ones have the least effect of spectral leakage. Therefore, the new windows are well suited for high accuracy harmonic analysis and parameter estimation for periodic signals. The error analysis and computer simulations show that the estimation errors, corresponding to frequency,amplitude and phase of every harmonic component of a signal, are proportional to thepth power of the relative frequency deviation in case of the pth-order convolution window is applied to windowing signal of approximately p cycles. By introducing real time adjustment in sampling interval, the proposed algorithm can adaptively trace signal frequency and lead to less sampling synchronous deviation. The proposed approach has the advantages of easy implementation and high measure precision and can be used in harmonic analysis of quasi-periodic signals whose fundamental frequency drifts slowly with time.
Inverse Problems for a Parabolic Integrodifferential Equation in a Convolutional Weak Form
Directory of Open Access Journals (Sweden)
Kairi Kasemets
2013-01-01
Full Text Available We deduce formulas for the Fréchet derivatives of cost functionals of several inverse problems for a parabolic integrodifferential equation in a weak formulation. The method consists in the application of an integrated convolutional form of the weak problem and all computations are implemented in regular Sobolev spaces.
Schuh, Fabian
2012-01-01
In this paper we propose a matched decoding scheme for convolutionally encoded transmission over intersymbol interference (ISI) channels and devise a nonlinear trellis description. As an application we show that for coded continuous phase modulation (CPM) using a non-coherent receiver the number of states of the super trellis can be significantly reduced by means of a matched non-linear trellis encoder.
Mapping the Relationship between Cortical Convolution and Intelligence: Effects of Gender
Luders, Eileen; Narr, Katherine L.; Bilder, Robert M.; Szeszko, Philip R.; Gurbani, Mala N.; Hamilton, Liberty; Gaser, Christian
2008-01-01
The pronounced convolution of the human cortex may be a morphological substrate that supports some of our species’ most distinctive cognitive abilities. Therefore, individual intelligence within humans might be modulated by the degree of folding in certain cortical regions. We applied advanced methods to analyze cortical convolution at high spatial resolution and correlated those measurements with intelligence quotients. Within a large sample of healthy adult subjects (n = 65), we detected the most prominent correlations in the left medial hemisphere. More specifically, intelligence scores were positively associated with the degree of folding in the temporo-occipital lobe, particularly in the outermost section of the posterior cingulate gyrus (retrosplenial areas). Thus, this region might be an important contributor toward individual intelligence, either via modulating pathways to (pre)frontal regions or by serving as a location for the convergence of information. Prominent gender differences within the right frontal cortex were observed; females showed uncorrected significant positive correlations and males showed a nonsignificant trend toward negative correlations. It is possible that formerly described gender differences in regional convolution are associated with differences in the underlying architecture. This might lead to the development of sexually dimorphic information processing strategies and affect the relationship between intelligence and cortical convolution. PMID:18089578
Producing data-based sensitivity kernels from convolution and correlation in exploration geophysics.
Chmiel, M. J.; Roux, P.; Herrmann, P.; Rondeleux, B.
2016-12-01
Many studies have shown that seismic interferometry can be used to estimate surface wave arrivals by correlation of seismic signals recorded at a pair of locations. In the case of ambient noise sources, the convergence towards the surface wave Green's functions is obtained with the criterion of equipartitioned energy. However, seismic acquisition with active, controlled sources gives more possibilities when it comes to interferometry. The use of controlled sources makes it possible to recover the surface wave Green's function between two points using either correlation or convolution. We investigate the convolutional and correlational approaches using land active-seismic data from exploration geophysics. The data were recorded on 10,710 vertical receivers using 51,808 sources (seismic vibrator trucks). The sources spacing is the same in both X and Y directions (30 m) which is known as a "carpet shooting". The receivers are placed in parallel lines with a spacing 150 m in the X direction and 30 m in the Y direction. Invoking spatial reciprocity between sources and receivers, correlation and convolution functions can thus be constructed between either pairs of receivers or pairs of sources. Benefiting from the dense acquisition, we extract sensitivity kernels from correlation and convolution measurements of the seismic data. These sensitivity kernels are subsequently used to produce phase-velocity dispersion curves between two points and to separate the higher mode from the fundamental mode for surface waves. Potential application to surface wave cancellation is also envisaged.
New molecular players facilitating Mg(2+) reabsorption in the distal convoluted tubule.
Glaudemans, B.; Knoers, N.V.A.M.; Hoenderop, J.G.J.; Bindels, R.J.M.
2010-01-01
The renal distal convoluted tubule (DCT) has an essential role in maintaining systemic magnesium (Mg(2+)) concentration. The DCT is the final determinant of plasma Mg(2+) levels, as the more distal nephron segments are largely impermeable to Mg(2+). In the past decade, positional candidate strategie
Directory of Open Access Journals (Sweden)
Husan Vokhidov
2016-12-01
Full Text Available Automobile driver information as displayed on marked road signs indicates the state of the road, traffic conditions, proximity to schools, etc. These signs are important to insure the safety of the driver and pedestrians. They are also important input to the automated advanced driver assistance system (ADAS, installed in many automobiles. Over time, the arrow-road markings may be eroded or otherwise damaged by automobile contact, making it difficult for the driver to correctly identify the marking. Failure to properly identify an arrow-road marker creates a dangerous situation that may result in traffic accidents or pedestrian injury. Very little research exists that studies the problem of automated identification of damaged arrow-road marking painted on the road. In this study, we propose a method that uses a convolutional neural network (CNN to recognize six types of arrow-road markings, possibly damaged, by visible light camera sensor. Experimental results with six databases of Road marking dataset, KITTI dataset, Málaga dataset 2009, Málaga urban dataset, Naver street view dataset, and Road/Lane detection evaluation 2013 dataset, show that our method outperforms conventional methods.
Vokhidov, Husan; Hong, Hyung Gil; Kang, Jin Kyu; Hoang, Toan Minh; Park, Kang Ryoung
2016-01-01
Automobile driver information as displayed on marked road signs indicates the state of the road, traffic conditions, proximity to schools, etc. These signs are important to insure the safety of the driver and pedestrians. They are also important input to the automated advanced driver assistance system (ADAS), installed in many automobiles. Over time, the arrow-road markings may be eroded or otherwise damaged by automobile contact, making it difficult for the driver to correctly identify the marking. Failure to properly identify an arrow-road marker creates a dangerous situation that may result in traffic accidents or pedestrian injury. Very little research exists that studies the problem of automated identification of damaged arrow-road marking painted on the road. In this study, we propose a method that uses a convolutional neural network (CNN) to recognize six types of arrow-road markings, possibly damaged, by visible light camera sensor. Experimental results with six databases of Road marking dataset, KITTI dataset, Málaga dataset 2009, Málaga urban dataset, Naver street view dataset, and Road/Lane detection evaluation 2013 dataset, show that our method outperforms conventional methods. PMID:27999301
Institute of Scientific and Technical Information of China (English)
戚林; 郝士琦; 王磊
2011-01-01
针对高误码率情况下删除卷积码的盲解码问题,提出了一种删除卷积码的盲解码算法.通过采用改进的沃尔什-哈达马变换(Walsh-Hadamard translate,WHT),识别出删除卷积码的高维校验矩阵;再基于源卷积码的生成矩阵与校验矩阵间的约束关系,求出了源卷积码的最佳生成多项式和删除模式,并提出了码字同步的方法,最终按识别出来的参数构建盲解码模型,实现了删除卷积码的盲解码.仿真结果表明,在误码率低于0.01,的情况下,该算法的译码正确率能达到93%.%In order to solve the problem of blind decoding punctured convolutional codes at high bit-error rates, this paper proposed a blind decoding algorithm. Through applying the advanced Walsh-Hadamard transform, recognized the high dimensionality parity-check matrix of punctured convolutional codes. Moreover, solved the generator matrix and puncture map of the source convolutional codes based on the constraint relationship between the parity-check matrix and the generator matrix. Then proposed a method of codeword synchronization. According to the recognition results, established a model of blind decoding of punctured convolutional codes. The simulation results show, the bit correct rate after decoding of this algorithm can be up to 93％ under the condition of BER lower than 0.01.
Directory of Open Access Journals (Sweden)
Abbasi Mahdi
2012-06-01
Full Text Available Abstract Background Electrical Impedance Tomography (EIT is used as a fast clinical imaging technique for monitoring the health of the human organs such as lungs, heart, brain and breast. Each practical EIT reconstruction algorithm should be efficient enough in terms of convergence rate, and accuracy. The main objective of this study is to investigate the feasibility of precise empirical conductivity imaging using a sinc-convolution algorithm in D-bar framework. Methods At the first step, synthetic and experimental data were used to compute an intermediate object named scattering transform. Next, this object was used in a two-dimensional integral equation which was precisely and rapidly solved via sinc-convolution algorithm to find the square root of the conductivity for each pixel of image. For the purpose of comparison, multigrid and NOSER algorithms were implemented under a similar setting. Quality of reconstructions of synthetic models was tested against GREIT approved quality measures. To validate the simulation results, reconstructions of a phantom chest and a human lung were used. Results Evaluation of synthetic reconstructions shows that the quality of sinc-convolution reconstructions is considerably better than that of each of its competitors in terms of amplitude response, position error, ringing, resolution and shape-deformation. In addition, the results confirm near-exponential and linear convergence rates for sinc-convolution and multigrid, respectively. Moreover, the least degree of relative errors and the most degree of truth were found in sinc-convolution reconstructions from experimental phantom data. Reconstructions of clinical lung data show that the related physiological effect is well recovered by sinc-convolution algorithm. Conclusions Parametric evaluation demonstrates the efficiency of sinc-convolution to reconstruct accurate conductivity images from experimental data. Excellent results in phantom and clinical
Institute of Scientific and Technical Information of China (English)
俞丹丽; 夏厚培
2012-01-01
针对Rice信道的特性，采用RS码与卷积码进行级联的纠错编码方案，使用matlab对其在Rice信道上的纠错性能进行仿真分析。结果表明，在Rice信道上，RS和卷积的级联码是一种易于实现且性能优越高效的纠错码，有较高的实用价值。%In terms of the characteristics of the Rice channel, coding using the cascaded codes of the RS and the convolution the scheme is adopted, the error correction in the Rice channel are simulated and analyzed through of the error correction whose performances of the Matlab. The results show that the cascaded codes of the RS and the convolution as the error correction codes are easy to be realized and have excellent and highly-efficient performances in the Rice channel with high appli- cation value.
Full Text Available ... Corrective Jaw Surgery Corrective Jaw Surgery Orthognathic surgery is performed to correct the misalignment of jaws and ... Implant Surgery Dental Implant Surgery Dental implant surgery is, of course, surgery, and is best performed by ...
Full Text Available ... and Craniofacial Surgery Cleft Lip/Palate and Craniofacial Surgery A cleft lip may require one or more ... find out more. Corrective Jaw Surgery Corrective Jaw Surgery Orthognathic surgery is performed to correct the misalignment ...
... and Craniofacial Surgery Cleft Lip/Palate and Craniofacial Surgery A cleft lip may require one or more ... find out more. Corrective Jaw Surgery Corrective Jaw Surgery Orthognathic surgery is performed to correct the misalignment ...
Higher order QCD corrections in small x physics
Energy Technology Data Exchange (ETDEWEB)
Chachamis, G.
2006-11-15
We study higher order QCD corrections in small x Physics. The numerical implementation of the full NLO photon impact factor is the remaining necessary piece for the testing of the NLO BFKL resummation against data from physical processes, such as {gamma}{sup *}{gamma}{sup *} collisions. We perform the numerical integration over phase space for the virtual corrections to the NLO photon impact factor. This, along with the previously calculated real corrections, makes feasible in the near future first estimates for the {gamma}*{gamma}* total cross section, since the convolution of the full impact factor with the NLO BFKL gluon Green's function is now straightforward. The NLO corrections for the photon impact factor are sizeable and negative. In the second part of this thesis, we estimate higher order correction to the BK equation. We are mainly interested in whether partonic saturation delays or not in rapidity when going beyond the leading order. In our investigation, we use the so called 'rapidity veto' which forbid two emissions to be very close in rapidity, to 'switch on' higher order corrections to the BK equation. From analytic and numerical analysis, we conclude that indeed saturation does delay in rapidity when higher order corrections are taken into account. In the last part, we investigate higher order QCD corrections as additional corrections to the Electroweak (EW) sector. The question of whether BFKL corrections are of any importance in the Regge limit for the EW sector seems natural; although they arise in higher loop level, the accumulation of logarithms in energy s at high energies, cannot be dismissed without an investigation. We focus on the process {gamma}{gamma}{yields}ZZ. We calculate the pQCD corrections in the forward region at leading logarithmic (LL) BFKL accuracy, which are of the order of few percent at the TeV energy scale. (orig.)
The Gaussian streaming model and convolution Lagrangian effective field theory
Vlah, Zvonimir; Castorina, Emanuele; White, Martin
2016-12-01
We update the ingredients of the Gaussian streaming model (GSM) for the redshift-space clustering of biased tracers using the techniques of Lagrangian perturbation theory, effective field theory (EFT) and a generalized Lagrangian bias expansion. After relating the GSM to the cumulant expansion, we present new results for the real-space correlation function, mean pairwise velocity and pairwise velocity dispersion including counter terms from EFT and bias terms through third order in the linear density, its leading derivatives and its shear up to second order. We discuss the connection to the Gaussian peaks formalism. We compare the ingredients of the GSM to a suite of large N-body simulations, and show the performance of the theory on the low order multipoles of the redshift-space correlation function and power spectrum. We highlight the importance of a general biasing scheme, which we find to be as important as higher-order corrections due to non-linear evolution for the halos we consider on the scales of interest to us.
Sá, Ana Cravo; Coelho, Carina Marques; Monsanto, Fátima
2014-01-01
Objectivo do estudo: comparar o desempenho dos algoritmos Pencil Beam Convolution (PBC) e do Analytical Anisotropic Algorithm (AAA) no planeamento do tratamento de tumores de mama com radioterapia conformacional a 3D.
National Research Council Canada - National Science Library
Cang, Zixuan; Wei, Guowei
2017-01-01
.... This representation reveals hidden structure-function relationships in biomolecules. We further integrate ESPH and deep convolutional neural networks to construct a multichannel topological neural network (TopologyNet...
Sim, K S; Teh, V; Tey, Y C; Kho, T K
2016-11-01
This paper introduces new development technique to improve the Scanning Electron Microscope (SEM) image quality and we name it as sub-blocking multiple peak histogram equalization (SUB-B-MPHE) with convolution operator. By using this new proposed technique, it shows that the new modified MPHE performs better than original MPHE. In addition, the sub-blocking method consists of convolution operator which can help to remove the blocking effect for SEM images after applying this new developed technique. Hence, by using the convolution operator, it effectively removes the blocking effect by properly distributing the suitable pixel value for the whole image. Overall, the SUB-B-MPHE with convolution outperforms the rest of methods. SCANNING 38:492-501, 2016. © 2015 Wiley Periodicals, Inc. © Wiley Periodicals, Inc.
Subleading Corrections To Thrust Using Effective Field Theory
Freedman, Simon M
2013-01-01
We calculate the subleading corrections to the thrust rate using Soft-Collinear Effective Theory to factorize the rate and match onto jet and soft operators that describe the degrees of freedom of the relevant scales. We work in the perturbative regime where all the scales are well above \\Lambda_QCD. The thrust rate involves an incomplete sum over final states that is enforced by a measurement operator. Subleading corrections require matching onto not only the higher dimensional dijet operators, but also matching onto subleading measurement operators in the effective theory. We explicitly show how to factorize the O(\\alpha_s \\tau) thrust rate into a hard function multiplied by the convolution of the vacuum expectation value of jet and soft operators. Our approach can be generalized to other jet shapes and rates.
A patient-specific scatter artifacts correction method
Zhao, Wei; Niu, Kai; Schafer, Sebastian; Royalty, Kevin; Chen, Guang-Hong
2015-01-01
This paper provides a fast and patient-specific scatter artifact correction method for cone-beam computed tomography (CBCT) used in image-guided interventional procedures. Due to increased irradiated volume of interest in CBCT imaging, scatter radiation has increased dramatically compared to 2D imaging, leading to a degradation of image quality. In this study, we propose a scatter artifact correction strategy using an analytical convolution-based model whose free parameters are estimated using a rough estimation of scatter profiles from the acquired cone-beam projections. It was evaluated using Monte Carlo simulations with both monochromatic and polychromatic X-ray sources. The results demonstrated that the proposed method significantly reduced the scatter-induced shading artifacts and recovered CT numbers.
Calculation of the reactor neutron time of flight spectrum by convolution technique
Institute of Scientific and Technical Information of China (English)
Cheng Jin-Xing; Ouyang Xiao-Ping; Zheng Yi; Zhang An-Hui; Ouyang Mao-Jie
2008-01-01
It is a very complex and tlme-consuming process to simulate the nuclear reactor neutron spectrum from the reactor core to the export channel by applying a Monte Carlo program. This paper presents a new method to calculate the neutron spectrum by using the convolution technique which considers the channel transportation as a linear system and the transportation scattering as the response function. It also applies Monte Carlo Neutron and Photon Transport Code (MCNP) to simulate the response function numerically. With the application of convolution technique to calculate thespectrum distribution from the core to the channel, the process is then much more convenient only with the simple numerical integral numeration. This saves computer time and reduces some trouble in re-writing of the MCNP program.
Shkolyar, Anat; Gefen, Amit; Benayahu, Dafna; Greenspan, Hayit
2015-08-01
We propose a semi-automated pipeline for the detection of possible cell divisions in live-imaging microscopy and the classification of these mitosis candidates using a Convolutional Neural Network (CNN). We use time-lapse images of NIH3T3 scratch assay cultures, extract patches around bright candidate regions that then undergo segmentation and binarization, followed by a classification of the binary patches into either containing or not containing cell division. The classification is performed by training a Convolutional Neural Network on a specially constructed database. We show strong results of AUC = 0.91 and F-score = 0.89, competitive with state-of-the-art methods in this field.
Video-based convolutional neural networks for activity recognition from robot-centric videos
Ryoo, M. S.; Matthies, Larry
2016-05-01
In this evaluation paper, we discuss convolutional neural network (CNN)-based approaches for human activity recognition. In particular, we investigate CNN architectures designed to capture temporal information in videos and their applications to the human activity recognition problem. There have been multiple previous works to use CNN-features for videos. These include CNNs using 3-D XYT convolutional filters, CNNs using pooling operations on top of per-frame image-based CNN descriptors, and recurrent neural networks to learn temporal changes in per-frame CNN descriptors. We experimentally compare some of these different representatives CNNs while using first-person human activity videos. We especially focus on videos from a robots viewpoint, captured during its operations and human-robot interactions.
Multi-Task Learning for Food Identification and Analysis with Deep Convolutional Neural Networks
Institute of Scientific and Technical Information of China (English)
Xi-Jin Zhang; Yi-Fan Lu; Song-Hai Zhang
2016-01-01
In this paper, we proposed a multi-task system that can identify dish types, food ingredients, and cooking methods from food images with deep convolutional neural networks. We built up a dataset of 360 classes of different foods with at least 500 images for each class. To reduce the noises of the data, which was collected from the Internet, outlier images were detected and eliminated through a one-class SVM trained with deep convolutional features. We simultaneously trained a dish identifier, a cooking method recognizer, and a multi-label ingredient detector. They share a few low-level layers in the deep network architecture. The proposed framework shows higher accuracy than traditional method with handcrafted features, and the cooking method recognizer and ingredient detector can be applied to dishes which are not included in the training dataset to provide reference information for users.
ARKCoS: Artifact-Suppressed Accelerated Radial Kernel Convolution on the Sphere
Elsner, Franz
2011-01-01
We describe a hybrid Fourier/direct space convolution algorithm for compact radial (azimuthally symmetric) kernels on the sphere. For high resolution maps covering a large fraction of the sky, our implementation takes advantage of the inexpensive massive parallelism afforded by consumer graphics processing units (GPUs). Applications involve modeling of instrumental beam shapes in terms of compact kernels, computation of fine-scale wavelet transformations, and optimal filtering for the detection of point sources. Our algorithm works for any pixelization where pixels are grouped into isolatitude rings. Even for kernels that are not bandwidth limited, ringing features are completely absent on an ECP grid. We demonstrate that they can be highly suppressed on the popular HEALPix pixelization, for which we develop a freely available implementation of the algorithm. As an example application, we show that running on a high-end consumer graphics card our method speeds up beam convolution for simulations of a characte...
Melchert, O; Roth, B
2016-01-01
We discuss efficient algorithms for the accurate forward and reverse evaluation of the discrete Fourier-Bessel transform (dFBT) as numerical tools to assist in the 2D polar convolution of two radially symmetric functions, relevant, e.g., to applications in computational biophotonics. In our survey of the numerical procedure we account for the circumstance that the objective function might result from a more complex measurement process and is, in the worst case, known on a finite sequence of coordinate values, only. We contrast the performance of the resulting algorithms with a procedure based on a straight forward numerical quadrature of the underlying integral transform and asses its efficienty for two benchmark Fourier-Bessel pairs. An application to the problem of finite-size beam-shape convolution in polar coordinates, relevant in the context of tissue optics and optoacoustics, is used to illustrate the versatility and computational efficiency of the numerical procedure.
On the Convolution Equation Related to the Diamond Klein-Gordon Operator
Directory of Open Access Journals (Sweden)
Amphon Liangprom
2011-01-01
Full Text Available We study the distribution eαx(♢+m2kδ for m≥0, where (♢+m2k is the diamond Klein-Gordon operator iterated k times, δ is the Dirac delta distribution, x=(x1,x2,…,xn is a variable in ℝn, and α=(α1,α2,…,αn is a constant. In particular, we study the application of eαx(♢+m2kδ for solving the solution of some convolution equation. We find that the types of solution of such convolution equation, such as the ordinary function and the singular distribution, depend on the relationship between k and M.
Finessing filter scarcity problem in face recognition via multi-fold filter convolution
Low, Cheng-Yaw; Teoh, Andrew Beng-Jin
2017-06-01
The deep convolutional neural networks for face recognition, from DeepFace to the recent FaceNet, demand a sufficiently large volume of filters for feature extraction, in addition to being deep. The shallow filter-bank approaches, e.g., principal component analysis network (PCANet), binarized statistical image features (BSIF), and other analogous variants, endure the filter scarcity problem that not all PCA and ICA filters available are discriminative to abstract noise-free features. This paper extends our previous work on multi-fold filter convolution (ℳ-FFC), where the pre-learned PCA and ICA filter sets are exponentially diversified by ℳ folds to instantiate PCA, ICA, and PCA-ICA offspring. The experimental results unveil that the 2-FFC operation solves the filter scarcity state. The 2-FFC descriptors are also evidenced to be superior to that of PCANet, BSIF, and other face descriptors, in terms of rank-1 identification rate (%).
Detection and recognition of bridge crack based on convolutional neural network
Directory of Open Access Journals (Sweden)
Honggong LIU
2016-10-01
Full Text Available Aiming at the backward artificial visual detection status of bridge crack in China, which has a great danger coefficient, a digital and intelligent detection method of improving the diagnostic efficiency and reducing the risk coefficient is studied. Combing with machine vision and convolutional neural network technology, Raspberry Pi is used to acquire and pre-process image, and the crack image is analyzed; the processing algorithm which has the best effect in detecting and recognizing is selected; the convolutional neural network(CNN for crack classification is optimized; finally, a new intelligent crack detection method is put forward. The experimental result shows that the system can find all cracks beyond the maximum limit, and effectively identify the type of fracture, and the recognition rate is above 90%. The study provides reference data for engineering detection.
Visual and Textual Sentiment Analysis of a Microblog Using Deep Convolutional Neural Networks
Directory of Open Access Journals (Sweden)
Yuhai Yu
2016-06-01
Full Text Available Sentiment analysis of online social media has attracted significant interest recently. Many studies have been performed, but most existing methods focus on either only textual content or only visual content. In this paper, we utilize deep learning models in a convolutional neural network (CNN to analyze the sentiment in Chinese microblogs from both textual and visual content. We first train a CNN on top of pre-trained word vectors for textual sentiment analysis and employ a deep convolutional neural network (DNN with generalized dropout for visual sentiment analysis. We then evaluate our sentiment prediction framework on a dataset collected from a famous Chinese social media network (Sina Weibo that includes text and related images and demonstrate state-of-the-art results on this Chinese sentiment analysis benchmark.
A comparison of Gamma and Gaussian dynamic convolution models of the fMRI BOLD response.
Chen, Huafu; Yao, Dezhong; Liu, Zuxiang
2005-01-01
Blood oxygenation level-dependent (BOLD) contrast-based functional magnetic resonance imaging (fMRI) has been widely utilized to detect brain neural activities and great efforts are now stressed on the hemodynamic processes of different brain regions activated by a stimulus. The focus of this paper is the comparison of Gamma and Gaussian dynamic convolution models of the fMRI BOLD response. The convolutions are between the perfusion function of the neural response to a stimulus and a Gaussian or Gamma function. The parameters of the two models are estimated by a nonlinear least-squares optimal algorithm for the fMRI data of eight subjects collected in a visual stimulus experiment. The results show that the Gaussian model is better than the Gamma model in fitting the data. The model parameters are different in the left and right occipital regions, which indicate that the dynamic processes seem different in various cerebral functional regions.
Brain tumor grading based on Neural Networks and Convolutional Neural Networks.
Yuehao Pan; Weimin Huang; Zhiping Lin; Wanzheng Zhu; Jiayin Zhou; Wong, Jocelyn; Zhongxiang Ding
2015-08-01
This paper studies brain tumor grading using multiphase MRI images and compares the results with various configurations of deep learning structure and baseline Neural Networks. The MRI images are used directly into the learning machine, with some combination operations between multiphase MRIs. Compared to other researches, which involve additional effort to design and choose feature sets, the approach used in this paper leverages the learning capability of deep learning machine. We present the grading performance on the testing data measured by the sensitivity and specificity. The results show a maximum improvement of 18% on grading performance of Convolutional Neural Networks based on sensitivity and specificity compared to Neural Networks. We also visualize the kernels trained in different layers and display some self-learned features obtained from Convolutional Neural Networks.
The statistical nature of the second order corrections to the thermal SZE
Sandoval-Villalbazo, A
2004-01-01
This paper shows that the accepted expressions for the second order corrections in the parameter $z$ to the thermal Sunyaev-Zel'dovich effect can be accurately reproduced by a simple convolution integral approach. This representation allows to separate the second order SZE corrections into two type of components. One associated to a single line broadening, directly related to the even derivative terms present in the distortion intensity curve, while the other is related to a frequency shift, which is in turn related to the first derivative term.
Niessner, R.; Schilling, H.; Jutzi, B.
2017-05-01
In recent years, there has been a significant improvement in the detection, identification and classification of objects and images using Convolutional Neural Networks. To study the potential of the Convolutional Neural Network, in this paper three approaches are investigated to train classifiers based on Convolutional Neural Networks. These approaches allow Convolutional Neural Networks to be trained on datasets containing only a few hundred training samples, which results in a successful classification. Two of these approaches are based on the concept of transfer learning. In the first approach features, created by a pretrained Convolutional Neural Network, are used for a classification using a support vector machine. In the second approach a pretrained Convolutional Neural Network gets fine-tuned on a different data set. The third approach includes the design and training for flat Convolutional Neural Networks from the scratch. The evaluation of the proposed approaches is based on a data set provided by the IEEE Geoscience and Remote Sensing Society (GRSS) which contains RGB and LiDAR data of an urban area. In this work it is shown that these Convolutional Neural Networks lead to classification results with high accuracy both on RGB and LiDAR data. Features which are derived by RGB data transferred into LiDAR data by transfer learning lead to better results in classification in contrast to RGB data. Using a neural network which contains fewer layers than common neural networks leads to the best classification results. In this framework, it can furthermore be shown that the practical application of LiDAR images results in a better data basis for classification of vehicles than the use of RGB images.
2016-01-01
The state-of-the-art methods for protein-protein interaction (PPI) extraction are primarily based on kernel methods, and their performances strongly depend on the handcraft features. In this paper, we tackle PPI extraction by using convolutional neural networks (CNN) and propose a shortest dependency path based CNN (sdpCNN) model. The proposed method (1) only takes the sdp and word embedding as input and (2) could avoid bias from feature selection by using CNN. We performed experiments on sta...
Learning Depth from Single Monocular Images Using Deep Convolutional Neural Fields.
Liu, Fayao; Shen, Chunhua; Lin, Guosheng; Reid, Ian
2016-10-01
In this article, we tackle the problem of depth estimation from single monocular images. Compared with depth estimation using multiple images such as stereo depth perception, depth from monocular images is much more challenging. Prior work typically focuses on exploiting geometric priors or additional sources of information, most using hand-crafted features. Recently, there is mounting evidence that features from deep convolutional neural networks (CNN) set new records for various vision applications. On the other hand, considering the continuous characteristic of the depth values, depth estimation can be naturally formulated as a continuous conditional random field (CRF) learning problem. Therefore, here we present a deep convolutional neural field model for estimating depths from single monocular images, aiming to jointly explore the capacity of deep CNN and continuous CRF. In particular, we propose a deep structured learning scheme which learns the unary and pairwise potentials of continuous CRF in a unified deep CNN framework. We then further propose an equally effective model based on fully convolutional networks and a novel superpixel pooling method, which is about 10 times faster, to speedup the patch-wise convolutions in the deep model. With this more efficient model, we are able to design deeper networks to pursue better performance. Our proposed method can be used for depth estimation of general scenes with no geometric priors nor any extra information injected. In our case, the integral of the partition function can be calculated in a closed form such that we can exactly solve the log-likelihood maximization. Moreover, solving the inference problem for predicting depths of a test image is highly efficient as closed-form solutions exist. Experiments on both indoor and outdoor scene datasets demonstrate that the proposed method outperforms state-of-the-art depth estimation approaches.
An All-In-One Convolutional Neural Network for Face Analysis
Ranjan, Rajeev; Sankaranarayanan, Swami; Castillo, Carlos D.; Chellappa, Rama
2016-01-01
We present a multi-purpose algorithm for simultaneous face detection, face alignment, pose estimation, gender recognition, smile detection, age estimation and face recognition using a single deep convolutional neural network (CNN). The proposed method employs a multi-task learning framework that regularizes the shared parameters of CNN and builds a synergy among different domains and tasks. Extensive experiments show that the network has a better understanding of face and achieves state-of-th...
Edelen, A. L.; Biedron, S. G.; Milton, S. V.; Edelen, J. P.
2016-01-01
At present, a variety of image-based diagnostics are used in particle accelerator systems. Often times, these are viewed by a human operator who then makes appropriate adjustments to the machine. Given recent advances in using convolutional neural networks (CNNs) for image processing, it should be possible to use image diagnostics directly in control routines (NN-based or otherwise). This is especially appealing for non-intercepting diagnostics that could run continuously during beam operatio...
A convolution integral representation of the thermal Sunyaev-Zel'dovich effect
Sandoval-Villalbazo, A
2003-01-01
Analytical expressions for the non-relativistic and relativistic Sunyaev-Zel'dovich effect (SZE) are derived by means of suitable convolution integrals. The establishment of these expressions is based on the fact that the SZE disturbed spectrum, at high frequencies, possesses the form of a Laplace transform of the single line distortion profile (structure factor). Implications of this description of the SZE related to light scattering in optically thin plasmas are discussed.
A convolution integral representation of the thermal Sunyaev-Zel'dovich effect
Energy Technology Data Exchange (ETDEWEB)
Sandoval-Villalbazo, A [Departamento de Fisica y Matematicas, Universidad Iberoamericana, Lomas de Santa Fe 01210 Mexico DF (Mexico); Garcia-Colin, L S [Departamento de Fisica, Universidad Autonoma Metropolitana, Mexico DF, 09340 (Mexico)
2003-04-25
Analytical expressions for the non-relativistic and relativistic Sunyaev-Zel'dovich effect (SZE) are derived by means of suitable convolution integrals. The establishment of these expressions is based on the fact that the SZE disturbed spectrum, at high frequencies, possesses the form of a Laplace transform of the single line distortion profile (structure factor). Implications of this description of the SZE related to light scattering in optically thin plasmas are discussed.
Deep Self-Convolutional Activations Descriptor for Dense Cross-Modal Correspondence
Kim, Seungryong; Min, Dongbo; Lin, Stephen; Sohn, Kwanghoon
2016-01-01
We present a novel descriptor, called deep self-convolutional activations (DeSCA), designed for establishing dense correspondences between images taken under different imaging modalities, such as different spectral ranges or lighting conditions. Motivated by descriptors based on local self-similarity (LSS), we formulate a novel descriptor by leveraging LSS in a deep architecture, leading to better discriminative power and greater robustness to non-rigid image deformations than state-of-the-ar...
Analytic continuation of solutions of some nonlinear convolution partial differential equations
Directory of Open Access Journals (Sweden)
Hidetoshi Tahara
2015-01-01
Full Text Available The paper considers a problem of analytic continuation of solutions of some nonlinear convolution partial differential equations which naturally appear in the summability theory of formal solutions of nonlinear partial differential equations. Under a suitable assumption it is proved that any local holomorphic solution has an analytic extension to a certain sector and its extension has exponential growth when the variable goes to infinity in the sector.
A Deep 3D Convolutional Neural Network Based Design for Manufacturability Framework
Balu, Aditya; Lore, Kin Gwn; Young, Gavin; Krishnamurthy, Adarsh; Sarkar, Soumik
2016-01-01
Deep 3D Convolutional Neural Networks (3D-CNN) are traditionally used for object recognition, video data analytics and human gesture recognition. In this paper, we present a novel application of 3D-CNNs in understanding difficult-to-manufacture features from computer-aided design (CAD) models to develop a decision support tool for cyber-enabled manufacturing. Traditionally, design for manufacturability (DFM) rules are hand-crafted and used to accelerate the engineering product design cycle by...
Fuzzy Logic Module of Convolutional Neural Network for Handwritten Digits Recognition
Popko, E. A.; Weinstein, I. A.
2016-08-01
Optical character recognition is one of the important issues in the field of pattern recognition. This paper presents a method for recognizing handwritten digits based on the modeling of convolutional neural network. The integrated fuzzy logic module based on a structural approach was developed. Used system architecture adjusted the output of the neural network to improve quality of symbol identification. It was shown that proposed algorithm was flexible and high recognition rate of 99.23% was achieved.
Automatic Graphic Logo Detection via Fast Region-based Convolutional Networks
Oliveira, Gonçalo; Frazão, Xavier; Pimentel, André; Ribeiro, Bernardete
2016-01-01
Brand recognition is a very challenging topic with many useful applications in localization recognition, advertisement and marketing. In this paper we present an automatic graphic logo detection system that robustly handles unconstrained imaging conditions. Our approach is based on Fast Region-based Convolutional Networks (FRCN) proposed by Ross Girshick, which have shown state-of-the-art performance in several generic object recognition tasks (PASCAL Visual Object Classes challenges). In par...
Institute of Scientific and Technical Information of China (English)
R. M. EL-ASHWAH
2013-01-01
Making use of the linear operator Lmp (λ,ℓ)f (z)= 1zp + ∞X k=1 ?ℓℓ+λk ?m akzk-p, where ℓ > 0,λ > 0, p ∈ N, m ∈ N0 = N∪{0}, z ∈ U* and f(z) ∈ Σp, we introduce two subclasses of meromorphic p-valent analytic functions and investigate convolution and inclusion properties for these classes.
Modelling of nonlinear bridge aerodynamics and aeroelasticity: a convolution based approach
Directory of Open Access Journals (Sweden)
Wu T.
2012-07-01
Full Text Available Innovative bridge decks exhibit nonlinear behaviour in wind tunnel studies which has placed increasing importance on the nonlinear bridge aerodynamics/aeroelasticity considerations for long-span bridges. The convolution scheme concerning the first-order kernels for linear analysis is reviewed, which is followed by an introduction to higher-order kernels for nonlinear analysis. A numerical example of a longspan suspension bridge is presented that demonstrates the efficacy of the proposed scheme.
Mollah, Ayatullah Faruk; Basu, Subhadip; Nasipuri, Mita
2012-01-01
One of the most important steps of document image processing is binarization. The computational requirements of locally adaptive binarization techniques make them unsuitable for devices with limited computing facilities. In this paper, we have presented a computationally efficient implementation of convolution based locally adaptive binarization techniques keeping the performance comparable to the original implementation. The computational complexity has been reduced from O(W2N2) to O(WN2) wh...
Shi, Wuzhen; Jiang, Feng; Zhao, Debin
2017-01-01
Traditional works have shown that patches in a natural image tend to redundantly recur many times inside the image, both within the same scale, as well as across different scales. Make full use of these multi-scale information can improve the image restoration performance. However, the current proposed deep learning based restoration methods do not take the multi-scale information into account. In this paper, we propose a dilated convolution based inception module to learn multi-scale informa...
Directory of Open Access Journals (Sweden)
HE Xiaofei
2016-09-01
Full Text Available The scene information existing in high resolution remote sensing images is important for image interpretation and understanding of the real world. Traditional scene classification methods often use middle and low-level artificial features, but high resolution images have rich information and complex scene configuration, which need high-level feature to express. A joint saliency and multi-convolutional neural network method is proposed in this paper. Firstly, we obtain meaningful patches that include dominant image information by saliency sampling. Secondly, these patches will be set as a sample input to the convolutional neural network for training, obtain feature expression on different levels. Finally, we embed the multi-layer features into the support vector machine (SVM for image classification. Experiments using two high resolution image scene data show that saliency sampling can effectively get the main target, weaken the impact of other unrelated targets, and reduce data redundancy; convolutional neural network can automatically learn the high-level feature, compared to existing methods, the proposed method can effectively improve the classification accuracy.
Using hybrid GPU/CPU kernel splitting to accelerate spherical convolutions
Sutter, P. M.; Wandelt, B. D.; Elsner, F.
2015-06-01
We present a general method for accelerating by more than an order of magnitude the convolution of pixelated functions on the sphere with a radially-symmetric kernel. Our method splits the kernel into a compact real-space component and a compact spherical harmonic space component. These components can then be convolved in parallel using an inexpensive commodity GPU and a CPU. We provide models for the computational cost of both real-space and Fourier space convolutions and an estimate for the approximation error. Using these models we can determine the optimum split that minimizes the wall clock time for the convolution while satisfying the desired error bounds. We apply this technique to the problem of simulating a cosmic microwave background (CMB) anisotropy sky map at the resolution typical of the high resolution maps produced by the Planck mission. For the main Planck CMB science channels we achieve a speedup of over a factor of ten, assuming an acceptable fractional rms error of order 10-5 in the power spectrum of the output map.
Deep Convolutional Extreme Learning Machine and Its Application in Handwritten Digit Classification.
Pang, Shan; Yang, Xinyi
2016-01-01
In recent years, some deep learning methods have been developed and applied to image classification applications, such as convolutional neuron network (CNN) and deep belief network (DBN). However they are suffering from some problems like local minima, slow convergence rate, and intensive human intervention. In this paper, we propose a rapid learning method, namely, deep convolutional extreme learning machine (DC-ELM), which combines the power of CNN and fast training of ELM. It uses multiple alternate convolution layers and pooling layers to effectively abstract high level features from input images. Then the abstracted features are fed to an ELM classifier, which leads to better generalization performance with faster learning speed. DC-ELM also introduces stochastic pooling in the last hidden layer to reduce dimensionality of features greatly, thus saving much training time and computation resources. We systematically evaluated the performance of DC-ELM on two handwritten digit data sets: MNIST and USPS. Experimental results show that our method achieved better testing accuracy with significantly shorter training time in comparison with deep learning methods and other ELM methods.
Organ detection in thorax abdomen CT using multi-label convolutional neural networks
Humpire Mamani, Gabriel Efrain; Setio, Arnaud Arindra Adiyoso; van Ginneken, Bram; Jacobs, Colin
2017-03-01
A convolutional network architecture is presented to determine bounding boxes around six organs in thoraxabdomen CT scans. A single network for each orthogonal view determines the presence of lungs, kidneys, spleen and liver. We show that an architecture that takes additional slices before and after the slice of interest as an additional input outperforms an architecture that processes single slices. From the slice-based analysis, a bounding box around the structures of interest can be computed. The system uses 6 convolutional, 4 pooling and one fully connected layer and uses 333 scans for training and 110 for validation. The test set contains 110 scans. The average Dice score of the proposed method was 0.95 and 0.95 for the lungs, 0.59 and 0.58 for the kidneys, 0.83 for the liver and 0.63 for the spleen. This paper shows that automatic localization of organs using multi-label convolution neural networks is possible. This architecture can likely be used to identify other organs of interest as well.
Xiong, Jun; Liu, J. G.; Cao, Li
2015-12-01
This paper presents hardware efficient designs for implementing the one-dimensional (1D) discrete Fourier transform (DFT). Once DFT is formulated as the cyclic convolution form, the improved first-order moments-based cyclic convolution structure can be used as the basic computing unit for the DFT computation, which only contains a control module, a barrel shifter and (N-1)/2 accumulation units. After decomposing and reordering the twiddle factors, all that remains to do is shifting the input data sequence and accumulating them under the control of the statistical results on the twiddle factors. The whole calculation process only contains shift operations and additions with no need for multipliers and large memory. Compared with the previous first-order moments-based structure for DFT, the proposed designs have the advantages of less hardware consumption, lower power consumption and the flexibility to achieve better performance in certain cases. A series of experiments have proven the high performance of the proposed designs in terms of the area time product and power consumption. Similar efficient designs can be obtained for other computations, such as DCT/IDCT, DST/IDST, digital filter and correlation by transforming them into the forms of the first-order moments based cyclic convolution.
Li, Siqi; Jiang, Huiyan; Pang, Wenbo
2017-05-01
Accurate cell grading of cancerous tissue pathological image is of great importance in medical diagnosis and treatment. This paper proposes a joint multiple fully connected convolutional neural network with extreme learning machine (MFC-CNN-ELM) architecture for hepatocellular carcinoma (HCC) nuclei grading. First, in preprocessing stage, each grayscale image patch with the fixed size is obtained using center-proliferation segmentation (CPS) method and the corresponding labels are marked under the guidance of three pathologists. Next, a multiple fully connected convolutional neural network (MFC-CNN) is designed to extract the multi-form feature vectors of each input image automatically, which considers multi-scale contextual information of deep layer maps sufficiently. After that, a convolutional neural network extreme learning machine (CNN-ELM) model is proposed to grade HCC nuclei. Finally, a back propagation (BP) algorithm, which contains a new up-sample method, is utilized to train MFC-CNN-ELM architecture. The experiment comparison results demonstrate that our proposed MFC-CNN-ELM has superior performance compared with related works for HCC nuclei grading. Meanwhile, external validation using ICPR 2014 HEp-2 cell dataset shows the good generalization of our MFC-CNN-ELM architecture. Copyright © 2017 Elsevier Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Jilin Zhang
2017-01-01
Full Text Available With the development of the mobile systems, we gain a lot of benefits and convenience by leveraging mobile devices; at the same time, the information gathered by smartphones, such as location and environment, is also valuable for business to provide more intelligent services for customers. More and more machine learning methods have been used in the field of mobile information systems to study user behavior and classify usage patterns, especially convolutional neural network. With the increasing of model training parameters and data scale, the traditional single machine training method cannot meet the requirements of time complexity in practical application scenarios. The current training framework often uses simple data parallel or model parallel method to speed up the training process, which is why heterogeneous computing resources have not been fully utilized. To solve these problems, our paper proposes a delay synchronization convolutional neural network parallel strategy, which leverages the heterogeneous system. The strategy is based on both synchronous parallel and asynchronous parallel approaches; the model training process can reduce the dependence on the heterogeneous architecture in the premise of ensuring the model convergence, so the convolution neural network framework is more adaptive to different heterogeneous system environments. The experimental results show that the proposed delay synchronization strategy can achieve at least three times the speedup compared to the traditional data parallelism.
Convolution power spectrum analysis for FMRI data based on prior image signal.
Zhang, Jiang; Chen, Huafu; Fang, Fang; Liao, Wei
2010-02-01
Functional MRI (fMRI) data-processing methods based on changes in the time domain involve, among other things, correlation analysis and use of the general linear model with statistical parametric mapping (SPM). Unlike conventional fMRI data analysis methods, which aim to model the blood-oxygen-level-dependent (BOLD) response of voxels as a function of time, the theory of power spectrum (PS) analysis focuses completely on understanding the dynamic energy change of interacting systems. We propose a new convolution PS (CPS) analysis of fMRI data, based on the theory of matched filtering, to detect brain functional activation for fMRI data. First, convolution signals are computed between the measured fMRI signals and the image signal of prior experimental pattern to suppress noise in the fMRI data. Then, the PS density analysis of the convolution signal is specified as the quantitative analysis energy index of BOLD signal change. The data from simulation studies and in vivo fMRI studies, including block-design experiments, reveal that the CPS method enables a more effective detection of some aspects of brain functional activation, as compared with the canonical PS SPM and the support vector machine methods. Our results demonstrate that the CPS method is useful as a complementary analysis in revealing brain functional information regarding the complex nature of fMRI time series.
Deep Convolutional Extreme Learning Machine and Its Application in Handwritten Digit Classification
Directory of Open Access Journals (Sweden)
Shan Pang
2016-01-01
Full Text Available In recent years, some deep learning methods have been developed and applied to image classification applications, such as convolutional neuron network (CNN and deep belief network (DBN. However they are suffering from some problems like local minima, slow convergence rate, and intensive human intervention. In this paper, we propose a rapid learning method, namely, deep convolutional extreme learning machine (DC-ELM, which combines the power of CNN and fast training of ELM. It uses multiple alternate convolution layers and pooling layers to effectively abstract high level features from input images. Then the abstracted features are fed to an ELM classifier, which leads to better generalization performance with faster learning speed. DC-ELM also introduces stochastic pooling in the last hidden layer to reduce dimensionality of features greatly, thus saving much training time and computation resources. We systematically evaluated the performance of DC-ELM on two handwritten digit data sets: MNIST and USPS. Experimental results show that our method achieved better testing accuracy with significantly shorter training time in comparison with deep learning methods and other ELM methods.
Accurate segmentation of lung fields on chest radiographs using deep convolutional networks
Arbabshirani, Mohammad R.; Dallal, Ahmed H.; Agarwal, Chirag; Patel, Aalpan; Moore, Gregory
2017-02-01
Accurate segmentation of lung fields on chest radiographs is the primary step for computer-aided detection of various conditions such as lung cancer and tuberculosis. The size, shape and texture of lung fields are key parameters for chest X-ray (CXR) based lung disease diagnosis in which the lung field segmentation is a significant primary step. Although many methods have been proposed for this problem, lung field segmentation remains as a challenge. In recent years, deep learning has shown state of the art performance in many visual tasks such as object detection, image classification and semantic image segmentation. In this study, we propose a deep convolutional neural network (CNN) framework for segmentation of lung fields. The algorithm was developed and tested on 167 clinical posterior-anterior (PA) CXR images collected retrospectively from picture archiving and communication system (PACS) of Geisinger Health System. The proposed multi-scale network is composed of five convolutional and two fully connected layers. The framework achieved IOU (intersection over union) of 0.96 on the testing dataset as compared to manual segmentation. The suggested framework outperforms state of the art registration-based segmentation by a significant margin. To our knowledge, this is the first deep learning based study of lung field segmentation on CXR images developed on a heterogeneous clinical dataset. The results suggest that convolutional neural networks could be employed reliably for lung field segmentation.
Deep Convolutional Extreme Learning Machine and Its Application in Handwritten Digit Classification
Yang, Xinyi
2016-01-01
In recent years, some deep learning methods have been developed and applied to image classification applications, such as convolutional neuron network (CNN) and deep belief network (DBN). However they are suffering from some problems like local minima, slow convergence rate, and intensive human intervention. In this paper, we propose a rapid learning method, namely, deep convolutional extreme learning machine (DC-ELM), which combines the power of CNN and fast training of ELM. It uses multiple alternate convolution layers and pooling layers to effectively abstract high level features from input images. Then the abstracted features are fed to an ELM classifier, which leads to better generalization performance with faster learning speed. DC-ELM also introduces stochastic pooling in the last hidden layer to reduce dimensionality of features greatly, thus saving much training time and computation resources. We systematically evaluated the performance of DC-ELM on two handwritten digit data sets: MNIST and USPS. Experimental results show that our method achieved better testing accuracy with significantly shorter training time in comparison with deep learning methods and other ELM methods. PMID:27610128
A Convolution-LSTM-Based Deep Neural Network for Cross-Domain MOOC Forum Post Classification
Directory of Open Access Journals (Sweden)
Xiaocong Wei
2017-07-01
Full Text Available Learners in a massive open online course often express feelings, exchange ideas and seek help by posting questions in discussion forums. Due to the very high learner-to-instructor ratios, it is unrealistic to expect instructors to adequately track the forums, find all of the issues that need resolution and understand their urgency and sentiment. In this paper, considering the biases among different courses, we propose a transfer learning framework based on a convolutional neural network and a long short-term memory model, called ConvL, to automatically identify whether a post expresses confusion, determine the urgency and classify the polarity of the sentiment. First, we learn the feature representation for each word by considering the local contextual feature via the convolution operation. Second, we learn the post representation from the features extracted through the convolution operation via the LSTM model, which considers the long-term temporal semantic relationships of features. Third, we investigate the possibility of transferring parameters from a model trained on one course to another course and the subsequent fine-tuning. Experiments on three real-world MOOC courses confirm the effectiveness of our framework. This work suggests that our model can potentially significantly increase the effectiveness of monitoring MOOC forums in real time.
On the Design of Error-Correcting Ciphers
Directory of Open Access Journals (Sweden)
Mathur Chetan Nanjunda
2006-01-01
Full Text Available Securing transmission over a wireless network is especially challenging, not only because of the inherently insecure nature of the medium, but also because of the highly error-prone nature of the wireless environment. In this paper, we take a joint encryption-error correction approach to ensure secure and robust communication over the wireless link. In particular, we design an error-correcting cipher (called the high diffusion cipher and prove bounds on its error-correcting capacity as well as its security. Towards this end, we propose a new class of error-correcting codes (HD-codes with built-in security features that we use in the diffusion layer of the proposed cipher. We construct an example, 128-bit cipher using the HD-codes, and compare it experimentally with two traditional concatenated systems: (a AES (Rijndael followed by Reed-Solomon codes, (b Rijndael followed by convolutional codes. We show that the HD-cipher is as resistant to linear and differential cryptanalysis as the Rijndael. We also show that any chosen plaintext attack that can be performed on the HD cipher can be transformed into a chosen plaintext attack on the Rijndael cipher. In terms of error correction capacity, the traditional systems using Reed-Solomon codes are comparable to the proposed joint error-correcting cipher and those that use convolutional codes require more data expansion in order to achieve similar error correction as the HD-cipher. The original contributions of this work are (1 design of a new joint error-correction-encryption system, (2 design of a new class of algebraic codes with built-in security criteria, called the high diffusion codes (HD-codes for use in the HD-cipher, (3 mathematical properties of these codes, (4 methods for construction of the codes, (5 bounds on the error-correcting capacity of the HD-cipher, (6 mathematical derivation of the bound on resistance of HD cipher to linear and differential cryptanalysis, (7 experimental comparison
Institute of Scientific and Technical Information of China (English)
黄咨; 刘琦; 陈致远; 赵宇明
2016-01-01
Pedestrian detection has become one of the hot research topics in various social fields.Convolutional neural networks have excellent learning ability.The characteristics of targets learned by these networks are more natural and more conducive to distinguishing different targets.However,traditional convolutional neural network models have to process entire target.Meanwhile,all the training samples need to be pre-labelled correctly,these hamper the development of convolutional neural network models.In this paper,we propose a convolutional neural network-based latent training model.The model reduces the computation complexity by integrating multiple part detection modules and learns the targets classification rules from unlabelled samples by adopting a latent training method.In the paper we also propose a two-stage learning scheme to overlay the size of the network step by step.Evaluation of the tests on public static pedestrian detection dataset,INRIA Person Dataset[1],demonstrates that our model achieves 98% of detection accuracy and 95% of average precision.%行人检测已经成为社会各领域里的热门研究课题之一。卷积神经网络 CNNs（Convolutional neural networks）良好的学习能力使其学习得到的目标特征更自然，更有利于区分不同目标。但传统的卷积神经网络模型需要对整体目标进行处理，同时要求所有训练样本预先正确标注，这些阻碍了卷积神经网络模型的发展。提出一种基于卷积神经网络的隐式训练模型，该模型通过结合多部件检测模块降低计算复杂度，并采用隐式学习方法从未标注的样本中学习目标的分类规则。还提出一种两段式学习方案来逐步叠加网络的规模。在公共的静态行人检测库 INRIA[1]上的试验评测中，所提模型获得98％的检测准确率和95％的平均准确率。
Convolution-based estimation of organ dose in tube current modulated CT
Tian, Xiaoyu; Segars, W. Paul; Dixon, Robert L.; Samei, Ehsan
2016-05-01
Estimating organ dose for clinical patients requires accurate modeling of the patient anatomy and the dose field of the CT exam. The modeling of patient anatomy can be achieved using a library of representative computational phantoms (Samei et al 2014 Pediatr. Radiol. 44 460-7). The modeling of the dose field can be challenging for CT exams performed with a tube current modulation (TCM) technique. The purpose of this work was to effectively model the dose field for TCM exams using a convolution-based method. A framework was further proposed for prospective and retrospective organ dose estimation in clinical practice. The study included 60 adult patients (age range: 18-70 years, weight range: 60-180 kg). Patient-specific computational phantoms were generated based on patient CT image datasets. A previously validated Monte Carlo simulation program was used to model a clinical CT scanner (SOMATOM Definition Flash, Siemens Healthcare, Forchheim, Germany). A practical strategy was developed to achieve real-time organ dose estimation for a given clinical patient. CTDIvol-normalized organ dose coefficients ({{h}\\text{Organ}} ) under constant tube current were estimated and modeled as a function of patient size. Each clinical patient in the library was optimally matched to another computational phantom to obtain a representation of organ location/distribution. The patient organ distribution was convolved with a dose distribution profile to generate {{≤ft(\\text{CTD}{{\\text{I}}\\text{vol}}\\right)}\\text{organ, \\text{convolution}}} values that quantified the regional dose field for each organ. The organ dose was estimated by multiplying {{≤ft(\\text{CTD}{{\\text{I}}\\text{vol}}\\right)}\\text{organ, \\text{convolution}}} with the organ dose coefficients ({{h}\\text{Organ}} ). To validate the accuracy of this dose estimation technique, the organ dose of the original clinical patient was estimated using Monte Carlo program with TCM profiles explicitly modeled. The
Patient-specific scatter correction for flat-panel detector-based cone-beam CT imaging.
Zhao, Wei; Brunner, Stephen; Niu, Kai; Schafer, Sebastian; Royalty, Kevin; Chen, Guang-Hong
2015-02-07
A patient-specific scatter correction algorithm is proposed to mitigate scatter artefacts in cone-beam CT (CBCT). The approach belongs to the category of convolution-based methods in which a scatter potential function is convolved with a convolution kernel to estimate the scatter profile. A key step in this method is to determine the free parameters introduced in both scatter potential and convolution kernel using a so-called calibration process, which is to seek for the optimal parameters such that the models for both scatter potential and convolution kernel is able to optimally fit the previously known coarse estimates of scatter profiles of the image object. Both direct measurements and Monte Carlo (MC) simulations have been proposed by other investigators to achieve the aforementioned rough estimates. In the present paper, a novel method has been proposed and validated to generate the needed coarse scatter profile for parameter calibration in the convolution method. The method is based upon an image segmentation of the scatter contaminated CBCT image volume, followed by a reprojection of the segmented image volume using a given x-ray spectrum. The reprojected data is subtracted from the scatter contaminated projection data to generate a coarse estimate of the needed scatter profile used in parameter calibration. The method was qualitatively and quantitatively evaluated using numerical simulations and experimental CBCT data acquired on a clinical CBCT imaging system. Results show that the proposed algorithm can significantly reduce scatter artefacts and recover the correct CT number. Numerical simulation results show the method is patient specific, can accurately estimate the scatter, and is robust with respect to segmentation procedure. For experimental and in vivo human data, the results show the CT number can be successfully recovered and anatomical structure visibility can be significantly improved.
Ashoor, Mansour; Asgari, Afrouz; Khorshidi, Abdollah; Rezaei, Ali
2015-01-01
Purpose: Estimation of Compton attenuation and the photoelectric absorption coefficients were explored at various depths. Methods: A new method was proposed for estimating the depth based on the convolution of two exponential functions, namely convolution of scattering and primary functions (CSPF), which the convolved result will conform to the photopeak region of energy spectrum with the variable energy-window widths (EWWs) and a theory on the scattering cross-section. The triple energy-windows (TEW) and extended triple energy-windows scatter correction (ETEW) methods were used to estimate the scattered and primary photons according to the energy spectra at various depths due to a better performance than the other methods in nuclear medicine. For this purpose, the energy spectra were employed, and a distinct phantom along with a technetium-99 m source was simulated by Monte Carlo method. Results: The simulated results indicate that the EWW, used to calculate the scattered and primary counts in terms of the integral operators on the functions, was proportional to the depth as an exponential function. The depth will be calculated by the combination of either TEW or ETEW and proposed method resulting in the distinct energy-window. The EWWs for primary photons were in good agreement with those of scattered photons at the same as depths. The average errors between these windows for both methods TEW, and ETEW were 7.25% and 6.03% at different depths, respectively. The EWW value for functions of scattered and primary photons was reduced by increasing the depth in the CSPF method. Conclusions: This coefficient may be an index for the scattering cross-section. PMID:26170567
Directory of Open Access Journals (Sweden)
Suresh B Rana
2013-01-01
Full Text Available Purpose: It is well known that photon beam radiation therapy requires dose calculation algorithms. The objective of this study was to measure and assess the ability of pencil beam convolution (PBC and anisotropic analytical algorithm (AAA to predict doses beyond high density heterogeneity. Materials and Methods: An inhomogeneous phantom of five layers was created in Eclipse planning system (version 8.6.15. Each layer of phantom was assigned in terms of water (first or top, air (second, water (third, bone (fourth, and water (fifth or bottom medium. Depth doses in water (bottom medium were calculated for 100 monitor units (MUs with 6 Megavoltage (MV photon beam for different field sizes using AAA and PBC with heterogeneity correction. Combinations of solid water, Poly Vinyl Chloride (PVC, and Styrofoam were then manufactured to mimic phantoms and doses for 100 MUs were acquired with cylindrical ionization chamber at selected depths beyond high density heterogeneity interface. The measured and calculated depth doses were then compared. Results: AAA′s values had better agreement with measurements at all measured depths. Dose overestimation by AAA (up to 5.3% and by PBC (up to 6.7% was found to be higher in proximity to the high-density heterogeneity interface, and the dose discrepancies were more pronounced for larger field sizes. The errors in dose estimation by AAA and PBC may be due to improper beam modeling of primary beam attenuation or lateral scatter contributions or combination of both in heterogeneous media that include low and high density materials. Conclusions: AAA is more accurate than PBC for dose calculations in treating deep-seated tumor beyond high-density heterogeneity interface.
NWS Corrections to Observations
National Oceanic and Atmospheric Administration, Department of Commerce — Form B-14 is the National Weather Service form entitled 'Notice of Corrections to Weather Records.' The forms are used to make corrections to observations on forms...
Institute of Scientific and Technical Information of China (English)
Dr. Grace Zhang
2000-01-01
Error correction is an important issue in foreign language acquisition. This paper investigates how students feel about the way in which error correction should take place in a Chinese-as-a foreign-language classroom, based on empirical data of a large scale. The study shows that there is a general consensus that error correction is necessary. In terms of correction strategy, the students preferred a combination of direct and indirect corrections, or a direct only correction. The former choice indicates that students would be happy to take either so long as the correction gets done.Most students didn't mind peer correcting provided it is conducted in a constructive way. More than halfofthe students would feel uncomfortable ifthe same error they make in class is corrected consecutively more than three times. Taking these findings into consideration, we may want to cncourage peer correcting, use a combination of correction strategies (direct only if suitable) and do it in a non-threatening and sensitive way. It is hoped that this study would contribute to the effectiveness of error correction in a Chinese language classroom and it may also have a wider implication on other languages.
Directory of Open Access Journals (Sweden)
F. Alidoost
2016-06-01
Full Text Available In recent years, with the development of the high resolution data acquisition technologies, many different approaches and algorithms have been presented to extract the accurate and timely updated 3D models of buildings as a key element of city structures for numerous applications in urban mapping. In this paper, a novel and model-based approach is proposed for automatic recognition of buildings’ roof models such as flat, gable, hip, and pyramid hip roof models based on deep structures for hierarchical learning of features that are extracted from both LiDAR and aerial ortho-photos. The main steps of this approach include building segmentation, feature extraction and learning, and finally building roof labeling in a supervised pre-trained Convolutional Neural Network (CNN framework to have an automatic recognition system for various types of buildings over an urban area. In this framework, the height information provides invariant geometric features for convolutional neural network to localize the boundary of each individual roofs. CNN is a kind of feed-forward neural network with the multilayer perceptron concept which consists of a number of convolutional and subsampling layers in an adaptable structure and it is widely used in pattern recognition and object detection application. Since the training dataset is a small library of labeled models for different shapes of roofs, the computation time of learning can be decreased significantly using the pre-trained models. The experimental results highlight the effectiveness of the deep learning approach to detect and extract the pattern of buildings’ roofs automatically considering the complementary nature of height and RGB information.
Alidoost, F.; Arefi, H.
2016-06-01
In recent years, with the development of the high resolution data acquisition technologies, many different approaches and algorithms have been presented to extract the accurate and timely updated 3D models of buildings as a key element of city structures for numerous applications in urban mapping. In this paper, a novel and model-based approach is proposed for automatic recognition of buildings' roof models such as flat, gable, hip, and pyramid hip roof models based on deep structures for hierarchical learning of features that are extracted from both LiDAR and aerial ortho-photos. The main steps of this approach include building segmentation, feature extraction and learning, and finally building roof labeling in a supervised pre-trained Convolutional Neural Network (CNN) framework to have an automatic recognition system for various types of buildings over an urban area. In this framework, the height information provides invariant geometric features for convolutional neural network to localize the boundary of each individual roofs. CNN is a kind of feed-forward neural network with the multilayer perceptron concept which consists of a number of convolutional and subsampling layers in an adaptable structure and it is widely used in pattern recognition and object detection application. Since the training dataset is a small library of labeled models for different shapes of roofs, the computation time of learning can be decreased significantly using the pre-trained models. The experimental results highlight the effectiveness of the deep learning approach to detect and extract the pattern of buildings' roofs automatically considering the complementary nature of height and RGB information.
BrainNetCNN: Convolutional neural networks for brain networks; towards predicting neurodevelopment.
Kawahara, Jeremy; Brown, Colin J; Miller, Steven P; Booth, Brian G; Chau, Vann; Grunau, Ruth E; Zwicker, Jill G; Hamarneh, Ghassan
2017-02-01
We propose BrainNetCNN, a convolutional neural network (CNN) framework to predict clinical neurodevelopmental outcomes from brain networks. In contrast to the spatially local convolutions done in traditional image-based CNNs, our BrainNetCNN is composed of novel edge-to-edge, edge-to-node and node-to-graph convolutional filters that leverage the topological locality of structural brain networks. We apply the BrainNetCNN framework to predict cognitive and motor developmental outcome scores from structural brain networks of infants born preterm. Diffusion tensor images (DTI) of preterm infants, acquired between 27 and 46 weeks gestational age, were used to construct a dataset of structural brain connectivity networks. We first demonstrate the predictive capabilities of BrainNetCNN on synthetic phantom networks with simulated injury patterns and added noise. BrainNetCNN outperforms a fully connected neural-network with the same number of model parameters on both phantoms with focal and diffuse injury patterns. We then apply our method to the task of joint prediction of Bayley-III cognitive and motor scores, assessed at 18 months of age, adjusted for prematurity. We show that our BrainNetCNN framework outperforms a variety of other methods on the same data. Furthermore, BrainNetCNN is able to identify an infant's postmenstrual age to within about 2 weeks. Finally, we explore the high-level features learned by BrainNetCNN by visualizing the importance of each connection in the brain with respect to predicting the outcome scores. These findings are then discussed in the context of the anatomy and function of the developing preterm infant brain.
Directory of Open Access Journals (Sweden)
Fan Hu
2015-11-01
Full Text Available Learning efficient image representations is at the core of the scene classification task of remote sensing imagery. The existing methods for solving the scene classification task, based on either feature coding approaches with low-level hand-engineered features or unsupervised feature learning, can only generate mid-level image features with limited representative ability, which essentially prevents them from achieving better performance. Recently, the deep convolutional neural networks (CNNs, which are hierarchical architectures trained on large-scale datasets, have shown astounding performance in object recognition and detection. However, it is still not clear how to use these deep convolutional neural networks for high-resolution remote sensing (HRRS scene classification. In this paper, we investigate how to transfer features from these successfully pre-trained CNNs for HRRS scene classification. We propose two scenarios for generating image features via extracting CNN features from different layers. In the first scenario, the activation vectors extracted from fully-connected layers are regarded as the final image features; in the second scenario, we extract dense features from the last convolutional layer at multiple scales and then encode the dense features into global image features through commonly used feature coding approaches. Extensive experiments on two public scene classification datasets demonstrate that the image features obtained by the two proposed scenarios, even with a simple linear classifier, can result in remarkable performance and improve the state-of-the-art by a significant margin. The results reveal that the features from pre-trained CNNs generalize well to HRRS datasets and are more expressive than the low- and mid-level features. Moreover, we tentatively combine features extracted from different CNN models for better performance.
Mao, H.; Bhaduri, B. L.
2016-12-01
Understanding public opinions on climate change is important for policy making. Public opinion, however, is typically measured with national surveys, which are often too expensive and thus being updated at a low frequency. Twitter has become a major platform for people to express their opinions on social and political issues. Our work attempts to understand if Twitter data can provide complimentary insights about climate change perceptions. Since the nature of social media is real-time, this data source can especially help us understand how public opinion changes over time in response to climate events and hazards, which though is very difficult to be captured by manual surveys. We use the Twitter Streaming API to collect tweets that contain keywords, "climate change" or "#climatechange". Traditional machine-learning based opinion mining algorithms require a significant amount of labeled data. Data labeling is notoriously time consuming. To address this problem, we use hashtags (a significant feature used to mark topics of tweets) to annotate tweets automatically. For example, hashtags, #climatedenial and #climatescam, are negative opinion labels, while #actonclimate and #climateaction are positive. Following this method, we can obtain a large amount of training data without human labor. This labeled dataset is used to train a deep convolutional neural network that classifies tweets into positive (i.e. believe in climate change) and negative (i.e. do not believe). Based on the positive/negative tweets obtained, we will further analyze risk perceptions and opinions towards policy support. In addition, we analyze twitter user profiles to understand the demographics of proponents and opponents of climate change. Deep learning techniques, especially convolutional deep neural networks, have achieved much success in computer vision. In this work, we propose a convolutional neural network architecture for understanding opinions within text. This method is compared with
De-convoluting mixed crude oil in Prudhoe Bay Field, North Slope, Alaska
Peters, K.E.; Scott, Ramos L.; Zumberge, J.E.; Valin, Z.C.; Bird, K.J.
2008-01-01
Seventy-four crude oil samples from the Barrow arch on the North Slope of Alaska were studied to assess the relative volumetric contributions from different source rocks to the giant Prudhoe Bay Field. We applied alternating least squares to concentration data (ALS-C) for 46 biomarkers in the range C19-C35 to de-convolute mixtures of oil generated from carbonate rich Triassic Shublik Formation and clay rich Jurassic Kingak Shale and Cretaceous Hue Shale-gamma ray zone (Hue-GRZ) source rocks. ALS-C results for 23 oil samples from the prolific Ivishak Formation reservoir of the Prudhoe Bay Field indicate approximately equal contributions from Shublik Formation and Hue-GRZ source rocks (37% each), less from the Kingak Shale (26%), and little or no contribution from other source rocks. These results differ from published interpretations that most oil in the Prudhoe Bay Field originated from the Shublik Formation source rock. With few exceptions, the relative contribution of oil from the Shublik Formation decreases, while that from the Hue-GRZ increases in reservoirs along the Barrow arch from Point Barrow in the northwest to Point Thomson in the southeast (???250 miles or 400 km). The Shublik contribution also decreases to a lesser degree between fault blocks within the Ivishak pool from west to east across the Prudhoe Bay Field. ALS-C provides a robust means to calculate the relative amounts of two or more oil types in a mixture. Furthermore, ALS-C does not require that pure end member oils be identified prior to analysis or that laboratory mixtures of these oils be prepared to evaluate mixing. ALS-C of biomarkers reliably de-convolutes mixtures because the concentrations of compounds in mixtures vary as linear functions of the amount of each oil type. ALS of biomarker ratios (ALS-R) cannot be used to de-convolute mixtures because compound ratios vary as nonlinear functions of the amount of each oil type.
Huang, Lin; Xia, Wei; Zhang, Bo; Qiu, Bensheng; Gao, Xin
2017-05-01
Automatic osteosarcoma tumor segmentation on computed tomography (CT) images is a challenging problem, as tumors have large spatial and structural variabilities. In this study, an automatic tumor segmentation method, which was based on a fully convolutional networks with multiple supervised side output layers (MSFCN), was presented. Image normalization is applied as a pre-processing step for decreasing the differences among images. In the frame of the fully convolutional networks, supervised side output layers were added to three layers in order to guide the multi-scale feature learning as a contracting structure, which was then able to capture both the local and global image features. Multiple feature channels were used in the up-sampling portion to capture more context information, for the assurance of accurate segmentation of the tumor, with low contrast around the soft tissue. The results of all the side outputs were fused to determine the final boundaries of the tumors. A quantitative comparison of the 405 osteosarcoma manual segmentation results from the CT images showed that the average Dice similarity coefficient (DSC), average sensitivity, average Hammoude distance (HM) and F1-measure were 87.80%, 86.88%, 19.81% and 0.908, respectively. It was determined that, when compared with the other learning-based algorithms (for example, the fully convolution networks (FCN), U-Net method, and holistically-nested edge detection (HED) method), the MSFCN had the best performances in terms of DSC, sensitivity, HM and F1-measure. The results indicated that the proposed algorithm contributed to the fast and accurate delineation of tumor boundaries, which could potentially assist doctors in making more precise treatment plans. Copyright © 2017 Elsevier B.V. All rights reserved.
A multiple circular path convolution neural network system for detection of mammographic masses.
Lo, Shih-Chung B; Li, Huai; Wang, Yue; Kinnard, Lisa; Freedman, Matthew T
2002-02-01
A multiple circular path convolution neural network (MCPCNN) architecture specifically designed for the analysis of tumor and tumor-like structures has been constructed. We first divided each suspected tumor area into sectors and computed the defined mass features for each sector independently. These sector features were used on the input layer and were coordinated by convolution kernels of different sizes that propagated signals to the second layer in the neural network system. The convolution kernels were trained, as required, by presenting the training cases to the neural network. In this study, randomly selected mammograms were processed by a dual morphological enhancement technique. Radiodense areas were isolated and were delineated using a region growing algorithm. The boundary of each region of interest was then divided into 36 sectors using 36 equi-angular dividers radiated from the center of the region. A total of 144 Breast Imaging-Reporting and Data System-based features (i.e., four features per sector for 36 sectors) were computed as input values for the evaluation of this newly invented neural network system. The overall performance was 0.78-0.80 for the areas (Az) under the receiver operating characteristic curves using the conventional feed-forward neural network in the detection of mammographic masses. The performance was markedly improved with Az values ranging from 0.84 to 0.89 using the MCPCNN. This paper does not intend to claim the best mass detection system. Instead it reports a potentially better neural network structure for analyzing a set of the mass features defined by an investigator.
Repeat-punctured superorthogonal convolutional turbo codes on AWGN and flat Rayleigh fading channels
Directory of Open Access Journals (Sweden)
Fambirai Takawira
2010-10-01
Full Text Available Repeat-punctured turbo codes, an extension of the conventional turbo-coding scheme, has shown a significant increase in bit-error rate performance at moderate to high signal-to-noise ratios for short frame lengths. Superorthogonal convolutional turbo codes (SCTC makes use of superorthogonal signals to improve the performance of the conventional turbo codes and a coding scheme that applies the repeat-punctured technique into SCTC has shown to perform better. We investigated two new low-rate coding schemes, repeat-punctured superorthogonal convolutional turbo codes (RPSCTC and dual-repeat-punctured superorthogonal convolutional turbo codes (DRPSCTC, that make use of superorthogonal signaling, together with repetition and puncturing, to improve the performance of SCTC for reliable and effective communications. Simulation results in the additive white Gaussian noise (AWGN channel and the frequency non-selective Rayleigh fading channel are presented together with analytical bounds of bit error probabilities, derived from transfer function bounding techniques. From the simulation results and the analytical bounds presented, it is evident that RPSCTC and DRPSCTC offer a more superior performance than SCTC in the AWGN channel, as well as in flat Rayleigh non-line-of-sight fading channels. The distance spectrum is also presented for the new schemes and accounts for the performance improvement rendered in simulations. It is important to note that the improved performance that SCTC, and consequently RPSCTC and DRPSCTC, exhibit is achieved at the expense of bandwidth expansion and complexity and would be ideal for power-limited satellite communication links or interference-limited systems.
Directory of Open Access Journals (Sweden)
Haoning Lin
2017-05-01
Full Text Available In current remote sensing literature, the problems of sea-land segmentation and ship detection (including in-dock ships are investigated separately despite the high correlation between them. This inhibits joint optimization and makes the implementation of the methods highly complicated. In this paper, we propose a novel fully convolutional network to accomplish the two tasks simultaneously, in a semantic labeling fashion, i.e., to label every pixel of the image into 3 classes, sea, land and ships. A multi-scale structure for the network is proposed to address the huge scale gap between different classes of targets, i.e., sea/land and ships. Conventional multi-scale structure utilizes shortcuts to connect low level, fine scale feature maps to high level ones to increase the network’s ability to produce finer results. In contrast, our proposed multi-scale structure focuses on increasing the receptive field of the network while maintaining the ability towards fine scale details. The multi-scale convolution network accommodates the huge scale difference between sea-land and ships and provides comprehensive features, and is able to accomplish the tasks in an end-to-end manner that is easy for implementation and feasible for joint optimization. In the network, the input forks into fine-scale and coarse-scale paths, which share the same convolution layers to minimize network parameter increase, and then are joined together to produce the final result. The experiments show that the network tackles the semantic labeling problem with improved performance.
Institute of Scientific and Technical Information of China (English)
D.C. Wan; G.W. Wei
2000-01-01
An efficient discrete singular convolution (DSC) method is introduced to the numerical solutions of incompressible Euler and Navier-Stokes equations with periodic boundary conditions. Two numerical tests of two-dimensional NavierStokes equations with periodic boundary conditions and Euler equations for doubly periodic shear layer flows are carried out by using the DSC method for spatial derivatives and fourth-order Runge-Kutta method for time advancement, respectively. The computational results show that the DSC method is efficient and robust for solving the problems of incompressible flows, and has the potential of being extended to numerically solve much broader problems in fluid dynamics.
The recording of digital hologram at short distance and reconstruction using convolution approach
Institute of Scientific and Technical Information of China (English)
Chen Li-Ping; LU Xiao-Xu
2009-01-01
By adopting in-line lensless Fourier setup and phase-shifting technique, we recorded the phase-shifting digital hologram at short distance. As the Fresnel diffraction condition is no longer valid, the convolution approach is chosen for the reconstruction. However, the simulated reference wave for the reconstruction would suffer from severe under-sampling due to the comparatively large pixel size. To solve this problem, sine-interpolation is introduced to get the pixel-size of the hologram reduced prior to the reconstruction. The experimental results show that an object image of high fidelity is obtained with this method.
Directory of Open Access Journals (Sweden)
Mercan Kadir
2016-01-01
Full Text Available In the present manuscript, free vibration response of circular cylindrical shells with functionally graded material (FGM is investigated. The method of discrete singular convolution (DSC is used for numerical solution of the related governing equation of motion of FGM cylindrical shell. The constitutive relations are based on the Love’s first approximation shell theory. The material properties are graded in the thickness direction according to a volume fraction power law indexes. Frequency values are calculated for different types of boundary conditions, material and geometric parameters. In general, close agreement between the obtained results and those of other researchers has been found.
Real-time minimal bit error probability decoding of convolutional codes
Lee, L. N.
1973-01-01
A recursive procedure is derived for decoding of rate R=1/n binary convolutional codes which minimizes the probability of the individual decoding decisions for each information bit subject to the constraint that the decoding delay be limited to Delta branches. This new decoding algorithm is similar to, but somewhat more complex than, the Viterbi decoding algorithm. A real-time, i.e. fixed decoding delay, version of the Viterbi algorithm is also developed and used for comparison to the new algorithm on simulated channels. It is shown that the new algorithm offers advantages over Viterbi decoding in soft-decision applications such as in the inner coding system for concatenated coding.
Real-time minimal-bit-error probability decoding of convolutional codes
Lee, L.-N.
1974-01-01
A recursive procedure is derived for decoding of rate R = 1/n binary convolutional codes which minimizes the probability of the individual decoding decisions for each information bit, subject to the constraint that the decoding delay be limited to Delta branches. This new decoding algorithm is similar to, but somewhat more complex than, the Viterbi decoding algorithm. A real-time, i.e., fixed decoding delay, version of the Viterbi algorithm is also developed and used for comparison to the new algorithm on simulated channels. It is shown that the new algorithm offers advantages over Viterbi decoding in soft-decision applications, such as in the inner coding system for concatenated coding.
Energy Technology Data Exchange (ETDEWEB)
Xu, Chen; Wang, Zhong Lin [School of Materials Science and Engineering, Georgia Institute of Technology, Atlanta, GA 30332 (United States)
2011-02-15
A fully integrated, solid-state, compact hybrid cell (CHC) that comprises ''convoluted'' ZnO nanowire structures for concurrent harvesting of both solar and mechanical energy is demonstrated. The compact hybrid cell is based on a conjunction design of an organic solid-state dye-sensitized solar cell (DSSC) and piezoelectric nanogenerator in one compact structure. The CHC shows a significant increase in output power, clearly demonstrating its potential for simultaneously harvesting multiple types of energy for powering small electronic devices for independent, sustainable, and mobile operation. (Copyright copyright 2011 WILEY-VCH Verlag GmbH and Co. KGaA, Weinheim)
Multi-Scale Rotation-Invariant Convolutional Neural Networks for Lung Texture Classification.
Wang, Qiangchang; Zheng, Yuanjie; Yang, Gongping; Jin, Weidong; Chen, Xinjian; Yin, Yilong
2017-03-21
We propose a new Multi-scale Rotation-invariant Convolutional Neural Network (MRCNN) model for classifying various lung tissue types on high-resolution computed tomography (HRCT). MRCNN employs Gabor-local binary pattern (Gabor-LBP) which introduces a good property in image analysis - invariance to image scales and rotations. In addition, we offer an approach to deal with the problems caused by imbalanced number of samples between different classes in most of the existing works, accomplished by changing the overlapping size between the adjacent patches. Experimental results on a public Interstitial Lung Disease (ILD) database show a superior performance of the proposed method to state-of-the-art.
Blind Source Separation in Farsi Language by Using Hermitian Angle in Convolutive Enviroment
Directory of Open Access Journals (Sweden)
Atefeh Soltani
2013-04-01
Full Text Available This paper presents a T-F masking method for convolutive blind source separation based on hermitian angle concept. The hermitian angle is calculated between T-F domain mixture vector and reference vector. Two different reference vectors are assumed for calculating two different hermitian angles, and then these angles are clustered with k-means or FCM method to estimate unmixing masks. The well-known permutation problem is solved based on k-means clustering of estimated masks which are partitioned to small groups. The experimental results show an improvement in performance when using two different reference vectors compared to only one.
Estimating the number of sources in a noisy convolutive mixture using BIC
DEFF Research Database (Denmark)
Olsson, Rasmus Kongsgaard; Hansen, Lars Kai
2004-01-01
posterior probability of the sources conditioned on the observations is obtained. The log-likelihood of the parameters is computed exactly in the process, which allows for model evidence comparison assisted by the BIC approximation. This is used to determine the activity pattern of two speakers......The number of source signals in a noisy convolutive mixture is determined based on the exact log-likelihoods of the candidate models. In (Olsson and Hansen, 2004), a novel probabilistic blind source separator was introduced that is based solely on the time-varying second-order statistics...
ON THE STRUCTURES OF RANDOM MEASURE AND POINT PROCESS CONVOLUTION SEMIGROUPS
Institute of Scientific and Technical Information of China (English)
HEYUANJIANG
1996-01-01
Let D be a convolution semigroup of random measures or point processes on a locally compact second countable T2 space. There is a topological isomorphism from D into a subsemigroup of product topological semigroup (R+, +)N. D is a sequentially stable and D-separableZH-semigroup, as well as a metrizable, stable and normable Hun semigroup, so it has the corresponding properties. In particular the author has a new and simple proof by ZH-semigroupapproach or Hun semigroup approach to show that D has property ILID (an infinitesimal arraylimit is infinitely divisible), and know the Bairn types which some subsets of D belong in.
Obeso, Abraham Montoya; Benois-Pineau, Jenny; Acosta, Alejandro Álvaro Ramirez; Vázquez, Mireya Saraí García
2017-01-01
We propose a convolutional neural network to classify images of buildings using sparse features at the network's input in conjunction with primary color pixel values. As a result, a trained neuronal model is obtained to classify Mexican buildings in three classes according to the architectural styles: prehispanic, colonial, and modern with an accuracy of 88.01%. The problem of poor information in a training dataset is faced due to the unequal availability of cultural material. We propose a data augmentation and oversampling method to solve this problem. The results are encouraging and allow for prefiltering of the content in the search tasks.
Response compaction for system-on-a-chip based on advanced convolutional codes
Institute of Scientific and Technical Information of China (English)
无
2006-01-01
This paper addresses the problem of test response compaction. In order to maximize compaction ratio, a single-output compactor based on a (n, n-1, m, 3) convolutional code is presented. When the proposed theorems are satisfied, the compactor can avoid two and any odd erroneous bits cancellations, and handle one unknown bit (X bit). When the X bits in response are clustered, multiple-weight check matrix design algorithm can be used to reduce the effect of massive X bits. Some extended experimental results show that the proposed encoder has an acceptable-level X tolerant capacity and low error cancellations probability.
Cygrid: Cython-powered convolution-based gridding module for Python
Winkel, B.; Lenz, D.; Flöer, L.
2016-06-01
The Python module Cygrid grids (resamples) data to any collection of spherical target coordinates, although its typical application involves FITS maps or data cubes. The module supports the FITS world coordinate system (WCS) standard; its underlying algorithm is based on the convolution of the original samples with a 2D Gaussian kernel. A lookup table scheme allows parallelization of the code and is combined with the HEALPix tessellation of the sphere for fast neighbor searches. Cygrid's runtime scales between O(n) and O(nlog n), with n being the number of input samples.
Robust and accurate transient light transport decomposition via convolutional sparse coding.
Hu, Xuemei; Deng, Yue; Lin, Xing; Suo, Jinli; Dai, Qionghai; Barsi, Christopher; Raskar, Ramesh
2014-06-01
Ultrafast sources and detectors have been used to record the time-resolved scattering of light propagating through macroscopic scenes. In the context of computational imaging, decomposition of this transient light transport (TLT) is useful for applications, such as characterizing materials, imaging through diffuser layers, and relighting scenes dynamically. Here, we demonstrate a method of convolutional sparse coding to decompose TLT into direct reflections, inter-reflections, and subsurface scattering. The method relies on the sparsity composition of the time-resolved kernel. We show that it is robust and accurate to noise during the acquisition process.
The Approximation Theorem of Convolution Operator in △p Set-valued Function Space
Institute of Scientific and Technical Information of China (English)
Pei-xin Ye
2002-01-01
The paper is a contribution to the problem of approximating random set with values in a separable Banach space. This class of set-valued function is widely used in many areas.We investigate the properties of p-bounded integrable random set. Based on this we endow it with △p metric which can be viewed as a integral type hausdorff metric and present some approximation theorem of a class of convolution operators with respect to △p metric. Moreover we also can establish analogous theorem for other integral type operator in △p space,
Cross-model convolutional neural network for multiple modality data representation
Wu, Yanbin; Wang, Li; Cui, Fan; Zhai, Hongbin; Dong, Baoming; Wang, Jim Jing-Yan
2016-01-01
A novel data representation method of convolutional neural net- work (CNN) is proposed in this paper to represent data of different modalities. We learn a CNN model for the data of each modality to map the data of differ- ent modalities to a common space, and regularize the new representations in the common space by a cross-model relevance matrix. We further impose that the class label of data points can also be predicted from the CNN representa- tions in the common space. The learning proble...
DEFF Research Database (Denmark)
Wulff-Jensen, Andreas; Rant, Niclas Nerup; Møller, Tobias Nordvig
2017-01-01
This paper proposes a novel framework for improving procedural generation of 3D landscapes using machine learning. We utilized a Deep Convolutional Generative Adversarial Network (DC-GAN) to generate heightmaps. The network was trained on a dataset consisting of Digital Elevation Maps (DEM......) of the alps. During map generation, the batch size and learning rate were optimized for the most efficient and satisfying map production. The diversity of the final output was tested against Perlin noise using Mean Square Error [1] and Structure Similarity Index [2]. Perlin noise is especially interesting...
A convolutional recursive modified Self Organizing Map for handwritten digits recognition.
Mohebi, Ehsan; Bagirov, Adil
2014-12-01
It is well known that the handwritten digits recognition is a challenging problem. Different classification algorithms have been applied to solve it. Among them, the Self Organizing Maps (SOM) produced promising results. In this paper, first we introduce a Modified SOM for the vector quantization problem with improved initialization process and topology preservation. Then we develop a Convolutional Recursive Modified SOM and apply it to the problem of handwritten digits recognition. The computational results obtained using the well known MNIST dataset demonstrate the superiority of the proposed algorithm over the existing SOM-based algorithms.
A low-memory intensive decoding architecture for double-binary convolutional turbo code
Zhan, Ming; Zhou, Liang; Wu, Jun
2014-01-01
Memory accesses take a large part of the power consumption in the iterative decoding of double-binary convolutional turbo code (DB-CTC). To deal with this, a low-memory intensive decoding architecture is proposed for DB-CTC in this paper. The new scheme is based on an improved maximum a posteriori probability algorithm, where instead of storing all of the state metrics, only a part of these state metrics is stored in the state metrics cache (SMC), and the memory size of the SMC is thus ...
Deep convolutional neural networks for dense non-uniform motion deblurring
CSIR Research Space (South Africa)
Cronje, J
2015-11-01
Full Text Available stream_source_info Cronje_2015_ABSTRACT.pdf.txt stream_content_type text/plain stream_size 885 Content-Encoding ISO-8859-1 stream_name Cronje_2015_ABSTRACT.pdf.txt Content-Type text/plain; charset=ISO-8859-1 Computer... Vision and Pattern Recognition Deep Convolutional Neural Networks for Dense Non- Uniform Motion Deblurring Jaco Cronje Council for Scientific and Industrial Research, Pretoria, South Africa Email: jcronje@csir.co.za Abstract The work...
Excursion sets of infinitely divisible random fields with convolution equivalent Lévy measure
DEFF Research Database (Denmark)
Rønn-Nielsen, Anders; Jensen, Eva B. Vedel
We consider a continuous, infinitely divisible random field in R d , d = 1, 2, 3, given as an integral of a kernel function with respect to a Lévy basis with convolution equivalent Lévy measure. For a large class of such random fields we compute the asymptotic probability that the excursion set a...... at level x contains some rotation of an object with fixed radius as x → ∞. Our main result is that the asymptotic probability is equivalent to the right tail of the underlying Lévy measure...
Real-time convolution method for generating light diffusion profiles of layered turbid media.
Kim, Hoe-Min; Ko, Kwang Hee; Lee, Kwan H
2011-06-01
In this paper we present a technique to obtain a diffusion profile of layered turbid media in real time by using the quasi fast Hankel transform (QFHT) and the latest graphics processing unit technique. We apply the QFHT to convolve the diffusion profiles of each layer so as to dramatically reduce the time for the convolution step while maintaining the accuracy. In addition, we also introduce an accelerated technique to generate individual discrete diffusion profiles for each layer through parallel processing. The proposed method is 2 orders of magnitude faster than the existing method, and we validate its efficiency by comparing it with Monte Carlo simulation and another relevant methods.
Firdaus; Arkeman, Y.; Buono, A.; Hermadi, I.
2017-01-01
Translating satellite imagery to a useful data for decision making during this time are usually done manually by human. In this research, we are going to translate satellite imagery by using artificial intelligence method specifically using convolutional neural network and genetic algorithm to become a useful data for decision making, especially for precision agriculture and agroindustry. In this research, we are focused on how to made a sustainable land use planning with 3 objectives. The first is maximizing economic factor. Second is minimizing CO2 emission and the last is minimizing land degradation. Results show that by using artificial intelligence method, can produced a good pareto optimum solutions in a short time.
Jordan, Tyler S.
2016-05-01
This paper presents the findings of using convolutional neural networks (CNNs) to classify human activity from micro-Doppler features. An emphasis on activities involving potential security threats such as holding a gun are explored. An automotive 24 GHz radar on chip was used to collect the data and a CNN (normally applied to image classification) was trained on the resulting spectrograms. The CNN achieves an error rate of 1.65 % on classifying running vs. walking, 17.3 % error on armed walking vs. unarmed walking, and 22 % on classifying six different actions.
tf_unet: Generic convolutional neural network U-Net implementation in Tensorflow
Akeret, Joel; Chang, Chihway; Lucchi, Aurelien; Refregier, Alexandre
2016-11-01
tf_unet mitigates radio frequency interference (RFI) signals in radio data using a special type of Convolutional Neural Network, the U-Net, that enables the classification of clean signal and RFI signatures in 2D time-ordered data acquired from a radio telescope. The code is not tied to a specific segmentation and can be used, for example, to detect radio frequency interference (RFI) in radio astronomy or galaxies and stars in widefield imaging data. This U-Net implementation can outperform classical RFI mitigation algorithms.
Yoon, Seungjong; Kim, Eungtae
2017-02-01
In this paper, we propose the method that classifies the traffic signs by using Convolutional Neural Network(CNN) and compensates the error rate of CNN using the temporal correlation between adjacent successive frames. Instead of applying a conventional CNN architecture with more layers, Temporal Classification Error Compensation(TCEC) is proposed to improve the error rate in the architecture which has less nodes and layers than a conventional CNN. Experimental results show that the complexity of the proposed method could be reduced by 50% compared with that of the conventional CNN with same layers, and the error rate could be improved by about 3%.
New Spatial Modulation with Convolutional Code-Aided Constellation%基于卷积编码的发射天线星座图空间调制系统
Institute of Scientific and Technical Information of China (English)
辛秀; 杨芷华; 方泽凯
2015-01-01
为了提高空间调制系统的可靠性，本文提出了一种基于卷积编码辅助设计发射天线星座图的空间调制系统，有效利用卷积码的纠错能力，提高MIMO通信系统的误码性能。仿真实验结果表明：在同一条件下，基于卷积编码辅助设计发射天线星座图的空间调制系统优于汉明码辅助设计星座图的空移键控调制系统和网格编码空间调制系统，分别有10 dB和8 dB的增益。%In order to improve the reliability of the space modulation system, this paper proposes a type of spatial modulating system based on convolution code-aided transmitting antenna constellation spatial modulation system. The proposed scheme employs the error-correcting ability of the convolution code to improve the performance of the MIMO communication system. Simulation results show that under the same conditions, the spatial modulation system based on the transmitting antenna constellation diagram aided by convolutional coding is superior to the shift keying modulation system based on the constellation graph aided by the Hamming code and to the spatial modulation system based on the trellis code, respectively with an extra gain of10dBand8dB.
Lee, L.-N.
1977-01-01
Concatenated coding systems utilizing a convolutional code as the inner code and a Reed-Solomon code as the outer code are considered. In order to obtain very reliable communications over a very noisy channel with relatively modest coding complexity, it is proposed to concatenate a byte-oriented unit-memory convolutional code with an RS outer code whose symbol size is one byte. It is further proposed to utilize a real-time minimal-byte-error probability decoding algorithm, together with feedback from the outer decoder, in the decoder for the inner convolutional code. The performance of the proposed concatenated coding system is studied, and the improvement over conventional concatenated systems due to each additional feature is isolated.
Keeling, Stephen L; Bammer, Roland; Stollberger, Rudolf
2007-09-01
Counterexamples are used to motivate the revision of the established theory of tracer transport. Then dynamic contrast enhanced magnetic resonance imaging in particular is conceptualized in terms of a fully distributed convection-diffusion model from which a widely used convolution model is derived using, alternatively, compartmental discretizations or semigroup theory. On this basis, applications and limitations of the convolution model are identified. For instance, it is proved that perfusion and tissue exchange states cannot be identified on the basis of a single convolution equation alone. Yet under certain assumptions, particularly that flux is purely convective at the boundary of a tissue region, physiological parameters such as mean transit time, effective volume fraction, and volumetric flow rate per unit tissue volume can be deduced from the kernel.
Some problems on the convolution of distributional Denjoy integral%广义Denjoy可积函数的卷积
Institute of Scientific and Technical Information of China (English)
刘巧玲; 叶国菊; 刘尉
2011-01-01
Some problems on the convolution of distributional Denjoy integrable functions are discussed. The definition of the convolution of two distributional Denjoy integrable functions is presented. It is proven that the convolution of distributional Denjoy integrable functions is distributional Denjoy integrable when one of their primitive is bounded variation function.%讨论广义Denjoy可积函数的卷积问题.给出两个广义Denjoy可积函数卷积的定义,并且证明当其中一个广义Denjoy可积函数的原函数为有界变差时,这时的卷积也是广义Denjoy可积函数.
Directory of Open Access Journals (Sweden)
Dong Hyun Cho
2017-01-01
Full Text Available Using a simple formula for conditional expectations over continuous paths, we will evaluate conditional expectations which are types of analytic conditional Fourier-Feynman transforms and conditional convolution products of generalized cylinder functions and the functions in a Banach algebra which is the space of generalized Fourier transforms of the measures on the Borel class of L2[0,T]. We will then investigate their relationships. Particularly, we prove that the conditional transform of the conditional convolution product can be expressed by the product of the conditional transforms of each function. Finally we will establish change of scale formulas for the conditional transforms and the conditional convolution products. In these evaluation formulas and change of scale formulas, we use multivariate normal distributions so that the conditioning function does not contain present positions of the paths.
An efficient de-convolution reconstruction method for spatiotemporal-encoding single-scan 2D MRI.
Cai, Congbo; Dong, Jiyang; Cai, Shuhui; Li, Jing; Chen, Ying; Bao, Lijun; Chen, Zhong
2013-03-01
Spatiotemporal-encoding single-scan MRI method is relatively insensitive to field inhomogeneity compared to EPI method. Conjugate gradient (CG) method has been used to reconstruct super-resolved images from the original blurred ones based on coarse magnitude-calculation. In this article, a new de-convolution reconstruction method is proposed. Through removing the quadratic phase modulation from the signal acquired with spatiotemporal-encoding MRI, the signal can be described as a convolution of desired super-resolved image and a point spread function. The de-convolution method proposed herein not only is simpler than the CG method, but also provides super-resolved images with better quality. This new reconstruction method may make the spatiotemporal-encoding 2D MRI technique more valuable for clinic applications. Copyright © 2013 Elsevier Inc. All rights reserved.
Correction method for line extraction in vision measurement.
Directory of Open Access Journals (Sweden)
Mingwei Shao
Full Text Available Over-exposure and perspective distortion are two of the main factors underlying inaccurate feature extraction. First, based on Steger's method, we propose a method for correcting curvilinear structures (lines extracted from over-exposed images. A new line model based on the Gaussian line profile is developed, and its description in the scale space is provided. The line position is analytically determined by the zero crossing of its first-order derivative, and the bias due to convolution with the normal Gaussian kernel function is eliminated on the basis of the related description. The model considers over-exposure features and is capable of detecting the line position in an over-exposed image. Simulations and experiments show that the proposed method is not significantly affected by the exposure level and is suitable for correcting lines extracted from an over-exposed image. In our experiments, the corrected result is found to be more precise than the uncorrected result by around 45.5%. Second, we analyze perspective distortion, which is inevitable during line extraction owing to the projective camera model. The perspective distortion can be rectified on the basis of the bias introduced as a function of related parameters. The properties of the proposed model and its application to vision measurement are discussed. In practice, the proposed model can be adopted to correct line extraction according to specific requirements by employing suitable parameters.
Segmentation of Bone Structure in X-ray Images using Convolutional Neural Network
Directory of Open Access Journals (Sweden)
CERNAZANU-GLAVAN, C.
2013-02-01
Full Text Available The segmentation process represents a first step necessary for any automatic method of extracting information from an image. In the case of X-ray images, through segmentation we can differentiate the bone tissue from the rest of the image. There are nowadays several segmentation techniques, but in general, they all require the human intervention in the segmentation process. Consequently, this article proposes a new segmentation method for the X-ray images using a Convolutional Neural Network (CNN. In present, the convolutional networks are the best techniques for image segmentation. This fact is demonstrated by their wide usage in all the fields, including the medical one. As the X-ray images have large dimensions, for reducing the training time, the method proposed by the present article selects only certain areas (maximum interest areas from the entire image. The neural network is used as pixel classifier thus causing the label of each pixel (bone or none-bone from a raw pixel values in a square area. We will also present the method through which the network final configuration was chosen and we will make a comparative analysis with other 3 CNN configurations. The network chosen by us obtained the best results for all the evaluation metrics used, i.e. warping error, rand error and pixel error.
Muon Neutrino Disappearance in NOvA with a Deep Convolutional Neural Network Classifier
Rocco, Dominick Rosario
The NuMI Off-axis Neutrino Appearance Experiment (NOvA) is designed to study neutrino oscillation in the NuMI (Neutrinos at the Main Injector) beam. NOvA observes neutrino oscillation using two detectors separated by a baseline of 810 km; a 14 kt Far Detector in Ash River, MN and a functionally identical 0.3 kt Near Detector at Fermilab. The experiment aims to provide new measurements of $[special characters omitted]. and theta23 and has potential to determine the neutrino mass hierarchy as well as observe CP violation in the neutrino sector. Essential to these analyses is the classification of neutrino interaction events in NOvA detectors. Raw detector output from NOvA is interpretable as a pair of images which provide orthogonal views of particle interactions. A recent advance in the field of computer vision is the advent of convolutional neural networks, which have delivered top results in the latest image recognition contests. This work presents an approach novel to particle physics analysis in which a convolutional neural network is used for classification of particle interactions. The approach has been demonstrated to improve the signal efficiency and purity of the event selection, and thus physics sensitivity. Early NOvA data has been analyzed (2.74 x 1020 POT, 14 kt equivalent) to provide new best-fit measurements of sin2(theta23) = 0.43 (with a statistically-degenerate compliment near 0.60) and [special characters omitted]..
Liu, Ziyi; Gao, Junfeng; Yang, Guoguo; Zhang, Huan; He, Yong
2016-01-01
We present a pipeline for the visual localization and classification of agricultural pest insects by computing a saliency map and applying deep convolutional neural network (DCNN) learning. First, we used a global contrast region-based approach to compute a saliency map for localizing pest insect objects. Bounding squares containing targets were then extracted, resized to a fixed size, and used to construct a large standard database called Pest ID. This database was then utilized for self-learning of local image features which were, in turn, used for classification by DCNN. DCNN learning optimized the critical parameters, including size, number and convolutional stride of local receptive fields, dropout ratio and the final loss function. To demonstrate the practical utility of using DCNN, we explored different architectures by shrinking depth and width, and found effective sizes that can act as alternatives for practical applications. On the test set of paddy field images, our architectures achieved a mean Accuracy Precision (mAP) of 0.951, a significant improvement over previous methods.
Strahl, Stefan B; Ramekers, Dyan; Nagelkerke, Marjolijn M B; Schwarz, Konrad E; Spitzer, Philipp; Klis, Sjaak F L; Grolman, Wilko; Versnel, Huib
2016-01-01
The electrically evoked compound action potential (eCAP) is a routinely performed measure of the auditory nerve in cochlear implant users. Using a convolution model of the eCAP, additional information about the neural firing properties can be obtained, which may provide relevant information about the health of the auditory nerve. In this study, guinea pigs with various degrees of nerve degeneration were used to directly relate firing properties to nerve histology. The same convolution model was applied on human eCAPs to examine similarities and ultimately to examine its clinical applicability. For most eCAPs, the estimated nerve firing probability was bimodal and could be parameterised by two Gaussian distributions with an average latency difference of 0.4 ms. The ratio of the scaling factors of the late and early component increased with neural degeneration in the guinea pig. This ratio decreased with stimulation intensity in humans. The latency of the early component decreased with neural degeneration in the guinea pig. Indirectly, this was observed in humans as well, assuming that the cochlear base exhibits more neural degeneration than the apex. Differences between guinea pigs and humans were observed, among other parameters, in the width of the early component: very robust in guinea pig, and dependent on stimulation intensity and cochlear region in humans. We conclude that the deconvolution of the eCAP is a valuable addition to existing analyses, in particular as it reveals two separate firing components in the auditory nerve.
Projection of fMRI data onto the cortical surface using anatomically-informed convolution kernels.
Operto, G; Bulot, R; Anton, J-L; Coulon, O
2008-01-01
As surface-based data analysis offer an attractive approach for intersubject matching and comparison, the projection of voxel-based 3D volumes onto the cortical surface is an essential problem. We present here a method that aims at producing representations of functional brain data on the cortical surface from functional MRI volumes. Such representations are for instance required for subsequent cortical-based functional analysis. We propose a projection technique based on the definition, around each node of the gray/white matter interface mesh, of convolution kernels whose shape and distribution rely on the geometry of the local anatomy. For one anatomy, a set of convolution kernels is computed that can be used to project any functional data registered with this anatomy. Therefore resulting in anatomically-informed projections of data onto the cortical surface, this kernel-based approach offers better sensitivity, specificity than other classical methods and robustness to misregistration errors. Influences of mesh and volumes spatial resolutions were also estimated for various projection techniques, using simulated functional maps.
Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition
Ordóñez, Francisco Javier; Roggen, Daniel
2016-01-01
Human activity recognition (HAR) tasks have traditionally been solved using engineered features obtained by heuristic processes. Current research suggests that deep convolutional neural networks are suited to automate feature extraction from raw sensor inputs. However, human activities are made of complex sequences of motor movements, and capturing this temporal dynamics is fundamental for successful HAR. Based on the recent success of recurrent neural networks for time series domains, we propose a generic deep framework for activity recognition based on convolutional and LSTM recurrent units, which: (i) is suitable for multimodal wearable sensors; (ii) can perform sensor fusion naturally; (iii) does not require expert knowledge in designing features; and (iv) explicitly models the temporal dynamics of feature activations. We evaluate our framework on two datasets, one of which has been used in a public activity recognition challenge. Our results show that our framework outperforms competing deep non-recurrent networks on the challenge dataset by 4% on average; outperforming some of the previous reported results by up to 9%. Our results show that the framework can be applied to homogeneous sensor modalities, but can also fuse multimodal sensors to improve performance. We characterise key architectural hyperparameters’ influence on performance to provide insights about their optimisation. PMID:26797612
Deep Convolutional Networks for Event Reconstruction and Particle Tagging on NOvA and DUNE
CERN. Geneva
2017-01-01
Deep Convolutional Neural Networks (CNNs) have been widely applied in computer vision to solve complex problems in image recognition and analysis. In recent years many efforts have emerged to extend the use of this technology to HEP applications, including the Convolutional Visual Network (CVN), our implementation for identification of neutrino events. In this presentation I will describe the core concepts of CNNs, the details of our particular implementation in the Caffe framework and our application to identify NOvA events. NOvA is a long baseline neutrino experiment whose main goal is the measurement of neutrino oscillations. This relies on the accurate identification and reconstruction of the neutrino flavor in the interactions we observe. In 2016 the NOvA experiment released results for the observation of oscillations in the ν μ → ν e channel, the first HEP result employing CNNs. I will also discuss our approach at event identification on NOvA as well as recent developments in the application of CNN...