Sample records for ultrasensitive imaging method

  1. Ultra-sensitive chemiluminescence imaging DNA hybridization method in the detection of mosquito-borne viruses and parasites. (United States)

    Zhang, Yingjie; Liu, Qiqi; Zhou, Biao; Wang, Xiaobo; Chen, Suhong; Wang, Shengqi


    Mosquito-borne viruses (MBVs) and parasites (MBPs) are transmitted through hematophagous arthropods-mosquitoes to homoiothermous vertebrates. This study aims at developing a detection method to monitor the spread of mosquito-borne diseases to new areas and diagnose the infections caused by MBVs and MBPs. In this assay, an ultra-sensitive chemiluminescence (CL) detection method was developed and used to simultaneously detect 19 common MBVs and MBPs. In vitro transcript RNA, virus-like particles (VLPs), and plasmids were established as positive or limit of detection (LOD) reference materials. MBVs and MBPs could be genotyped with high sensitivity and specificity. The cut-off values of probes were calculated. The absolute LODs of this strategy to detect serially diluted in vitro transcribed RNAs of MBVs and serially diluted plasmids of MBPs were 10 2 -10 3 copies/μl and 10 1 -10 2 copies/μl, respectively. Further, the LOD of detecting a strain of pre-quantified JEV was 10 1.8 -10 0.8 PFU/ml, fitted well in a linear regression model (coefficient of determination = 0.9678). Ultra-sensitive CL imaging DNA hybridization was developed and could simultaneously detect various MBVs and MBPs. The method described here has the potential to provide considerable labor savings due to its ability to screen for 19 mosquito-borne pathogens simultaneously.

  2. An ultrasensitive method of real time pH monitoring with complementary metal oxide semiconductor image sensor. (United States)

    Devadhasan, Jasmine Pramila; Kim, Sanghyo


    CMOS sensors are becoming a powerful tool in the biological and chemical field. In this work, we introduce a new approach on quantifying various pH solutions with a CMOS image sensor. The CMOS image sensor based pH measurement produces high-accuracy analysis, making it a truly portable and user friendly system. pH indicator blended hydrogel matrix was fabricated as a thin film to the accurate color development. A distinct color change of red, green and blue (RGB) develops in the hydrogel film by applying various pH solutions (pH 1-14). The semi-quantitative pH evolution was acquired by visual read out. Further, CMOS image sensor absorbs the RGB color intensity of the film and hue value converted into digital numbers with the aid of an analog-to-digital converter (ADC) to determine the pH ranges of solutions. Chromaticity diagram and Euclidean distance represent the RGB color space and differentiation of pH ranges, respectively. This technique is applicable to sense the various toxic chemicals and chemical vapors by situ sensing. Ultimately, the entire approach can be integrated into smartphone and operable with the user friendly manner. Copyright © 2014 Elsevier B.V. All rights reserved.

  3. ATTA - A new method of ultrasensitive isotope trace analysis

    International Nuclear Information System (INIS)

    Bailey, K.; Chen, C.Y.; Du, X.; Li, Y.M.; Lu, Z.-T.; O'Connor, T.P.; Young, L.


    A new method of ultrasensitive isotope trace analysis has been developed. This method, based on the technique of laser manipulation of neutral atoms, has been used to count individual 85 Kr and 81 Kr atoms present in a natural krypton gas sample with isotopic abundances in the range of 10 -11 and 10 -13 , respectively. This method is free of contamination from other isotopes and elements and can be applied to various different isotope tracers for a wide range of applications. The demonstrated detection efficiency is 1x10 -7 . System improvements could increase the efficiency by many orders of magnitude

  4. Smart nanoprobes for ultrasensitive detection of breast cancer via magnetic resonance imaging

    International Nuclear Information System (INIS)

    Lee, Jaemin; Yang, Jaemoon; Seo, Sung-Baek; Haam, Seungjoo; Ko, Hyun-Ju; Suh, Jin-Suck; Huh, Yong-Min


    Antibody-conjugated hydrophilic magnetic nanocrystals for use as smart nanoprobes were developed for ultrasensitive detection of breast cancer via magnetic resonance (MR) imaging. MnFe 2 O 4 nanocrystals (MNCs) for use as MR imaging contrast agents were synthesized by thermal decomposition to take advantage of their MR signal enhancement effect. The MNC surfaces were then modified with amphiphilic tri-block copolymers (dicarboxy poly(ethylene glycol)-block-poly(propylene glycol)-block-poly(ethylene glycol)), not only allowing the MNCs to transfer from the organic to the aqueous phase, but also increasing the colloidal stability of the MNCs by masking poly(ethylene glycol). The physicochemical properties of the synthesized hydrophilic magnetic nanocrystals (HMNCs) were fully investigated. Trastuzumab (TZ), a monoclonal antibody against human epidermal growth factor receptor (HER2/neu), was further conjugated on the surface of HMNCs to specifically target HER2/neu over-expressed breast cancer cells. MR imaging analysis of target cells treated with TZ-conjugated HMNCs (TZ-HMNCs) clearly demonstrated their potential as high-performance nanoprobes for selective imaging.

  5. Mechanical modulation method for ultrasensitive phase measurements in photonics biosensing. (United States)

    Patskovsky, S; Maisonneuve, M; Meunier, M; Kabashin, A V


    A novel polarimetry methodology for phase-sensitive measurements in single reflection geometry is proposed for applications in optical transduction-based biological sensing. The methodology uses altering step-like chopper-based mechanical phase modulation for orthogonal s- and p- polarizations of light reflected from the sensing interface and the extraction of phase information at different harmonics of the modulation. We show that even under a relatively simple experimental arrangement, the methodology provides the resolution of phase measurements as low as 0.007 deg. We also examine the proposed approach using Total Internal Reflection (TIR) and Surface Plasmon Resonance (SPR) geometries. For TIR geometry, the response appears to be strongly dependent on the prism material with the best values for high refractive index Si. The detection limit for Si-based TIR is estimated as 10(-5) in terms Refractive Index Units (RIU) change. SPR geometry offers much stronger phase response due to a much sharper phase characteristics. With the detection limit of 3.2*10(-7) RIU, the proposed methodology provides one of best sensitivities for phase-sensitive SPR devices. Advantages of the proposed method include high sensitivity, simplicity of experimental setup and noise immunity as a result of a high stability modulation.

  6. A Rapid, Onsite, Ultrasensitive Melamine Quantitation Method for Protein Beverages Using Time-Resolved Fluorescence Detection Paper. (United States)

    Li, Guanghua; Wang, Du; Zhou, Aijun; Sun, Yimin; Zhang, Qi; Poapolathep, Amnart; Zhang, Li; Fan, Zhiyong; Zhang, Zhaowei; Li, Peiwu


    To ensure protein beverage safety and prevent illegal melamine use to artificially increase protein content, a rapid, onsite, ultrasensitive detection method for melamine must be developed because melamine is detrimental to human health and life. Herein, an ultrasensitive time-resolved fluorescence detection paper (TFDP) was developed to detect melamine in protein beverages within 15 min using a one-step sample preparation. The lower limits of detection were 0.89, 0.94, and 1.05 ng/mL, and the linear ranges were 2.67-150, 2.82-150, and 3.15-150 ng/mL (R2>0.982) for peanut, walnut, and coconut beverages, respectively. The recovery rates were 85.86-110.60% with a coefficient of variation beverage samples, the TFDP and ultra-performance liquid chromatography-tandem mass spectrometer (UPLC-MS/MS) results were consistent. This method is a promising alternative for rapid, onsite detection of melamine in beverages.

  7. Ultra-sensitive "turn-on" detection method for Hg(2+) based on mispairing biosensor and emulsion PCR. (United States)

    Zhu, Pengyu; Tian, Wenying; Cheng, Nan; Huang, Kunlun; Luo, Yunbo; Xu, Wentao


    Sensor-based detection methods have inspired the idea that chemical or physical signals could be converted to nucleic acid signals to be quantitatively detected using a combination of appropriate detection tools. To achieve ultra-sensitive and absolute quantitative detection of mercury ion (Hg(2+)), we have combined a mispairing biosensor for Hg(2+) and emulsion PCR. The parameters that might influence the biosensor step, such as the duration of isothermal amplification and the concentration of the sensor oligonucleotide, have been firstly optimized in our study to achieve the most efficient biosensor detection. The evaluation results of secondary structures between the biosensors with different number of T-Hg-T structures achieved by Circular Dichroism have indicated that the secondary hairpin structure would be varied according to the change of number of T-Hg-T structures, which could influence the quantitative detection results. Further optimization of number of T-Hg-T within the biosensor sequences showed that 5 T-Hg-T structures could generate the most efficient amplification. After the above optimizations, the emulsion PCR has been employed to achieve the absolute quantitation of nucleic acid signals. The final results have shown that the limit of quantitation (LOQ) in our study was as low as 40fmol, and the limit of detection (LOD) was 10fmol. The practical detection tests showed that the quantitative results were stable and accurate for all substrates. In conclusion, by combining a mispairing biosensor with emulsion PCR, we developed a flexible and stable quantitative "turn-on" detection method with ultra-sensitivity that can detect trace amounts Hg(2+) within different substrates. Copyright © 2016 Elsevier B.V. All rights reserved.

  8. Ultrasensitive near-infrared fluorescence-enhanced probe for in vivo nitroreductase imaging. (United States)

    Li, Yuhao; Sun, Yun; Li, Jiachang; Su, Qianqian; Yuan, Wei; Dai, Yu; Han, Chunmiao; Wang, Qiuhong; Feng, Wei; Li, Fuyou


    Nitroreductase (NTR) can be overexpressed in hypoxic tumors, thus the selective and efficient detection of NTR is of great importance. To date, although a few optical methods have been reported for the detection of NTR in solution, an effective optical probe for NTR monitoring in vivo is still lacking. Therefore, it is necessary to develop a near-infrared (NIR) fluorescent detection probe for NTR. In this study, five NIR cyanine dyes with fluorescence reporting structure decorated with different nitro aromatic groups, Cy7-1-5, have been designed and explored for possible rapid detection of NTR. Our experimental results presented that only a para-nitro benzoate group modified cyanine probe (Cy7-1) could serve as a rapid NIR fluorescence-enhanced probe for monitoring and bioimaging of NTR. The structure-function relationship has been revealed by theoretical study. The linker connecting the detecting and fluorescence reporting groups and the nitro group position is a key factor for the formation of hydrogen bonds and spatial structure match, inducing the NTR catalytic ability enhancement. The in vitro response and mechanism of the enzyme-catalyzed reduction of Cy7-1 have been investigated through kinetic optical studies and other methods. The results have indicated that an electro-withdrawing group induced electron-transfer process becomes blocked when Cy7-1 is catalytically reduced to Cy7-NH2 by NTR, which is manifested in enhanced fluorescence intensity during the detection process. Confocal fluorescence imaging of hypoxic A549 cells has confirmed the NTR detection ability of Cy7-1 at the cellular level. Importantly, Cy7-1 can detect tumor hypoxia in a murine hypoxic tumor model, showing a rapid and significant enhancement of its NIR fluorescence characteristics suitable for fluorescence bioimaging. This method may potentially be used for tumor hypoxia diagnosis.

  9. A novel method for extracting nucleic acids from dried blood spots for ultrasensitive detection of low-density Plasmodium falciparum and Plasmodium vivax infections. (United States)

    Zainabadi, Kayvan; Adams, Matthew; Han, Zay Yar; Lwin, Hnin Wai; Han, Kay Thwe; Ouattara, Amed; Thura, Si; Plowe, Christopher V; Nyunt, Myaing M


    Greater Mekong Subregion countries are committed to eliminating Plasmodium falciparum malaria by 2025. Current elimination interventions target infections at parasite densities that can be detected by standard microscopy or rapid diagnostic tests (RDTs). More sensitive detection methods have been developed to detect lower density "asymptomatic" infections that may represent an important transmission reservoir. These ultrasensitive polymerase chain reaction (usPCR) tests have been used to identify target populations for mass drug administration (MDA). To date, malaria usPCR tests have used either venous or capillary blood sampling, which entails complex sample collection, processing and shipping requirements. An ultrasensitive method performed on standard dried blood spots (DBS) would greatly facilitate the molecular surveillance studies needed for targeting elimination interventions. A highly sensitive method for detecting Plasmodium falciparum and P. vivax 18S ribosomal RNA from DBS was developed by empirically optimizing nucleic acid extraction conditions. The limit of detection (LoD) was determined using spiked DBS samples that were dried and stored under simulated field conditions. Further, to assess its utility for routine molecular surveillance, two cross-sectional surveys were performed in Myanmar during the wet and dry seasons. The lower LoD of the DBS-based ultrasensitive assay was 20 parasites/mL for DBS collected on Whatman 3MM filter paper and 23 parasites/mL for Whatman 903 Protein Saver cards-equivalent to 1 parasite per 50 µL DBS. This is about 5000-fold more sensitive than standard RDTs and similar to the LoD of ≤16-22 parasites/mL reported for other ultrasensitive methods based on whole blood. In two cross-sectional surveys in Myanmar, nearly identical prevalence estimates were obtained from contemporaneous DBS samples and capillary blood samples collected during the wet and dry season. The DBS-based ultrasensitive method described in this

  10. Magnetically engineered smart thin films: toward lab-on-chip ultra-sensitive molecular imaging. (United States)

    Hassan, Muhammad A; Saqib, Mudassara; Shaikh, Haseeb; Ahmad, Nasir M; Elaissari, Abdelhamid


    Magnetically responsive engineered smart thin films of nanoferrites as contrast agent are employed to develop surface based magnetic resonance imaging to acquire simple yet fast molecular imaging. The work presented here can be of significant potential for future lab-on-chip point-of-care diagnostics from the whole blood pool on almost any substrates to reduce or even prevent clinical studies involve a living organism to enhance the non-invasive imaging to advance the '3Rs' of work in animals-replacement, refinement and reduction.

  11. Terahertz composite imaging method

    Institute of Scientific and Technical Information of China (English)

    QIAO Xiaoli; REN Jiaojiao; ZHANG Dandan; CAO Guohua; LI Lijuan; ZHANG Xinming


    In order to improve the imaging quality of terahertz(THz) spectroscopy, Terahertz Composite Imaging Method(TCIM) is proposed. The traditional methods of improving THz spectroscopy image quality are mainly from the aspects of de-noising and image enhancement. TCIM breaks through this limitation. A set of images, reconstructed in a single data collection, can be utilized to construct two kinds of composite images. One algorithm, called Function Superposition Imaging Algorithm(FSIA), is to construct a new gray image utilizing multiple gray images through a certain function. The features of the Region Of Interest (ROI) are more obvious after operating, and it has capability of merging ROIs in multiple images. The other, called Multi-characteristics Pseudo-color Imaging Algorithm(McPcIA), is to construct a pseudo-color image by combining multiple reconstructed gray images in a single data collection. The features of ROI are enhanced by color differences. Two algorithms can not only improve the contrast of ROIs, but also increase the amount of information resulting in analysis convenience. The experimental results show that TCIM is a simple and effective tool for THz spectroscopy image analysis.

  12. Development of an ultra-sensitive detection method for transuranium elements with respect to ocean water samples from Fukushima

    Energy Technology Data Exchange (ETDEWEB)

    Hain, Karin Margarete


    The development of a detection method for the identification of a possible emission of actinides into the Pacific Ocean by the Fukushima Daiichi Nuclear Power Plant (FDNPP) accident, was the major aim of this PhD thesis. It is shown in this work that in addition to the isotopic plutonium ratio {sup 240}Pu/{sup 239}Pu, which is an important signature for different contamination sources, the {sup 241}Pu/{sup 239}Pu ratio has to be determined, to distinguish between a Fukushima entry of actinides and nuclear weapon fallout in the Pacific Ocean. Furthermore, this study was aiming for the improvement of the data on the neptunium ({sup 237}Np) distribution in the ocean. For the detection of the trace concentrations of actinides in the order of mBq/m{sup 3} in a small sample volume of 20 L ocean water, the ultra-sensitive method accelerator mass spectrometry (AMS) was chosen. A chemical separation procedure for Pu and Np based on extraction chromatography was developed using mass spectrometry and radiometric analysis to determine the chemical recovery. {sup 241}Am, which causes isobaric background to {sup 241}Pu in mass spectrometric measurements, was separated with a suppression of at least 10{sup -4} from Pu by this method. The detection method adjusted to Pu and Np was successfully verified by analyzing the concentration of Pu in certified reference material with AMS at the Maier-Leibnitz-Laboratory (MLL) in Munich. Due to a temporary closure of the MLL, 12 sea water samples, collected in autumn 2012, were prepared at the Radiochemie Muenchen and measured at the Vienna Environmental Research Laboratory (VERA). The sampling station closest to the FDNPP was located at a distance of 39.6 km. Three of the samples showed a slightly elevated {sup 240}Pu/{sup 239}Pu ratio of up to 0.23{sup +0.04}{sub -0.03} compared to global fallout ({sup 240}Pu/{sup 239}Pu = 0.180±0.007), whereas all measured {sup 241}Pu/{sup 239}Pu ratios were clearly consistent with nuclear weapon

  13. Ultrasensitive and selective gold film-based detection of mercury (II) in tap water using a laser scanning confocal imaging-surface plasmon resonance system in real time. (United States)

    Zhang, Hongyan; Yang, Liquan; Zhou, Bingjiang; Liu, Weimin; Ge, Jiechao; Wu, Jiasheng; Wang, Ying; Wang, Pengfei


    An ultrasensitive and selective detection of mercury (II) was investigated using a laser scanning confocal imaging-surface plasmon resonance system (LSCI-SPR). The detection limit was as low as 0.01ng/ml for Hg(2+) ions in ultrapure and tap water based on a T-rich, single-stranded DNA (ssDNA)-modified gold film, which can be individually manipulated using specific T-Hg(2+)-T complex formation. The quenching intensity of the fluorescence images for rhodamine-labeled ssDNA fitted well with the changes in SPR. The changes varied with the Hg(2+) ion concentration, which is unaffected by the presence of other metal ions. The coefficients obtained for ultrapure and tap water were 0.99902 and 0.99512, respectively, for the linear part over a range of 0.01-100ng/ml. The results show that the double-effect sensor has potential for practical applications with ultra sensitivity and selectivity, especially in online or real-time monitoring of Hg(2+) ions pollution in tap water with the further improvement of portable LSCI-SPR instrument. Copyright © 2013 Elsevier B.V. All rights reserved.

  14. Ultrasensitive Detection of Proteins on Western Blots with Semiconducting Polymer Dots


    Ye, Fangmao; Smith, Polina B.; Wu, Changfeng; Chiu, Daniel T.


    We demonstrate ultrasensitive fluorescence imaging of proteins on Western blots using a bright, compact, and orange-emitting semiconducting polymer dot (CN-PPV). We achieved a detection limit at the single-picogram level in dot blots; with conventional Western blotting, we detected 50 pg of transferrin and trypsin inhibitor after SDS-PAGE and transfer onto a PVDF membrane. Our method does not require any additional equipment or time compared to the conventional procedure with traditional fluo...

  15. Rapid flow imaging method

    International Nuclear Information System (INIS)

    Pelc, N.J.; Spritzer, C.E.; Lee, J.N.


    A rapid, phase-contrast, MR imaging method of imaging flow has been implemented. The method, called VIGRE (velocity imaging with gradient recalled echoes), consists of two interleaved, narrow flip angle, gradient-recalled acquisitions. One is flow compensated while the second has a specified flow encoding (both peak velocity and direction) that causes signals to contain additional phase in proportion to velocity in the specified direction. Complex image data from the first acquisition are used as a phase reference for the second, yielding immunity from phase accumulation due to causes other than motion. Images with pixel values equal to MΔΘ where M is the magnitude of the flow compensated image and ΔΘ is the phase difference at the pixel, are produced. The magnitude weighting provides additional vessel contrast, suppresses background noise, maintains the flow direction information, and still allows quantitative data to be retrieved. The method has been validated with phantoms and is undergoing initial clinical evaluation. Early results are extremely encouraging

  16. Magnetic imager and method (United States)

    Powell, James; Reich, Morris; Danby, Gordon


    A magnetic imager 10 includes a generator 18 for practicing a method of applying a background magnetic field over a concealed object, with the object being effective to locally perturb the background field. The imager 10 also includes a sensor 20 for measuring perturbations of the background field to detect the object. In one embodiment, the background field is applied quasi-statically. And, the magnitude or rate of change of the perturbations may be measured for determining location, size, and/or condition of the object.

  17. Methods in Astronomical Image Processing (United States)

    Jörsäter, S.

    A Brief Introductory Note History of Astronomical Imaging Astronomical Image Data Images in Various Formats Digitized Image Data Digital Image Data Philosophy of Astronomical Image Processing Properties of Digital Astronomical Images Human Image Processing Astronomical vs. Computer Science Image Processing Basic Tools of Astronomical Image Processing Display Applications Calibration of Intensity Scales Calibration of Length Scales Image Re-shaping Feature Enhancement Noise Suppression Noise and Error Analysis Image Processing Packages: Design of AIPS and MIDAS AIPS MIDAS Reduction of CCD Data Bias Subtraction Clipping Preflash Subtraction Dark Subtraction Flat Fielding Sky Subtraction Extinction Correction Deconvolution Methods Rebinning/Combining Summary and Prospects for the Future

  18. Advances in ultrasensitive mass spectrometry of organic molecules. (United States)

    Kandiah, Mathivathani; Urban, Pawel L


    Ultrasensitive mass spectrometric analysis of organic molecules is important for various branches of chemistry, and other fields including physics, earth and environmental sciences, archaeology, biomedicine, and materials science. It finds applications--as an enabling tool--in systems biology, biological imaging, clinical analysis, and forensics. Although there are a number of technical obstacles associated with the analysis of samples by mass spectrometry at ultratrace level (for example analyte losses during sample preparation, insufficient sensitivity, ion suppression), several noteworthy developments have been made over the years. They include: sensitive ion sources, loss-free interfaces, ion optics components, efficient mass analyzers and detectors, as well as "smart" sample preparation strategies. Some of the mass spectrometric methods published to date can achieve sensitivity which is by several orders of magnitude higher than that of alternative approaches. Femto- and attomole level limits of detection are nowadays common, while zepto- and yoctomole level limits of detection have also been reported. We envision that the ultrasensitive mass spectrometric assays will soon contribute to new discoveries in bioscience and other areas.

  19. Universal Image Steganalytic Method

    Directory of Open Access Journals (Sweden)

    V. Banoci


    Full Text Available In the paper we introduce a new universal steganalytic method in JPEG file format that is detecting well-known and also newly developed steganographic methods. The steganalytic model is trained by MHF-DZ steganographic algorithm previously designed by the same authors. The calibration technique with the Feature Based Steganalysis (FBS was employed in order to identify statistical changes caused by embedding a secret data into original image. The steganalyzer concept utilizes Support Vector Machine (SVM classification for training a model that is later used by the same steganalyzer in order to identify between a clean (cover and steganographic image. The aim of the paper was to analyze the variety in accuracy of detection results (ACR while detecting testing steganographic algorithms as F5, Outguess, Model Based Steganography without deblocking, JP Hide and Seek which represent the generally used steganographic tools. The comparison of four feature vectors with different lengths FBS (22, FBS (66 FBS(274 and FBS(285 shows promising results of proposed universal steganalytic method comparing to binary methods.

  20. Hyperspectral image processing methods (United States)

    Hyperspectral image processing refers to the use of computer algorithms to extract, store and manipulate both spatial and spectral information contained in hyperspectral images across the visible and near-infrared portion of the electromagnetic spectrum. A typical hyperspectral image processing work...

  1. Image registration method for medical image sequences (United States)

    Gee, Timothy F.; Goddard, James S.


    Image registration of low contrast image sequences is provided. In one aspect, a desired region of an image is automatically segmented and only the desired region is registered. Active contours and adaptive thresholding of intensity or edge information may be used to segment the desired regions. A transform function is defined to register the segmented region, and sub-pixel information may be determined using one or more interpolation methods.

  2. Ultrasensitive and selective detection of mercury (II) in serum based on the gold film sensor using a laser scanning confocal imaging-surface plasmon resonance system in real time (United States)

    Liu, Sha; Zhang, Hongyan; Liu, Weimin; Wang, Pengfei


    Hg2+ ions are one of the most toxic heavy metal ion pollutants, and are caustic and carcinogenic materials with high cellular toxicity. The Hg2+ ions can accumulate in the human body through the food chain and cause serious and permanent damage to the brain with both acute and chronic toxicity. According to the US Environment Protection Agency (EPA) guidelines, Hg2+ ions must be at concentrations below 1 ng/ml (10 nM) in drinking water. If the Hg2+ ions are higher than 2.5 ng/ml in serum, that will bring mercury poisoning. The traditional testing for Hg2+ ions includes atomic absorption, atomic fluorescence, and inductively coupled plasma mass spectrometry. These methods are usually coupled with gas chromatography, high-performance liquid chromatography, and capillary electrophoresis. However, these instrument-based techniques are rather complicated, time-consuming, costly, and unsuitable for online and portable use. An ultrasensitive and selective detection of mercury (II) in serum was investigated using a laser scanning confocal imaging-surface plasmon resonance system (LSCI-SPR). The detection limit was as low as 0.01 ng/ml for Hg2+ ions in fetal calf serum and that is lower than that was required Hg2+ ions must be at concentrations below 1 ng/ml by the US Environment Protection Agency (EPA) guidelines. This sensor was designed on a T-rich, single-stranded DNA (ssDNA)-modified gold film, which can be individually manipulated using specific T-Hg2+-T complex formation. The quenching intensity of the fluorescence images for rhodamine-labeled ssDNA fitted well with the changes in SPR. The changes varied with the Hg2+ ion concentration, which is unaffected by the presence of other metal ions. A good liner relation was got with the coefficients of 0.9116 in 30% fetal calf serums with the linear part over a range of 0.01 ng/ml to10 ng/ml.

  3. Soft tissue tumors - imaging methods

    International Nuclear Information System (INIS)

    Arlart, I.P.


    Soft Tissue Tumors - Imaging Methods: Imaging methods play an important diagnostic role in soft tissue tumors concerning a preoperative evaluation of localization, size, topographic relationship, dignity, and metastatic disease. The present paper gives an overview about diagnostic methods available today such as ultrasound, thermography, roentgenographic plain films and xeroradiography, radionuclide methods, computed tomography, lymphography, angiography, and magnetic resonance imaging. Besides sonography particularly computed tomography has the most important diagnostic value in soft tissue tumors. The application of a recently developed method, the magnetic resonance imaging, cannot yet be assessed in its significance. (orig.) [de

  4. Computational methods for molecular imaging

    CERN Document Server

    Shi, Kuangyu; Li, Shuo


    This volume contains original submissions on the development and application of molecular imaging computing. The editors invited authors to submit high-quality contributions on a wide range of topics including, but not limited to: • Image Synthesis & Reconstruction of Emission Tomography (PET, SPECT) and other Molecular Imaging Modalities • Molecular Imaging Enhancement • Data Analysis of Clinical & Pre-clinical Molecular Imaging • Multi-Modal Image Processing (PET/CT, PET/MR, SPECT/CT, etc.) • Machine Learning and Data Mining in Molecular Imaging. Molecular imaging is an evolving clinical and research discipline enabling the visualization, characterization and quantification of biological processes taking place at the cellular and subcellular levels within intact living subjects. Computational methods play an important role in the development of molecular imaging, from image synthesis to data analysis and from clinical diagnosis to therapy individualization. This work will bring readers fro...

  5. Imaging Apparatus And Method

    NARCIS (Netherlands)

    Manohar, Srirang; van Leeuwen, A.G.J.M.


    A thermoacoustic imaging apparatus comprises an electromagnetic radiation source configured to irradiate a sample area and an acoustic signal detection probe arrangement for detecting acoustic signals. A radiation responsive acoustic signal generator is added outside the sample area. The detection


    NARCIS (Netherlands)

    Manohar, Srirang; van Leeuwen, A.G.J.M.


    A thermoacoustic imaging apparatus comprises an electromagnetic radiation source configured to irradiate a sample area and an acoustic signal detection probe arrangement for detecting acoustic signals. A radiation responsive acoustic signal generator is added outside the sample area. The detection

  7. Methods of digital image processing

    International Nuclear Information System (INIS)

    Doeler, W.


    Increasing use of computerized methods for diagnostical imaging of radiological problems will open up a wide field of applications for digital image processing. The requirements set by routine diagnostics in medical radiology point to picture data storage and documentation and communication as the main points of interest for application of digital image processing. As to the purely radiological problems, the value of digital image processing is to be sought in the improved interpretability of the image information in those cases where the expert's experience and image interpretation by human visual capacities do not suffice. There are many other domains of imaging in medical physics where digital image processing and evaluation is very useful. The paper reviews the various methods available for a variety of problem solutions, and explains the hardware available for the tasks discussed. (orig.) [de

  8. An Image Registration Method for Colposcopic Images

    Directory of Open Access Journals (Sweden)

    Efrén Mezura-Montes


    sequence and a division of such image into small windows. A search process is then carried out to find the window with the highest affinity in each image of the sequence and replace it with the window in the reference image. The affinity value is based on polynomial approximation of the time series computed and the search is bounded by a search radius which defines the neighborhood of each window. The proposed approach is tested in ten 310-frame real cases in two experiments: the first one to determine the best values for the window size and the search radius and the second one to compare the best obtained results with respect to four registration methods found in the specialized literature. The obtained results show a robust and competitive performance of the proposed approach with a significant lower time with respect to the compared methods.

  9. Ultrasensitivity in phosphorylation-dephosphorylation cycles with little substrate.

    Directory of Open Access Journals (Sweden)

    Bruno M C Martins

    Full Text Available Cellular decision-making is driven by dynamic behaviours, such as the preparations for sunrise enabled by circadian rhythms and the choice of cell fates enabled by positive feedback. Such behaviours are often built upon ultrasensitive responses where a linear change in input generates a sigmoidal change in output. Phosphorylation-dephosphorylation cycles are one means to generate ultrasensitivity. Using bioinformatics, we show that in vivo levels of kinases and phosphatases frequently exceed the levels of their corresponding substrates in budding yeast. This result is in contrast to the conditions often required by zero-order ultrasensitivity, perhaps the most well known means for how such cycles become ultrasensitive. We therefore introduce a mechanism to generate ultrasensitivity when numbers of enzymes are higher than numbers of substrates. Our model combines distributive and non-distributive actions of the enzymes with two-stage binding and concerted allosteric transitions of the substrate. We use analytical and numerical methods to calculate the Hill number of the response. For a substrate with [Formula: see text] phosphosites, we find an upper bound of the Hill number of [Formula: see text], and so even systems with a single phosphosite can be ultrasensitive. Two-stage binding, where an enzyme must first bind to a binding site on the substrate before it can access the substrate's phosphosites, allows the enzymes to sequester the substrate. Such sequestration combined with competition for each phosphosite provides an intuitive explanation for the sigmoidal shifts in levels of phosphorylated substrate. Additionally, we find cases for which the response is not monotonic, but shows instead a peak at intermediate levels of input. Given its generality, we expect the mechanism described by our model to often underlay decision-making circuits in eukaryotic cells.

  10. Imaging methods in otorhinolaryngology

    International Nuclear Information System (INIS)

    Frey, K.W.; Mees, K.; Vogl, T.


    This book is the work of an otorhinolaryngologist and two radiologists, who combined their experience and efforts in order to solve a great variety and number of problems encountered in practical work, taking into account the latest technical potentials and the practical feasibility, which is determined by the equipment available. Every chapter presents the full range of diagnostic methods applicable, starting with the suitable plain radiography methods and proceeding to the various tomographic scanning methods, including conventional tomography. Every technique is assessed in terms of diagnostic value and drawbacks. (orig./MG) With 778 figs [de

  11. Image restoration and processing methods

    International Nuclear Information System (INIS)

    Daniell, G.J.


    This review will stress the importance of using image restoration techniques that deal with incomplete, inconsistent, and noisy data and do not introduce spurious features into the processed image. No single image is equally suitable for both the resolution of detail and the accurate measurement of intensities. A good general purpose technique is the maximum entropy method and the basis and use of this will be explained. (orig.)

  12. Methods in quantitative image analysis. (United States)

    Oberholzer, M; Ostreicher, M; Christen, H; Brühlmann, M


    histogram of an existing image (input image) into a new grey value histogram (output image) are most quickly handled by a look-up table (LUT). The histogram of an image can be influenced by gain, offset and gamma of the camera. Gain defines the voltage range, offset defines the reference voltage and gamma the slope of the regression line between the light intensity and the voltage of the camera. A very important descriptor of neighbourhood relations in an image is the co-occurrence matrix. The distance between the pixels (original pixel and its neighbouring pixel) can influence the various parameters calculated from the co-occurrence matrix. The main goals of image enhancement are elimination of surface roughness in an image (smoothing), correction of defects (e.g. noise), extraction of edges, identification of points, strengthening texture elements and improving contrast. In enhancement, two types of operations can be distinguished: pixel-based (point operations) and neighbourhood-based (matrix operations). The most important pixel-based operations are linear stretching of grey values, application of pre-stored LUTs and histogram equalisation. The neighbourhood-based operations work with so-called filters. These are organising elements with an original or initial point in their centre. Filters can be used to accentuate or to suppress specific structures within the image. Filters can work either in the spatial or in the frequency domain. The method used for analysing alterations of grey value intensities in the frequency domain is the Hartley transform. Filter operations in the spatial domain can be based on averaging or ranking the grey values occurring in the organising element. The most important filters, which are usually applied, are the Gaussian filter and the Laplace filter (both averaging filters), and the median filter, the top hat filter and the range operator (all ranking filters). Segmentation of objects is traditionally based on threshold grey values. (AB

  13. Numerical methods for image registration

    CERN Document Server

    Modersitzki, Jan


    Based on the author's lecture notes and research, this well-illustrated and comprehensive text is one of the first to provide an introduction to image registration with particular emphasis on numerical methods in medical imaging. Ideal for researchers in industry and academia, it is also a suitable study guide for graduate mathematicians, computer scientists, engineers, medical physicists, and radiologists.Image registration is utilised whenever information obtained from different viewpoints needs to be combined or compared and unwanted distortion needs to be eliminated. For example, CCTV imag

  14. Twin-Foucault imaging method (United States)

    Harada, Ken


    A method of Lorentz electron microscopy, which enables observation two Foucault images simultaneously by using an electron biprism instead of an objective aperture, was developed. The electron biprism is installed between two electron beams deflected by 180° magnetic domains. Potential applied to the biprism deflects the two electron beams further, and two Foucault images with reversed contrast are then obtained in one visual field. The twin Foucault images are able to extract the magnetic domain structures and to reconstruct an ordinary electron micrograph. The developed Foucault method was demonstrated with a 180° domain structure of manganite La0.825Sr0.175MnO3.

  15. Image portion identification methods, image parsing methods, image parsing systems, and articles of manufacture (United States)

    Lassahn, Gordon D.; Lancaster, Gregory D.; Apel, William A.; Thompson, Vicki S.


    Image portion identification methods, image parsing methods, image parsing systems, and articles of manufacture are described. According to one embodiment, an image portion identification method includes accessing data regarding an image depicting a plurality of biological substrates corresponding to at least one biological sample and indicating presence of at least one biological indicator within the biological sample and, using processing circuitry, automatically identifying a portion of the image depicting one of the biological substrates but not others of the biological substrates.

  16. Mathematical methods in elasticity imaging

    CERN Document Server

    Ammari, Habib; Garnier, Josselin; Kang, Hyeonbae; Lee, Hyundae; Wahab, Abdul


    This book is the first to comprehensively explore elasticity imaging and examines recent, important developments in asymptotic imaging, modeling, and analysis of deterministic and stochastic elastic wave propagation phenomena. It derives the best possible functional images for small inclusions and cracks within the context of stability and resolution, and introduces a topological derivative-based imaging framework for detecting elastic inclusions in the time-harmonic regime. For imaging extended elastic inclusions, accurate optimal control methodologies are designed and the effects of uncertainties of the geometric or physical parameters on stability and resolution properties are evaluated. In particular, the book shows how localized damage to a mechanical structure affects its dynamic characteristics, and how measured eigenparameters are linked to elastic inclusion or crack location, orientation, and size. Demonstrating a novel method for identifying, locating, and estimating inclusions and cracks in elastic...

  17. Flux dynamics in ultrasensitive superconducting focal planes (United States)

    National Aeronautics and Space Administration — The performance of superconducting focal planes will drive the achievable specifications of ultrasensitive instruments for NASA astrophysics missions, yet they have...

  18. Quantitative imaging methods in osteoporosis. (United States)

    Oei, Ling; Koromani, Fjorda; Rivadeneira, Fernando; Zillikens, M Carola; Oei, Edwin H G


    Osteoporosis is characterized by a decreased bone mass and quality resulting in an increased fracture risk. Quantitative imaging methods are critical in the diagnosis and follow-up of treatment effects in osteoporosis. Prior radiographic vertebral fractures and bone mineral density (BMD) as a quantitative parameter derived from dual-energy X-ray absorptiometry (DXA) are among the strongest known predictors of future osteoporotic fractures. Therefore, current clinical decision making relies heavily on accurate assessment of these imaging features. Further, novel quantitative techniques are being developed to appraise additional characteristics of osteoporosis including three-dimensional bone architecture with quantitative computed tomography (QCT). Dedicated high-resolution (HR) CT equipment is available to enhance image quality. At the other end of the spectrum, by utilizing post-processing techniques such as the trabecular bone score (TBS) information on three-dimensional architecture can be derived from DXA images. Further developments in magnetic resonance imaging (MRI) seem promising to not only capture bone micro-architecture but also characterize processes at the molecular level. This review provides an overview of various quantitative imaging techniques based on different radiological modalities utilized in clinical osteoporosis care and research.

  19. Methods of producing luminescent images

    International Nuclear Information System (INIS)

    Broadhead, P.; Newman, G.A.


    A method is described for producing a luminescent image in a layer of a binding material in which is dispersed a thermoluminescent material. The layer is heated uniformly to a temperature of 80 to 300 0 C and is exposed to luminescence inducing radiation whilst so heated. The preferred exposing radiation is X-rays and preferably the thermoluminescent material is insensitive to electromagnetic radiation of wavelength longer than 300 mm. Information concerning preparation of the luminescent material is given in BP 1,347,672; this material has the advantage that at elevated temperatures it shows increased sensitivity compared with room temperature. At temperatures in the range 80 to 150 0 C the thermoluminescent material exhibits 'afterglow', allowing the image to persist for several seconds after the X-radiation has ceased, thus allowing the image to be retained for visual inspection in this temperature range. At higher temperatures, however, there is negligible 'afterglow'. The thermoluminescent layers so produced are particularly useful as fluoroscopic screens. The preferred method of heating the thermoluminescent material is described in BP 1,354,149. An example is given of the application of the method. (U.K.)

  20. Digital image processing mathematical and computational methods

    CERN Document Server

    Blackledge, J M


    This authoritative text (the second part of a complete MSc course) provides mathematical methods required to describe images, image formation and different imaging systems, coupled with the principle techniques used for processing digital images. It is based on a course for postgraduates reading physics, electronic engineering, telecommunications engineering, information technology and computer science. This book relates the methods of processing and interpreting digital images to the 'physics' of imaging systems. Case studies reinforce the methods discussed, with examples of current research

  1. A method of image improvement in three-dimensional imaging

    International Nuclear Information System (INIS)

    Suto, Yasuzo; Huang, Tewen; Furuhata, Kentaro; Uchino, Masafumi.


    In general, image interpolation is required when the surface configurations of such structures as bones and organs are three-dimensionally constructed from the multi-sliced images obtained by CT. Image interpolation is a processing method whereby an artificial image is inserted between two adjacent slices to make spatial resolution equal to slice resolution in appearance. Such image interpolation makes it possible to increase the image quality of the constructed three-dimensional image. In our newly-developed algorithm, we have converted the presently and subsequently sliced images to distance images, and generated the interpolation images from these two distance images. As a result, compared with the previous method, three-dimensional images with better image quality have been constructed. (author)

  2. Ultrasensitivity in signaling cascades revisited: Linking local and global ultrasensitivity estimations.

    Directory of Open Access Journals (Sweden)

    Edgar Altszyler

    Full Text Available Ultrasensitive response motifs, capable of converting graded stimuli into binary responses, are well-conserved in signal transduction networks. Although it has been shown that a cascade arrangement of multiple ultrasensitive modules can enhance the system's ultrasensitivity, how a given combination of layers affects a cascade's ultrasensitivity remains an open question for the general case. Here, we introduce a methodology that allows us to determine the presence of sequestration effects and to quantify the relative contribution of each module to the overall cascade's ultrasensitivity. The proposed analysis framework provides a natural link between global and local ultrasensitivity descriptors and it is particularly well-suited to characterize and understand mathematical models used to study real biological systems. As a case study, we have considered three mathematical models introduced by O'Shaughnessy et al. to study a tunable synthetic MAPK cascade, and we show how our methodology can help modelers better understand alternative models.

  3. Method of assessing heterogeneity in images (United States)

    Jacob, Richard E.; Carson, James P.


    A method of assessing heterogeneity in images is disclosed. 3D images of an object are acquired. The acquired images may be filtered and masked. Iterative decomposition is performed on the masked images to obtain image subdivisions that are relatively homogeneous. Comparative analysis, such as variogram analysis or correlogram analysis, is performed of the decomposed images to determine spatial relationships between regions of the images that are relatively homogeneous.

  4. Compact, Ultrasensitive Formaldehyde Monitor, Phase I (United States)

    National Aeronautics and Space Administration — This Small Business Innovative Research Phase I proposal seeks to develop an ultrasensitive, laser-based formaldehyde gas sensor system for airborne and ground-based...


    Directory of Open Access Journals (Sweden)

    Dewa Made Sri Arsa


    Full Text Available The chromosome is a set of DNA structure that carry information about our life. The information can be obtained through Karyotyping. The process requires a clear image so the chromosome can be evaluate well. Preprocessing have to be done on chromosome images that is image enhancement. The process starts with image background removing. The image will be cleaned background color. The next step is image enhancement. This paper compares several methods for image enhancement. We evaluate some method in image enhancement like Histogram Equalization (HE, Contrast-limiting Adaptive Histogram Equalization (CLAHE, Histogram Equalization with 3D Block Matching (HE+BM3D, and basic image enhancement, unsharp masking. We examine and discuss the best method for enhancing chromosome image. Therefore, to evaluate the methods, the original image was manipulated by the addition of some noise and blur. Peak Signal-to-noise Ratio (PSNR and Structural Similarity Index (SSIM are used to examine method performance. The output of enhancement method will be compared with result of Professional software for karyotyping analysis named Ikaros MetasystemT M . Based on experimental results, HE+BM3D method gets a stable result on both scenario noised and blur image.

  6. Double-compression method for biomedical images (United States)

    Antonenko, Yevhenii A.; Mustetsov, Timofey N.; Hamdi, Rami R.; Małecka-Massalska, Teresa; Orshubekov, Nurbek; DzierŻak, RóŻa; Uvaysova, Svetlana


    This paper describes a double compression method (DCM) of biomedical images. A comparison of image compression factors in size JPEG, PNG and developed DCM was carried out. The main purpose of the DCM - compression of medical images while maintaining the key points that carry diagnostic information. To estimate the minimum compression factor an analysis of the coding of random noise image is presented.

  7. Optoelectronic imaging of speckle using image processing method (United States)

    Wang, Jinjiang; Wang, Pengfei


    A detailed image processing of laser speckle interferometry is proposed as an example for the course of postgraduate student. Several image processing methods were used together for dealing with optoelectronic imaging system, such as the partial differential equations (PDEs) are used to reduce the effect of noise, the thresholding segmentation also based on heat equation with PDEs, the central line is extracted based on image skeleton, and the branch is removed automatically, the phase level is calculated by spline interpolation method, and the fringe phase can be unwrapped. Finally, the imaging processing method was used to automatically measure the bubble in rubber with negative pressure which could be used in the tire detection.

  8. Color image definition evaluation method based on deep learning method (United States)

    Liu, Di; Li, YingChun


    In order to evaluate different blurring levels of color image and improve the method of image definition evaluation, this paper proposed a method based on the depth learning framework and BP neural network classification model, and presents a non-reference color image clarity evaluation method. Firstly, using VGG16 net as the feature extractor to extract 4,096 dimensions features of the images, then the extracted features and labeled images are employed in BP neural network to train. And finally achieve the color image definition evaluation. The method in this paper are experimented by using images from the CSIQ database. The images are blurred at different levels. There are 4,000 images after the processing. Dividing the 4,000 images into three categories, each category represents a blur level. 300 out of 400 high-dimensional features are trained in VGG16 net and BP neural network, and the rest of 100 samples are tested. The experimental results show that the method can take full advantage of the learning and characterization capability of deep learning. Referring to the current shortcomings of the major existing image clarity evaluation methods, which manually design and extract features. The method in this paper can extract the images features automatically, and has got excellent image quality classification accuracy for the test data set. The accuracy rate is 96%. Moreover, the predicted quality levels of original color images are similar to the perception of the human visual system.

  9. Ultrasensitive leak detection

    International Nuclear Information System (INIS)

    Winkelman, C.R.; Davidson, H.G.


    The objective of this investigation was to develop a method of detecting leaks to a sensitivity of 1.0 x 10 -13 std/cm 3 /s in vacuum devices and to develop a qualifiable standard leak to provide system calibration at this leak rate. The development work demonstrated that minimum detectable leak rates of 6.5 x 10 -14 std/cm 3 /s and 5.5 x 10 -15 std/cm 3 /s are possible for respective analog and digital measurement modes

  10. Nuclear magnetic resonance imaging method

    International Nuclear Information System (INIS)

    Johnson, G.; MacDonald, J.; Hutchison, S.; Eastwood, L.M.; Redpath, T.W.T.; Mallard, J.R.


    A method of deriving three dimensional image information from an object using nuclear magnetic resonance signals comprises subjecting the object to a continuous, static magnetic field and carrying out the following set of sequential steps: 1) exciting nuclear spins in a selected volume (90deg pulse); 2) applying non-aligned first, second and third gradients of the magnetic field; 3) causing the spins to rephase periodically by reversal of the first gradient to produce spin echoes, and applying pulses of the second gradient prior to every read-out of an echo signal from the object, to differently encode the spin in the second gradient direction for each read-out signal. The above steps 1-3 are then successively repeated with different values of gradient of the third gradient, there being a recovery interval between the repetition of successive sets of steps. Alternate echoes only are read out, the other echoes being time-reversed and ignored for convenience. The resulting signals are appropriately sampled, set out in an array and subjected to three dimensional Fourier transformation. (author)

  11. Computational methods in molecular imaging technologies

    CERN Document Server

    Gunjan, Vinit Kumar; Venkatesh, C; Amarnath, M


    This book highlights the experimental investigations that have been carried out on magnetic resonance imaging and computed tomography (MRI & CT) images using state-of-the-art Computational Image processing techniques, and tabulates the statistical values wherever necessary. In a very simple and straightforward way, it explains how image processing methods are used to improve the quality of medical images and facilitate analysis. It offers a valuable resource for researchers, engineers, medical doctors and bioinformatics experts alike.

  12. Review methods for image segmentation from computed tomography images

    International Nuclear Information System (INIS)

    Mamat, Nurwahidah; Rahman, Wan Eny Zarina Wan Abdul; Soh, Shaharuddin Cik; Mahmud, Rozi


    Image segmentation is a challenging process in order to get the accuracy of segmentation, automation and robustness especially in medical images. There exist many segmentation methods that can be implemented to medical images but not all methods are suitable. For the medical purposes, the aims of image segmentation are to study the anatomical structure, identify the region of interest, measure tissue volume to measure growth of tumor and help in treatment planning prior to radiation therapy. In this paper, we present a review method for segmentation purposes using Computed Tomography (CT) images. CT images has their own characteristics that affect the ability to visualize anatomic structures and pathologic features such as blurring of the image and visual noise. The details about the methods, the goodness and the problem incurred in the methods will be defined and explained. It is necessary to know the suitable segmentation method in order to get accurate segmentation. This paper can be a guide to researcher to choose the suitable segmentation method especially in segmenting the images from CT scan

  13. Gamma-ray Imaging Methods

    Energy Technology Data Exchange (ETDEWEB)

    Vetter, K; Mihailescu, L; Nelson, K; Valentine, J; Wright, D


    In this document we discuss specific implementations for gamma-ray imaging instruments including the principle of operation and describe systems which have been built and demonstrated as well as systems currently under development. There are several fundamentally different technologies each with specific operational requirements and performance trade offs. We provide an overview of the different gamma-ray imaging techniques and briefly discuss challenges and limitations associated with each modality (in the appendix we give detailed descriptions of specific implementations for many of these technologies). In Section 3 we summarize the performance and operational aspects in tabular form as an aid for comparing technologies and mapping technologies to potential applications.

  14. A new ultrasensitive scanning calorimeter. (United States)

    Plotnikov, V V; Brandts, J M; Lin, L N; Brandts, J F


    A new ultrasensitive differential scanning calorimeter is described, having a number of novel features arising from integration between hardware and software. It is capable of high performance in either a scanning or isothermal mode of operation. Upscanning is carried out adiabatically while downscanning is nonadiabatic. By using software-controlled signals sent continuously to appropriate hardware devices, it is possible to improve adiabaticity and constancy of scan rate through use of empirical prerun information stored in memory rather than by using feedback systems which respond in real time and generate thermal noise. Also, instrument response time is software-selectable, maximizing performance for both slow- and fast-transient systems. While these and other sophisticated functionalities have been introduced into the instrument to improve performance and data analysis, they are virtually invisible and add no additional complexities into operation of the instrument. Noise and baseline repeatability are an order of magnitude better than published raw data from other instruments so that high-quality results can be obtained on protein solutions, for example, using as little as 50 microg of protein in the sample cell.

  15. Magnetic resonance spectroscopy as an imaging method

    International Nuclear Information System (INIS)

    Bomsdorf, H.; Imme, M.; Jensen, D.; Kunz, D.; Menhardt, W.; Ottenberg, K.; Roeschmann, P.; Schmidt, K.H.; Tschendel, O.; Wieland, J.


    An experimental Magnetic Resonance (MR) system with 4 tesla flux density was set up. For that purpose a data acquisition system and RF coils for resonance frequencies up to 170 MHz were developed. Methods for image guided spectroscopy as well as spectroscopic imaging focussing on the nuclei 1 H and 13 C were developed and tested on volunteers and selected patients. The advantages of the high field strength with respect to spectroscopic studies were demonstrated. Developments of a new fast imaging technique for the acquisition of scout images as well as a method for mapping and displaying the magnetic field inhomogeneity in-vivo represent contributions to the optimisation of the experimental procedure in spectroscopic studies. Investigations on the interaction of RF radiation with the exposed tissue allowed conclusions regarding the applicability of MR methods at high field strengths. Methods for display and processing of multi-dimensional spectroscopic imaging data sets were developed and existing methods for real-time image synthesis were extended. Results achieved in the field of computer aided analysis of MR images comprised new techniques for image background detection, contour detection and automatic image interpretation as well as knowledge bases for textural representation of medical knowledge for diagnosis. (orig.) With 82 refs., 3 tabs., 75 figs [de

  16. Image splitting and remapping method for radiological image compression (United States)

    Lo, Shih-Chung B.; Shen, Ellen L.; Mun, Seong K.


    A new decomposition method using image splitting and gray-level remapping has been proposed for image compression, particularly for images with high contrast resolution. The effects of this method are especially evident in our radiological image compression study. In our experiments, we tested the impact of this decomposition method on image compression by employing it with two coding techniques on a set of clinically used CT images and several laser film digitized chest radiographs. One of the compression techniques used was full-frame bit-allocation in the discrete cosine transform domain, which has been proven to be an effective technique for radiological image compression. The other compression technique used was vector quantization with pruned tree-structured encoding, which through recent research has also been found to produce a low mean-square-error and a high compression ratio. The parameters we used in this study were mean-square-error and the bit rate required for the compressed file. In addition to these parameters, the difference between the original and reconstructed images will be presented so that the specific artifacts generated by both techniques can be discerned by visual perception.

  17. Method for position emission mammography image reconstruction (United States)

    Smith, Mark Frederick


    An image reconstruction method comprising accepting coincidence datat from either a data file or in real time from a pair of detector heads, culling event data that is outside a desired energy range, optionally saving the desired data for each detector position or for each pair of detector pixels on the two detector heads, and then reconstructing the image either by backprojection image reconstruction or by iterative image reconstruction. In the backprojection image reconstruction mode, rays are traced between centers of lines of response (LOR's), counts are then either allocated by nearest pixel interpolation or allocated by an overlap method and then corrected for geometric effects and attenuation and the data file updated. If the iterative image reconstruction option is selected, one implementation is to compute a grid Siddon retracing, and to perform maximum likelihood expectation maiximization (MLEM) computed by either: a) tracing parallel rays between subpixels on opposite detector heads; or b) tracing rays between randomized endpoint locations on opposite detector heads.

  18. Linear Methods for Image Interpolation


    Pascal Getreuer


    We discuss linear methods for interpolation, including nearest neighbor, bilinear, bicubic, splines, and sinc interpolation. We focus on separable interpolation, so most of what is said applies to one-dimensional interpolation as well as N-dimensional separable interpolation.

  19. Digital image envelope: method and evaluation (United States)

    Huang, H. K.; Cao, Fei; Zhou, Michael Z.; Mogel, Greg T.; Liu, Brent J.; Zhou, Xiaoqiang


    Health data security, characterized in terms of data privacy, authenticity, and integrity, is a vital issue when digital images and other patient information are transmitted through public networks in telehealth applications such as teleradiology. Mandates for ensuring health data security have been extensively discussed (for example The Health Insurance Portability and Accountability Act, HIPAA) and health informatics guidelines (such as the DICOM standard) are beginning to focus on issues of data continue to be published by organizing bodies in healthcare; however, there has not been a systematic method developed to ensure data security in medical imaging Because data privacy and authenticity are often managed primarily with firewall and password protection, we have focused our research and development on data integrity. We have developed a systematic method of ensuring medical image data integrity across public networks using the concept of the digital envelope. When a medical image is generated regardless of the modality, three processes are performed: the image signature is obtained, the DICOM image header is encrypted, and a digital envelope is formed by combining the signature and the encrypted header. The envelope is encrypted and embedded in the original image. This assures the security of both the image and the patient ID. The embedded image is encrypted again and transmitted across the network. The reverse process is performed at the receiving site. The result is two digital signatures, one from the original image before transmission, and second from the image after transmission. If the signatures are identical, there has been no alteration of the image. This paper concentrates in the method and evaluation of the digital image envelope.

  20. On an image reconstruction method for ECT (United States)

    Sasamoto, Akira; Suzuki, Takayuki; Nishimura, Yoshihiro


    An image by Eddy Current Testing(ECT) is a blurred image to original flaw shape. In order to reconstruct fine flaw image, a new image reconstruction method has been proposed. This method is based on an assumption that a very simple relationship between measured data and source were described by a convolution of response function and flaw shape. This assumption leads to a simple inverse analysis method with deconvolution.In this method, Point Spread Function (PSF) and Line Spread Function(LSF) play a key role in deconvolution processing. This study proposes a simple data processing to determine PSF and LSF from ECT data of machined hole and line flaw. In order to verify its validity, ECT data for SUS316 plate(200x200x10mm) with artificial machined hole and notch flaw had been acquired by differential coil type sensors(produced by ZETEC Inc). Those data were analyzed by the proposed method. The proposed method restored sharp discrete multiple hole image from interfered data by multiple holes. Also the estimated width of line flaw has been much improved compared with original experimental data. Although proposed inverse analysis strategy is simple and easy to implement, its validity to holes and line flaw have been shown by many results that much finer image than original image have been reconstructed.

  1. Method for nuclear magnetic resonance imaging (United States)

    Kehayias, J.J.; Joel, D.D.; Adams, W.H.; Stein, H.L.


    A method for in vivo NMR imaging of the blood vessels and organs of a patient characterized by using a dark dye-like imaging substance consisting essentially of a stable, high-purity concentration of D/sub 2/O in a solution with water.

  2. Linear Methods for Image Interpolation

    Directory of Open Access Journals (Sweden)

    Pascal Getreuer


    Full Text Available We discuss linear methods for interpolation, including nearest neighbor, bilinear, bicubic, splines, and sinc interpolation. We focus on separable interpolation, so most of what is said applies to one-dimensional interpolation as well as N-dimensional separable interpolation.

  3. Imaging methods in medical diagnosis

    International Nuclear Information System (INIS)

    Krestel, E.


    Pictures of parts of the human body or of the human body (views, superposition pictures, pictures of body layers, or photographs) are considerable helps for the medical diagnostics. Physics, electrotechnique, and machine construction make the picture production possible. Modern electronics and optics offer facilities of picture processing which influences the picture quality. Picture interpretation is the the physican's task. The picture-delivering methods applied in medicine include the conventional X-ray diagnostics, X-ray computer tomography, nuclear diagnostics, sonography with ultas sound, and endoscopy. Their rapid development and immprovement was caused by the development of electronics during the past 20 years. A method presently in discussion and development is the Kernspin-tomography. (orig./MG) [de


    Institute of Scientific and Technical Information of China (English)

    Nie Xuan; Zhao Rongchun; Jiang Zetao


    The proposed algorithm relies on a group of new formulas for calculating tangent slope so as to address angle feature of edge curves of image. It can utilize tangent angle features to estimate automatically and fully the rotation parameters of geometric transform and enable rough matching of images with huge rotation difference. After angle compensation, it can search for matching point sets by correlation criterion, then calculate parameters of affine transform, enable higher-precision emendation of rotation and transferring. Finally, it fulfills precise matching for images with relax-tense iteration method. Compared with the registration approach based on wavelet direction-angle features, the matching algorithm with tangent feature of image edge is more robust and realizes precise registration of various images. Furthermore, it is also helpful in graphics matching.

  5. Historic Methods for Capturing Magnetic Field Images (United States)

    Kwan, Alistair


    I investigated two late 19th-century methods for capturing magnetic field images from iron filings for historical insight into the pedagogy of hands-on physics education methods, and to flesh out teaching and learning practicalities tacit in the historical record. Both methods offer opportunities for close sensory engagement in data-collection…

  6. Methods for evaluating imaging methods of limited reproducibility

    International Nuclear Information System (INIS)

    Krummenauer, F.


    Just like new drugs, new or modified imaging methods must be subjected to objective clinical tests, including tests on humans. In this, it must be ensured that the principle of Good Clinical Practice (GCP) are followed with regard to medical, administrative and methodical quality. Innovative methods fo clinical epidemiology and medical biometry should be applied from the planning stage to the final statistical evaluation. The author presents established and new methods for planning, evaluation and reporting of clinical tests of diagnostic methods, and especially imaging methods, in clinical medicine and illustrates these by means of current research projects in the various medical disciplines. The strategies presented are summarized in a recommendation based on the concept of phases I - IV of clinical drug testing in order to enable standardisation of the clinical evaluation of imaging methods. (orig.)

  7. Comparison of two ultra-sensitive methods for the determination of 232Th by recovery corrected pre-concentration radiochemical neutron activation analysis

    International Nuclear Information System (INIS)

    Glover, S.E.; Qu, H.; LaMont, S.P.; Grimm, C.A.; Filby, R.H.


    The determination of isotopic thorium by alpha spectrometric methods is a routine practice for bioassay and environmental measurement programs. Alpha-spectrometry has excellent detection limits (by mass) for all isotopes of thorium except 232 Th due to its extremely long half-life. Improvements in the detection limit an sensitivity over previously reported methods of pre-concentration neutron activation analysis (PCNAA) for the recovery corrected, isotopic determination of thorium in various matrices is discussed. Following irradiation, the samples were dissolved, 231 Pa added as a tracer, and Pa isolated by two different methods and compared (extraction chromatography and anion exchange chromatography) followed by alpha spectrometry for recovery correction. Ion exchange chromatography was found to be superior for this application at this time, principally for reliability. The detection limit for 232 Th of 3.5 x 10 -7 Bq is almost three orders of magnitude lower than for alpha spectrometry using the PCRNAA method and one order of magnitude below previously reported PCNAA methods. (author)

  8. Systematic approach to optimize a pretreatment method for ultrasensitive liquid chromatography with tandem mass spectrometry analysis of multiple target compounds in biological samples. (United States)

    Togashi, Kazutaka; Mutaguchi, Kuninori; Komuro, Setsuko; Kataoka, Makoto; Yamazaki, Hiroshi; Yamashita, Shinji


    In current approaches for new drug development, highly sensitive and robust analytical methods for the determination of test compounds in biological samples are essential. These analytical methods should be optimized for every target compound. However, for biological samples that contain multiple compounds as new drug candidates obtained by cassette dosing tests, it would be preferable to develop a single method that allows the determination of all compounds at once. This study aims to establish a systematic approach that enables a selection of the most appropriate pretreatment method for multiple target compounds without the use of their chemical information. We investigated the retention times of 27 known compounds under different mobile phase conditions and determined the required pretreatment of human plasma samples using several solid-phase and liquid-liquid extractions. From the relationship between retention time and recovery in a principal component analysis, appropriate pretreatments were categorized into several types. Based on the category, we have optimized a pretreatment method for the identification of three calcium channel blockers in human plasma. Plasma concentrations of these drugs in a cassette-dose clinical study at microdose level were successfully determined with a lower limit of quantitation of 0.2 pg/mL for diltiazem, 1 pg/mL for nicardipine, and 2 pg/mL for nifedipine. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  9. Prognostic aspects of imaging method development

    International Nuclear Information System (INIS)

    Steinhart, L.


    A survey is presented of X-ray diagnostic methods and techniques and possibilities of their further development. Promising methods include direct imaging using digital radiography. In connection with computer technology these methods achieve higher resolution. The storage of obtained images in the computer memory will allow automated processing and evaluation and the use of expert systems. Development is expected to take place especially in computerized tomography using magnetic resonance, and positron computed tomography and other non-radioactive diagnostic methods. (J.B.). 5 figs., 1 tab., 1 ref

  10. Matrix Krylov subspace methods for image restoration

    Directory of Open Access Journals (Sweden)

    khalide jbilou


    Full Text Available In the present paper, we consider some matrix Krylov subspace methods for solving ill-posed linear matrix equations and in those problems coming from the restoration of blurred and noisy images. Applying the well known Tikhonov regularization procedure leads to a Sylvester matrix equation depending the Tikhonov regularized parameter. We apply the matrix versions of the well known Krylov subspace methods, namely the Least Squared (LSQR and the conjugate gradient (CG methods to get approximate solutions representing the restored images. Some numerical tests are presented to show the effectiveness of the proposed methods.

  11. Nanobody medicated immunoassay for ultrasensitive detection of cancer biomarker alpha-fetoprotein. (United States)

    Chen, Jing; He, Qing-hua; Xu, Yang; Fu, Jin-heng; Li, Yan-ping; Tu, Zhui; Wang, Dan; Shu, Mei; Qiu, Yu-lou; Yang, Hong-wei; Liu, Yuan-yuan


    Immunoassay for cancer biomarkers plays an important role in cancer prevention and early diagnosis. To the development of immunoassay, the quality and stability of applied antibody is one of the key points to obtain reliability and high sensitivity for immunoassay. The main purpose of this study was to develop a novel immunoassay for ultrasensitive detection of cancer biomarker alpha-fetoprotein (AFP) based on nanobody against AFP. Two nanobodies which bind to AFP were selected from a phage display nanobody library by biopanning strategy. The prepared nanobodies are clonable, thermally stable and applied in both sandwich enzyme linked immunoassay (ELISA) and immuno-PCR assay for ultrasensitive detection of AFP. The limit detection of sandwich ELISA setup with optimized nanobodies was 0.48ng mL(-1), and the half of saturation concentration (SC50) value was 6.68±0.56ng mL(-1). These nanobodies were also used to develop an immuno-PCR assay for ultrasensitive detection of AFP, its limit detection values was 0.005ng mL(-1), and the linear range was 0.01-10,000ng mL(-1). These established immunoassays based on nanobodies were highly specific to AFP and with negligible cross reactivity with other tested caner biomarkers. Furthermore, this novel concept of nanobodies mediated immunoassay may provide potential applications in a general method for the ultrasensitive detection of various cancer biomarkers. Copyright © 2015 Elsevier B.V. All rights reserved.

  12. Ultra-sensitive LC-MS/MS method for the quantification of gemcitabine and its metabolite 2',2'-difluorodeoxyuridine in human plasma for a microdose clinical trial. (United States)

    van Nuland, M; Hillebrand, M J X; Rosing, H; Burgers, J A; Schellens, J H M; Beijnen, J H


    In microdose clinical trials a maximum of 100 μg of drug substance is administered to participants, in order to determine the pharmacokinetic properties of the agents. Measuring low plasma concentrations after administration of a microdose is challenging and requires the use of ulta-sensitive equipment. Novel liquid chromatography-mass spectrometry (LC-MS/MS) platforms can be used for quantification of low drug plasma levels. Here we describe the development and validation of an LC-MS/MS method for quantification of gemcitabine and its metabolite 2',2'-difluorodeoxyuridine (dFdU) in the low picogram per milliliter range to support a microdose trial. The validated assay ranges from 2.5-500 pg/mL for gemcitabine and 250-50,000 pg/mL for dFdU were linear, with a correlation coefficient (r 2 ) of 0.996 or better. Sample preparation with solid phase extraction provided a good and reproducible recovery. All results were within the acceptance criteria of the latest US FDA guidance and EMA guidelines. In addition, the method was successfully applied to measure plasma concentrations of gemcitabine in a patient after administration of a microdose of gemcitabine. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Analysis of Non Local Image Denoising Methods (United States)

    Pardo, Álvaro

    Image denoising is probably one of the most studied problems in the image processing community. Recently a new paradigm on non local denoising was introduced. The Non Local Means method proposed by Buades, Morel and Coll attracted the attention of other researches who proposed improvements and modifications to their proposal. In this work we analyze those methods trying to understand their properties while connecting them to segmentation based on spectral graph properties. We also propose some improvements to automatically estimate the parameters used on these methods.

  14. Handbook of mathematical methods in imaging

    CERN Document Server


    The Handbook of Mathematical Methods in Imaging provides a comprehensive treatment of the mathematical techniques used in imaging science. The material is grouped into two central themes, namely, Inverse Problems (Algorithmic Reconstruction) and Signal and Image Processing. Each section within the themes covers applications (modeling), mathematics, numerical methods (using a case example) and open questions. Written by experts in the area, the presentation is mathematically rigorous. This expanded and revised second edition contains updates to existing chapters and 16 additional entries on important mathematical methods such as graph cuts, morphology, discrete geometry, PDEs, conformal methods, to name a few. The entries are cross-referenced for easy navigation through connected topics. Available in both print and electronic forms, the handbook is enhanced by more than 200 illustrations and an extended bibliography. It will benefit students, scientists and researchers in applied mathematics. Engineers and com...

  15. Accelerated gradient methods for constrained image deblurring

    International Nuclear Information System (INIS)

    Bonettini, S; Zanella, R; Zanni, L; Bertero, M


    In this paper we propose a special gradient projection method for the image deblurring problem, in the framework of the maximum likelihood approach. We present the method in a very general form and we give convergence results under standard assumptions. Then we consider the deblurring problem and the generality of the proposed algorithm allows us to add a energy conservation constraint to the maximum likelihood problem. In order to improve the convergence rate, we devise appropriate scaling strategies and steplength updating rules, especially designed for this application. The effectiveness of the method is evaluated by means of a computational study on astronomical images corrupted by Poisson noise. Comparisons with standard methods for image restoration, such as the expectation maximization algorithm, are also reported.

  16. Image change detection systems, methods, and articles of manufacture (United States)

    Jones, James L.; Lassahn, Gordon D.; Lancaster, Gregory D.


    Aspects of the invention relate to image change detection systems, methods, and articles of manufacture. According to one aspect, a method of identifying differences between a plurality of images is described. The method includes loading a source image and a target image into memory of a computer, constructing source and target edge images from the source and target images to enable processing of multiband images, displaying the source and target images on a display device of the computer, aligning the source and target edge images, switching displaying of the source image and the target image on the display device, to enable identification of differences between the source image and the target image.

  17. Quality assessment in radiological imaging methods

    International Nuclear Information System (INIS)

    Herstel, W.


    The equipment used in diagnostic radiology is becoming more and more complicated. In the imaging process four components are distinguished, each of which can introduce loss in essential information: the X-ray source, the human body, the imaging system and the observer. In nearly all imaging methods the X-ray quantum fluctuations are a limitation to observation. But there are also technical factors. As an illustration it is shown how in a television scanning process the resolution is restricted by the system parameters. A short review is given of test devices and the results are given of an image comparison based on regular bar patterns. Although this method has the disadvantage of measuring mainly the limiting resolution, the results of the test correlate reasonably well with the subjective appreciations of radiographs of bony structures made by a group of trained radiologists. Fluoroscopic systems should preferably be tested using moving structures under dynamic conditions. (author)

  18. Circular SAR Optimization Imaging Method of Buildings

    Directory of Open Access Journals (Sweden)

    Wang Jian-feng


    Full Text Available The Circular Synthetic Aperture Radar (CSAR can obtain the entire scattering properties of targets because of its great ability of 360° observation. In this study, an optimal orientation of the CSAR imaging algorithm of buildings is proposed by applying a combination of coherent and incoherent processing techniques. FEKO software is used to construct the electromagnetic scattering modes and simulate the radar echo. The FEKO imaging results are compared with the isotropic scattering results. On comparison, the optimal azimuth coherent accumulation angle of CSAR imaging of buildings is obtained. Practically, the scattering directions of buildings are unknown; therefore, we divide the 360° echo of CSAR into many overlapped and few angle echoes corresponding to the sub-aperture and then perform an imaging procedure on each sub-aperture. Sub-aperture imaging results are applied to obtain the all-around image using incoherent fusion techniques. The polarimetry decomposition method is used to decompose the all-around image and further retrieve the edge information of buildings successfully. The proposed method is validated with P-band airborne CSAR data from Sichuan, China.

  19. Method of orthogonally splitting imaging pose measurement (United States)

    Zhao, Na; Sun, Changku; Wang, Peng; Yang, Qian; Liu, Xintong


    In order to meet the aviation's and machinery manufacturing's pose measurement need of high precision, fast speed and wide measurement range, and to resolve the contradiction between measurement range and resolution of vision sensor, this paper proposes an orthogonally splitting imaging pose measurement method. This paper designs and realizes an orthogonally splitting imaging vision sensor and establishes a pose measurement system. The vision sensor consists of one imaging lens, a beam splitter prism, cylindrical lenses and dual linear CCD. Dual linear CCD respectively acquire one dimensional image coordinate data of the target point, and two data can restore the two dimensional image coordinates of the target point. According to the characteristics of imaging system, this paper establishes the nonlinear distortion model to correct distortion. Based on cross ratio invariability, polynomial equation is established and solved by the least square fitting method. After completing distortion correction, this paper establishes the measurement mathematical model of vision sensor, and determines intrinsic parameters to calibrate. An array of feature points for calibration is built by placing a planar target in any different positions for a few times. An terative optimization method is presented to solve the parameters of model. The experimental results show that the field angle is 52 °, the focus distance is 27.40 mm, image resolution is 5185×5117 pixels, displacement measurement error is less than 0.1mm, and rotation angle measurement error is less than 0.15°. The method of orthogonally splitting imaging pose measurement can satisfy the pose measurement requirement of high precision, fast speed and wide measurement range.


    Directory of Open Access Journals (Sweden)

    S. A. Seyyedi


    Full Text Available Steganography is a method of hiding information in other information of different format (container. There are many steganography techniques with various types of container. In the Internet, digital images are the most popular and frequently used containers. We consider main image steganography techniques and their advantages and disadvantages. We also identify the requirements of a good steganography algorithm and compare various such algorithms.

  1. Study on Processing Method of Image Shadow

    Directory of Open Access Journals (Sweden)

    Wang Bo


    Full Text Available In order to effectively remove disturbance of shadow and enhance robustness of information processing of computer visual image, this paper makes study on inspection and removal of image shadow. It makes study the continual removal algorithm of shadow based on integration, the illumination surface and texture, it respectively introduces their work principles and realization method, it can effectively carrying processing for shadow by test.

  2. Coherent diffractive imaging methods for semiconductor manufacturing (United States)

    Helfenstein, Patrick; Mochi, Iacopo; Rajeev, Rajendran; Fernandez, Sara; Ekinci, Yasin


    The paradigm shift of the semiconductor industry moving from deep ultraviolet to extreme ultraviolet lithography (EUVL) brought about new challenges in the fabrication of illumination and projection optics, which constitute one of the core sources of cost of ownership for many of the metrology tools needed in the lithography process. For this reason, lensless imaging techniques based on coherent diffractive imaging started to raise interest in the EUVL community. This paper presents an overview of currently on-going research endeavors that use a number of methods based on lensless imaging with coherent light.

  3. Green method for ultrasensitive determination of Hg in natural waters by electrothermal-atomic absorption spectrometry following sono-induced cold vapor generation and 'in-atomizer trapping'

    International Nuclear Information System (INIS)

    Gil, Sandra; Lavilla, Isela; Bendicho, Carlos


    Sono-induced cold vapor generation (SI-CVG) has been used for the first time in combination with a graphite furnace atomizer for determination of Hg in natural waters by electrothermal-atomic absorption spectrometry after in situ trapping onto a noble metal-pretreated platform (Pd, Pt or Rh) inserted into a graphite tube. The system allows 'in-atomizer trapping' of Hg without the use of conventional reduction reactions based on sodium borohydride or tin chloride in acid medium for cold vapor generation. The sono-induced reaction is accomplished by applying ultrasound irradiation to the sample solution containing Hg(II) in the presence of an organic compound such as formic acid. As this organic acid is partly degraded upon ultrasound irradiation to yield CO, CO 2 , H 2 and H 2 O, the amount of lab wastes is minimized and a green methodology is achieved. For this purpose, experimental variables influencing the generation/trapping process are fully investigated. The limit of detection for a 10 min trapping time and 10 mL sample volume was 0.03 μg L -1 (Integrated absorbance) and the repeatability expressed as relative standard deviation was about 3%. Carbonates and chlorides at 100 mg L -1 level caused a signal depression by 20-30%. The enhanced trapping efficiency observed with the sono-induced cold vapor generation as compared with 'in-atomizer trapping' methods employing chemical vapor generation is discussed. A reaction pathway for SI-CVG is proposed on the basis of the current knowledge for synthesis of noble metal nanoparticles by ultrasound

  4. Improved image alignment method in application to X-ray images and biological images. (United States)

    Wang, Ching-Wei; Chen, Hsiang-Chou


    Alignment of medical images is a vital component of a large number of applications throughout the clinical track of events; not only within clinical diagnostic settings, but prominently so in the area of planning, consummation and evaluation of surgical and radiotherapeutical procedures. However, image registration of medical images is challenging because of variations on data appearance, imaging artifacts and complex data deformation problems. Hence, the aim of this study is to develop a robust image alignment method for medical images. An improved image registration method is proposed, and the method is evaluated with two types of medical data, including biological microscopic tissue images and dental X-ray images and compared with five state-of-the-art image registration techniques. The experimental results show that the presented method consistently performs well on both types of medical images, achieving 88.44 and 88.93% averaged registration accuracies for biological tissue images and X-ray images, respectively, and outperforms the benchmark methods. Based on the Tukey's honestly significant difference test and Fisher's least square difference test tests, the presented method performs significantly better than all existing methods (P ≤ 0.001) for tissue image alignment, and for the X-ray image registration, the proposed method performs significantly better than the two benchmark b-spline approaches (P < 0.001). The software implementation of the presented method and the data used in this study are made publicly available for scientific communities to use (∼cweiwang/ImprovedImageRegistration/).

  5. An Improved Image Contrast Assessment Method

    Directory of Open Access Journals (Sweden)

    Yuanyuan Fan


    Full Text Available Contrast is an important factor affecting the image quality. In order to overcome the problems of local band-limited contrast, a novel image contrast assessment method based on the property of HVS is proposed. Firstly, the image by low-pass filter is performed fast wavelet decomposition. Secondly, all levels of band-pass filtered image and its corresponding low-pass filtered image are obtained by processing wavelet coefficients. Thirdly, local band-limited contrast is calculated, and the local band-limited contrast entropy is calculated according to the definition of entropy, Finally, the contrast entropy of image is obtained by averaging the local band-limited contrast entropy weighed using CSF coefficient. The experiment results show that the best contrast image can be accurately identified in the sequence images obtained by adjusting the exposure time and stretching gray respectively, the assessment results accord with human visual characteristics and make up the lack of local band-limited contrast.

  6. NMR blood vessel imaging method and apparatus

    International Nuclear Information System (INIS)

    Riederer, S.J.


    A high speed method of forming computed images of blood vessels based on measurements of characteristics of a body is described comprising the steps of: subjecting a predetermined body area containing blood vessels of interest to, successively, applications of a short repetition time (TR) NMR pulse sequence during the period of high blood velocity and then to corresponding applications during the period of low blood velocity for successive heart beat cycles; weighting the collected imaging data from each application of the NMR pulse sequence according to whether the data was acquired during the period of high blood velocity or a period of low blood velocity of the corresponding heart beat cycle; accumulating weighted imaging data from a plurality of NMR pulse sequences corresponding to high blood velocity periods and from a plurality of NMR pulse sequences corresponding to low blood velocity periods; subtracting the weighted imaging data corresponding to each specific phase encoding acquired during the high blood velocity periods from the weighted imaging data for the same phase encoding corresponding to low blood velocity periods in order to compute blood vessel imaging data; and forming an image of the blood vessels of interest from the blood vessel imaging data

  7. Medical Imaging Image Quality Assessment with Monte Carlo Methods

    International Nuclear Information System (INIS)

    Michail, C M; Fountos, G P; Kalyvas, N I; Valais, I G; Kandarakis, I S; Karpetas, G E; Martini, Niki; Koukou, Vaia


    The aim of the present study was to assess image quality of PET scanners through a thin layer chromatography (TLC) plane source. The source was simulated using a previously validated Monte Carlo model. The model was developed by using the GATE MC package and reconstructed images obtained with the STIR software for tomographic image reconstruction, with cluster computing. The PET scanner simulated in this study was the GE DiscoveryST. A plane source consisted of a TLC plate, was simulated by a layer of silica gel on aluminum (Al) foil substrates, immersed in 18F-FDG bath solution (1MBq). Image quality was assessed in terms of the Modulation Transfer Function (MTF). MTF curves were estimated from transverse reconstructed images of the plane source. Images were reconstructed by the maximum likelihood estimation (MLE)-OSMAPOSL algorithm. OSMAPOSL reconstruction was assessed by using various subsets (3 to 21) and iterations (1 to 20), as well as by using various beta (hyper) parameter values. MTF values were found to increase up to the 12th iteration whereas remain almost constant thereafter. MTF improves by using lower beta values. The simulated PET evaluation method based on the TLC plane source can be also useful in research for the further development of PET and SPECT scanners though GATE simulations. (paper)

  8. Development of Quantification Method for Bioluminescence Imaging

    International Nuclear Information System (INIS)

    Kim, Hyeon Sik; Min, Jung Joon; Lee, Byeong Il; Choi, Eun Seo; Tak, Yoon O; Choi, Heung Kook; Lee, Ju Young


    Optical molecular luminescence imaging is widely used for detection and imaging of bio-photons emitted by luminescent luciferase activation. The measured photons in this method provide the degree of molecular alteration or cell numbers with the advantage of high signal-to-noise ratio. To extract useful information from the measured results, the analysis based on a proper quantification method is necessary. In this research, we propose a quantification method presenting linear response of measured light signal to measurement time. We detected the luminescence signal by using lab-made optical imaging equipment of animal light imaging system (ALIS) and different two kinds of light sources. One is three bacterial light-emitting sources containing different number of bacteria. The other is three different non-bacterial light sources emitting very weak light. By using the concept of the candela and the flux, we could derive simplified linear quantification formula. After experimentally measuring light intensity, the data was processed with the proposed quantification function. We could obtain linear response of photon counts to measurement time by applying the pre-determined quantification function. The ratio of the re-calculated photon counts and measurement time present a constant value although different light source was applied. The quantification function for linear response could be applicable to the standard quantification process. The proposed method could be used for the exact quantitative analysis in various light imaging equipment with presenting linear response behavior of constant light emitting sources to measurement time

  9. Blind image deconvolution methods and convergence

    CERN Document Server

    Chaudhuri, Subhasis; Rameshan, Renu


    Blind deconvolution is a classical image processing problem which has been investigated by a large number of researchers over the last four decades. The purpose of this monograph is not to propose yet another method for blind image restoration. Rather the basic issue of deconvolvability has been explored from a theoretical view point. Some authors claim very good results while quite a few claim that blind restoration does not work. The authors clearly detail when such methods are expected to work and when they will not. In order to avoid the assumptions needed for convergence analysis in the

  10. Ultrasensitivity of the Bacillus subtilis sporulation decision. (United States)

    Narula, Jatin; Devi, Seram N; Fujita, Masaya; Igoshin, Oleg A


    Starving Bacillus subtilis cells execute a gene expression program resulting in the formation of stress-resistant spores. Sporulation master regulator, Spo0A, is activated by a phosphorelay and controls the expression of a multitude of genes, including the forespore-specific sigma factor σ(F) and the mother cell-specific sigma factor σ(E). Identification of the system-level mechanism of the sporulation decision is hindered by a lack of direct control over Spo0A activity. This limitation can be overcome by using a synthetic system in which Spo0A activation is controlled by inducing expression of phosphorelay kinase KinA. This induction results in a switch-like increase in the number of sporulating cells at a threshold of KinA. Using a combination of mathematical modeling and single-cell microscopy, we investigate the origin and physiological significance of this ultrasensitive threshold. The results indicate that the phosphorelay is unable to achieve a sufficiently fast and ultrasensitive response via its positive feedback architecture, suggesting that the sporulation decision is made downstream. In contrast, activation of σ(F) in the forespore and of σ(E) in the mother cell compartments occurs via a cascade of coherent feed-forward loops, and thereby can produce fast and ultrasensitive responses as a result of KinA induction. Unlike σ(F) activation, σ(E) activation in the mother cell compartment only occurs above the KinA threshold, resulting in completion of sporulation. Thus, ultrasensitive σ(E) activation explains the KinA threshold for sporulation induction. We therefore infer that under uncertain conditions, cells initiate sporulation but postpone making the sporulation decision to average stochastic fluctuations and to achieve a robust population response.

  11. Radiographic imaging method by gas ionisation

    International Nuclear Information System (INIS)

    Eickel, R.; Rheude, A.


    The search for a substitute of the silver halide film has been intensified worldwide due to the shortage and price increase of silver metal. Gasionography could be an alternative to the wellknown silver film imaging techniques in roentgenology. Therefore the practical basis of the imaging process and the electrophoretic development was investigated. The technical realisation of this method was demonstrated for two different types of X-ray examen by developing a fully automatic chest changer and a mammography system that can be adapted to commercially available imaging stands. The image quality achieved with these apparatus was evaluated in comparison with conventional film techniques in the laboratory as well as in a clinical trial. (orig.) [de

  12. A multicore based parallel image registration method. (United States)

    Yang, Lin; Gong, Leiguang; Zhang, Hong; Nosher, John L; Foran, David J


    Image registration is a crucial step for many image-assisted clinical applications such as surgery planning and treatment evaluation. In this paper we proposed a landmark based nonlinear image registration algorithm for matching 2D image pairs. The algorithm was shown to be effective and robust under conditions of large deformations. In landmark based registration, the most important step is establishing the correspondence among the selected landmark points. This usually requires an extensive search which is often computationally expensive. We introduced a nonregular data partition algorithm using the K-means clustering algorithm to group the landmarks based on the number of available processing cores. The step optimizes the memory usage and data transfer. We have tested our method using IBM Cell Broadband Engine (Cell/B.E.) platform.

  13. Image-reconstruction methods in positron tomography

    CERN Document Server

    Townsend, David W; CERN. Geneva


    Physics and mathematics for medical imaging In the two decades since the introduction of the X-ray scanner into radiology, medical imaging techniques have become widely established as essential tools in the diagnosis of disease. As a consequence of recent technological and mathematical advances, the non-invasive, three-dimensional imaging of internal organs such as the brain and the heart is now possible, not only for anatomical investigations using X-rays but also for studies which explore the functional status of the body using positron-emitting radioisotopes and nuclear magnetic resonance. Mathematical methods which enable three-dimentional distributions to be reconstructed from projection data acquired by radiation detectors suitably positioned around the patient will be described in detail. The lectures will trace the development of medical imaging from simpleradiographs to the present-day non-invasive measurement of in vivo boichemistry. Powerful techniques to correlate anatomy and function that are cur...

  14. 3D Seismic Imaging using Marchenko Methods (United States)

    Lomas, A.; Curtis, A.


    Marchenko methods are novel, data driven techniques that allow seismic wavefields from sources and receivers on the Earth's surface to be redatumed to construct wavefields with sources in the subsurface - including complex multiply-reflected waves, and without the need for a complex reference model. In turn, this allows subsurface images to be constructed at any such subsurface redatuming points (image or virtual receiver points). Such images are then free of artefacts from multiply-scattered waves that usually contaminate migrated seismic images. Marchenko algorithms require as input the same information as standard migration methods: the full reflection response from sources and receivers at the Earth's surface, and an estimate of the first arriving wave between the chosen image point and the surface. The latter can be calculated using a smooth velocity model estimated using standard methods. The algorithm iteratively calculates a signal that focuses at the image point to create a virtual source at that point, and this can be used to retrieve the signal between the virtual source and the surface. A feature of these methods is that the retrieved signals are naturally decomposed into up- and down-going components. That is, we obtain both the signal that initially propagated upwards from the virtual source and arrived at the surface, separated from the signal that initially propagated downwards. Figure (a) shows a 3D subsurface model with a variable density but a constant velocity (3000m/s). Along the surface of this model (z=0) in both the x and y directions are co-located sources and receivers at 20-meter intervals. The redatumed signal in figure (b) has been calculated using Marchenko methods from a virtual source (1200m, 500m and 400m) to the surface. For comparison the true solution is given in figure (c), and shows a good match when compared to figure (b). While these 2D redatuming and imaging methods are still in their infancy having first been developed in

  15. Radiopharmaceutical chelates and method of external imaging

    International Nuclear Information System (INIS)

    Loberg, M.D.; Callery, P.S.; Cooper, M.


    A chelate of technetium-99m, cobalt-57, gallium-67, gallium-68, indium-111 or indium-113m and a substituted iminodiacetic acid or an 8-hydroxyquinoline useful as a radiopharmaceutical external imaging agent. The invention also includes preparative methods therefor

  16. Image correlation method for DNA sequence alignment. (United States)

    Curilem Saldías, Millaray; Villarroel Sassarini, Felipe; Muñoz Poblete, Carlos; Vargas Vásquez, Asticio; Maureira Butler, Iván


    The complexity of searches and the volume of genomic data make sequence alignment one of bioinformatics most active research areas. New alignment approaches have incorporated digital signal processing techniques. Among these, correlation methods are highly sensitive. This paper proposes a novel sequence alignment method based on 2-dimensional images, where each nucleic acid base is represented as a fixed gray intensity pixel. Query and known database sequences are coded to their pixel representation and sequence alignment is handled as object recognition in a scene problem. Query and database become object and scene, respectively. An image correlation process is carried out in order to search for the best match between them. Given that this procedure can be implemented in an optical correlator, the correlation could eventually be accomplished at light speed. This paper shows an initial research stage where results were "digitally" obtained by simulating an optical correlation of DNA sequences represented as images. A total of 303 queries (variable lengths from 50 to 4500 base pairs) and 100 scenes represented by 100 x 100 images each (in total, one million base pair database) were considered for the image correlation analysis. The results showed that correlations reached very high sensitivity (99.01%), specificity (98.99%) and outperformed BLAST when mutation numbers increased. However, digital correlation processes were hundred times slower than BLAST. We are currently starting an initiative to evaluate the correlation speed process of a real experimental optical correlator. By doing this, we expect to fully exploit optical correlation light properties. As the optical correlator works jointly with the computer, digital algorithms should also be optimized. The results presented in this paper are encouraging and support the study of image correlation methods on sequence alignment.

  17. New magnetic resonance imaging methods in nephrology (United States)

    Zhang, Jeff L.; Morrell, Glen; Rusinek, Henry; Sigmund, Eric; Chandarana, Hersh; Lerman, Lilach O.; Prasad, Pottumarthi Vara; Niles, David; Artz, Nathan; Fain, Sean; Vivier, Pierre H.; Cheung, Alfred K.; Lee, Vivian S.


    Established as a method to study anatomic changes, such as renal tumors or atherosclerotic vascular disease, magnetic resonance imaging (MRI) to interrogate renal function has only recently begun to come of age. In this review, we briefly introduce some of the most important MRI techniques for renal functional imaging, and then review current findings on their use for diagnosis and monitoring of major kidney diseases. Specific applications include renovascular disease, diabetic nephropathy, renal transplants, renal masses, acute kidney injury and pediatric anomalies. With this review, we hope to encourage more collaboration between nephrologists and radiologists to accelerate the development and application of modern MRI tools in nephrology clinics. PMID:24067433

  18. Development of Ultrasensitive Plasmonic Nanosensors (United States)

    Joshi, Gayatribahen K.

    -based LSPR nanosensors ("plasmonic nanosensors") have been developed for different sensing applications. Specifically, these plasmonic nanosensors displayed capacity to detect streptavidine, glucose, microRNA (cancer biomarker), as well as molecular and stimuli-responsive polymers conformational changes. In this study we found that the plasmonic nanosensors are exceptionally sensitive compared to other NSs and the sensitivity is highly edge length dependent. An ultrasensitive plasmonic nanosensor has been developed for the detection of microRNAs in crude plasma collected from pancreatic cancer patients. It shows that the LSPR-based nanosensor has the ability to detect and quantify the microRNA concentrations in clinical samples without any purification. The results presented here show potential for patients to commence treatment in early stage cancer diagnosis. The effect of various physiological medias and edge length of nanoprisms on the sensitivity of this nanosensor has been discussed. Second, molecular sensors have been developed by functionalization of azobenzene molecule contain alkanethiols onto the nanoprisms surface. Molecular conformational changes basis on a very less dielectric thickness changes have been detected through lambdaLSPR shift of nanoprisms and confirmed through surface enhanced Raman spectroscopy (SERS). In this study, the influence of resonance energy transfer between the molecule and nanoprisms onto the lambda LSPR shift and Raman intensity has been investigated by changing the distance between them. Finally, utilization of stimuli-responsive polymers structural change in the development of stimuli-responsive such as pH and temperature-responsive plasmonic nanosensors has been demonstrated. It was found that the stimuli-responsive nanosensors were able to detect very small R.I. change due to the polymers structural change. The enzymatic reaction between glucose and glucose oxidase has been used to detect glucose in bovine plasma using p

  19. Three-dimensional image signals: processing methods (United States)

    Schiopu, Paul; Manea, Adrian; Craciun, Anca-Ileana; Craciun, Alexandru


    Over the years extensive studies have been carried out to apply coherent optics methods in real-time processing, communications and transmission image. This is especially true when a large amount of information needs to be processed, e.g., in high-resolution imaging. The recent progress in data-processing networks and communication systems has considerably increased the capacity of information exchange. We describe the results of literature investigation research of processing methods for the signals of the three-dimensional images. All commercially available 3D technologies today are based on stereoscopic viewing. 3D technology was once the exclusive domain of skilled computer-graphics developers with high-end machines and software. The images capture from the advanced 3D digital camera can be displayed onto screen of the 3D digital viewer with/ without special glasses. For this is needed considerable processing power and memory to create and render the complex mix of colors, textures, and virtual lighting and perspective necessary to make figures appear three-dimensional. Also, using a standard digital camera and a technique called phase-shift interferometry we can capture "digital holograms." These are holograms that can be stored on computer and transmitted over conventional networks. We present some research methods to process "digital holograms" for the Internet transmission and results.

  20. Image reconstruction methods in positron tomography

    International Nuclear Information System (INIS)

    Townsend, D.W.; Defrise, M.


    In the two decades since the introduction of the X-ray scanner into radiology, medical imaging techniques have become widely established as essential tools in the diagnosis of disease. As a consequence of recent technological and mathematical advances, the non-invasive, three-dimensional imaging of internal organs such as the brain and the heart is now possible, not only for anatomical investigations using X-ray but also for studies which explore the functional status of the body using positron-emitting radioisotopes. This report reviews the historical and physical basis of medical imaging techniques using positron-emitting radioisotopes. Mathematical methods which enable three-dimensional distributions of radioisotopes to be reconstructed from projection data (sinograms) acquired by detectors suitably positioned around the patient are discussed. The extension of conventional two-dimensional tomographic reconstruction algorithms to fully three-dimensional reconstruction is described in detail. (orig.)

  1. Rational Design of an Ultrasensitive Quorum-Sensing Switch. (United States)

    Zeng, Weiqian; Du, Pei; Lou, Qiuli; Wu, Lili; Zhang, Haoqian M; Lou, Chunbo; Wang, Hongli; Ouyang, Qi


    One of the purposes of synthetic biology is to develop rational methods that accelerate the design of genetic circuits, saving time and effort spent on experiments and providing reliably predictable circuit performance. We applied a reverse engineering approach to design an ultrasensitive transcriptional quorum-sensing switch. We want to explore how systems biology can guide synthetic biology in the choice of specific DNA sequences and their regulatory relations to achieve a targeted function. The workflow comprises network enumeration that achieves the target function robustly, experimental restriction of the obtained candidate networks, global parameter optimization via mathematical analysis, selection and engineering of parts based on these calculations, and finally, circuit construction based on the principles of standardization and modularization. The performance of realized quorum-sensing switches was in good qualitative agreement with the computational predictions. This study provides practical principles for the rational design of genetic circuits with targeted functions.

  2. Ultrasensitive chemiluminescence of tetracyclines in the presence of MCLA

    Energy Technology Data Exchange (ETDEWEB)

    Zeng, Wangsheng; Zhu, Chenyao [School of Public Health, Nanchang University, Nanchang 330006 (China); Liu, Hongcheng [Institute of Quality Standard and Testing Technology, Yunnan Academy of Agriculture Science, 650223 Kunming (China); Liu, Jing; Cai, Hongping [School of Public Health, Nanchang University, Nanchang 330006 (China); Cheng, Xianglei, E-mail: [School of Public Health, Nanchang University, Nanchang 330006 (China); Wei, Lijun, E-mail: [School of Public Health, Nanchang University, Nanchang 330006 (China)


    In this article, five tetracyclines (TCs) showed ultrasensitive chemiluminescence (CL) based on the novel CL system of methoxylated Cypridina luciferin analogues (MCLA) and Ce(IV). It was found that the CL intensity of MCLA-Ce(IV) system was enhanced 30 times by adding 10 pmol chlortetracycline. Mechanisms of MCLA-Ce(IV)-TCs were investigated by CL spectra, radical scavengers and UV spectra. The results indicated that singlet oxygen, resulting from the reaction between Ce(IV) and TCs, was involved in the enhancement of CL phenomenon. This CL system coupling with flow injection analysis was developed for the determination of five common TCs. Linear ranged from 100 fmol to 20 pmol (r>0.999) with limit of detection of 20 to 50 fmol (S/N=3). The sensitivity of the present method is comparable to that of UPLC-MS/MS detection for TCs.

  3. Method and apparatus for enhancing radiometric imaging

    International Nuclear Information System (INIS)

    Logan, R. H.; Paradish, F. J.


    Disclosed is a method and apparatus for enhancing target detection, particularly in the millimeter wave frequency range, through the utilization of an imaging radiometer. The radiometer, which is a passive thermal receiver, detects the reflected and emitted thermal radiation of targets within a predetermined antenna/receiver beamwidth. By scanning the radiometer over a target area, a thermal image is created. At millimeter wave frequencies, the received emissions from the target area are highly dependent on the emissivity of the target of interest. Foliage will appear ''hot'' due to its high emissivity and metals will appear cold due to their low emissivities. A noise power illuminator is periodically actuated to illuminate the target of interest. When the illuminator is actuated, the role of emissivity is reversed, namely poorly emissive targets will generally be good reflectors which in the presence of an illuminator will appear ''hot''. The highly emissive targets (such as foliage and dirt) which absorb most of the transmitted energy will appear almost the same as in a nonilluminated, passive image. Using a data processor, the intensity of the passive image is subtracted from the intensity of the illuminated, active image which thereby cancels the background foliage, dirt, etc. and the reflective metallic targets are enhanced

  4. Active learning methods for interactive image retrieval. (United States)

    Gosselin, Philippe Henri; Cord, Matthieu


    Active learning methods have been considered with increased interest in the statistical learning community. Initially developed within a classification framework, a lot of extensions are now being proposed to handle multimedia applications. This paper provides algorithms within a statistical framework to extend active learning for online content-based image retrieval (CBIR). The classification framework is presented with experiments to compare several powerful classification techniques in this information retrieval context. Focusing on interactive methods, active learning strategy is then described. The limitations of this approach for CBIR are emphasized before presenting our new active selection process RETIN. First, as any active method is sensitive to the boundary estimation between classes, the RETIN strategy carries out a boundary correction to make the retrieval process more robust. Second, the criterion of generalization error to optimize the active learning selection is modified to better represent the CBIR objective of database ranking. Third, a batch processing of images is proposed. Our strategy leads to a fast and efficient active learning scheme to retrieve sets of online images (query concept). Experiments on large databases show that the RETIN method performs well in comparison to several other active strategies.

  5. Ultrasensitive FRET-based DNA sensor using PNA/DNA hybridization. (United States)

    Yang, Lan-Hee; Ahn, Dong June; Koo, Eunhae


    In the diagnosis of genetic diseases, rapid and highly sensitive DNA detection is crucial. Therefore, many strategies for detecting target DNA have been developed, including electrical, optical, and mechanical methods. Herein, a highly sensitive FRET based sensor was developed by using PNA (Peptide Nucleic Acid) probe and QD, in which red color QDs are hybridized with capture probes, reporter probes and target DNAs by EDC-NHS coupling. The hybridized probe with target DNA gives off fluorescent signal due to the energy transfer from QD to Cy5 dye in the reporter probe. Compared to the conventional DNA sensor using DNA probes, the DNA sensor using PNA probes shows higher FRET factor and efficiency due to the higher reactivity between PNA and target DNA. In addition, to elicit the effect of the distance between the donor and the acceptor, we have investigated two types of the reporter probes having Cy5 dyes attached at the different positions of the reporter probes. Results show that the shorter the distance between QDs and Cy5s, the stronger the signal intensity. Furthermore, based on the fluorescence microscopy images using microcapillary chips, the FRET signal is enhanced to be up to 276% times stronger than the signal obtained using the cuvette by the fluorescence spectrometer. These results suggest that the PNA probe system conjugated with QDs can be used as ultrasensitive DNA nanosensors. Copyright © 2016. Published by Elsevier B.V.

  6. Enhancing the (MSLDIP) image steganographic method (ESLDIP method) (United States)

    Seddik Saad, Al-hussien


    Message transmissions over the Internet still have data security problem. Therefore, secure and secret communication methods are needed for transmitting messages over the Internet. Cryptography scrambles the message so that it cannot be understood. However, it makes the message suspicious enough to attract eavesdropper's attention. Steganography hides the secret message within other innocuous-looking cover files (i.e. images, music and video files) so that it cannot be observed [1].The term steganography originates from the Greek root words "steganos'' and "graphein'' which literally mean "covered writing''. It is defined as the science that involves communicating secret data in an appropriate multimedia carrier, e.g., image, audio text and video files [3].Steganographic techniques allow one party to communicate information to another without a third party even knowing that the communication is occurring. The ways to deliver these "secret messages" vary greatly [3].Our proposed method called Enhanced SLDIP (ESLDIP). In which the maximmum hiding capacity (MHC) of proposed ESLDIP method is higher than the previously proposed MSLDIP methods and the PSNR of the ESLDIP method is higher than the MSLDIP PSNR values', which means that the image quality of the ESLDIP method will be better than MSLDIP method and the maximmum hiding capacity (MHC) also improved. The rest of this paper is organized as follows. In section 2, steganography has been discussed; lingo, carriers and types. In section 3, related works are introduced. In section 4, the proposed method will be discussed in details. In section 5, the simulation results are given and Section 6 concludes the paper.

  7. A Frequency Splitting Method For CFM Imaging

    DEFF Research Database (Denmark)

    Udesen, Jesper; Gran, Fredrik; Jensen, Jørgen Arendt


    The performance of conventional CFM imaging will often be degraded due to the relatively low number of pulses (4-10) used for each velocity estimate. To circumvent this problem we propose a new method using frequency splitting (FS). The FS method uses broad band chirps as excitation pulses instead...... of narrow band pulses as in conventional CFM imaging. By appropriate filtration, the returned signals are divided into a number of narrow band signals which are approximately disjoint. After clutter filtering the velocities are found from each frequency band using a conventional autocorrelation estimator......, a 5 MHz linear array transducer was used to scan a vessel situated at 30 mm depth with a maximum flow velocity of 0.1 m/s. The pulse repetition frequency was 1.8 kHz and the angle between the flow and the beam was 60 deg. A 15 mus chirp was used as excitation pulse and 40 independent velocity...

  8. In-electrode vs. on-electrode: ultrasensitive Faraday cage-type electrochemiluminescence immunoassay. (United States)

    Guo, Zhiyong; Sha, Yuhong; Hu, Yufang; Wang, Sui


    A new-concept of an "in-electrode" Faraday cage-type electrochemiluminescence immunoassay (ECLIA) method for the ultrasensitive detection of neurotensin (NT) was reported with capture antibody (Ab1)-nanoFe3O4@graphene (GO) and detector antibody (Ab2)&N-(4-aminobutyl)-N-ethylisoluminol (ABEI)@GO, which led to about 1000-fold improvement in sensitivity by extending the Helmholtz plane (OHP) of the proposed electrode assembly effectively.

  9. A Method for Denoising Image Contours

    Directory of Open Access Journals (Sweden)

    Ovidiu COSMA


    Full Text Available The edge detection techniques have to compromise between sensitivity and noise. In order for the main contours to be uninterrupted, the level of sensitivity has to be raised, which however has the negative effect of producing a multitude of insignificant contours (noise. This article proposes a method of removing this noise, which acts directly on the binary representation of the image contours.

  10. Diffusion weighted imaging by MR method

    International Nuclear Information System (INIS)

    Horikawa, Yoshiharu; Naruse, Shoji; Ebisu, Toshihiko; Tokumitsu, Takuaki; Ueda, Satoshi; Tanaka, Chuzo; Higuchi, Toshihiro; Umeda, Masahiro.


    Diffusion weighted magnetic resonance imaging is a recently developed technique used to examine the micromovement of water molecules in vivo. We have applied this technique to examine various kinds of brain diseases, both experimentally and clinically. The calculated apparent diffusion coefficient (ADC) in vivo showed reliable values. In experimentally induced brain edema in rats, the pathophysiological difference of the type of edema (such as cytotoxic, and vasogenic) could be differentiated on the diffusion weighted MR images. Cytotoxic brain edema showed high intensity (slower diffusion) on the diffusion weighted images. On the other hand, vasogenic brain edema showed a low intensity image (faster diffusion). Diffusion anisotropy was demonstrated according to the direction of myelinated fibers and applied motion proving gradient (MPG). This anisotropy was also demonstrated in human brain tissue along the course of the corpus callosum, pyramidal tract and optic radiation. In brain ischemia cases, lesions were detected as high signal intensity areas, even one hour after the onset of ischemia. Diffusion was faster in brain tumor compared with normal brain. Histological differences were not clearly reflected by the ADC value. In epidermoid tumor cases, the intensity was characteristically high, was demonstrated, and the cerebrospinal fluid border was clearly demonstrated. New clinical information obtainable with this molecular diffusion method will prove to be useful in various clinical studies. (author)

  11. A nuclear method to authenticate Buddha images

    International Nuclear Information System (INIS)

    Khaweerat, S; Ratanatongchai, W; Channuie, J; Wonglee, S; Picha, R; Promping, J; Silva, K; Liamsuwan, T


    The value of Buddha images in Thailand varies dramatically depending on authentication and provenance. In general, people use their individual skills to make the justification which frequently leads to obscurity, deception and illegal activities. Here, we propose two non-destructive techniques of neutron radiography (NR) and neutron activation autoradiography (NAAR) to reveal respectively structural and elemental profiles of small Buddha images. For NR, a thermal neutron flux of 10 5 n cm -2 s -1 was applied. NAAR needed a higher neutron flux of 10 12 n cm -2 s -1 to activate the samples. Results from NR and NAAR revealed unique characteristic of the samples. Similarity of the profile played a key role in the classification of the samples. The results provided visual evidence to enhance the reliability of authenticity approval. The method can be further developed for routine practice which impact thousands of customers in Thailand. (paper)

  12. A nuclear method to authenticate Buddha images (United States)

    Khaweerat, S.; Ratanatongchai, W.; Channuie, J.; Wonglee, S.; Picha, R.; Promping, J.; Silva, K.; Liamsuwan, T.


    The value of Buddha images in Thailand varies dramatically depending on authentication and provenance. In general, people use their individual skills to make the justification which frequently leads to obscurity, deception and illegal activities. Here, we propose two non-destructive techniques of neutron radiography (NR) and neutron activation autoradiography (NAAR) to reveal respectively structural and elemental profiles of small Buddha images. For NR, a thermal neutron flux of 105 n cm-2s-1 was applied. NAAR needed a higher neutron flux of 1012 n cm-2 s-1 to activate the samples. Results from NR and NAAR revealed unique characteristic of the samples. Similarity of the profile played a key role in the classification of the samples. The results provided visual evidence to enhance the reliability of authenticity approval. The method can be further developed for routine practice which impact thousands of customers in Thailand.

  13. Research of ART method in CT image reconstruction

    International Nuclear Information System (INIS)

    Li Zhipeng; Cong Peng; Wu Haifeng


    This paper studied Algebraic Reconstruction Technique (ART) in CT image reconstruction. Discussed the ray number influence on image quality. And the adopting of smooth method got high quality CT image. (authors)


    Directory of Open Access Journals (Sweden)

    H. Shen


    Full Text Available Data fusion techniques have been widely researched and applied in remote sensing field. In this paper, an integrated fusion method for remotely sensed images is presented. Differently from the existed methods, the proposed method has the performance to integrate the complementary information in multiple temporal-spatial-spectral images. In order to represent and process the images in one unified framework, two general image observation models are firstly presented, and then the maximum a posteriori (MAP framework is used to set up the fusion model. The gradient descent method is employed to solve the fused image. The efficacy of the proposed method is validated using simulated images.


    Directory of Open Access Journals (Sweden)

    J. Zhao


    Full Text Available The image segmentation method based on two-dimensional histogram segments the image according to the thresholds of the intensity of the target pixel and the average intensity of its neighborhood. This method is essentially a hard-decision method. Due to the uncertainties when labeling the pixels around the threshold, the hard-decision method can easily get the wrong segmentation result. Therefore, a fusion segmentation method based on fuzzy theory is proposed in this paper. We use membership function to model the uncertainties on each color channel of the color image. Then, we segment the color image according to the fuzzy reasoning. The experiment results show that our proposed method can get better segmentation results both on the natural scene images and optical remote sensing images compared with the traditional thresholding method. The fusion method in this paper can provide new ideas for the information extraction of optical remote sensing images and polarization SAR images.

  16. Method and apparatus for imaging volume data

    International Nuclear Information System (INIS)

    Drebin, R.; Carpenter, L.C.


    An imaging system projects a two dimensional representation of three dimensional volumes where surface boundaries and objects internal to the volumes are readily shown, and hidden surfaces and the surface boundaries themselves are accurately rendered by determining volume elements or voxels. An image volume representing a volume object or data structure is written into memory. A color and opacity is assigned to each voxel within the volume and stored as a red (R), green (G), blue (B), and opacity (A) component, three dimensional data volume. The RGBA assignment for each voxel is determined based on the percentage component composition of the materials represented in the volume, and thus, the percentage of color and transparency associated with those materials. The voxels in the RGBA volume are used as mathematical filters such that each successive voxel filter is overlayed over a prior background voxel filter. Through a linear interpolation, a new background filter is determined and generated. The interpolation is successively performed for all voxels up to the front most voxel for the plane of view. The method is repeated until all display voxels are determined for the plane of view. (author)

  17. Subsurface imaging by electrical and EM methods

    Energy Technology Data Exchange (ETDEWEB)



    This report consists of 3 subjects. 1) Three dimensional inversion of resistivity data with topography : In this study, we developed a 3-D inversion method based on the finite element calculation of model responses, which can effectively accommodate the irregular topography. In solving the inverse problem, the iterative least-squares approach comprising the smoothness-constraints was taken along with the reciprocity approach in the calculation of Jacobian. Furthermore the Active Constraint Balancing, which has been recently developed by ourselves to enhance the resolving power of the inverse problem, was also employed. Since our new algorithm accounts for the topography in the inversion step, topography correction is not necessary as a preliminary processing and we can expect a more accurate image of the earth. 2) Electromagnetic responses due to a source in the borehole : The effects of borehole fluid and casing on the borehole EM responses should thoroughly be analyzed since they may affect the resultant image of the earth. In this study, we developed an accurate algorithm for calculating the EM responses containing the effects of borehole fluid and casing when a current-carrying ring is located on the borehole axis. An analytic expression for primary vertical magnetic field along the borehole axis was first formulated and the fast Fourier transform is to be applied to get the EM fields at any location in whole space. 3) High frequency electromagnetic impedance survey : At high frequencies the EM impedance becomes a function of the angle of incidence or the horizontal wavenumber, so the electrical properties cannot be readily extracted without first eliminating the effect of horizontal wavenumber on the impedance. For this purpose, this paper considers two independent methods for accurately determining the horizontal wavenumber, which in turn is used to correct the impedance data. The 'apparent' electrical properties derived from the corrected impedance

  18. Image Registration Methode in Radar Interferometry

    Directory of Open Access Journals (Sweden)

    S. Chelbi


    Full Text Available This article presents a methodology for the determination of the registration of an Interferometric Synthetic radar (InSAR pair images with half pixel precision. Using the two superposed radar images Single Look complexes (SLC [1-4], we developed an iterative process to superpose these two images according to their correlation coefficient with a high coherence area. This work concerns the exploitation of ERS Tandem pair of radar images SLC of the Algiers area acquired on 03 January and 04 January 1994. The former is taken as a master image and the latter as a slave image.

  19. Numerical methods in image processing for applications in jewellery industry


    Petrla, Martin


    Presented thesis deals with a problem from the field of image processing for application in multiple scanning of jewelery stones. The aim is to develop a method for preprocessing and subsequent mathematical registration of images in order to increase the effectivity and reliability of the output quality control. For these purposes the thesis summerizes mathematical definition of digital image as well as theoretical base of image registration. It proposes a method adjusting every single image ...

  20. Keyhole imaging method for dynamic objects behind the occlusion area (United States)

    Hao, Conghui; Chen, Xi; Dong, Liquan; Zhao, Yuejin; Liu, Ming; Kong, Lingqin; Hui, Mei; Liu, Xiaohua; Wu, Hong


    A method of keyhole imaging based on camera array is realized to obtain the video image behind a keyhole in shielded space at a relatively long distance. We get the multi-angle video images by using a 2×2 CCD camera array to take the images behind the keyhole in four directions. The multi-angle video images are saved in the form of frame sequences. This paper presents a method of video frame alignment. In order to remove the non-target area outside the aperture, we use the canny operator and morphological method to realize the edge detection of images and fill the images. The image stitching of four images is accomplished on the basis of the image stitching algorithm of two images. In the image stitching algorithm of two images, the SIFT method is adopted to accomplish the initial matching of images, and then the RANSAC algorithm is applied to eliminate the wrong matching points and to obtain a homography matrix. A method of optimizing transformation matrix is proposed in this paper. Finally, the video image with larger field of view behind the keyhole can be synthesized with image frame sequence in which every single frame is stitched. The results show that the screen of the video is clear and natural, the brightness transition is smooth. There is no obvious artificial stitching marks in the video, and it can be applied in different engineering environment .

  1. [Multimodal medical image registration using cubic spline interpolation method]. (United States)

    He, Yuanlie; Tian, Lianfang; Chen, Ping; Wang, Lifei; Ye, Guangchun; Mao, Zongyuan


    Based on the characteristic of the PET-CT multimodal image series, a novel image registration and fusion method is proposed, in which the cubic spline interpolation method is applied to realize the interpolation of PET-CT image series, then registration is carried out by using mutual information algorithm and finally the improved principal component analysis method is used for the fusion of PET-CT multimodal images to enhance the visual effect of PET image, thus satisfied registration and fusion results are obtained. The cubic spline interpolation method is used for reconstruction to restore the missed information between image slices, which can compensate for the shortage of previous registration methods, improve the accuracy of the registration, and make the fused multimodal images more similar to the real image. Finally, the cubic spline interpolation method has been successfully applied in developing 3D-CRT (3D Conformal Radiation Therapy) system.

  2. Evaluation of processing methods for static radioisotope scan images

    International Nuclear Information System (INIS)

    Oakberg, J.A.


    Radioisotope scanning in the field of nuclear medicine provides a method for the mapping of a radioactive drug in the human body to produce maps (images) which prove useful in detecting abnormalities in vital organs. At best, radioisotope scanning methods produce images with poor counting statistics. One solution to improving the body scan images is using dedicated small computers with appropriate software to process the scan data. Eleven methods for processing image data are compared

  3. Enhancement of Electroluminescence (EL) image measurements for failure quantification methods

    DEFF Research Database (Denmark)

    Parikh, Harsh; Spataru, Sergiu; Sera, Dezso


    Enhanced quality images are necessary for EL image analysis and failure quantification. A method is proposed which determines image quality in terms of more accurate failure detection of solar panels through electroluminescence (EL) imaging technique. The goal of the paper is to determine the most...

  4. Method for Surface Scanning in Medical Imaging and Related Apparatus

    DEFF Research Database (Denmark)


    A method and apparatus for surface scanning in medical imaging is provided. The surface scanning apparatus comprises an image source, a first optical fiber bundle comprising first optical fibers having proximal ends and distal ends, and a first optical coupler for coupling an image from the image...

  5. Quantum dots as optical labels for ultrasensitive detection of polyphenols. (United States)

    Akshath, Uchangi Satyaprasad; Shubha, Likitha R; Bhatt, Praveena; Thakur, Munna Singh


    Considering the fact that polyphenols have versatile activity in-vivo, its detection and quantification is very much important for a healthy diet. Laccase enzyme can convert polyphenols to yield mono/polyquinones which can quench Quantum dots fluorescence. This phenomenon of charge transfer from quinones to QDs was exploited as optical labels to detect polyphenols. CdTe QD may undergo dipolar interaction with quinones as a result of broad spectral absorption due to multiple excitonic states resulting from quantum confinement effects. Thus, "turn-off" fluorescence method was applied for ultrasensitive detection of polyphenols by using laccase. We observed proportionate quenching of QDs fluorescence with respect to polyphenol concentration in the range of 100 µg to 1 ng/mL. Also, quenching of the photoluminescence was highly efficient and stable and could detect individual and total polyphenols with high sensitivity (LOD-1 ng/mL). Moreover, proposed method was highly efficient than any other reported methods in terms of sensitivity, specificity and selectivity. Therefore, a novel optical sensor was developed for the detection of polyphenols at a sensitive level based on the charge transfer mechanism. Copyright © 2014 Elsevier B.V. All rights reserved.

  6. Alternate method for to realize image fusion

    International Nuclear Information System (INIS)

    Vargas, L.; Hernandez, F.; Fernandez, R.


    At the present time the image departments have the necessity of carrying out image fusion obtained of diverse apparatuses. Conventionally its fuse resonance or tomography images by X-rays with functional images as the gammagrams and PET images. The fusion technology is for sale with the modern image equipment and not all the cabinets of nuclear medicine have access to it. By this reason we analyze, study and we find a solution so that all the cabinets of nuclear medicine can benefit of the image fusion. The first indispensable requirement is to have a personal computer with capacity to put up image digitizer cards. It is also possible, if one has a gamma camera that can export images in JPG, GIF, TIFF or BMP formats, to do without of the digitizer card and to record the images in a disk to be able to use them in the personal computer. It is required of one of the following commercially available graph design programs: Corel Draw, Photo Shop, FreeHand, Illustrator or Macromedia Flash that are those that we evaluate and that its allow to make the images fusion. Anyone of them works well and a short training is required to be able to manage them. It is necessary a photographic digital camera with a resolution of at least 3.0 mega pixel. The procedure consists on taking photographic images of the radiological studies that the patient already has, selecting those demonstrative images of the pathology in study and that its can also be concordant with the images that we have created in the gammagraphic studies, whether for planar or tomographic. We transfer the images to the personal computer and we read them with the graph design program. To continuation also reads the gammagraphic images. We use those digital tools to make transparent the images, to clip them, to adjust the sizes and to create the fused images. The process is manual and it is requires of ability and experience to choose the images, the cuts, those sizes and the transparency grade. (Author)

  7. 3D Interpolation Method for CT Images of the Lung

    Directory of Open Access Journals (Sweden)

    Noriaki Asada


    Full Text Available A 3-D image can be reconstructed from numerous CT images of the lung. The procedure reconstructs a solid from multiple cross section images, which are collected during pulsation of the heart. Thus the motion of the heart is a special factor that must be taken into consideration during reconstruction. The lung exhibits a repeating transformation synchronized to the beating of the heart as an elastic body. There are discontinuities among neighboring CT images due to the beating of the heart, if no special techniques are used in taking CT images. The 3-D heart image is reconstructed from numerous CT images in which both the heart and the lung are taken. Although the outline shape of the reconstructed 3-D heart is quite unnatural, the envelope of the 3-D unnatural heart is fit to the shape of the standard heart. The envelopes of the lung in the CT images are calculated after the section images of the best fitting standard heart are located at the same positions of the CT images. Thus the CT images are geometrically transformed to the optimal CT images fitting best to the standard heart. Since correct transformation of images is required, an Area oriented interpolation method proposed by us is used for interpolation of transformed images. An attempt to reconstruct a 3-D lung image by a series of such operations without discontinuity is shown. Additionally, the same geometrical transformation method to the original projection images is proposed as a more advanced method.

  8. An efficient direct method for image registration of flat objects (United States)

    Nikolaev, Dmitry; Tihonkih, Dmitrii; Makovetskii, Artyom; Voronin, Sergei


    Image alignment of rigid surfaces is a rapidly developing area of research and has many practical applications. Alignment methods can be roughly divided into two types: feature-based methods and direct methods. Known SURF and SIFT algorithms are examples of the feature-based methods. Direct methods refer to those that exploit the pixel intensities without resorting to image features and image-based deformations are general direct method to align images of deformable objects in 3D space. Nevertheless, it is not good for the registration of images of 3D rigid objects since the underlying structure cannot be directly evaluated. In the article, we propose a model that is suitable for image alignment of rigid flat objects under various illumination models. The brightness consistency assumptions used for reconstruction of optimal geometrical transformation. Computer simulation results are provided to illustrate the performance of the proposed algorithm for computing of an accordance between pixels of two images.

  9. Image Processing Methods Usable for Object Detection on the Chessboard

    Directory of Open Access Journals (Sweden)

    Beran Ladislav


    Full Text Available Image segmentation and object detection is challenging problem in many research. Although many algorithms for image segmentation have been invented, there is no simple algorithm for image segmentation and object detection. Our research is based on combination of several methods for object detection. The first method suitable for image segmentation and object detection is colour detection. This method is very simply, but there is problem with different colours. For this method it is necessary to have precisely determined colour of segmented object before all calculations. In many cases it is necessary to determine this colour manually. Alternative simply method is method based on background removal. This method is based on difference between reference image and detected image. In this paper several methods suitable for object detection are described. Thisresearch is focused on coloured object detection on chessboard. The results from this research with fusion of neural networks for user-computer game checkers will be applied.

  10. Finite element formulation for a digital image correlation method

    International Nuclear Information System (INIS)

    Sun Yaofeng; Pang, John H. L.; Wong, Chee Khuen; Su Fei


    A finite element formulation for a digital image correlation method is presented that will determine directly the complete, two-dimensional displacement field during the image correlation process on digital images. The entire interested image area is discretized into finite elements that are involved in the common image correlation process by use of our algorithms. This image correlation method with finite element formulation has an advantage over subset-based image correlation methods because it satisfies the requirements of displacement continuity and derivative continuity among elements on images. Numerical studies and a real experiment are used to verify the proposed formulation. Results have shown that the image correlation with the finite element formulation is computationally efficient, accurate, and robust

  11. Perceptual digital imaging methods and applications

    CERN Document Server

    Lukac, Rastislav


    Visual perception is a complex process requiring interaction between the receptors in the eye that sense the stimulus and the neural system and the brain that are responsible for communicating and interpreting the sensed visual information. This process involves several physical, neural, and cognitive phenomena whose understanding is essential to design effective and computationally efficient imaging solutions. Building on advances in computer vision, image and video processing, neuroscience, and information engineering, perceptual digital imaging greatly enhances the capabilities of tradition

  12. Developing optical traps for ultra-sensitive analysis

    International Nuclear Information System (INIS)

    Zhao, X.; Vieira, D.J.; Guckert, R.; Crane, S.


    The authors describe the coupling of a magneto-optical trap to a mass separator for the ultra-sensitive detection of selected radioactive species. As a proof of principle test, they have demonstrated the trapping of ∼ 6 million 82 Rb (t 1/2 = 75 s) atoms using an ion implantation and heated foil release method for introducing the sample into a trapping cell with minimal gas loading. Gamma-ray counting techniques were used to determine the efficiencies of each step in the process. By far the weakest step in the process is the efficiency of the optical trap itself (0.3%). Further improvements in the quality of the nonstick dryfilm coating on the inside of the trapping cell and the possible use of larger diameter laser beams are indicated. In the presence of a large background of scattered light, this initial work achieved a detection sensitivity of ∼ 4,000 trapped atoms. Improved detection schemes using a pulsed trap and gated photon detection method are outlined. Application of this technology to the areas of environmental monitoring and nuclear proliferation are foreseen

  13. New LSB-based colour image steganography method to enhance ...

    Indian Academy of Sciences (India)

    Mustafa Cem kasapbaşi


    Apr 27, 2018 ... evaluate the proposed method, comparative performance tests are carried out against different spatial image ... image steganography applications based on LSB are ..... worst case scenario could occur when having highest.

  14. ISAR imaging using the instantaneous range instantaneous Doppler method

    CSIR Research Space (South Africa)

    Wazna, TM


    Full Text Available In Inverse Synthetic Aperture Radar (ISAR) imaging, the Range Instantaneous Doppler (RID) method is used to compensate for the nonuniform rotational motion of the target that degrades the Doppler resolution of the ISAR image. The Instantaneous Range...

  15. Ultra-sensitive DNA assay based on single-molecule detection coupled with fluorescent quantum dot-labeling and its application to determination of messenger RNA

    International Nuclear Information System (INIS)

    Li Li; Li Xincang; Li Lu; Wang Jinxing; Jin Wenrui


    An ultra-sensitive single-molecule detection (SMD) method for quantification of DNA using total internal reflection fluorescence microscopy (TIRFM) coupled with fluorescent quantum dot (QD)-labeling was developed. In this method, the target DNA (tDNA) was captured by the capture DNA immobilized on the silanized coverslip blocked with ethanolamine and bovine serum albumin. Then, the QD-labeled probe DNA was hybridized to the tDNA. Ten fluorescent images of the QD-labeled sandwich DNA hybrids on the coverslip were taken by a high-sensitive CCD. The tDNA was quantified by counting the bright spots on the images using a calibration curve. The LOD of the method was 1 x 10 -14 mol L -1 . Several key factors, including image acquirement, fluorescence probe, substrate preparation, noise elimination from solutions and glass coverslips, and nonspecific adsorption and binding of solution-phase detection probes were discussed in detail. The method could be applied to quantify messenger RNA (mRNA) in cells. In order to determine mRNA, double-stranded RNA-DNA hybrids consisting of mRNA and corresponding cDNA were synthesized from the cellular mRNA template using reverse transcription in the presence of reverse transcriptase. After removing the mRNA in the double-stranded hybrids using ribonuclease, cDNA was quantified using the SMD-based TIRFM. Osteopontin mRNA in decidual stromal cells was chosen as the model analyte.

  16. Ultra-sensitive DNA assay based on single-molecule detection coupled with fluorescent quantum dot-labeling and its application to determination of messenger RNA

    Energy Technology Data Exchange (ETDEWEB)

    Li Li [School of Chemistry and Chemical Engineering, Shandong University, Jinan 250100 (China); Li Xincang [School of Life Sciences, Shandong University, Jinan 250100 (China); Li Lu [School of Chemistry and Chemical Engineering, Shandong University, Jinan 250100 (China); Wang Jinxing [School of Life Sciences, Shandong University, Jinan 250100 (China); Jin Wenrui, E-mail: [School of Chemistry and Chemical Engineering, Shandong University, Jinan 250100 (China)


    An ultra-sensitive single-molecule detection (SMD) method for quantification of DNA using total internal reflection fluorescence microscopy (TIRFM) coupled with fluorescent quantum dot (QD)-labeling was developed. In this method, the target DNA (tDNA) was captured by the capture DNA immobilized on the silanized coverslip blocked with ethanolamine and bovine serum albumin. Then, the QD-labeled probe DNA was hybridized to the tDNA. Ten fluorescent images of the QD-labeled sandwich DNA hybrids on the coverslip were taken by a high-sensitive CCD. The tDNA was quantified by counting the bright spots on the images using a calibration curve. The LOD of the method was 1 x 10{sup -14} mol L{sup -1}. Several key factors, including image acquirement, fluorescence probe, substrate preparation, noise elimination from solutions and glass coverslips, and nonspecific adsorption and binding of solution-phase detection probes were discussed in detail. The method could be applied to quantify messenger RNA (mRNA) in cells. In order to determine mRNA, double-stranded RNA-DNA hybrids consisting of mRNA and corresponding cDNA were synthesized from the cellular mRNA template using reverse transcription in the presence of reverse transcriptase. After removing the mRNA in the double-stranded hybrids using ribonuclease, cDNA was quantified using the SMD-based TIRFM. Osteopontin mRNA in decidual stromal cells was chosen as the model analyte.

  17. Improved radionuclide bone imaging agent injection needle withdrawal method can improve image quality

    International Nuclear Information System (INIS)

    Qin Yongmei; Wang Laihao; Zhao Lihua; Guo Xiaogang; Kong Qingfeng


    Objective: To investigate the improvement of radionuclide bone imaging agent injection needle withdrawal method on whole body bone scan image quality. Methods: Elbow vein injection syringe needle directly into the bone imaging agent in the routine group of 117 cases, with a cotton swab needle injection method for the rapid pull out the needle puncture point pressing, pressing moment. Improvement of 117 cases of needle injection method to put two needles into the skin swabs and blood vessels, pull out the needle while pressing two or more entry point 5min. After 2 hours underwent whole body bone SPECT imaging plane. Results: The conventional group at the injection site imaging agents uptake rate was 16.24%, improved group was 2.56%. Conclusion: The modified bone imaging agent injection needle withdrawal method, injection-site imaging agent uptake were significantly decreased whole body bone imaging can improve image quality. (authors)

  18. Comparative analysis of different methods for image enhancement

    Institute of Scientific and Technical Information of China (English)

    吴笑峰; 胡仕刚; 赵瑾; 李志明; 李劲; 唐志军; 席在芳


    Image enhancement technology plays a very important role to improve image quality in image processing. By enhancing some information and restraining other information selectively, it can improve image visual effect. The objective of this work is to implement the image enhancement to gray scale images using different techniques. After the fundamental methods of image enhancement processing are demonstrated, image enhancement algorithms based on space and frequency domains are systematically investigated and compared. The advantage and defect of the above-mentioned algorithms are analyzed. The algorithms of wavelet based image enhancement are also deduced and generalized. Wavelet transform modulus maxima (WTMM) is a method for detecting the fractal dimension of a signal, it is well used for image enhancement. The image techniques are compared by using the mean (μ), standard deviation (s), mean square error (MSE) and PSNR (peak signal to noise ratio). A group of experimental results demonstrate that the image enhancement algorithm based on wavelet transform is effective for image de-noising and enhancement. Wavelet transform modulus maxima method is one of the best methods for image enhancement.

  19. System and method for image mapping and visual attention (United States)

    Peters, II, Richard A. (Inventor)


    A method is described for mapping dense sensory data to a Sensory Ego Sphere (SES). Methods are also described for finding and ranking areas of interest in the images that form a complete visual scene on an SES. Further, attentional processing of image data is best done by performing attentional processing on individual full-size images from the image sequence, mapping each attentional location to the nearest node, and then summing all attentional locations at each node.

  20. Apparatus and method X-ray image processing

    International Nuclear Information System (INIS)


    The invention relates to a method for X-ray image processing. The radiation passed through the object is transformed into an electric image signal from which the logarithmic value is determined and displayed by a display device. Its main objective is to provide a method and apparatus that renders X-ray images or X-ray subtraction images with strong reduction of stray radiation. (Auth.)

  1. A comparative study on medical image segmentation methods

    Directory of Open Access Journals (Sweden)

    Praylin Selva Blessy SELVARAJ ASSLEY


    Full Text Available Image segmentation plays an important role in medical images. It has been a relevant research area in computer vision and image analysis. Many segmentation algorithms have been proposed for medical images. This paper makes a review on segmentation methods for medical images. In this survey, segmentation methods are divided into five categories: region based, boundary based, model based, hybrid based and atlas based. The five different categories with their principle ideas, advantages and disadvantages in segmenting different medical images are discussed.

  2. An Image Encryption Method Based on Bit Plane Hiding Technology

    Institute of Scientific and Technical Information of China (English)

    LIU Bin; LI Zhitang; TU Hao


    A novel image hiding method based on the correlation analysis of bit plane is described in this paper. Firstly, based on the correlation analysis, different bit plane of a secret image is hided in different bit plane of several different open images. And then a new hiding image is acquired by a nesting "Exclusive-OR" operation on those images obtained from the first step. At last, by employing image fusion technique, the final hiding result is achieved. The experimental result shows that the method proposed in this paper is effective.

  3. Comparison of whole-body-imaging methods

    International Nuclear Information System (INIS)

    Rollo, F.D.; Hoffer, P.


    Currently there are four different devices that have found clinical utility in whole-body imaging. These are the rectilinear scanner, the multicrystal whole-body scanner, the Anger-type camera with a whole-body-imaging table, and the tomoscanner. In this text, the basic theory of operation and a discussion of the advantages and disadvantages in whole-body imaging is presented for each device. When applicable, a comparative assessment of the various devices is also presented. As with all else in life, there is no simple answer to the question ''which total body imaging device is best.'' Institutions with a very heavy total-body-imaging load may prefer to use an already available dual-headed rectilinear scanner system for these studies, rather than invest in a new instrument. Institutions with moderate total-body-imaging loads may wish to invest in moving table or moving camera devices which make total body imaging more convenient but retain the basic flexibility of the camera. The large-field Anger camera with or without motion offers another flexible option to these institutions. The laboratory with a very heavy total body imaging load may select efficiency over flexibility, thereby freeing up other instruments for additional studies. Finally, reliability as well as availability and quality of local service must be considered. After all, design features of an instrument become irrelevant when it is broken down and awaiting repair

  4. Methods of fetal MR: beyond T2-weighted imaging

    Energy Technology Data Exchange (ETDEWEB)

    Brugger, Peter C. [Center of Anatomy and Cell Biology, Integrative Morphology Group, Medical University of Vienna, Waehringerstrasse 13, 1090 Vienna (Austria)]. E-mail:; Stuhr, Fritz [Department of Radiology, Medical University of Vienna, Waehringerguertel 18-20, 1090 Vienna (Austria); Lindner, Christian [Department of Radiology, Medical University of Vienna, Waehringerguertel 18-20, 1090 Vienna (Austria); Prayer, Daniela [Department of Radiology, Medical University of Vienna, Waehringerguertel 18-20, 1090 Vienna (Austria)


    The present work reviews the basic methods of performing fetal magnetic resonance imaging (MRI). Since fetal MRI differs in many respects from a postnatal study, several factors have to be taken into account to achieve satisfying image quality. Image quality depends on adequate positioning of the pregnant woman in the magnet, use of appropriate coils and the selection of sequences. Ultrafast T2-weighted sequences are regarded as the mainstay of fetal MR-imaging. However, additional sequences, such as T1-weighted images, diffusion-weighted images, echoplanar imaging may provide further information, especially in extra- central-nervous system regions of the fetal body.

  5. Methods of fetal MR: beyond T2-weighted imaging

    International Nuclear Information System (INIS)

    Brugger, Peter C.; Stuhr, Fritz; Lindner, Christian; Prayer, Daniela


    The present work reviews the basic methods of performing fetal magnetic resonance imaging (MRI). Since fetal MRI differs in many respects from a postnatal study, several factors have to be taken into account to achieve satisfying image quality. Image quality depends on adequate positioning of the pregnant woman in the magnet, use of appropriate coils and the selection of sequences. Ultrafast T2-weighted sequences are regarded as the mainstay of fetal MR-imaging. However, additional sequences, such as T1-weighted images, diffusion-weighted images, echoplanar imaging may provide further information, especially in extra- central-nervous system regions of the fetal body

  6. Hiding a Covert Digital Image by Assembling the RSA Encryption Method and the Binary Encoding Method

    Directory of Open Access Journals (Sweden)

    Kuang Tsan Lin


    Full Text Available The Rivest-Shamir-Adleman (RSA encryption method and the binary encoding method are assembled to form a hybrid hiding method to hide a covert digital image into a dot-matrix holographic image. First, the RSA encryption method is used to transform the covert image to form a RSA encryption data string. Then, all the elements of the RSA encryption data string are transferred into binary data. Finally, the binary data are encoded into the dot-matrix holographic image. The pixels of the dot-matrix holographic image contain seven groups of codes used for reconstructing the covert image. The seven groups of codes are identification codes, covert-image dimension codes, covert-image graylevel codes, pre-RSA bit number codes, RSA key codes, post-RSA bit number codes, and information codes. The reconstructed covert image derived from the dot-matrix holographic image and the original covert image are exactly the same.

  7. Research on image complexity evaluation method based on color information (United States)

    Wang, Hao; Duan, Jin; Han, Xue-hui; Xiao, Bo


    In order to evaluate the complexity of a color image more effectively and find the connection between image complexity and image information, this paper presents a method to compute the complexity of image based on color information.Under the complexity ,the theoretical analysis first divides the complexity from the subjective level, divides into three levels: low complexity, medium complexity and high complexity, and then carries on the image feature extraction, finally establishes the function between the complexity value and the color characteristic model. The experimental results show that this kind of evaluation method can objectively reconstruct the complexity of the image from the image feature research. The experimental results obtained by the method of this paper are in good agreement with the results of human visual perception complexity,Color image complexity has a certain reference value.

  8. A new method for mobile phone image denoising (United States)

    Jin, Lianghai; Jin, Min; Li, Xiang; Xu, Xiangyang


    Images captured by mobile phone cameras via pipeline processing usually contain various kinds of noises, especially granular noise with different shapes and sizes in both luminance and chrominance channels. In chrominance channels, noise is closely related to image brightness. To improve image quality, this paper presents a new method to denoise such mobile phone images. The proposed scheme converts the noisy RGB image to luminance and chrominance images, which are then denoised by a common filtering framework. The common filtering framework processes a noisy pixel by first excluding the neighborhood pixels that significantly deviate from the (vector) median and then utilizing the other neighborhood pixels to restore the current pixel. In the framework, the strength of chrominance image denoising is controlled by image brightness. The experimental results show that the proposed method obviously outperforms some other representative denoising methods in terms of both objective measure and visual evaluation.

  9. New mobile methods for dietary assessment: review of image-assisted and image-based dietary assessment methods. (United States)

    Boushey, C J; Spoden, M; Zhu, F M; Delp, E J; Kerr, D A


    For nutrition practitioners and researchers, assessing dietary intake of children and adults with a high level of accuracy continues to be a challenge. Developments in mobile technologies have created a role for images in the assessment of dietary intake. The objective of this review was to examine peer-reviewed published papers covering development, evaluation and/or validation of image-assisted or image-based dietary assessment methods from December 2013 to January 2016. Images taken with handheld devices or wearable cameras have been used to assist traditional dietary assessment methods for portion size estimations made by dietitians (image-assisted methods). Image-assisted approaches can supplement either dietary records or 24-h dietary recalls. In recent years, image-based approaches integrating application technology for mobile devices have been developed (image-based methods). Image-based approaches aim at capturing all eating occasions by images as the primary record of dietary intake, and therefore follow the methodology of food records. The present paper reviews several image-assisted and image-based methods, their benefits and challenges; followed by details on an image-based mobile food record. Mobile technology offers a wide range of feasible options for dietary assessment, which are easier to incorporate into daily routines. The presented studies illustrate that image-assisted methods can improve the accuracy of conventional dietary assessment methods by adding eating occasion detail via pictures captured by an individual (dynamic images). All of the studies reduced underreporting with the help of images compared with results with traditional assessment methods. Studies with larger sample sizes are needed to better delineate attributes with regards to age of user, degree of error and cost.

  10. Separation method of heavy-ion particle image from gamma-ray mixed images using an imaging plate

    CERN Document Server

    Yamadera, A; Ohuchi, H; Nakamura, T; Fukumura, A


    We have developed a separation method of alpha-ray and gamma-ray images using the imaging plate (IP). The IP from which the first image was read out by an image reader was annealed at 50 deg. C for 2 h in a drying oven and the second image was read out by the image reader. It was found out that an annealing ratio, k, which is defined as a ratio of the photo-stimulated luminescence (PSL) density at the first measurement to that at the second measurement, was different for alpha rays and gamma rays. By subtracting the second image multiplied by a factor of k from the first image, the alpha-ray image was separated from the alpha and gamma-ray mixed images. This method was applied to identify the images of helium, carbon and neon particles of high energies using the heavy-ion medical accelerator, HIMAC. (author)

  11. Quantum dynamic imaging theoretical and numerical methods

    CERN Document Server

    Ivanov, Misha


    Studying and using light or "photons" to image and then to control and transmit molecular information is among the most challenging and significant research fields to emerge in recent years. One of the fastest growing areas involves research in the temporal imaging of quantum phenomena, ranging from molecular dynamics in the femto (10-15s) time regime for atomic motion to the atto (10-18s) time scale of electron motion. In fact, the attosecond "revolution" is now recognized as one of the most important recent breakthroughs and innovations in the science of the 21st century. A major participant in the development of ultrafast femto and attosecond temporal imaging of molecular quantum phenomena has been theory and numerical simulation of the nonlinear, non-perturbative response of atoms and molecules to ultrashort laser pulses. Therefore, imaging quantum dynamics is a new frontier of science requiring advanced mathematical approaches for analyzing and solving spatial and temporal multidimensional partial differ...

  12. Blind compressed sensing image reconstruction based on alternating direction method (United States)

    Liu, Qinan; Guo, Shuxu


    In order to solve the problem of how to reconstruct the original image under the condition of unknown sparse basis, this paper proposes an image reconstruction method based on blind compressed sensing model. In this model, the image signal is regarded as the product of a sparse coefficient matrix and a dictionary matrix. Based on the existing blind compressed sensing theory, the optimal solution is solved by the alternative minimization method. The proposed method solves the problem that the sparse basis in compressed sensing is difficult to represent, which restrains the noise and improves the quality of reconstructed image. This method ensures that the blind compressed sensing theory has a unique solution and can recover the reconstructed original image signal from a complex environment with a stronger self-adaptability. The experimental results show that the image reconstruction algorithm based on blind compressed sensing proposed in this paper can recover high quality image signals under the condition of under-sampling.

  13. Novel welding image processing method based on fractal theory

    Institute of Scientific and Technical Information of China (English)

    陈强; 孙振国; 肖勇; 路井荣


    Computer vision has come into used in the fields of welding process control and automation. In order to improve precision and rapidity of welding image processing, a novel method based on fractal theory has been put forward in this paper. Compared with traditional methods, the image is preliminarily processed in the macroscopic regions then thoroughly analyzed in the microscopic regions in the new method. With which, an image is divided up to some regions according to the different fractal characters of image edge, and the fuzzy regions including image edges are detected out, then image edges are identified with Sobel operator and curved by LSM (Lease Square Method). Since the data to be processed have been decreased and the noise of image has been reduced, it has been testified through experiments that edges of weld seam or weld pool could be recognized correctly and quickly.

  14. Blind Methods for Detecting Image Fakery

    Czech Academy of Sciences Publication Activity Database

    Mahdian, Babak; Saic, Stanislav


    Roč. 25, č. 4 (2010), s. 18-24 ISSN 0885-8985 R&D Projects: GA ČR GA102/08/0470 Institutional research plan: CEZ:AV0Z10750506 Keywords : Image forensics * Image Fakery * Forgery detection * Authentication Subject RIV: BD - Theory of Information Impact factor: 0.179, year: 2010

  15. Speckle imaging using the principle value decomposition method

    International Nuclear Information System (INIS)

    Sherman, J.W.


    Obtaining diffraction-limited images in the presence of atmospheric turbulence is a topic of current interest. Two types of approaches have evolved: real-time correction and speckle imaging. A speckle imaging reconstruction method was developed by use of an ''optimal'' filtering approach. This method is based on a nonlinear integral equation which is solved by principle value decomposition. The method was implemented on a CDC 7600 for study. The restoration algorithm is discussed and its performance is illustrated. 7 figures

  16. Survey: interpolation methods for whole slide image processing. (United States)

    Roszkowiak, L; Korzynska, A; Zak, J; Pijanowska, D; Swiderska-Chadaj, Z; Markiewicz, T


    Evaluating whole slide images of histological and cytological samples is used in pathology for diagnostics, grading and prognosis . It is often necessary to rescale whole slide images of a very large size. Image resizing is one of the most common applications of interpolation. We collect the advantages and drawbacks of nine interpolation methods, and as a result of our analysis, we try to select one interpolation method as the preferred solution. To compare the performance of interpolation methods, test images were scaled and then rescaled to the original size using the same algorithm. The modified image was compared to the original image in various aspects. The time needed for calculations and results of quantification performance on modified images were also compared. For evaluation purposes, we used four general test images and 12 specialized biological immunohistochemically stained tissue sample images. The purpose of this survey is to determine which method of interpolation is the best to resize whole slide images, so they can be further processed using quantification methods. As a result, the interpolation method has to be selected depending on the task involving whole slide images. © 2016 The Authors Journal of Microscopy © 2016 Royal Microscopical Society.

  17. On a selection method of imaging condition in scintigraphy

    International Nuclear Information System (INIS)

    Ikeda, Hozumi; Kishimoto, Kenji; Shimonishi, Yoshihiro; Ohmura, Masahiro; Kosakai, Kazuhisa; Ochi, Hironobu


    Selection of imaging condition in scintigraphy was evaluated using analytic hierarchy process. First, a method of the selection was led by determining at the points of image quantity and imaging time. Influence of image quality was thought to depend on changes of system resolution, count density, image size, and image density. Also influence of imaging time was thought to depend on changes of system sensitivity and data acquisition time. Phantom study was done for paired comparison of these selection factors, and relations of sample data and the factors, that is Rollo phantom images were taken by changing count density, image size, and image density. Image quality was shown by calculating the score of visual evaluation that done by comparing of a pair of images in clearer cold lesion on the scintigrams. Imaging time was shown by relative values for changes of count density. However, system resolution and system sensitivity were constant in this study. Next, using these values analytic hierarchy process was adapted for this selection of imaging conditions. We conclude that this selection of imaging conditions can be analyzed quantitatively using analytic hierarchy process and this analysis develops theoretical consideration of imaging technique. (author)

  18. Development of Electrochemical Biosensors for Ultrasensitive Detection of Bacteria in the Environment

    DEFF Research Database (Denmark)

    Fapyane, Deby


    to those conventional methods, are intensively studied. Biosensor technology is one of the strategies for rapid monitoring of pathogens such as bacteria, virus, and parasites in the environment. Among them, the electrochemical biosensor offers simple, rapid, cost-effective and possibility...... for ultrasensitive detection of bacterial cells, DNA and rRNA. Several key operational parameters were assessed such as the optimization of probe design and labeling molecules. Here, more specifically we used two novel labels for the development of the electrochemical biosensor for bacteria detection; cellulase...

  19. Methods of filtering the graph images of the functions

    Directory of Open Access Journals (Sweden)

    Олександр Григорович Бурса


    Full Text Available The theoretical aspects of cleaning raster images of scanned graphs of functions from digital, chromatic and luminance distortions by using computer graphics techniques have been considered. The basic types of distortions characteristic of graph images of functions have been stated. To suppress the distortion several methods, providing for high-quality of the resulting images and saving their topological features, were suggested. The paper describes the techniques developed and improved by the authors: the method of cleaning the image of distortions by means of iterative contrasting, based on the step-by-step increase in image contrast in the graph by 1%; the method of small entities distortion restoring, based on the thinning of the known matrix of contrast increase filter (the allowable dimensions of the nucleus dilution radius convolution matrix, which provide for the retention of the graph lines have been established; integration technique of the noise reduction method by means of contrasting and distortion restoring method of small entities with known σ-filter. Each method in the complex has been theoretically substantiated. The developed methods involve treatment of graph images as the entire image (global processing and its fragments (local processing. The metrics assessing the quality of the resulting image with the global and local processing have been chosen, the substantiation of the choice as well as the formulas have been given. The proposed complex methods of cleaning the graphs images of functions from grayscale image distortions is adaptive to the form of an image carrier, the distortion level in the image and its distribution. The presented results of testing the developed complex of methods for a representative sample of images confirm its effectiveness

  20. Wavelet imaging cleaning method for atmospheric Cherenkov telescopes (United States)

    Lessard, R. W.; Cayón, L.; Sembroski, G. H.; Gaidos, J. A.


    We present a new method of image cleaning for imaging atmospheric Cherenkov telescopes. The method is based on the utilization of wavelets to identify noise pixels in images of gamma-ray and hadronic induced air showers. This method selects more signal pixels with Cherenkov photons than traditional image processing techniques. In addition, the method is equally efficient at rejecting pixels with noise alone. The inclusion of more signal pixels in an image of an air shower allows for a more accurate reconstruction, especially at lower gamma-ray energies that produce low levels of light. We present the results of Monte Carlo simulations of gamma-ray and hadronic air showers which show improved angular resolution using this cleaning procedure. Data from the Whipple Observatory's 10-m telescope are utilized to show the efficacy of the method for extracting a gamma-ray signal from the background of hadronic generated images.

  1. Training Methods for Image Noise Level Estimation on Wavelet Components

    Directory of Open Access Journals (Sweden)

    A. De Stefano


    Full Text Available The estimation of the standard deviation of noise contaminating an image is a fundamental step in wavelet-based noise reduction techniques. The method widely used is based on the mean absolute deviation (MAD. This model-based method assumes specific characteristics of the noise-contaminated image component. Three novel and alternative methods for estimating the noise standard deviation are proposed in this work and compared with the MAD method. Two of these methods rely on a preliminary training stage in order to extract parameters which are then used in the application stage. The sets used for training and testing, 13 and 5 images, respectively, are fully disjoint. The third method assumes specific statistical distributions for image and noise components. Results showed the prevalence of the training-based methods for the images and the range of noise levels considered.

  2. Distributed MIMO-ISAR Sub-image Fusion Method

    Directory of Open Access Journals (Sweden)

    Gu Wenkun


    Full Text Available The fast fluctuation associated with maneuvering a target’s radar cross-section often affects the imaging performance stability of traditional monostatic Inverse Synthetic Aperture Radar (ISAR. To address this problem, in this study, we propose an imaging method based on the fusion of sub-images of frequencydiversity-distributed multiple Input-Multiple Output-Inverse Synthetic Aperture Radar (MIMO-ISAR. First, we establish the analytic expression of a two-dimensional ISAR sub-image acquired by different channels of distributed MIMO-ISAR. Then, we derive the distance and azimuth distortion factors of the image acquired by the different channels. By compensating for the distortion of the ISAR image, we ultimately realize distributed MIMO-ISAR fusion imaging. Simulations verify the validity of this imaging method using distributed MIMO-ISAR.

  3. Investigation of Optimal Integrated Circuit Raster Image Vectorization Method

    Directory of Open Access Journals (Sweden)

    Leonas Jasevičius


    Full Text Available Visual analysis of integrated circuit layer requires raster image vectorization stage to extract layer topology data to CAD tools. In this paper vectorization problems of raster IC layer images are presented. Various line extraction from raster images algorithms and their properties are discussed. Optimal raster image vectorization method was developed which allows utilization of common vectorization algorithms to achieve the best possible extracted vector data match with perfect manual vectorization results. To develop the optimal method, vectorized data quality dependence on initial raster image skeleton filter selection was assessed.Article in Lithuanian


    Directory of Open Access Journals (Sweden)

    W. Lu


    Full Text Available In order to improve the stability and rapidity of synthetic aperture radar (SAR images matching, an effective method was presented. Firstly, the adaptive smoothing filtering was employed for image denoising in image processing based on Wallis filtering to avoid the follow-up noise is amplified. Secondly, feature points were extracted by a simplified SIFT algorithm. Finally, the exact matching of the images was achieved with these points. Compared with the existing methods, it not only maintains the richness of features, but a-lso reduces the noise of the image. The simulation results show that the proposed algorithm can achieve better matching effect.

  5. A method for fast automated microscope image stitching. (United States)

    Yang, Fan; Deng, Zhen-Sheng; Fan, Qiu-Hong


    Image stitching is an important technology to produce a panorama or larger image by combining several images with overlapped areas. In many biomedical researches, image stitching is highly desirable to acquire a panoramic image which represents large areas of certain structures or whole sections, while retaining microscopic resolution. In this study, we develop a fast normal light microscope image stitching algorithm based on feature extraction. At first, an algorithm of scale-space reconstruction of speeded-up robust features (SURF) was proposed to extract features from the images to be stitched with a short time and higher repeatability. Then, the histogram equalization (HE) method was employed to preprocess the images to enhance their contrast for extracting more features. Thirdly, the rough overlapping zones of the images preprocessed were calculated by phase correlation, and the improved SURF was used to extract the image features in the rough overlapping areas. Fourthly, the features were corresponded by matching algorithm and the transformation parameters were estimated, then the images were blended seamlessly. Finally, this procedure was applied to stitch normal light microscope images to verify its validity. Our experimental results demonstrate that the improved SURF algorithm is very robust to viewpoint, illumination, blur, rotation and zoom of the images and our method is able to stitch microscope images automatically with high precision and high speed. Also, the method proposed in this paper is applicable to registration and stitching of common images as well as stitching the microscope images in the field of virtual microscope for the purpose of observing, exchanging, saving, and establishing a database of microscope images. Copyright © 2013 Elsevier Ltd. All rights reserved.

  6. On two methods of statistical image analysis

    NARCIS (Netherlands)

    Missimer, J; Knorr, U; Maguire, RP; Herzog, H; Seitz, RJ; Tellman, L; Leenders, K.L.


    The computerized brain atlas (CBA) and statistical parametric mapping (SPM) are two procedures for voxel-based statistical evaluation of PET activation studies. Each includes spatial standardization of image volumes, computation of a statistic, and evaluation of its significance. In addition,

  7. Ultrasound Imaging Methods for Breast Cancer Detection

    NARCIS (Netherlands)

    Ozmen, N.


    The main focus of this thesis is on modeling acoustic wavefield propagation and implementing imaging algorithms for breast cancer detection using ultrasound. As a starting point, we use an integral equation formulation, which can be used to solve both the forward and inverse problems. This thesis

  8. Human body region enhancement method based on Kinect infrared imaging (United States)

    Yang, Lei; Fan, Yubo; Song, Xiaowei; Cai, Wenjing


    To effectively improve the low contrast of human body region in the infrared images, a combing method of several enhancement methods is utilized to enhance the human body region. Firstly, for the infrared images acquired by Kinect, in order to improve the overall contrast of the infrared images, an Optimal Contrast-Tone Mapping (OCTM) method with multi-iterations is applied to balance the contrast of low-luminosity infrared images. Secondly, to enhance the human body region better, a Level Set algorithm is employed to improve the contour edges of human body region. Finally, to further improve the human body region in infrared images, Laplacian Pyramid decomposition is adopted to enhance the contour-improved human body region. Meanwhile, the background area without human body region is processed by bilateral filtering to improve the overall effect. With theoretical analysis and experimental verification, the results show that the proposed method could effectively enhance the human body region of such infrared images.

  9. Imaging systems and methods for obtaining and using biometric information (United States)

    McMakin, Douglas L [Richland, WA; Kennedy, Mike O [Richland, WA


    Disclosed herein are exemplary embodiments of imaging systems and methods of using such systems. In one exemplary embodiment, one or more direct images of the body of a clothed subject are received, and a motion signature is determined from the one or more images. In this embodiment, the one or more images show movement of the body of the subject over time, and the motion signature is associated with the movement of the subject's body. In certain implementations, the subject can be identified based at least in part on the motion signature. Imaging systems for performing any of the disclosed methods are also disclosed herein. Furthermore, the disclosed imaging, rendering, and analysis methods can be implemented, at least in part, as one or more computer-readable media comprising computer-executable instructions for causing a computer to perform the respective methods.

  10. Imaging method of brain surface anatomy structures using conventional T2-weighted MR images

    International Nuclear Information System (INIS)

    Hatanaka, Masahiko; Machida, Yoshio; Yoshida, Tadatoki; Katada, Kazuhiro.


    As a non-invasive technique for visualizing the brain surface structure by MRI, surface anatomy scanning (SAS) and the multislice SAS methods have been developed. Both techniques require additional MRI scanning to obtain images for the brain surface. In this paper, we report an alternative method to obtain the brain surface image using conventional T2-weighted multislice images without any additional scanning. The power calculation of the image pixel values, which is incorporated in the routine processing, has been applied in order to enhance the cerebrospinal fluid (CSF) contrast. We think that this method is one of practical approaches for imaging the surface anatomy of the brain. (author)

  11. Brain diagnosis with imaging methods: Psychical changes made visible

    International Nuclear Information System (INIS)



    The First International Symposium on Imaging Methods in Psychiatry, held in May 1988 in Wuerzburg, very impressively has shown that imaging methods are on advance not only in medical diagnostics, but also in psychiatric diagnostics, where they already proved to be a valuable tool. (orig./MG) [de

  12. Radiopharmaceutical chelates and method of external imaging

    International Nuclear Information System (INIS)


    The preparation of the following chemicals is described: chelates of technetium-99m, cobalt-57, gallium-67, gallium-68, indium-111 or indium-113m and a substituted iminodiacetic acid or an 8-hydroxyquinoline useful as a radiopharmaceutical external imaging agent. The compounds described are suitable for intravenous injection, have an excellent in vivo stability and are good organ seekers. Tin(II) choride or other tin(II) compounds are used as chelating agents

  13. Soft Shadow Removal and Image Evaluation Methods


    Gryka, M.


    High-level image manipulation techniques are in increasing demand as they allow users to intuitively edit photographs to achieve desired effects quickly. As opposed to low-level manipulations, which provide complete freedom, but also require specialized skills and significant effort, high-level editing operations, such as removing objects (inpainting), relighting and material editing, need to respect semantic constraints. As such they shift the burden from the user to the algorithm to only al...

  14. An FPGA-based heterogeneous image fusion system design method (United States)

    Song, Le; Lin, Yu-chi; Chen, Yan-hua; Zhao, Mei-rong


    Taking the advantages of FPGA's low cost and compact structure, an FPGA-based heterogeneous image fusion platform is established in this study. Altera's Cyclone IV series FPGA is adopted as the core processor of the platform, and the visible light CCD camera and infrared thermal imager are used as the image-capturing device in order to obtain dualchannel heterogeneous video images. Tailor-made image fusion algorithms such as gray-scale weighted averaging, maximum selection and minimum selection methods are analyzed and compared. VHDL language and the synchronous design method are utilized to perform a reliable RTL-level description. Altera's Quartus II 9.0 software is applied to simulate and implement the algorithm modules. The contrast experiments of various fusion algorithms show that, preferably image quality of the heterogeneous image fusion can be obtained on top of the proposed system. The applied range of the different fusion algorithms is also discussed.

  15. Level set method for image segmentation based on moment competition (United States)

    Min, Hai; Wang, Xiao-Feng; Huang, De-Shuang; Jin, Jing; Wang, Hong-Zhi; Li, Hai


    We propose a level set method for image segmentation which introduces the moment competition and weakly supervised information into the energy functional construction. Different from the region-based level set methods which use force competition, the moment competition is adopted to drive the contour evolution. Here, a so-called three-point labeling scheme is proposed to manually label three independent points (weakly supervised information) on the image. Then the intensity differences between the three points and the unlabeled pixels are used to construct the force arms for each image pixel. The corresponding force is generated from the global statistical information of a region-based method and weighted by the force arm. As a result, the moment can be constructed and incorporated into the energy functional to drive the evolving contour to approach the object boundary. In our method, the force arm can take full advantage of the three-point labeling scheme to constrain the moment competition. Additionally, the global statistical information and weakly supervised information are successfully integrated, which makes the proposed method more robust than traditional methods for initial contour placement and parameter setting. Experimental results with performance analysis also show the superiority of the proposed method on segmenting different types of complicated images, such as noisy images, three-phase images, images with intensity inhomogeneity, and texture images.

  16. Analysis of live cell images: Methods, tools and opportunities. (United States)

    Nketia, Thomas A; Sailem, Heba; Rohde, Gustavo; Machiraju, Raghu; Rittscher, Jens


    Advances in optical microscopy, biosensors and cell culturing technologies have transformed live cell imaging. Thanks to these advances live cell imaging plays an increasingly important role in basic biology research as well as at all stages of drug development. Image analysis methods are needed to extract quantitative information from these vast and complex data sets. The aim of this review is to provide an overview of available image analysis methods for live cell imaging, in particular required preprocessing image segmentation, cell tracking and data visualisation methods. The potential opportunities recent advances in machine learning, especially deep learning, and computer vision provide are being discussed. This review includes overview of the different available software packages and toolkits. Copyright © 2017. Published by Elsevier Inc.

  17. Ultra-sensitive and selective Hg{sup 2+} detection based on fluorescent carbon dots

    Energy Technology Data Exchange (ETDEWEB)

    Liu, Ruihua; Li, Haitao; Kong, Weiqian; Liu, Juan [Institute of Functional Nano and Soft Materials (FUNSOM) and Jiangsu Key Laboratory for Carbon-Based Functional Materials and Devices, Soochow University, Suzhou 215123 (China); Liu, Yang, E-mail: [Institute of Functional Nano and Soft Materials (FUNSOM) and Jiangsu Key Laboratory for Carbon-Based Functional Materials and Devices, Soochow University, Suzhou 215123 (China); Tong, Cuiyan, E-mail: [Chemisty Department, Northeast Normal University, Changchun 130024 (China); Zhang, Xing [Institute of Functional Nano and Soft Materials (FUNSOM) and Jiangsu Key Laboratory for Carbon-Based Functional Materials and Devices, Soochow University, Suzhou 215123 (China); Kang, Zhenhui, E-mail: [Institute of Functional Nano and Soft Materials (FUNSOM) and Jiangsu Key Laboratory for Carbon-Based Functional Materials and Devices, Soochow University, Suzhou 215123 (China)


    Graphical abstract: Fluorescent carbon dots were efficiently synthesized by one-step sodium hydroxide-assisted reflux method from PEG and demonstrated to show high selectivity toward Hg2+ ions detection. - Highlights: • FCDs were synthesized by one-step sodium hydroxide-assisted reflux method from PEG. • The FCDs emit blue photoluminescence and have upconversion fluorescent property. • The FCDs show ultra-sensitive detective ability for Hg{sup 2+} ions. - Abstract: Fluorescent carbon dots (FCDs) were efficiently synthesized by one-step sodium hydroxide-assisted reflux method from poly(ethylene glycol) (PEG). The obtained FCDs exhibit excellent water-solubility and high stability. Under the UV irradiation, the FCDs could emit bright blue photoluminescence, and also they were found to show excellent up-conversion fluorescence. It was further demonstrated that such FCDs can serve as effective fluorescent sensing platform for Hg{sup 2+} ions detection with ultra-sensitivity and selectivity. The sensing system achieved a limit of detection as low as 1 fM, which is much lower than all the previous reported sensing systems for Hg{sup 2+} ions detection. This FCDs sensing system has been successfully applied for the analysis of Hg{sup 2+} ions in water samples from river, lake, and tap water, showing good practical feasibility.

  18. An Ultrasensitive Mechanism Regulates Influenza Virus-Induced Inflammation.

    Directory of Open Access Journals (Sweden)

    Jason E Shoemaker


    Full Text Available Influenza viruses present major challenges to public health, evident by the 2009 influenza pandemic. Highly pathogenic influenza virus infections generally coincide with early, high levels of inflammatory cytokines that some studies have suggested may be regulated in a strain-dependent manner. However, a comprehensive characterization of the complex dynamics of the inflammatory response induced by virulent influenza strains is lacking. Here, we applied gene co-expression and nonlinear regression analysis to time-course, microarray data developed from influenza-infected mouse lung to create mathematical models of the host inflammatory response. We found that the dynamics of inflammation-associated gene expression are regulated by an ultrasensitive-like mechanism in which low levels of virus induce minimal gene expression but expression is strongly induced once a threshold virus titer is exceeded. Cytokine assays confirmed that the production of several key inflammatory cytokines, such as interleukin 6 and monocyte chemotactic protein 1, exhibit ultrasensitive behavior. A systematic exploration of the pathways regulating the inflammatory-associated gene response suggests that the molecular origins of this ultrasensitive response mechanism lie within the branch of the Toll-like receptor pathway that regulates STAT1 phosphorylation. This study provides the first evidence of an ultrasensitive mechanism regulating influenza virus-induced inflammation in whole lungs and provides insight into how different virus strains can induce distinct temporal inflammation response profiles. The approach developed here should facilitate the construction of gene regulatory models of other infectious diseases.

  19. Quantitative Analysis of Range Image Patches by NEB Method

    Directory of Open Access Journals (Sweden)

    Wang Wen


    Full Text Available In this paper we analyze sampled high dimensional data with the NEB method from a range image database. Select a large random sample of log-valued, high contrast, normalized, 8×8 range image patches from the Brown database. We make a density estimator and we establish 1-dimensional cell complexes from the range image patch data. We find topological properties of 8×8 range image patches, prove that there exist two types of subsets of 8×8 range image patches modelled as a circle.

  20. New Finger Biometric Method Using Near Infrared Imaging (United States)

    Lee, Eui Chul; Jung, Hyunwoo; Kim, Daeyeoul


    In this paper, we propose a new finger biometric method. Infrared finger images are first captured, and then feature extraction is performed using a modified Gaussian high-pass filter through binarization, local binary pattern (LBP), and local derivative pattern (LDP) methods. Infrared finger images include the multimodal features of finger veins and finger geometries. Instead of extracting each feature using different methods, the modified Gaussian high-pass filter is fully convolved. Therefore, the extracted binary patterns of finger images include the multimodal features of veins and finger geometries. Experimental results show that the proposed method has an error rate of 0.13%. PMID:22163741

  1. Image Classification Workflow Using Machine Learning Methods (United States)

    Christoffersen, M. S.; Roser, M.; Valadez-Vergara, R.; Fernández-Vega, J. A.; Pierce, S. A.; Arora, R.


    Recent increases in the availability and quality of remote sensing datasets have fueled an increasing number of scientifically significant discoveries based on land use classification and land use change analysis. However, much of the software made to work with remote sensing data products, specifically multispectral images, is commercial and often prohibitively expensive. The free to use solutions that are currently available come bundled up as small parts of much larger programs that are very susceptible to bugs and difficult to install and configure. What is needed is a compact, easy to use set of tools to perform land use analysis on multispectral images. To address this need, we have developed software using the Python programming language with the sole function of land use classification and land use change analysis. We chose Python to develop our software because it is relatively readable, has a large body of relevant third party libraries such as GDAL and Spectral Python, and is free to install and use on Windows, Linux, and Macintosh operating systems. In order to test our classification software, we performed a K-means unsupervised classification, Gaussian Maximum Likelihood supervised classification, and a Mahalanobis Distance based supervised classification. The images used for testing were three Landsat rasters of Austin, Texas with a spatial resolution of 60 meters for the years of 1984 and 1999, and 30 meters for the year 2015. The testing dataset was easily downloaded using the Earth Explorer application produced by the USGS. The software should be able to perform classification based on any set of multispectral rasters with little to no modification. Our software makes the ease of land use classification using commercial software available without an expensive license.


    Directory of Open Access Journals (Sweden)

    O. Akcay


    Full Text Available A successful image matching is essential to provide an automatic photogrammetric process accurately. Feature detection, extraction and matching algorithms have performed on the high resolution images perfectly. However, images of cameras, which are equipped with low-resolution thermal sensors are problematic with the current algorithms. In this paper, some digital image processing techniques were applied to the low-resolution images taken with Optris PI 450 382 x 288 pixel optical resolution lightweight thermal camera to increase extraction and matching performance. Image enhancement methods that adjust low quality digital thermal images, were used to produce more suitable images for detection and extraction. Three main digital image process techniques: histogram equalization, high pass and low pass filters were considered to increase the signal-to-noise ratio, sharpen image, remove noise, respectively. Later on, the pre-processed images were evaluated using current image detection and feature extraction methods Maximally Stable Extremal Regions (MSER and Speeded Up Robust Features (SURF algorithms. Obtained results showed that some enhancement methods increased number of extracted features and decreased blunder errors during image matching. Consequently, the effects of different pre-process techniques were compared in the paper.

  3. Multi-band Image Registration Method Based on Fourier Transform

    Institute of Scientific and Technical Information of China (English)

    庹红娅; 刘允才


    This paper presented a registration method based on Fourier transform for multi-band images which is involved in translation and small rotation. Although different band images differ a lot in the intensity and features,they contain certain common information which we can exploit. A model was given that the multi-band images have linear correlations under the least-square sense. It is proved that the coefficients have no effect on the registration progress if two images have linear correlations. Finally, the steps of the registration method were proposed. The experiments show that the model is reasonable and the results are satisfying.

  4. Image Mosaic Method Based on SIFT Features of Line Segment

    Directory of Open Access Journals (Sweden)

    Jun Zhu


    Full Text Available This paper proposes a novel image mosaic method based on SIFT (Scale Invariant Feature Transform feature of line segment, aiming to resolve incident scaling, rotation, changes in lighting condition, and so on between two images in the panoramic image mosaic process. This method firstly uses Harris corner detection operator to detect key points. Secondly, it constructs directed line segments, describes them with SIFT feature, and matches those directed segments to acquire rough point matching. Finally, Ransac method is used to eliminate wrong pairs in order to accomplish image mosaic. The results from experiment based on four pairs of images show that our method has strong robustness for resolution, lighting, rotation, and scaling.

  5. Development of motion image prediction method using principal component analysis

    International Nuclear Information System (INIS)

    Chhatkuli, Ritu Bhusal; Demachi, Kazuyuki; Kawai, Masaki; Sakakibara, Hiroshi; Kamiaka, Kazuma


    Respiratory motion can induce the limit in the accuracy of area irradiated during lung cancer radiation therapy. Many methods have been introduced to minimize the impact of healthy tissue irradiation due to the lung tumor motion. The purpose of this research is to develop an algorithm for the improvement of image guided radiation therapy by the prediction of motion images. We predict the motion images by using principal component analysis (PCA) and multi-channel singular spectral analysis (MSSA) method. The images/movies were successfully predicted and verified using the developed algorithm. With the proposed prediction method it is possible to forecast the tumor images over the next breathing period. The implementation of this method in real time is believed to be significant for higher level of tumor tracking including the detection of sudden abdominal changes during radiation therapy. (author)

  6. Iterative methods for dose reduction and image enhancement in tomography (United States)

    Miao, Jianwei; Fahimian, Benjamin Pooya


    A system and method for creating a three dimensional cross sectional image of an object by the reconstruction of its projections that have been iteratively refined through modification in object space and Fourier space is disclosed. The invention provides systems and methods for use with any tomographic imaging system that reconstructs an object from its projections. In one embodiment, the invention presents a method to eliminate interpolations present in conventional tomography. The method has been experimentally shown to provide higher resolution and improved image quality parameters over existing approaches. A primary benefit of the method is radiation dose reduction since the invention can produce an image of a desired quality with a fewer number projections than seen with conventional methods.

  7. Thresholding methods for PET imaging: A review

    International Nuclear Information System (INIS)

    Dewalle-Vignion, A.S.; Betrouni, N.; Huglo, D.; Vermandel, M.; Dewalle-Vignion, A.S.; Hossein-Foucher, C.; Huglo, D.; Vermandel, M.; Dewalle-Vignion, A.S.; Hossein-Foucher, C.; Huglo, D.; Vermandel, M.; El Abiad, A.


    This work deals with positron emission tomography segmentation methods for tumor volume determination. We propose a state of art techniques based on fixed or adaptive threshold. Methods found in literature are analysed with an objective point of view on their methodology, advantages and limitations. Finally, a comparative study is presented. (authors)

  8. Generalized Row-Action Methods for Tomographic Imaging

    DEFF Research Database (Denmark)

    Andersen, Martin Skovgaard; Hansen, Per Christian


    Row-action methods play an important role in tomographic image reconstruction. Many such methods can be viewed as incremental gradient methods for minimizing a sum of a large number of convex functions, and despite their relatively poor global rate of convergence, these methods often exhibit fast...... initial convergence which is desirable in applications where a low-accuracy solution is acceptable. In this paper, we propose relaxed variants of a class of incremental proximal gradient methods, and these variants generalize many existing row-action methods for tomographic imaging. Moreover, they allow...

  9. Image based method for aberration measurement of lithographic tools (United States)

    Xu, Shuang; Tao, Bo; Guo, Yongxing; Li, Gongfa


    Information of lens aberration of lithographic tools is important as it directly affects the intensity distribution in the image plane. Zernike polynomials are commonly used for a mathematical description of lens aberrations. Due to the advantage of lower cost and easier implementation of tools, image based measurement techniques have been widely used. Lithographic tools are typically partially coherent systems that can be described by a bilinear model, which entails time consuming calculations and does not lend a simple and intuitive relationship between lens aberrations and the resulted images. Previous methods for retrieving lens aberrations in such partially coherent systems involve through-focus image measurements and time-consuming iterative algorithms. In this work, we propose a method for aberration measurement in lithographic tools, which only requires measuring two images of intensity distribution. Two linear formulations are derived in matrix forms that directly relate the measured images to the unknown Zernike coefficients. Consequently, an efficient non-iterative solution is obtained.

  10. A NDVI assisted remote sensing image adaptive scale segmentation method (United States)

    Zhang, Hong; Shen, Jinxiang; Ma, Yanmei


    Multiscale segmentation of images can effectively form boundaries of different objects with different scales. However, for the remote sensing image which widely coverage with complicated ground objects, the number of suitable segmentation scales, and each of the scale size is still difficult to be accurately determined, which severely restricts the rapid information extraction of the remote sensing image. A great deal of experiments showed that the normalized difference vegetation index (NDVI) can effectively express the spectral characteristics of a variety of ground objects in remote sensing images. This paper presents a method using NDVI assisted adaptive segmentation of remote sensing images, which segment the local area by using NDVI similarity threshold to iteratively select segmentation scales. According to the different regions which consist of different targets, different segmentation scale boundaries could be created. The experimental results showed that the adaptive segmentation method based on NDVI can effectively create the objects boundaries for different ground objects of remote sensing images.

  11. An efficient method for facial component detection in thermal images (United States)

    Paul, Michael; Blanik, Nikolai; Blazek, Vladimir; Leonhardt, Steffen


    A method to detect certain regions in thermal images of human faces is presented. In this approach, the following steps are necessary to locate the periorbital and the nose regions: First, the face is segmented from the background by thresholding and morphological filtering. Subsequently, a search region within the face, around its center of mass, is evaluated. Automatically computed temperature thresholds are used per subject and image or image sequence to generate binary images, in which the periorbital regions are located by integral projections. Then, the located positions are used to approximate the nose position. It is possible to track features in the located regions. Therefore, these regions are interesting for different applications like human-machine interaction, biometrics and biomedical imaging. The method is easy to implement and does not rely on any training images or templates. Furthermore, the approach saves processing resources due to simple computations and restricted search regions.

  12. New diffusion imaging method with a single acquisition sequence

    International Nuclear Information System (INIS)

    Melki, Ph.S.; Bittoun, J.; Lefevre, J.E.


    The apparent diffusion coefficient (ADC) is related to the molecular diffusion coefficient and to physiologic information: microcirculation in the capillary network, incoherent slow flow, and restricted diffusion. The authors present a new MR imaging sequence that yields computed ADC images in only one acquisition of 9-minutes with a 1.5-T imager (GE Signa). Compared to the previous method, this sequence is at least two times faster and thus can be used as a routine examination to supplement T1-, T2-, and density-weighted images. The method was assessed by measurement of the molecular diffusion in liquids, and the first clinical images obtained in neurologic diseases demonstrate its efficiency for clinical investigation. The possibility of separately imaging diffusion and perfusion is supported by an algorithm

  13. Profiling pleural effusion cells by a diffraction imaging method (United States)

    Al-Qaysi, Safaa; Hong, Heng; Wen, Yuhua; Lu, Jun Q.; Feng, Yuanming; Hu, Xin-Hua


    Assay of cells in pleural effusion (PE) is an important means of disease diagnosis. Conventional cytology of effusion samples, however, has low sensitivity and depends heavily on the expertise of cytopathologists. We applied a polarization diffraction imaging flow cytometry method on effusion cells to investigate their features. Diffraction imaging of the PE cell samples has been performed on 6000 to 12000 cells for each effusion cell sample of three patients. After prescreening to remove images by cellular debris and aggregated non-cellular particles, the image textures were extracted with a gray level co-occurrence matrix (GLCM) algorithm. The distribution of the imaged cells in the GLCM parameters space was analyzed by a Gaussian Mixture Model (GMM) to determine the number of clusters among the effusion cells. These results yield insight on textural features of diffraction images and related cellular morphology in effusion samples and can be used toward the development of a label-free method for effusion cells assay.

  14. Imaging methods for detection of infectious foci

    International Nuclear Information System (INIS)

    Couret, I.; Rossi, M.; Weinemann, P.; Moretti, J.L.


    Several tracers can be used for imaging infection. None is a worthwhile agent for all infectious foci, but each one has preferential applications, depending on its uptake mechanism by the infectious and/or inflammatory focus. Autologous leucocytes labeled in vitro with indium-111 (In-111) or with technetium-99-hexamethylpropyleneamine oxime (Tc-99m HMPAO) were applied with success in the detection of peripheral bone infection, focal vascular graft infection and inflammatory bowel disease. Labeling with In-111 is of interest in chronic bone infection, while labeling with Tc-99m HMPAO gets the advantage of a better dosimetry and imaging. The interest of in vivo labeled leucocytes with a Tc-99m labeled monoclonal antigranulocyte antibody anti-NCA 95 (BW 250/183) was proved in the same principal type of infectious foci than in vitro labeled leucocytes. Sites of chronic infection in the spine and the pelvis, whether active or healed, appear as photopenic defects on both in vitro labeled leucocytes and Tc-99m monoclonal antigranulocyte antibody (BW 250/183) scintigraphies. With gallium-67 results showed a high sensitivity with a low specificity. This tracer demonstrated good performance to delineate foci of infectious spondylitis. In-111 and Tc-99m labeled polyclonal human immunoglobulin (HIG) was applied with success in the assessment of various infectious foci, particularly in chronic sepsis. As labeled leucocytes, labeled HIG showed cold defects in infectious sepsis of the spine. Research in nuclear medicine is very active in the development of more specific tracers of infection, mainly involved in Tc-99m or In-111 labeled chemotactic peptides, antigranulocyte antibody fragments, antibiotic derivatives and interleukins. (authors). 70 refs

  15. The best printing methods to print satellite images


    G.A. Yousif; R.Sh. Mohamed


    Printing systems operate in general as a system of color its color scale is limited as compared with the system color satellite images. Satellite image is building from very small cell named pixel, which represents the picture element and the unity of color when the image is displayed on the screen, this unit becomes lesser in size and called screen point. This unit posseses different size and shape from the method of printing to another, depending on the output resolution, tools and material...

  16. Mathematical methods in time series analysis and digital image processing

    CERN Document Server

    Kurths, J; Maass, P; Timmer, J


    The aim of this volume is to bring together research directions in theoretical signal and imaging processing developed rather independently in electrical engineering, theoretical physics, mathematics and the computer sciences. In particular, mathematically justified algorithms and methods, the mathematical analysis of these algorithms, and methods as well as the investigation of connections between methods from time series analysis and image processing are reviewed. An interdisciplinary comparison of these methods, drawing upon common sets of test problems from medicine and geophysical/enviromental sciences, is also addressed. This volume coherently summarizes work carried out in the field of theoretical signal and image processing. It focuses on non-linear and non-parametric models for time series as well as on adaptive methods in image processing.

  17. Fingerprint image reconstruction for swipe sensor using Predictive Overlap Method

    Directory of Open Access Journals (Sweden)

    Mardiansyah Ahmad Zafrullah


    Full Text Available Swipe sensor is one of many biometric authentication sensor types that widely applied to embedded devices. The sensor produces an overlap on every pixel block of the image, so the picture requires a reconstruction process before heading to the feature extraction process. Conventional reconstruction methods require extensive computation, causing difficult to apply to embedded devices that have limited computing process. In this paper, image reconstruction is proposed using predictive overlap method, which determines the image block shift from the previous set of change data. The experiments were performed using 36 images generated by a swipe sensor with 128 x 8 pixels size of the area, where each image has an overlap in each block. The results reveal computation can increase up to 86.44% compared with conventional methods, with accuracy decreasing to 0.008% in average.

  18. Texture recognition of medical images with the ICM method

    International Nuclear Information System (INIS)

    Kinser, Jason M.; Wang Guisong


    The Integrated Cortical Model (ICM) is based upon several models of the mammalian visual cortex and produces pulse images over several iterations. These pulse images tend to isolate segments, edges, and textures that are inherent in the input image. To create a texture recognition engine the pulse spectrum of individual pixels are collected and used to develop a recognition library. Recognition is performed by comparing pulse spectra of unclassified regions of images with the known regions. Because signatures are smaller than images, signature-based computation is quite efficient and parasites can be recognized quickly. The precision of this method depends on the representative of signatures and classification. Our experiment results support the theoretical findings and show perspectives of practical applications of ICM-based method. The advantage of ICM method is using signatures to represent objects. ICM can extract the internal features of objects and represent them with signatures. Signature classification is critical for the precision of recognition

  19. Some selected quantitative methods of thermal image analysis in Matlab. (United States)

    Koprowski, Robert


    The paper presents a new algorithm based on some selected automatic quantitative methods for analysing thermal images. It shows the practical implementation of these image analysis methods in Matlab. It enables to perform fully automated and reproducible measurements of selected parameters in thermal images. The paper also shows two examples of the use of the proposed image analysis methods for the area of ​​the skin of a human foot and face. The full source code of the developed application is also provided as an attachment. The main window of the program during dynamic analysis of the foot thermal image. © 2016 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  20. Matrix-based image reconstruction methods for tomography

    International Nuclear Information System (INIS)

    Llacer, J.; Meng, J.D.


    Matrix methods of image reconstruction have not been used, in general, because of the large size of practical matrices, ill condition upon inversion and the success of Fourier-based techniques. An exception is the work that has been done at the Lawrence Berkeley Laboratory for imaging with accelerated radioactive ions. An extension of that work into more general imaging problems shows that, with a correct formulation of the problem, positron tomography with ring geometries results in well behaved matrices which can be used for image reconstruction with no distortion of the point response in the field of view and flexibility in the design of the instrument. Maximum Likelihood Estimator methods of reconstruction, which use the system matrices tailored to specific instruments and do not need matrix inversion, are shown to result in good preliminary images. A parallel processing computer structure based on multiple inexpensive microprocessors is proposed as a system to implement the matrix-MLE methods. 14 references, 7 figures

  1. Method and apparatus for improving the alignment of radiographic images

    International Nuclear Information System (INIS)

    Schuller, P.D.; Hatcher, D.C.; Caelli, T.M.; Eggert, F.M.; Yuzyk, J.


    This invention relates generally to the field of radiology, and has to do particularly with a method and apparatus for improving the alignment of radiographic images taken at different times of the same tissue structure, so that the images can be sequentially shown in aligned condition, whereby changes in the structure can be noted. (author). 10 figs

  2. Method for analysis of failure of material employing imaging

    Energy Technology Data Exchange (ETDEWEB)

    Vinegar, H.J.; Wellington, S.L.; de Waal, J.A.


    This patent describes a method for determining at least one preselected property of a sample of material employing an imaging apparatus. It comprises: imaging the sample during the application of known preselected forces to the sample, and determining density in the sample responsive to the preselected forces.

  3. An attenuation correction method for PET/CT images

    International Nuclear Information System (INIS)

    Ue, Hidenori; Yamazaki, Tomohiro; Haneishi, Hideaki


    In PET/CT systems, accurate attenuation correction can be achieved by creating an attenuation map from an X-ray CT image. On the other hand, respiratory-gated PET acquisition is an effective method for avoiding motion blurring of the thoracic and abdominal organs caused by respiratory motion. In PET/CT systems employing respiratory-gated PET, using an X-ray CT image acquired during breath-holding for attenuation correction may have a large effect on the voxel values, especially in regions with substantial respiratory motion. In this report, we propose an attenuation correction method in which, as the first step, a set of respiratory-gated PET images is reconstructed without attenuation correction, as the second step, the motion of each phase PET image from the PET image in the same phase as the CT acquisition timing is estimated by the previously proposed method, as the third step, the CT image corresponding to each respiratory phase is generated from the original CT image by deformation according to the motion vector maps, and as the final step, attenuation correction using these CT images and reconstruction are performed. The effectiveness of the proposed method was evaluated using 4D-NCAT phantoms, and good stability of the voxel values near the diaphragm was observed. (author)

  4. Method and Apparatus for Computed Imaging Backscatter Radiography (United States)

    Shedlock, Daniel (Inventor); Meng, Christopher (Inventor); Sabri, Nissia (Inventor); Dugan, Edward T. (Inventor); Jacobs, Alan M. (Inventor)


    Systems and methods of x-ray backscatter radiography are provided. A single-sided, non-destructive imaging technique utilizing x-ray radiation to image subsurface features is disclosed, capable of scanning a region using a fan beam aperture and gathering data using rotational motion.

  5. Image segmentation with a finite element method

    DEFF Research Database (Denmark)

    Bourdin, Blaise


    regularization results, make possible to imagine a finite element resolution method.In a first time, the Mumford-Shah functional is introduced and some existing results are quoted. Then, a discrete formulation for the Mumford-Shah problem is proposed and its $\\Gamma$-convergence is proved. Finally, some...

  6. Method and apparatus for producing tomographic images

    International Nuclear Information System (INIS)

    Annis, M.


    A device useful in producing a tomographic image of a selected slice of an object to be examined is described comprising: a source of penetrating radiation, sweep means for forming energy from the source into a pencil beam and repeatedly sweeping the pencil beam over a line in space to define a sweep plane, first means for supporting an object to be examined so that the pencil beam intersections the object along a path passing through the object and the selected slice, line collimating means for filtering radiation scattered by the object, the line collimating means having a field of view which intersects and sweep plane in a bounded line so that the line collimating means passes only radiation scattered by elementary volumes of the object lying along the bounded line, and line collimating means including a plurality of channels such substantially planar in form to collectively define the field of view, the channels oriented so that pencil beam sweeps along the bounded line as a function of time, and radiation detector means responsive to radiation passed by the line collimating means

  7. Method for imaging pulmonary arterial hypoplasia

    International Nuclear Information System (INIS)

    Triantafillou, M.


    Full text: Pulmonary hypoplasia represents an incomplete development of the lung, resulting in the reduction of distended lung volume. This is associated with small or absent number of airway divisions, alveoli, arteries and veins. Unilateral pulmonary Hypoplasia is often asymptomatic and may be demonstrated as a hypodense lung on a chest X-ray. Computer Tomography (CT) scanning would show anatomical detail and proximal vessels. Magnetic Resonance Imaging (MRI) will show no more detail than which the CT scan has already demonstrated. It is, also, difficult to visualise collateral vessels from systemic and/or bronchial vessels on both these modalities. Pulmonary Angiography would give the definitive answer, but it is time consuming and has significant risks associated with the procedure. There are high costs associated with these modalities. Nuclear Medicine Ventilation/Perfusion (V/Q) scan performed on these patients would demonstrate diminished ventilation due to reduced lung volume and absence of perfusion to the hypoplastic lung. To date, we have performed V/Q lung scan on two children in our department. Both cases demonstrate diminished ventilation with no perfusion to the hypoplastic lung. Though the gold standard is Pulmonary Angiography, V/Q scanning is cost effective, less time consuming and a non invasive procedure that can be performed as an outpatient. It is accurate as it demonstrates absent lung perfusion, confirming the patient has pulmonary arterial hypoplasia. Copyright (2000) The Australian and New Zealand Society of Nuclear Medicine Inc


    DEFF Research Database (Denmark)

    Jensen, Andreas Kryger; Savarimuthu, Thiusius Rajeeth; Sørensen, Anders Stengaard


    We investigate methods for improving the visual quality of in vivo images of blood vessels in the human forearm. Using a near-infrared light source and a dual CCD chip camera system capable of capturing images at visual and nearinfrared spectra, we evaluate three fusion methods in terms...... of their capability of enhancing the blood vessels while preserving the spectral signature of the original color image. Furthermore, we investigate a possibility of removing hair in the images using a fusion rule based on the "a trous" stationary wavelet decomposition. The method with the best overall performance...... with both speed and quality in mind is the Intensity Injection method. Using the developed system and the methods presented in this article, it is possible to create images of high visual quality with highly emphasized blood vessels....

  9. Image quality enhancement in low-light-level ghost imaging using modified compressive sensing method (United States)

    Shi, Xiaohui; Huang, Xianwei; Nan, Suqin; Li, Hengxing; Bai, Yanfeng; Fu, Xiquan


    Detector noise has a significantly negative impact on ghost imaging at low light levels, especially for existing recovery algorithm. Based on the characteristics of the additive detector noise, a method named modified compressive sensing ghost imaging is proposed to reduce the background imposed by the randomly distributed detector noise at signal path. Experimental results show that, with an appropriate choice of threshold value, modified compressive sensing ghost imaging algorithm can dramatically enhance the contrast-to-noise ratio of the object reconstruction significantly compared with traditional ghost imaging and compressive sensing ghost imaging methods. The relationship between the contrast-to-noise ratio of the reconstruction image and the intensity ratio (namely, the average signal intensity to average noise intensity ratio) for the three reconstruction algorithms are also discussed. This noise suppression imaging technique will have great applications in remote-sensing and security areas.

  10. SERS and fluorescence-based ultrasensitive detection of mercury in water. (United States)

    Makam, Pandeeswar; Shilpa, Rohilla; Kandjani, Ahmad Esmaielzadeh; Periasamy, Selvakannan R; Sabri, Ylias Mohammad; Madhu, Chilakapati; Bhargava, Suresh Kumar; Govindaraju, Thimmaiah


    The development of reliable and ultrasensitive detection marker for mercury ions (Hg 2+ ) in drinking water is of great interest for toxicology assessment, environmental protection and human health. Although many Hg 2+ detection methods have been developed, only few offer sensitivities below 1pM. Herein, we describe a simple histidine (H) conjugated perylene diimide (PDI) bolaamphiphile (HPH) as a dual-responsive optical marker to develop highly selective and sensitive probe as visible (sol-to-gel transformation), fluorescence and SERS-based Hg 2+ sensor platform in the water. Remarkably, HPH as a SERS marker supported on Au deposited monodispersed nanospheres monolayers (Au-MNM) of polystyrene offers an unprecedented selectivity and the best ever reported detection limit (LOD) of 60 attomolar (aM, 0.01 parts-per-quadrillion (ppq)) for Hg 2+ in water. This is ten orders of magnitude lower than the United States Environmental Protection Agency (USEPA) tolerance limit of Hg 2+ in drinking water (10nM, 2 ppb). This simple and effective design principle of host-guest interactions driven fluorescence and SERS-based detection may inspire the future molecular engineering strategies for the development of ultrasensitive toxic analyte sensor platforms. Copyright © 2017 Elsevier B.V. All rights reserved.

  11. Dual-signal amplification strategy for ultrasensitive chemiluminescence detection of PDGF-BB in capillary electrophoresis. (United States)

    Cao, Jun-Tao; Wang, Hui; Ren, Shu-Wei; Chen, Yong-Hong; Liu, Yan-Ming


    Many efforts have been made toward the achievement of high sensitivity in capillary electrophoresis coupled with chemiluminescence detection (CE-CL). This work describes a novel dual-signal amplification strategy for highly specific and ultrasensitive CL detection of human platelet-derived growth factor-BB (PDGF-BB) using both aptamer and horseradish peroxidase (HRP) modified gold nanoparticles (HRP-AuNPs-aptamer) as nanoprobes in CE. Both AuNPs and HRP in the nanoprobes could amplify the CL signals in the luminol-H2 O2 CL system, owing to the excellent catalytic behavior of AuNPs and HRP in the CL system. Meanwhile, the high affinity of aptamer modified on the AuNPs allows detection with high specificity. As proof-of-concept, the proposed method was employed to quantify the concentration of PDGF-BB from 0.50 to 250 fm with a detection limit of 0.21 fm. The applicability of the assay was further demonstrated in the analysis of PDGF-BB in human serum samples with acceptable accuracy and reliability. The result of this study exhibits distinct advantages, such as high sensitivity, good specificity, simplicity, and very small sample consumption. The good performances of the proposed strategy provide a powerful avenue for ultrasensitive detection of rare proteins in biological sample, showing great promise in biochemical analysis. Copyright © 2015 John Wiley & Sons, Ltd.

  12. Ultrasensitive leak detection during ultrahigh vacuum evacuation by quadrupole mass spectrometer

    International Nuclear Information System (INIS)

    Chen Xu; Huang Tianbin; Wang Ligong; Jin Qiji; Cha Liangzhen


    One must do ultrasensitive leak detection during ultrahigh-vacuum (UHV) evacuation, especially just before the device is sealed off from the vacuum system, to guarantee the longevity of the sealed high-vacuum or even UHV devices with small volume. A quadrupole mass spectrometer (QMS) with an UHV evacuation system can be used under accumulation mode to do the testing. Possible accumulate modes, as well as their advantages and shortcomings, are studied experimentally and discussed in this paper. We found that the opening action of the metal valve during accumulation mode always severely affects the height of the peak indicated by QMS and causes considerable errors. If we determine the leak rate by the peak area instead of the peak height, the situation is much improved. This method has proven quite useful in ensuring the tightness quality for complex sealed UHV devices with small volumes. Ultrasensitive leak detection has been carried out for such real evacuating devices, and a leak rate of 2x10 -14 Pa·m 3 /s was detected, which is far lower than its dynamic mode and the detection limit of the current advanced commercial leak detectors

  13. A method of fast mosaic for massive UAV images (United States)

    Xiang, Ren; Sun, Min; Jiang, Cheng; Liu, Lei; Zheng, Hui; Li, Xiaodong


    With the development of UAV technology, UAVs are used widely in multiple fields such as agriculture, forest protection, mineral exploration, natural disaster management and surveillances of public security events. In contrast of traditional manned aerial remote sensing platforms, UAVs are cheaper and more flexible to use. So users can obtain massive image data with UAVs, but this requires a lot of time to process the image data, for example, Pix4UAV need approximately 10 hours to process 1000 images in a high performance PC. But disaster management and many other fields require quick respond which is hard to realize with massive image data. Aiming at improving the disadvantage of high time consumption and manual interaction, in this article a solution of fast UAV image stitching is raised. GPS and POS data are used to pre-process the original images from UAV, belts and relation between belts and images are recognized automatically by the program, in the same time useless images are picked out. This can boost the progress of finding match points between images. Levenberg-Marquard algorithm is improved so that parallel computing can be applied to shorten the time of global optimization notably. Besides traditional mosaic result, it can also generate superoverlay result for Google Earth, which can provide a fast and easy way to show the result data. In order to verify the feasibility of this method, a fast mosaic system of massive UAV images is developed, which is fully automated and no manual interaction is needed after original images and GPS data are provided. A test using 800 images of Kelan River in Xinjiang Province shows that this system can reduce 35%-50% time consumption in contrast of traditional methods, and increases respond speed of UAV image processing rapidly.

  14. Hiding a Covert Digital Image by Assembling the RSA Encryption Method and the Binary Encoding Method


    Kuang Tsan Lin; Sheng Lih Yeh


    The Rivest-Shamir-Adleman (RSA) encryption method and the binary encoding method are assembled to form a hybrid hiding method to hide a covert digital image into a dot-matrix holographic image. First, the RSA encryption method is used to transform the covert image to form a RSA encryption data string. Then, all the elements of the RSA encryption data string are transferred into binary data. Finally, the binary data are encoded into the dot-matrix holographic image. The pixels of the dot-matri...

  15. Method and algorithm for image processing (United States)

    He, George G.; Moon, Brain D.


    The present invention is a modified Radon transform. It is similar to the traditional Radon transform for the extraction of line parameters and similar to traditional slant stack for the intensity summation of pixels away from a given pixel, for example ray paths that spans 360 degree at a given grid in the time and offset domain. However, the present invention differs from these methods in that the intensity and direction of a composite intensity for each pixel are maintained separately instead of combined after the transformation. An advantage of this approach is elimination of the work required to extract the line parameters in the transformed domain. The advantage of the modified Radon Transform method is amplified when many lines are present in the imagery or when the lines are just short segments which both occur in actual imagery.

  16. A Method for Improving the Progressive Image Coding Algorithms

    Directory of Open Access Journals (Sweden)

    Ovidiu COSMA


    Full Text Available This article presents a method for increasing the performance of the progressive coding algorithms for the subbands of images, by representing the coefficients with a code that reduces the truncation error.

  17. Development of digital image correlation method to analyse crack ...

    Indian Academy of Sciences (India)

    samples were performed to verify the performance of the digital image correlation method. ... development cannot be measured accurately. ..... Mendelson A 1983 Plasticity: Theory and application (USA: Krieger Publishing company Malabar,.

  18. Quantitative Methods for Molecular Diagnostic and Therapeutic Imaging


    Li, Quanzheng


    This theme issue provides an overview on the basic quantitative methods, an in-depth discussion on the cutting-edge quantitative analysis approaches as well as their applications for both static and dynamic molecular diagnostic and therapeutic imaging.

  19. Discrete gradient methods for solving variational image regularisation models

    International Nuclear Information System (INIS)

    Grimm, V; McLachlan, Robert I; McLaren, David I; Quispel, G R W; Schönlieb, C-B


    Discrete gradient methods are well-known methods of geometric numerical integration, which preserve the dissipation of gradient systems. In this paper we show that this property of discrete gradient methods can be interesting in the context of variational models for image processing, that is where the processed image is computed as a minimiser of an energy functional. Numerical schemes for computing minimisers of such energies are desired to inherit the dissipative property of the gradient system associated to the energy and consequently guarantee a monotonic decrease of the energy along iterations, avoiding situations in which more computational work might lead to less optimal solutions. Under appropriate smoothness assumptions on the energy functional we prove that discrete gradient methods guarantee a monotonic decrease of the energy towards stationary states, and we promote their use in image processing by exhibiting experiments with convex and non-convex variational models for image deblurring, denoising, and inpainting. (paper)

  20. Double Minimum Variance Beamforming Method to Enhance Photoacoustic Imaging


    Paridar, Roya; Mozaffarzadeh, Moein; Nasiriavanaki, Mohammadreza; Orooji, Mahdi


    One of the common algorithms used to reconstruct photoacoustic (PA) images is the non-adaptive Delay-and-Sum (DAS) beamformer. However, the quality of the reconstructed PA images obtained by DAS is not satisfying due to its high level of sidelobes and wide mainlobe. In contrast, adaptive beamformers, such as minimum variance (MV), result in an improved image compared to DAS. In this paper, a novel beamforming method, called Double MV (D-MV) is proposed to enhance the image quality compared to...

  1. Beam imaging sensor and method for using same

    Energy Technology Data Exchange (ETDEWEB)

    McAninch, Michael D.; Root, Jeffrey J.


    The present invention relates generally to the field of sensors for beam imaging and, in particular, to a new and useful beam imaging sensor for use in determining, for example, the power density distribution of a beam including, but not limited to, an electron beam or an ion beam. In one embodiment, the beam imaging sensor of the present invention comprises, among other items, a circumferential slit that is either circular, elliptical or polygonal in nature. In another embodiment, the beam imaging sensor of the present invention comprises, among other things, a discontinuous partially circumferential slit. Also disclosed is a method for using the various beams sensor embodiments of the present invention.

  2. Quantitative methods for the analysis of electron microscope images

    DEFF Research Database (Denmark)

    Skands, Peter Ulrik Vallø


    The topic of this thesis is an general introduction to quantitative methods for the analysis of digital microscope images. The images presented are primarily been acquired from Scanning Electron Microscopes (SEM) and interfermeter microscopes (IFM). The topic is approached though several examples...... foundation of the thesis fall in the areas of: 1) Mathematical Morphology; 2) Distance transforms and applications; and 3) Fractal geometry. Image analysis opens in general the possibility of a quantitative and statistical well founded measurement of digital microscope images. Herein lies also the conditions...

  3. System and method for image registration of multiple video streams (United States)

    Dillavou, Marcus W.; Shum, Phillip Corey; Guthrie, Baron L.; Shenai, Mahesh B.; Deaton, Drew Steven; May, Matthew Benton


    Provided herein are methods and systems for image registration from multiple sources. A method for image registration includes rendering a common field of interest that reflects a presence of a plurality of elements, wherein at least one of the elements is a remote element located remotely from another of the elements and updating the common field of interest such that the presence of the at least one of the elements is registered relative to another of the elements.


    Directory of Open Access Journals (Sweden)

    A. G. Chibunichev


    Full Text Available This article describes the method of constructing 3D models of objects (buildings, monuments based on digital images and a point cloud obtained by terrestrial laser scanner. The first step is the automated determination of exterior orientation parameters of digital image. We have to find the corresponding points of the image and point cloud to provide this operation. Before the corresponding points searching quasi image of point cloud is generated. After that SIFT algorithm is applied to quasi image and real image. SIFT algorithm allows to find corresponding points. Exterior orientation parameters of image are calculated from corresponding points. The second step is construction of the vector object model. Vectorization is performed by operator of PC in an interactive mode using single image. Spatial coordinates of the model are calculated automatically by cloud points. In addition, there is automatic edge detection with interactive editing available. Edge detection is performed on point cloud and on image with subsequent identification of correct edges. Experimental studies of the method have demonstrated its efficiency in case of building facade modeling.

  5. Analysis and Comparison of Objective Methods for Image Quality Assessment

    Directory of Open Access Journals (Sweden)

    P. S. Babkin


    Full Text Available The purpose of this work is research and modification of the reference objective methods for image quality assessment. The ultimate goal is to obtain a modification of formal assessments that more closely corresponds to the subjective expert estimates (MOS.In considering the formal reference objective methods for image quality assessment we used the results of other authors, which offer results and comparative analyzes of the most effective algorithms. Based on these investigations we have chosen two of the most successful algorithm for which was made a further analysis in the MATLAB 7.8 R 2009 a (PQS and MSSSIM. The publication focuses on the features of the algorithms, which have great importance in practical implementation, but are insufficiently covered in the publications by other authors.In the implemented modification of the algorithm PQS boundary detector Kirsch was replaced by the boundary detector Canny. Further experiments were carried out according to the method of the ITU-R VT.500-13 (01/2012 using monochrome images treated with different types of filters (should be emphasized that an objective assessment of image quality PQS is applicable only to monochrome images. Images were obtained with a thermal imaging surveillance system. The experimental results proved the effectiveness of this modification.In the specialized literature in the field of formal to evaluation methods pictures, this type of modification was not mentioned.The method described in the publication can be applied to various practical implementations of digital image processing.Advisability and effectiveness of using the modified method of PQS to assess the structural differences between the images are shown in the article and this will be used in solving the problems of identification and automatic control.

  6. Ultra-sensitive quantification of lysozyme based on element chelate labeling and capillary electrophoresis–inductively coupled plasma mass spectrometry

    International Nuclear Information System (INIS)

    Yang, MingWei; Wu, WeiHua; Ruan, YaJuan; Huang, LiMei; Wu, Zujian; Cai, Yong; Fu, FengFu


    Graphical abstract: An ultra-sensitive method for the determination of lysozyme was developed based on the Gd 3+ chelate labeling and CE–ICP–MS. The proposed method has an extremely low detection limit of 3.89 attomole and has been successfully used to detect lysozyme in saliva sample, showing excellent reliability. The success of the present method provides a new possibility for biological assays and clinical diagnoses. -- Highlights: •An ultra-sensitive method for detecting lysozyme based on CE–ICP–MS was described. •The proposed method has an extremely low detection limit of 3.89 attomole. •It can be used to detect trace lysozyme in saliva sample with a satisfied recovery. •The method provides a new potential for sensitive detection of low-abundant proteins. -- Abstract: In this study, an ultra-sensitive method for the quantification of lysozyme based on the Gd 3+ diethylenetriamine-N,N,N′,N″,N″-pentaacetic acid labeling and capillary electrophoresis–inductively coupled plasma mass spectrometry (CE–ICP–MS) was described. The Gd 3+ -tagged lysozyme was effectively separated by capillary electrophoresis (CE) and sensitively determined by inductively coupled plasma mass spectrometry (ICP–MS). Based on the gadolinium-tagging and CE–ICP–MS, the lysozyme was determined within 12 min with an extremely low detection limit of 3.89 attomole (3.89 × 10 −11 mol L −1 for 100 nL of sample injection) and a RSD < 6% (n = 5). The proposed method has been successfully used to detect lysozyme in saliva samples with a recovery of 91–106%, suggesting that our method is sensitive and reliable. The success of the present method provides a new potential for the biological assays and sensitive detection of low-abundant proteins

  7. Method and apparatus to image biological interactions in plants (United States)

    Weisenberger, Andrew; Bonito, Gregory M.; Reid, Chantal D.; Smith, Mark Frederick


    A method to dynamically image the actual translocation of molecular compounds of interest in a plant root, root system, and rhizosphere without disturbing the root or the soil. The technique makes use of radioactive isotopes as tracers to label molecules of interest and to image their distribution in the plant and/or soil. The method allows for the study and imaging of various biological and biochemical interactions in the rhizosphere of a plant, including, but not limited to, mycorrhizal associations in such regions.

  8. Molecular system analysis, multidimensional, dynamic, ultra-sensitive exploration of proteomes

    International Nuclear Information System (INIS)

    Scharattenholz, A.; Soski, V.; Stegmann, W.; Schroer, K.; Godovac-Zimmermann, J.; Cabuk, A.; Pejovi, V.; Wozny, W.; Cahill, M.A.; Drukier, A.K.; Volkovitsky, P.


    ProteoSys AG's holistic proteomics strategy extends beyond classical proteome research as a new paradigm. Our concept of multidimensional molecular systems analysis of complex model systems employs the innovative ProteoDyn TM approach. This enables us to correlate dynamic changes of proteomes with their biophysical and biochemical environment. Our supersensitive Multi Photon Detection (MPD) technology enables ultra-sensitive detection of proteins, deep into the low abundance domain. Our technology platform includes the affinity analysis of phospho- and glyco-proteomes, and with our 'fish hook' methods we can capture and fully characterize even serpentine G-coupled receptors and associated proteins, including routine comprehensive post-translational analyses performed by a well equipped mass spectrometry group. Throughput and quality is obtained by automation and high end robotics, with data management handled by a dedicated bioinformatics department. Thus ProteoSys AG has a range of state of the art and proprietary tools at its disposal to analyse even the most difficult complex model systems. MPD is an isotopic detection method proprietary to ProteoSys For MPD analysis we have implemented protocols where over 99% of proteins can be iodinated, and where the iodinated proteins can be identified by mass spectrometry. Because MPD measures the energy of detected particles, it can discriminate between signals originating from different isotopes co-electrophoresed by 2D-PAGE. Thus MPD imagers have a 'multicolour' functionality suitable for differential display and improved throughput, eliminating inter-gel variations. Importantly, MPD opens up not only the world of detection of low abundance proteins, but also identification and characterization. Radioactive low abundance protein spots containing less than one attomole of protein can be excised from a 2D-gel, mixed with unlabelled proteins, and 'tracked' by MPD. The identity of the labeled protein is determined by

  9. A new optimal seam method for seamless image stitching (United States)

    Xue, Jiale; Chen, Shengyong; Cheng, Xu; Han, Ying; Zhao, Meng


    A novel optimal seam method which aims to stitch those images with overlapping area more seamlessly has been propos ed. Considering the traditional gradient domain optimal seam method and fusion algorithm result in bad color difference measurement and taking a long time respectively, the input images would be converted to HSV space and a new energy function is designed to seek optimal stitching path. To smooth the optimal stitching path, a simplified pixel correction and weighted average method are utilized individually. The proposed methods exhibit performance in eliminating the stitching seam compared with the traditional gradient optimal seam and high efficiency with multi-band blending algorithm.

  10. Methods for processing and analysis functional and anatomical brain images: computerized tomography, emission tomography and nuclear resonance imaging

    International Nuclear Information System (INIS)

    Mazoyer, B.M.


    The various methods for brain image processing and analysis are presented and compared. The following topics are developed: the physical basis of brain image comparison (nature and formation of signals intrinsic performance of the methods image characteristics); mathematical methods for image processing and analysis (filtering, functional parameter extraction, morphological analysis, robotics and artificial intelligence); methods for anatomical localization (neuro-anatomy atlas, proportional stereotaxic atlas, numerized atlas); methodology of cerebral image superposition (normalization, retiming); image networks [fr

  11. A distortion correction method for image intensifier and electronic portal images used in radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Ioannidis, G T; Geramani, K N; Zamboglou, N [Strahlenklinik, Stadtische Kliniken Offenbach, Offenbach (Germany); Uzunoglu, N [Department of Electrical and Computer Engineering, National Technical University of Athens, Athens (Greece)


    At the most of radiation departments a simulator and an `on line` verification system of the treated volume, in form of an electronic portal imaging device (EPID), are available. Networking and digital handling (saving, archiving etc.) of the image information is a necessity in the image processing procedures in order to evaluate verification and simulation recordings at the computer screen. Distortion is on the other hand prerequisite for quantitative comparison of both image modalities. Another limitation factor, in order to make quantitative assertions, is the fact that the irradiation fields in radiotherapy are usually bigger than the field of view of an image intensifier. Several segments of the irradiation field must therefore be acquired. Using pattern recognition techniques these segments can be composed into a single image. In this paper a distortion correction method will be presented. The method is based upon a well defined Grid which is embedded during the registration process on the image. The video signal from the image intensifier is acquired and processed. The grid is then recognised using image processing techniques. Ideally if all grid points are recognised, various methods can be applied in order to correct the distortion. But in practice this is not the case. Overlapping structures (bones etc.) have as a consequence that not all of the grid points can be recognised. Mathematical models from the Graph theory are applied in order to reconstruct the whole grid. The deviation of the grid points positions from the rated value is then used to calculate correction coefficients. This method (well defined grid, grid recognition, correction factors) can also be applied in verification images from the EPID or in other image modalities, and therefore a quantitative comparison in radiation treatment is possible. The distortion correction method and the application on simulator images will be presented. (authors)

  12. Image Registration Using Single Cluster PHD Methods (United States)

    Campbell, M.; Schlangen, I.; Delande, E.; Clark, D.

    Cadets in the Department of Physics at the United States Air Force Academy are using the technique of slitless spectroscopy to analyze the spectra from geostationary satellites during glint season. The equinox periods of the year are particularly favorable for earth-based observers to detect specular reflections off satellites (glints), which have been observed in the past using broadband photometry techniques. Three seasons of glints were observed and analyzed for multiple satellites, as measured across the visible spectrum using a diffraction grating on the Academy’s 16-inch, f/8.2 telescope. It is clear from the results that the glint maximum wavelength decreases relative to the time periods before and after the glint, and that the spectral reflectance during the glint is less like a blackbody. These results are consistent with the presumption that solar panels are the predominant source of specular reflection. The glint spectra are also quantitatively compared to different blackbody curves and the solar spectrum by means of absolute differences and standard deviations. Our initial analysis appears to indicate a potential method of determining relative power capacity.

  13. New adaptive sampling method in particle image velocimetry

    International Nuclear Information System (INIS)

    Yu, Kaikai; Xu, Jinglei; Tang, Lan; Mo, Jianwei


    This study proposes a new adaptive method to enable the number of interrogation windows and their positions in a particle image velocimetry (PIV) image interrogation algorithm to become self-adapted according to the seeding density. The proposed method can relax the constraint of uniform sampling rate and uniform window size commonly adopted in the traditional PIV algorithm. In addition, the positions of the sampling points are redistributed on the basis of the spring force generated by the sampling points. The advantages include control of the number of interrogation windows according to the local seeding density and smoother distribution of sampling points. The reliability of the adaptive sampling method is illustrated by processing synthetic and experimental images. The synthetic example attests to the advantages of the sampling method. Compared with that of the uniform interrogation technique in the experimental application, the spatial resolution is locally enhanced when using the proposed sampling method. (technical design note)

  14. Research on interpolation methods in medical image processing. (United States)

    Pan, Mei-Sen; Yang, Xiao-Li; Tang, Jing-Tian


    Image interpolation is widely used for the field of medical image processing. In this paper, interpolation methods are divided into three groups: filter interpolation, ordinary interpolation and general partial volume interpolation. Some commonly-used filter methods for image interpolation are pioneered, but the interpolation effects need to be further improved. When analyzing and discussing ordinary interpolation, many asymmetrical kernel interpolation methods are proposed. Compared with symmetrical kernel ones, the former are have some advantages. After analyzing the partial volume and generalized partial volume estimation interpolations, the new concept and constraint conditions of the general partial volume interpolation are defined, and several new partial volume interpolation functions are derived. By performing the experiments of image scaling, rotation and self-registration, the interpolation methods mentioned in this paper are compared in the entropy, peak signal-to-noise ratio, cross entropy, normalized cross-correlation coefficient and running time. Among the filter interpolation methods, the median and B-spline filter interpolations have a relatively better interpolating performance. Among the ordinary interpolation methods, on the whole, the symmetrical cubic kernel interpolations demonstrate a strong advantage, especially the symmetrical cubic B-spline interpolation. However, we have to mention that they are very time-consuming and have lower time efficiency. As for the general partial volume interpolation methods, from the total error of image self-registration, the symmetrical interpolations provide certain superiority; but considering the processing efficiency, the asymmetrical interpolations are better.

  15. Advanced methods for image registration applied to JET videos

    Energy Technology Data Exchange (ETDEWEB)

    Craciunescu, Teddy, E-mail: [EURATOM-MEdC Association, NILPRP, Bucharest (Romania); Murari, Andrea [Consorzio RFX, Associazione EURATOM-ENEA per la Fusione, Padova (Italy); Gelfusa, Michela [Associazione EURATOM-ENEA – University of Rome “Tor Vergata”, Roma (Italy); Tiseanu, Ion; Zoita, Vasile [EURATOM-MEdC Association, NILPRP, Bucharest (Romania); Arnoux, Gilles [EURATOM/CCFE Fusion Association, Culham Science Centre, Abingdon, Oxon (United Kingdom)


    Graphical abstract: - Highlights: • Development of an image registration method for JET IR and fast visible cameras. • Method based on SIFT descriptors and coherent point drift points set registration technique. • Method able to deal with extremely noisy images and very low luminosity images. • Computation time compatible with the inter-shot analysis. - Abstract: The last years have witnessed a significant increase in the use of digital cameras on JET. They are routinely applied for imaging in the IR and visible spectral regions. One of the main technical difficulties in interpreting the data of camera based diagnostics is the presence of movements of the field of view. Small movements occur due to machine shaking during normal pulses while large ones may arise during disruptions. Some cameras show a correlation of image movement with change of magnetic field strength. For deriving unaltered information from the videos and for allowing correct interpretation an image registration method, based on highly distinctive scale invariant feature transform (SIFT) descriptors and on the coherent point drift (CPD) points set registration technique, has been developed. The algorithm incorporates a complex procedure for rejecting outliers. The method has been applied for vibrations correction to videos collected by the JET wide angle infrared camera and for the correction of spurious rotations in the case of the JET fast visible camera (which is equipped with an image intensifier). The method has proved to be able to deal with the images provided by this camera frequently characterized by low contrast and a high level of blurring and noise.

  16. Method for estimating modulation transfer function from sample images. (United States)

    Saiga, Rino; Takeuchi, Akihisa; Uesugi, Kentaro; Terada, Yasuko; Suzuki, Yoshio; Mizutani, Ryuta


    The modulation transfer function (MTF) represents the frequency domain response of imaging modalities. Here, we report a method for estimating the MTF from sample images. Test images were generated from a number of images, including those taken with an electron microscope and with an observation satellite. These original images were convolved with point spread functions (PSFs) including those of circular apertures. The resultant test images were subjected to a Fourier transformation. The logarithm of the squared norm of the Fourier transform was plotted against the squared distance from the origin. Linear correlations were observed in the logarithmic plots, indicating that the PSF of the test images can be approximated with a Gaussian. The MTF was then calculated from the Gaussian-approximated PSF. The obtained MTF closely coincided with the MTF predicted from the original PSF. The MTF of an x-ray microtomographic section of a fly brain was also estimated with this method. The obtained MTF showed good agreement with the MTF determined from an edge profile of an aluminum test object. We suggest that this approach is an alternative way of estimating the MTF, independently of the image type. Copyright © 2017 Elsevier Ltd. All rights reserved.

  17. Research of x-ray automatic image mosaic method (United States)

    Liu, Bin; Chen, Shunan; Guo, Lianpeng; Xu, Wanpeng


    Image mosaic has widely applications value in the fields of medical image analysis, and it is a technology that carries on the spatial matching to a series of image which are overlapped with each other, and finally builds a seamless and high quality image which has high resolution and big eyeshot. In this paper, the method of grayscale cutting pseudo-color enhancement was firstly used to complete the mapping transformation from gray to the pseudo-color, and to extract SIFT features from the images. And then by making use of a similar measure of NCC (normalized cross correlation - Normalized cross-correlation), the method of RANSAC (Random Sample Consensus) was used to exclude the pseudofeature points right in order to complete the exact match of feature points. Finally, seamless mosaic and color fusion were completed by using wavelet multi-decomposition. The experiment shows that the method we used can effectively improve the precision and automation of the medical image mosaic, and provide an effective technical approach for automatic medical image mosaic.

  18. [An Improved Spectral Quaternion Interpolation Method of Diffusion Tensor Imaging]. (United States)

    Xu, Yonghong; Gao, Shangce; Hao, Xiaofei


    Diffusion tensor imaging(DTI)is a rapid development technology in recent years of magnetic resonance imaging.The diffusion tensor interpolation is a very important procedure in DTI image processing.The traditional spectral quaternion interpolation method revises the direction of the interpolation tensor and can preserve tensors anisotropy,but the method does not revise the size of tensors.The present study puts forward an improved spectral quaternion interpolation method on the basis of traditional spectral quaternion interpolation.Firstly,we decomposed diffusion tensors with the direction of tensors being represented by quaternion.Then we revised the size and direction of the tensor respectively according to different situations.Finally,we acquired the tensor of interpolation point by calculating the weighted average.We compared the improved method with the spectral quaternion method and the Log-Euclidean method by the simulation data and the real data.The results showed that the improved method could not only keep the monotonicity of the fractional anisotropy(FA)and the determinant of tensors,but also preserve the tensor anisotropy at the same time.In conclusion,the improved method provides a kind of important interpolation method for diffusion tensor image processing.

  19. Plasma Enhanced Growth of Carbon Nanotubes For Ultrasensitive Biosensors (United States)

    Cassell, Alan M.; Li, J.; Ye, Q.; Koehne, J.; Chen, H.; Meyyappan, M.


    The multitude of considerations facing nanostructure growth and integration lends itself to combinatorial optimization approaches. Rapid optimization becomes even more important with wafer-scale growth and integration processes. Here we discuss methodology for developing plasma enhanced CVD growth techniques for achieving individual, vertically aligned carbon nanostructures that show excellent properties as ultrasensitive electrodes for nucleic acid detection. We utilize high throughput strategies for optimizing the upstream and downstream processing and integration of carbon nanotube electrodes as functional elements in various device types. An overview of ultrasensitive carbon nanotube based sensor arrays for electrochemical biosensing applications and the high throughput methodology utilized to combine novel electrode technology with conventional MEMS processing will be presented.


    Directory of Open Access Journals (Sweden)

    G. A. Kukharev


    Full Text Available In the paper a method of generating of standard type linear barcodes from facial images is proposed. The method is based on use of the histogram of facial image brightness, averaging the histogram on a limited number of intervals, quantization of results in a range of decimal numbers from 0 to 9 and table conversion into the final barcode. The proposed solution is computationally low-cost and not requires the use of specialized software on image processing that allows generating of facial barcodes in mobile systems, and thus the proposed method can be interpreted as an express method. Results of tests on the Face94 and CUHK Face Sketch FERET Databases showed that the proposed method is a new solution for use in the real-world practice and ensures the stability of generated barcodes in changes of scale, pose and mirroring of a facial image, and also changes of a facial expression and shadows on faces from local lighting. The proposed method is based on generating of a standard barcode directly from the facial image, and thus contains the subjective information about a person's face.

  1. Spectral analysis of mammographic images using a multitaper method

    International Nuclear Information System (INIS)

    Wu Gang; Mainprize, James G.; Yaffe, Martin J.


    Purpose: Power spectral analysis in radiographic images is conventionally performed using a windowed overlapping averaging periodogram. This study describes an alternative approach using a multitaper technique and compares its performance with that of the standard method. This tool will be valuable in power spectrum estimation of images, whose content deviates significantly from uniform white noise. The performance of the multitaper approach will be evaluated in terms of spectral stability, variance reduction, bias, and frequency precision. The ultimate goal is the development of a useful tool for image quality assurance. Methods: A multitaper approach uses successive data windows of increasing order. This mitigates spectral leakage allowing one to calculate a reduced-variance power spectrum. The multitaper approach will be compared with the conventional power spectrum method in several typical situations, including the noise power spectra (NPS) measurements of simulated projection images of a uniform phantom, NPS measurement of real detector images of a uniform phantom for two clinical digital mammography systems, and the estimation of the anatomic noise in mammographic images (simulated images and clinical mammograms). Results: Examination of spectrum variance versus frequency resolution and bias indicates that the multitaper approach is superior to the conventional single taper methods in the prevention of spectrum leakage and variance reduction. More than four times finer frequency precision can be achieved with equivalent or less variance and bias. Conclusions: Without any shortening of the image data length, the bias is smaller and the frequency resolution is higher with the multitaper method, and the need to compromise in the choice of regions of interest size to balance between the reduction of variance and the loss of frequency resolution is largely eliminated.

  2. Reconstruction of CT images by the Bayes- back projection method

    CERN Document Server

    Haruyama, M; Takase, M; Tobita, H


    In the course of research on quantitative assay of non-destructive measurement of radioactive waste, the have developed a unique program based on the Bayesian theory for reconstruction of transmission computed tomography (TCT) image. The reconstruction of cross-section images in the CT technology usually employs the Filtered Back Projection method. The new imaging reconstruction program reported here is based on the Bayesian Back Projection method, and it has a function of iterative improvement images by every step of measurement. Namely, this method has the capability of prompt display of a cross-section image corresponding to each angled projection data from every measurement. Hence, it is possible to observe an improved cross-section view by reflecting each projection data in almost real time. From the basic theory of Baysian Back Projection method, it can be not only applied to CT types of 1st, 2nd, and 3rd generation. This reported deals with a reconstruction program of cross-section images in the CT of ...

  3. Multi-crack imaging using nonclassical nonlinear acoustic method

    International Nuclear Information System (INIS)

    Zhang Lue; Zhang Ying; Liu Xiao-Zhou; Gong Xiu-Fen


    Solid materials with cracks exhibit the nonclassical nonlinear acoustical behavior. The micro-defects in solid materials can be detected by nonlinear elastic wave spectroscopy (NEWS) method with a time-reversal (TR) mirror. While defects lie in viscoelastic solid material with different distances from one another, the nonlinear and hysteretic stress—strain relation is established with Preisach—Mayergoyz (PM) model in crack zone. Pulse inversion (PI) and TR methods are used in numerical simulation and defect locations can be determined from images obtained by the maximum value. Since false-positive defects might appear and degrade the imaging when the defects are located quite closely, the maximum value imaging with a time window is introduced to analyze how defects affect each other and how the fake one occurs. Furthermore, NEWS-TR-NEWS method is put forward to improve NEWS-TR scheme, with another forward propagation (NEWS) added to the existing phases (NEWS and TR). In the added phase, scanner locations are determined by locations of all defects imaged in previous phases, so that whether an imaged defect is real can be deduced. NEWS-TR-NEWS method is proved to be effective to distinguish real defects from the false-positive ones. Moreover, it is also helpful to detect the crack that is weaker than others during imaging procedure. (electromagnetism, optics, acoustics, heat transfer, classical mechanics, and fluid dynamics)

  4. Multi-crack imaging using nonclassical nonlinear acoustic method (United States)

    Zhang, Lue; Zhang, Ying; Liu, Xiao-Zhou; Gong, Xiu-Fen


    Solid materials with cracks exhibit the nonclassical nonlinear acoustical behavior. The micro-defects in solid materials can be detected by nonlinear elastic wave spectroscopy (NEWS) method with a time-reversal (TR) mirror. While defects lie in viscoelastic solid material with different distances from one another, the nonlinear and hysteretic stress—strain relation is established with Preisach—Mayergoyz (PM) model in crack zone. Pulse inversion (PI) and TR methods are used in numerical simulation and defect locations can be determined from images obtained by the maximum value. Since false-positive defects might appear and degrade the imaging when the defects are located quite closely, the maximum value imaging with a time window is introduced to analyze how defects affect each other and how the fake one occurs. Furthermore, NEWS-TR-NEWS method is put forward to improve NEWS-TR scheme, with another forward propagation (NEWS) added to the existing phases (NEWS and TR). In the added phase, scanner locations are determined by locations of all defects imaged in previous phases, so that whether an imaged defect is real can be deduced. NEWS-TR-NEWS method is proved to be effective to distinguish real defects from the false-positive ones. Moreover, it is also helpful to detect the crack that is weaker than others during imaging procedure.

  5. A Quick and Affine Invariance Matching Method for Oblique Images

    Directory of Open Access Journals (Sweden)

    XIAO Xiongwu


    Full Text Available This paper proposed a quick, affine invariance matching method for oblique images. It calculated the initial affine matrix by making full use of the two estimated camera axis orientation parameters of an oblique image, then recovered the oblique image to a rectified image by doing the inverse affine transform, and left over by the SIFT method. We used the nearest neighbor distance ratio(NNDR, normalized cross correlation(NCC measure constraints and consistency check to get the coarse matches, then used RANSAC method to calculate the fundamental matrix and the homography matrix. And we got the matches that they were interior points when calculating the homography matrix, then calculated the average value of the matches' principal direction differences. During the matching process, we got the initial matching features by the nearest neighbor(NN matching strategy, then used the epipolar constrains, homography constrains, NCC measure constrains and consistency check of the initial matches' principal direction differences with the calculated average value of the interior matches' principal direction differences to eliminate false matches. Experiments conducted on three pairs of typical oblique images demonstrate that our method takes about the same time as SIFT to match a pair of oblique images with a plenty of corresponding points distributed evenly and an extremely low mismatching rate.

  6. A Study on the Improvement of Digital Periapical Images using Image Interpolation Methods

    International Nuclear Information System (INIS)

    Song, Nam Kyu; Koh, Kwang Joon


    Image resampling is of particular interest in digital radiology. When resampling an image to a new set of coordinate, there appears blocking artifacts and image changes. To enhance image quality, interpolation algorithms have been used. Resampling is used to increase the number of points in an image to improve its appearance for display. The process of interpolation is fitting a continuous function to the discrete points in the digital image. The purpose of this study was to determine the effects of the seven interpolation functions when image resampling in digital periapical images. The images were obtained by Digora, CDR and scanning of Ektaspeed plus periapical radiograms on the dry skull and human subject. The subjects were exposed to intraoral X-ray machine at 60 kVp and 70 kVp with exposure time varying between 0.01 and 0.50 second. To determine which interpolation method would provide the better image, seven functions were compared ; (1) nearest neighbor (2) linear (3) non-linear (4) facet model (5) cubic convolution (6) cubic spline (7) gray segment expansion. And resampled images were compared in terms of SNR (Signal to Noise Ratio) and MTF (Modulation Transfer Function) coefficient value. The obtained results were as follows ; 1. The highest SNR value (75.96 dB) was obtained with cubic convolution method and the lowest SNR value (72.44 dB) was obtained with facet model method among seven interpolation methods. 2. There were significant differences of SNR values among CDR, Digora and film scan (P 0.05). 4. There were significant differences of MTF coefficient values between linear interpolation method and the other six interpolation methods (P<0.05). 5. The speed of computation time was the fastest with nearest neighbor method and the slowest with non-linear method. 6. The better image was obtained with cubic convolution, cubic spline and gray segment method in ROC analysis. 7. The better sharpness of edge was obtained with gray segment expansion method

  7. A new method of the light irradiation image by the computed radiography (imaging plate) system

    International Nuclear Information System (INIS)

    Aiba, Susumu; Nishi, Katsuki.


    There are two method for the purpose of diagnosing medically by using gamma-ray light irradiation image. One is to use of the scintillation camera for gamma-ray, the other is to use of the photostimulable luminescence point by the secondary excitation of the image plate (IP) system for X-ray. The standpoint of the spatial resolution at the total medical image, using gamma-ray, the first can get the image on a short time, but the first is a poor image quality, and the second is good image quality, but the second can get the image on a long time, because of insensitive to gamma-ray. We report on the improvement for IP's week point by our proposal method, and by our clinical and quantitative analysis data, to use the highly efficient IP (ST-III). We make the improvement on the imaging time (from 30 minutes to 20 minutes), and the inprocessing time (from 33-50 minutes to 27 minutes) for a former method on an organism. We strongly believe that our convenience improvement method, and our clinical quantitative analysis data can contribute to the wide application as well as the quality up for the clinical diagnosis to use gamma-ray. (author)

  8. Dynamic PET Image reconstruction for parametric imaging using the HYPR kernel method (United States)

    Spencer, Benjamin; Qi, Jinyi; Badawi, Ramsey D.; Wang, Guobao


    Dynamic PET image reconstruction is a challenging problem because of the ill-conditioned nature of PET and the lowcounting statistics resulted from short time-frames in dynamic imaging. The kernel method for image reconstruction has been developed to improve image reconstruction of low-count PET data by incorporating prior information derived from high-count composite data. In contrast to most of the existing regularization-based methods, the kernel method embeds image prior information in the forward projection model and does not require an explicit regularization term in the reconstruction formula. Inspired by the existing highly constrained back-projection (HYPR) algorithm for dynamic PET image denoising, we propose in this work a new type of kernel that is simpler to implement and further improves the kernel-based dynamic PET image reconstruction. Our evaluation study using a physical phantom scan with synthetic FDG tracer kinetics has demonstrated that the new HYPR kernel-based reconstruction can achieve a better region-of-interest (ROI) bias versus standard deviation trade-off for dynamic PET parametric imaging than the post-reconstruction HYPR denoising method and the previously used nonlocal-means kernel.

  9. Development of an Ultrasensitive Immunoassay for Detecting Tartrazine


    Li, Zhuokun; Song, Shanshan; Xu, Liguang; Kuang, Hua; Guo, Shidong; Xu, Chuanlai


    We have developed an ultrasensitive indirect competitive enzyme-linked immunosorbent assay for the determination of tartrazine. Two carboxylated analogues of tartrazine with different spacer lengths, and one derivative from commercial tartrazine after a little chemical modification, were synthesized as haptens in order to produce antibodies specific to tartrazine. The effect of sulfonic acid groups on the hapten structure of tartrazine was also studied carefully for the first time. A most spe...

  10. Denoising imaging polarimetry by adapted BM3D method. (United States)

    Tibbs, Alexander B; Daly, Ilse M; Roberts, Nicholas W; Bull, David R


    In addition to the visual information contained in intensity and color, imaging polarimetry allows visual information to be extracted from the polarization of light. However, a major challenge of imaging polarimetry is image degradation due to noise. This paper investigates the mitigation of noise through denoising algorithms and compares existing denoising algorithms with a new method, based on BM3D (Block Matching 3D). This algorithm, Polarization-BM3D (PBM3D), gives visual quality superior to the state of the art across all images and noise standard deviations tested. We show that denoising polarization images using PBM3D allows the degree of polarization to be more accurately calculated by comparing it with spectral polarimetry measurements.

  11. Metal artifact reduction method using metal streaks image subtraction

    International Nuclear Information System (INIS)

    Pua, Rizza D.; Cho, Seung Ryong


    Many studies have been dedicated for metal artifact reduction (MAR); however, the methods are successful to varying degrees depending on situations. Sinogram in-painting, filtering, iterative method are some of the major categories of MAR. Each has its own merits and weaknesses. A combination of these methods or hybrid methods have also been developed to make use of the different benefits of two techniques and minimize the unfavorable results. Our method focuses on the in-paitning approach and a hybrid MAR described by Xia et al. Although in-painting scheme is an effective technique in reducing the primary metal artifacts, a major drawback is the reintroduction of new artifacts that can be caused by an inaccurate interpolation process. Furthermore, combining the segmented metal image to the corrected nonmetal image in the final step of a conventional inpainting approach causes an issue of incorrect metal pixel values. Our proposed method begins with a sinogram in-painting approach and ends with an image-based metal artifact reduction scheme. This work provides a simple, yet effective solution for reducing metal artifacts and acquiring the original metal pixel information. The proposed method demonstrated its effectiveness in a simulation setting. The proposed method showed image quality that is comparable to the standard MAR; however, quantitatively more accurate than the standard MAR

  12. Combination of acoustical radiosity and the image source method

    DEFF Research Database (Denmark)

    Koutsouris, Georgios I; Brunskog, Jonas; Jeong, Cheol-Ho


    A combined model for room acoustic predictions is developed, aiming to treat both diffuse and specular reflections in a unified way. Two established methods are incorporated: acoustical radiosity, accounting for the diffuse part, and the image source method, accounting for the specular part...

  13. Multi-spectral lifetime imaging: methods and applications

    NARCIS (Netherlands)

    Fereidouni, F.


    The aim of this PhD project is to further develop multispectral life time imaging hardware and analyses methods. The hardware system, Lambda-Tau, generates a considerable amount of data at high speed. To fully exploit the power of this new hardware, fast and reliable data analyses methods are

  14. An effective method on pornographic images realtime recognition (United States)

    Wang, Baosong; Lv, Xueqiang; Wang, Tao; Wang, Chengrui


    In this paper, skin detection, texture filtering and face detection are used to extract feature on an image library, training them with the decision tree arithmetic to create some rules as a decision tree classifier to distinguish an unknown image. Experiment based on more than twenty thousand images, the precision rate can get 76.21% when testing on 13025 pornographic images and elapsed time is less than 0.2s. This experiment shows it has a good popularity. Among the steps mentioned above, proposing a new skin detection model which called irregular polygon region skin detection model based on YCbCr color space. This skin detection model can lower the false detection rate on skin detection. A new method called sequence region labeling on binary connected area can calculate features on connected area, it is faster and needs less memory than other recursive methods.

  15. A Single Image Dehazing Method Using Average Saturation Prior

    Directory of Open Access Journals (Sweden)

    Zhenfei Gu


    Full Text Available Outdoor images captured in bad weather are prone to yield poor visibility, which is a fatal problem for most computer vision applications. The majority of existing dehazing methods rely on an atmospheric scattering model and therefore share a common limitation; that is, the model is only valid when the atmosphere is homogeneous. In this paper, we propose an improved atmospheric scattering model to overcome this inherent limitation. By adopting the proposed model, a corresponding dehazing method is also presented. In this method, we first create a haze density distribution map of a hazy image, which enables us to segment the hazy image into scenes according to the haze density similarity. Then, in order to improve the atmospheric light estimation accuracy, we define an effective weight assignment function to locate a candidate scene based on the scene segmentation results and therefore avoid most potential errors. Next, we propose a simple but powerful prior named the average saturation prior (ASP, which is a statistic of extensive high-definition outdoor images. Using this prior combined with the improved atmospheric scattering model, we can directly estimate the scene atmospheric scattering coefficient and restore the scene albedo. The experimental results verify that our model is physically valid, and the proposed method outperforms several state-of-the-art single image dehazing methods in terms of both robustness and effectiveness.

  16. Method of Poisson's ratio imaging within a material part (United States)

    Roth, Don J. (Inventor)


    The present invention is directed to a method of displaying the Poisson's ratio image of a material part. In the present invention longitudinal data is produced using a longitudinal wave transducer and shear wave data is produced using a shear wave transducer. The respective data is then used to calculate the Poisson's ratio for the entire material part. The Poisson's ratio approximations are then used to displayed the image.

  17. Interpretation of the method of images in estimating superconducting levitation

    International Nuclear Information System (INIS)

    Perez-Diaz, Jose Luis; Garcia-Prada, Juan Carlos


    Among different papers devoted to superconducting levitation of a permanent magnet over a superconductor using the method of images, there is a discrepancy of a factor of two when estimating the lift force. This is not a minor matter but an interesting fundamental question that contributes to understanding the physical phenomena of 'imaging' on a superconductor surface. We solve it, make clear the physical behavior underlying it, and suggest the reinterpretation of some previous experiments

  18. A developed unsharp masking method for images contrast enhancement

    International Nuclear Information System (INIS)

    Zaafouri, A.; Sayadi, M.; Fnaiech, F.


    In this paper, we propose a developed unsharp masking process for contrast image enhancement. The main idea here is to enhance the dark and bright area in the same way which matches the response of human visual system well. Then in order to reduce the noise effect, a mean weighted high pass filter is used for edge extraction. The proposed method gives satisfactory results for wide range of low contrast images compared with others known approaches.

  19. Bin mode estimation methods for Compton camera imaging

    International Nuclear Information System (INIS)

    Ikeda, S.; Odaka, H.; Uemura, M.; Takahashi, T.; Watanabe, S.; Takeda, S.


    We study the image reconstruction problem of a Compton camera which consists of semiconductor detectors. The image reconstruction is formulated as a statistical estimation problem. We employ a bin-mode estimation (BME) and extend an existing framework to a Compton camera with multiple scatterers and absorbers. Two estimation algorithms are proposed: an accelerated EM algorithm for the maximum likelihood estimation (MLE) and a modified EM algorithm for the maximum a posteriori (MAP) estimation. Numerical simulations demonstrate the potential of the proposed methods

  20. Cross-relaxation imaging:methods, challenges and applications

    International Nuclear Information System (INIS)

    Stikov, Nikola


    An overview of quantitative magnetization transfer (qMT) is given, with focus on cross relaxation imaging (CRI) as a fast method for quantifying the proportion of protons bound to complex macromolecules in tissue. The procedure for generating CRI maps is outlined, showing examples in the human brain and knee, and discussing the caveats and challenges in generating precise and accurate CRI maps. Finally, several applications of CRI for imaging tissue microstructure are presented.(Author)

  1. Classification Method in Integrated Information Network Using Vector Image Comparison

    Directory of Open Access Journals (Sweden)

    Zhou Yuan


    Full Text Available Wireless Integrated Information Network (WMN consists of integrated information that can get data from its surrounding, such as image, voice. To transmit information, large resource is required which decreases the service time of the network. In this paper we present a Classification Approach based on Vector Image Comparison (VIC for WMN that improve the service time of the network. The available methods for sub-region selection and conversion are also proposed.

  2. An Efficient Evolutionary Based Method For Image Segmentation


    Aslanzadeh, Roohollah; Qazanfari, Kazem; Rahmati, Mohammad


    The goal of this paper is to present a new efficient image segmentation method based on evolutionary computation which is a model inspired from human behavior. Based on this model, a four layer process for image segmentation is proposed using the split/merge approach. In the first layer, an image is split into numerous regions using the watershed algorithm. In the second layer, a co-evolutionary process is applied to form centers of finals segments by merging similar primary regions. In the t...

  3. Sharpening methods for images captured through Bayer matrix (United States)

    Kalevo, Ossi; Rantanen, Henry, Jr.


    Image resolution and sharpness are essential criteria for a human observer when estimating the image quality. Typically cheap small-sized, low-resolution CMOS-camera sensors do not provide sharp enough images, at least when comparing to high-end digital cameras. Sharpening function can be used to increase the subjective sharpness seen by the observer. In this paper, few methods to apply sharpening for images captured by CMOS imaging sensors through color filter array (CFA) are compared. The sharpening easily adds also the visibility of noise, pixel-cross talk and interpolation artifacts. Necessary arrangements to avoid the amplification of these unwanted phenomenon are discussed. By applying the sharpening only to the green component the processing power requirements can be clearly reduced. By adjusting the red and blue component sharpness, according to the green component sharpening, creation of false colors are reduced highly. Direction search sharpening method can be used to reduce the amplification of the artifacts caused by the CFA interpolation (CFAI). The comparison of the presented methods is based mainly on subjective image quality. Also the processing power and memory requirements are considered.

  4. Hyperspectral image compressing using wavelet-based method (United States)

    Yu, Hui; Zhang, Zhi-jie; Lei, Bo; Wang, Chen-sheng


    Hyperspectral imaging sensors can acquire images in hundreds of continuous narrow spectral bands. Therefore each object presented in the image can be identified from their spectral response. However, such kind of imaging brings a huge amount of data, which requires transmission, processing, and storage resources for both airborne and space borne imaging. Due to the high volume of hyperspectral image data, the exploration of compression strategies has received a lot of attention in recent years. Compression of hyperspectral data cubes is an effective solution for these problems. Lossless compression of the hyperspectral data usually results in low compression ratio, which may not meet the available resources; on the other hand, lossy compression may give the desired ratio, but with a significant degradation effect on object identification performance of the hyperspectral data. Moreover, most hyperspectral data compression techniques exploits the similarities in spectral dimensions; which requires bands reordering or regrouping, to make use of the spectral redundancy. In this paper, we explored the spectral cross correlation between different bands, and proposed an adaptive band selection method to obtain the spectral bands which contain most of the information of the acquired hyperspectral data cube. The proposed method mainly consist three steps: First, the algorithm decomposes the original hyperspectral imagery into a series of subspaces based on the hyper correlation matrix of the hyperspectral images between different bands. And then the Wavelet-based algorithm is applied to the each subspaces. At last the PCA method is applied to the wavelet coefficients to produce the chosen number of components. The performance of the proposed method was tested by using ISODATA classification method.

  5. Immunoliposome-PCR: a generic ultrasensitive quantitative antigen detection system

    Directory of Open Access Journals (Sweden)

    He Junkun


    Full Text Available Abstract Background The accurate quantification of antigens at low concentrations over a wide dynamic range is needed for identifying biomarkers associated with disease and detecting protein interactions in high-throughput microarrays used in proteomics. Here we report the development of an ultrasensitive quantitative assay format called immunoliposome polymerase chain reaction (ILPCR that fulfills these requirements. This method uses a liposome, with reporter DNA encapsulated inside and biotin-labeled polyethylene glycol (PEG phospholipid conjugates incorporated into the outer surface of the liposome, as a detection reagent. The antigenic target is immobilized in the well of a microplate by a capture antibody and the liposome detection reagent is then coupled to a biotin-labeled second antibody through a NeutrAvidin bridge. The liposome is ruptured to release the reporter DNA, which serves as a surrogate to quantify the protein target using real-time PCR. Results A liposome detection reagent was prepared, which consisted of a population of liposomes ~120 nm in diameter with each liposome possessing ~800 accessible biotin receptors and ~220 encapsulated reporters. This liposome detection reagent was used in an assay to quantify the concentration of carcinoembryonic antigen (CEA in human serum. This ILPCR assay exhibited a linear dose–response curve from 10-10 M to 10-16 M CEA. Within this range the assay coefficient of variance was Conclusions The ILPCR assay has several advantages over other immuno-PCR methods. The reporter DNA and biotin-labeled PEG phospholipids spontaneously incorporate into the liposomes as they form, simplifying preparation of the detection reagent. Encapsulation of the reporter inside the liposomes allows nonspecific DNA in the assay medium to be degraded with DNase I prior to quantification of the encapsulated reporter by PCR, which reduces false-positive results and improves quantitative accuracy. The ability to

  6. On the pinned field image binarization for signature generation in image ownership verification method

    Directory of Open Access Journals (Sweden)

    Chang Hsuan


    Full Text Available Abstract The issue of pinned field image binarization for signature generation in the ownership verification of the protected image is investigated. The pinned field explores the texture information of the protected image and can be employed to enhance the watermark robustness. In the proposed method, four optimization schemes are utilized to determine the threshold values for transforming the pinned field into a binary feature image, which is then utilized to generate an effective signature image. Experimental results show that the utilization of optimization schemes can significantly improve the signature robustness from the previous method (Lee and Chang, Opt. Eng. 49 (9, 097005, 2010. While considering both the watermark retrieval rate and the computation speed, the genetic algorithm is strongly recommended. In addition, compared with Chang and Lin's scheme (J. Syst. Softw. 81 (7, 1118-1129, 2008, the proposed scheme also has better performance.

  7. Fast method of constructing image correlations to build a free network based on image multivocabulary trees (United States)

    Zhan, Zongqian; Wang, Xin; Wei, Minglu


    In image-based three-dimensional (3-D) reconstruction, one topic of growing importance is how to quickly obtain a 3-D model from a large number of images. The retrieval of the correct and relevant images for the model poses a considerable technological challenge. The "image vocabulary tree" has been proposed as a method to search for similar images. However, a significant drawback of this approach is identified in its low time efficiency and barely satisfactory classification result. The method proposed is inspired by, and improves upon, some recent methods. Specifically, vocabulary quality is considered and multivocabulary trees are designed to improve the classification result. A marked improvement was, indeed, observed in our evaluation of the proposed method. To improve time efficiency, graphics processing unit (GPU) computer unified device architecture parallel computation is applied in the multivocabulary trees. The results of the experiments showed that the GPU was three to four times more efficient than the enumeration matching and CPU methods when the number of images is large. This paper presents a reliable reference method for the rapid construction of a free network to be used for the computing of 3-D information.

  8. Splitting methods in communication, imaging, science, and engineering

    CERN Document Server

    Osher, Stanley; Yin, Wotao


    This book is about computational methods based on operator splitting. It consists of twenty-three chapters written by recognized splitting method contributors and practitioners, and covers a vast spectrum of topics and application areas, including computational mechanics, computational physics, image processing, wireless communication, nonlinear optics, and finance. Therefore, the book presents very versatile aspects of splitting methods and their applications, motivating the cross-fertilization of ideas. .

  9. Domain decomposition methods for solving an image problem

    Energy Technology Data Exchange (ETDEWEB)

    Tsui, W.K.; Tong, C.S. [Hong Kong Baptist College (Hong Kong)


    The domain decomposition method is a technique to break up a problem so that ensuing sub-problems can be solved on a parallel computer. In order to improve the convergence rate of the capacitance systems, pre-conditioned conjugate gradient methods are commonly used. In the last decade, most of the efficient preconditioners are based on elliptic partial differential equations which are particularly useful for solving elliptic partial differential equations. In this paper, the authors apply the so called covering preconditioner, which is based on the information of the operator under investigation. Therefore, it is good for various kinds of applications, specifically, they shall apply the preconditioned domain decomposition method for solving an image restoration problem. The image restoration problem is to extract an original image which has been degraded by a known convolution process and additive Gaussian noise.

  10. A novel optical gating method for laser gated imaging (United States)

    Ginat, Ran; Schneider, Ron; Zohar, Eyal; Nesher, Ofer


    For the past 15 years, Elbit Systems is developing time-resolved active laser-gated imaging (LGI) systems for various applications. Traditional LGI systems are based on high sensitive gated sensors, synchronized to pulsed laser sources. Elbit propriety multi-pulse per frame method, which is being implemented in LGI systems, improves significantly the imaging quality. A significant characteristic of the LGI is its ability to penetrate a disturbing media, such as rain, haze and some fog types. Current LGI systems are based on image intensifier (II) sensors, limiting the system in spectral response, image quality, reliability and cost. A novel propriety optical gating module was developed in Elbit, untying the dependency of LGI system on II. The optical gating module is not bounded to the radiance wavelength and positioned between the system optics and the sensor. This optical gating method supports the use of conventional solid state sensors. By selecting the appropriate solid state sensor, the new LGI systems can operate at any desired wavelength. In this paper we present the new gating method characteristics, performance and its advantages over the II gating method. The use of the gated imaging systems is described in a variety of applications, including results from latest field experiments.

  11. Feature extraction from mammographic images using fast marching methods

    International Nuclear Information System (INIS)

    Bottigli, U.; Golosio, B.


    Features extraction from medical images represents a fundamental step for shape recognition and diagnostic support. The present work faces the problem of the detection of large features, such as massive lesions and organ contours, from mammographic images. The regions of interest are often characterized by an average grayness intensity that is different from the surrounding. In most cases, however, the desired features cannot be extracted by simple gray level thresholding, because of image noise and non-uniform density of the surrounding tissue. In this work, edge detection is achieved through the fast marching method (Level Set Methods and Fast Marching Methods, Cambridge University Press, Cambridge, 1999), which is based on the theory of interface evolution. Starting from a seed point in the shape of interest, a front is generated which evolves according to an appropriate speed function. Such function is expressed in terms of geometric properties of the evolving interface and of image properties, and should become zero when the front reaches the desired boundary. Some examples of application of such method to mammographic images from the CALMA database (Nucl. Instr. and Meth. A 460 (2001) 107) are presented here and discussed

  12. SAR Data Fusion Imaging Method Oriented to Target Feature Extraction

    Directory of Open Access Journals (Sweden)

    Yang Wei


    Full Text Available To deal with the difficulty for target outlines extracting precisely due to neglect of target scattering characteristic variation during the processing of high-resolution space-borne SAR data, a novel fusion imaging method is proposed oriented to target feature extraction. Firstly, several important aspects that affect target feature extraction and SAR image quality are analyzed, including curved orbit, stop-and-go approximation, atmospheric delay, and high-order residual phase error. Furthermore, the corresponding compensation methods are addressed as well. Based on the analysis, the mathematical model of SAR echo combined with target space-time spectrum is established for explaining the space-time-frequency change rule of target scattering characteristic. Moreover, a fusion imaging strategy and method under high-resolution and ultra-large observation angle range conditions are put forward to improve SAR quality by fusion processing in range-doppler and image domain. Finally, simulations based on typical military targets are used to verify the effectiveness of the fusion imaging method.

  13. Images Encryption Method using Steganographic LSB Method, AES and RSA algorithm (United States)

    Moumen, Abdelkader; Sissaoui, Hocine


    Vulnerability of communication of digital images is an extremely important issue nowadays, particularly when the images are communicated through insecure channels. To improve communication security, many cryptosystems have been presented in the image encryption literature. This paper proposes a novel image encryption technique based on an algorithm that is faster than current methods. The proposed algorithm eliminates the step in which the secrete key is shared during the encryption process. It is formulated based on the symmetric encryption, asymmetric encryption and steganography theories. The image is encrypted using a symmetric algorithm, then, the secret key is encrypted by means of an asymmetrical algorithm and it is hidden in the ciphered image using a least significant bits steganographic scheme. The analysis results show that while enjoying the faster computation, our method performs close to optimal in terms of accuracy.

  14. Ortho Image and DTM Generation with Intelligent Methods (United States)

    Bagheri, H.; Sadeghian, S.


    Nowadays the artificial intelligent algorithms has considered in GIS and remote sensing. Genetic algorithm and artificial neural network are two intelligent methods that are used for optimizing of image processing programs such as edge extraction and etc. these algorithms are very useful for solving of complex program. In this paper, the ability and application of genetic algorithm and artificial neural network in geospatial production process like geometric modelling of satellite images for ortho photo generation and height interpolation in raster Digital Terrain Model production process is discussed. In first, the geometric potential of Ikonos-2 and Worldview-2 with rational functions, 2D & 3D polynomials were tested. Also comprehensive experiments have been carried out to evaluate the viability of the genetic algorithm for optimization of rational function, 2D & 3D polynomials. Considering the quality of Ground Control Points, the accuracy (RMSE) with genetic algorithm and 3D polynomials method for Ikonos-2 Geo image was 0.508 pixel sizes and the accuracy (RMSE) with GA algorithm and rational function method for Worldview-2 image was 0.930 pixel sizes. For more another optimization artificial intelligent methods, neural networks were used. With the use of perceptron network in Worldview-2 image, a result of 0.84 pixel sizes with 4 neurons in middle layer was gained. The final conclusion was that with artificial intelligent algorithms it is possible to optimize the existing models and have better results than usual ones. Finally the artificial intelligence methods, like genetic algorithms as well as neural networks, were examined on sample data for optimizing interpolation and for generating Digital Terrain Models. The results then were compared with existing conventional methods and it appeared that these methods have a high capacity in heights interpolation and that using these networks for interpolating and optimizing the weighting methods based on inverse

  15. A systematic desaturation method for images from the Atmospheric Imaging Assembly in the Solar Dynamics Observatory. (United States)

    Torre, Gabriele; Schwartz, Richard; Piana, Michele; Massone, Anna Maria; Benvenuto, Federico


    The fine spatial resolution of the SDO AIA CCD's is often destroyed by the charge in saturated pixels overflowing into a swath of neighboring cells during fast rising solar flares. Automated exposure control can only mitigate this issue to a degree and it has other deleterious effects. Our method addresses the desaturation problem for AIA images as an image reconstruction problem in which the information content of the diffraction fringes, generated by the interaction between the incoming radiation and the hardware of the spacecraft, is exploited to recover the true image intensities within the primary saturated core of the image. This methodology takes advantage of some well defined techniques like cross-correlation and the Expectation Maximization method to invert the direct relation between the diffraction fringes intensities and the true flux intensities. During this talk a complete overview on the structure of the method will be provided, besides some reliability tests obtained by its application against synthetic and real data.

  16. Deep kernel learning method for SAR image target recognition (United States)

    Chen, Xiuyuan; Peng, Xiyuan; Duan, Ran; Li, Junbao


    With the development of deep learning, research on image target recognition has made great progress in recent years. Remote sensing detection urgently requires target recognition for military, geographic, and other scientific research. This paper aims to solve the synthetic aperture radar image target recognition problem by combining deep and kernel learning. The model, which has a multilayer multiple kernel structure, is optimized layer by layer with the parameters of Support Vector Machine and a gradient descent algorithm. This new deep kernel learning method improves accuracy and achieves competitive recognition results compared with other learning methods.

  17. Studying depression using imaging and machine learning methods

    Directory of Open Access Journals (Sweden)

    Meenal J. Patel


    Full Text Available Depression is a complex clinical entity that can pose challenges for clinicians regarding both accurate diagnosis and effective timely treatment. These challenges have prompted the development of multiple machine learning methods to help improve the management of this disease. These methods utilize anatomical and physiological data acquired from neuroimaging to create models that can identify depressed patients vs. non-depressed patients and predict treatment outcomes. This article (1 presents a background on depression, imaging, and machine learning methodologies; (2 reviews methodologies of past studies that have used imaging and machine learning to study depression; and (3 suggests directions for future depression-related studies.

  18. Studying depression using imaging and machine learning methods. (United States)

    Patel, Meenal J; Khalaf, Alexander; Aizenstein, Howard J


    Depression is a complex clinical entity that can pose challenges for clinicians regarding both accurate diagnosis and effective timely treatment. These challenges have prompted the development of multiple machine learning methods to help improve the management of this disease. These methods utilize anatomical and physiological data acquired from neuroimaging to create models that can identify depressed patients vs. non-depressed patients and predict treatment outcomes. This article (1) presents a background on depression, imaging, and machine learning methodologies; (2) reviews methodologies of past studies that have used imaging and machine learning to study depression; and (3) suggests directions for future depression-related studies.

  19. Quantitative Nuclear Medicine Imaging: Concepts, Requirements and Methods

    Energy Technology Data Exchange (ETDEWEB)



    The absolute quantification of radionuclide distribution has been a goal since the early days of nuclear medicine. Nevertheless, the apparent complexity and sometimes limited accuracy of these methods have prevented them from being widely used in important applications such as targeted radionuclide therapy or kinetic analysis. The intricacy of the effects degrading nuclear medicine images and the lack of availability of adequate methods to compensate for these effects have frequently been seen as insurmountable obstacles in the use of quantitative nuclear medicine in clinical institutions. In the last few decades, several research groups have consistently devoted their efforts to the filling of these gaps. As a result, many efficient methods are now available that make quantification a clinical reality, provided appropriate compensation tools are used. Despite these efforts, many clinical institutions still lack the knowledge and tools to adequately measure and estimate the accumulated activities in the human body, thereby using potentially outdated protocols and procedures. The purpose of the present publication is to review the current state of the art of image quantification and to provide medical physicists and other related professionals facing quantification tasks with a solid background of tools and methods. It describes and analyses the physical effects that degrade image quality and affect the accuracy of quantification, and describes methods to compensate for them in planar, single photon emission computed tomography (SPECT) and positron emission tomography (PET) images. The fast paced development of the computational infrastructure, both hardware and software, has made drastic changes in the ways image quantification is now performed. The measuring equipment has evolved from the simple blind probes to planar and three dimensional imaging, supported by SPECT, PET and hybrid equipment. Methods of iterative reconstruction have been developed to allow for

  20. Robust and efficient method for matching features in omnidirectional images (United States)

    Zhu, Qinyi; Zhang, Zhijiang; Zeng, Dan


    Binary descriptors have been widely used in many real-time applications due to their efficiency. These descriptors are commonly designed for perspective images but perform poorly on omnidirectional images, which are severely distorted. To address this issue, this paper proposes tangent plane BRIEF (TPBRIEF) and adapted log polar grid-based motion statistics (ALPGMS). TPBRIEF projects keypoints to a unit sphere and applies the fixed test set in BRIEF descriptor on the tangent plane of the unit sphere. The fixed test set is then backprojected onto the original distorted images to construct the distortion invariant descriptor. TPBRIEF directly enables keypoint detecting and feature describing on original distorted images, whereas other approaches correct the distortion through image resampling, which introduces artifacts and adds time cost. With ALPGMS, omnidirectional images are divided into circular arches named adapted log polar grids. Whether a match is true or false is then determined by simply thresholding the match numbers in a grid pair where the two matched points located. Experiments show that TPBRIEF greatly improves the feature matching accuracy and ALPGMS robustly removes wrong matches. Our proposed method outperforms the state-of-the-art methods.

  1. MR imaging methods for assessing fetal brain development. (United States)

    Rutherford, Mary; Jiang, Shuzhou; Allsop, Joanna; Perkins, Lucinda; Srinivasan, Latha; Hayat, Tayyib; Kumar, Sailesh; Hajnal, Jo


    Fetal magnetic resonance imaging provides an ideal tool for investigating growth and development of the brain in vivo. Current imaging methods have been hampered by fetal motion but recent advances in image acquisition can produce high signal to noise, high resolution 3-dimensional datasets suitable for objective quantification by state of the art post acquisition computer programs. Continuing development of imaging techniques will allow a unique insight into the developing brain, more specifically process of cell migration, axonal pathway formation, and cortical maturation. Accurate quantification of these developmental processes in the normal fetus will allow us to identify subtle deviations from normal during the second and third trimester of pregnancy either in the compromised fetus or in infants born prematurely.

  2. A proposed assessment method for image of regional educational institutions

    Directory of Open Access Journals (Sweden)

    Kataeva Natalya


    Full Text Available Market of educational services in the current Russian economic conditions is a complex of a huge variety of educational institutions. Market of educational services is already experiencing a significant influence of the demographic situation in Russia. This means that higher education institutions are forced to fight in a tough competition for high school students. Increased competition in the educational market forces universities to find new methods of non-price competition in attraction of potential students and throughout own educational and economic activities. Commercialization of education places universities in a single plane with commercial companies who study a positive perception of the image and reputation as a competitive advantage, which is quite acceptable for use in strategic and current activities of higher education institutions to ensure the competitiveness of educational services and educational institution in whole. Nevertheless, due to lack of evidence-based proposals in this area there is a need for scientific research in terms of justification of organizational and methodological aspects of image use as a factor in the competitiveness of the higher education institution. Theoretically and practically there are different methods and ways of evaluating the company’s image. The article provides a comparative assessment of the existing valuation methods of corporate image and the author’s method of estimating the image of higher education institutions based on the key influencing factors. The method has been tested on the Vyatka State Agricultural Academy (Russia. The results also indicate the strengths and weaknesses of the institution, highlights ways of improving, and adjusts the efforts for image improvement.

  3. Guidance for Methods Descriptions Used in Preclinical Imaging Papers

    Directory of Open Access Journals (Sweden)

    David Stout


    Full Text Available Preclinical molecular imaging is a rapidly growing field, where new imaging systems, methods, and biological findings are constantly being developed or discovered. Imaging systems and the associated software usually have multiple options for generating data, which is often overlooked but is essential when reporting the methods used to create and analyze data. Similarly, the ways in which animals are housed, handled, and treated to create physiologically based data must be well described in order that the findings be relevant, useful, and reproducible. There are frequently new developments for metabolic imaging methods. Thus, specific reporting requirements are difficult to establish; however, it remains essential to adequately report how the data have been collected, processed, and analyzed. To assist with future manuscript submissions, this article aims to provide guidelines of what details to report for several of the most common imaging modalities. Examples are provided in an attempt to give comprehensive, succinct descriptions of the essential items to report about the experimental process.

  4. Novel axolotl cardiac function analysis method using magnetic resonance imaging.

    Directory of Open Access Journals (Sweden)

    Pedro Gomes Sanches

    Full Text Available The salamander axolotl is capable of complete regeneration of amputated heart tissue. However, non-invasive imaging tools for assessing its cardiac function were so far not employed. In this study, cardiac magnetic resonance imaging is introduced as a non-invasive technique to image heart function of axolotls. Three axolotls were imaged with magnetic resonance imaging using a retrospectively gated Fast Low Angle Shot cine sequence. Within one scanning session the axolotl heart was imaged three times in all planes, consecutively. Heart rate, ejection fraction, stroke volume and cardiac output were calculated using three techniques: (1 combined long-axis, (2 short-axis series, and (3 ultrasound (control for heart rate only. All values are presented as mean ± standard deviation. Heart rate (beats per minute among different animals was 32.2±6.0 (long axis, 30.4±5.5 (short axis and 32.7±4.9 (ultrasound and statistically similar regardless of the imaging method (p > 0.05. Ejection fraction (% was 59.6±10.8 (long axis and 48.1±11.3 (short axis and it differed significantly (p = 0.019. Stroke volume (μl/beat was 133.7±33.7 (long axis and 93.2±31.2 (short axis, also differed significantly (p = 0.015. Calculations were consistent among the animals and over three repeated measurements. The heart rate varied depending on depth of anaesthesia. We described a new method for defining and imaging the anatomical planes of the axolotl heart and propose one of our techniques (long axis analysis may prove useful in defining cardiac function in regenerating axolotl hearts.

  5. Peptide-assembled graphene oxide as fluorescent turn-on sensor for ultrasensitive Lipopolysaccharide (Endotoxin detection

    Directory of Open Access Journals (Sweden)

    Seng Koon Lim


    Full Text Available Introduction: Lipopolysaccharide (LPS, or endotoxin, a major component in the outer cell membrane of Gram-negative bacteria is a very powerful and toxic inflammatory stimulator, resulting in sepsis or septic shock, a significant medical problem affecting about 700 000 patients and causing 250 000 casualties annually in the United States itself. The detection of LPS is highly importance. However, the currently used enzymatic limulus amebocyte lysate assay is highly susceptible to changes in temperature and pH, interference factors, and requires cumbersome sample preparation. A more cost-effective, sensitive and robust detection method is needed. Objective: To design and develop biosensor for LPS detection by assembling a LPS-binding peptide (as LPS receptor with graphene oxide (GO, as fluorescence quencher. Methods: GO was synthesized using a modified Hummer’s method. A synthetic LPS-binding peptide was designed, fluorescent labelled, and assembled with GO in PBS buffer solution. The fluorescence recovery of the peptide-GO was measured upon addition of LPS from Gram negative bacteria: E. coli, K. pneumoniae, Samonella Thyphosa, P. aeruginosa, as well as living pathogenic bacteria. Specificity tests were conducted with various biological molecules to evaluate the sensing performance. Results & Discussion: Specific binding of LPS with peptide release the peptides from GO, resulting in fluorescence recovery, allowing ultrasensitive detection of LPS with the limit of detection of 130 pM, the most sensitive synthetic LPS sensors to-date. The LPS sensor is highly selective to LPS than other biological species. Conclusion: We developed a peptide-GO assembled fluorescence sensor for ultrasensitive and specific LPS/endotoxin detection. This is the most sensitive synthetic LPS sensor reported in the world.

  6. Informatics methods to enable sharing of quantitative imaging research data. (United States)

    Levy, Mia A; Freymann, John B; Kirby, Justin S; Fedorov, Andriy; Fennessy, Fiona M; Eschrich, Steven A; Berglund, Anders E; Fenstermacher, David A; Tan, Yongqiang; Guo, Xiaotao; Casavant, Thomas L; Brown, Bartley J; Braun, Terry A; Dekker, Andre; Roelofs, Erik; Mountz, James M; Boada, Fernando; Laymon, Charles; Oborski, Matt; Rubin, Daniel L


    The National Cancer Institute Quantitative Research Network (QIN) is a collaborative research network whose goal is to share data, algorithms and research tools to accelerate quantitative imaging research. A challenge is the variability in tools and analysis platforms used in quantitative imaging. Our goal was to understand the extent of this variation and to develop an approach to enable sharing data and to promote reuse of quantitative imaging data in the community. We performed a survey of the current tools in use by the QIN member sites for representation and storage of their QIN research data including images, image meta-data and clinical data. We identified existing systems and standards for data sharing and their gaps for the QIN use case. We then proposed a system architecture to enable data sharing and collaborative experimentation within the QIN. There are a variety of tools currently used by each QIN institution. We developed a general information system architecture to support the QIN goals. We also describe the remaining architecture gaps we are developing to enable members to share research images and image meta-data across the network. As a research network, the QIN will stimulate quantitative imaging research by pooling data, algorithms and research tools. However, there are gaps in current functional requirements that will need to be met by future informatics development. Special attention must be given to the technical requirements needed to translate these methods into the clinical research workflow to enable validation and qualification of these novel imaging biomarkers. Copyright © 2012 Elsevier Inc. All rights reserved.

  7. Automatic intra-modality brain image registration method

    International Nuclear Information System (INIS)

    Whitaker, J.M.; Ardekani, B.A.; Braun, M.


    Full text: Registration of 3D images of brain of the same or different subjects has potential importance in clinical diagnosis, treatment planning and neurological research. The broad aim of our work is to produce an automatic and robust intra-modality, brain image registration algorithm for intra-subject and inter-subject studies. Our algorithm is composed of two stages. Initial alignment is achieved by finding the values of nine transformation parameters (representing translation, rotation and scale) that minimise the nonoverlapping regions of the head. This is achieved by minimisation of the sum of the exclusive OR of two binary head images, produced using the head extraction procedure described by Ardekani et al. (J Comput Assist Tomogr, 19:613-623, 1995). The initial alignment successfully determines the scale parameters and gross translation and rotation parameters. Fine alignment uses an objective function described for inter-modality registration in Ardekani et al. (ibid.). The algorithm segments one of the images to be aligned into a set of connected components using K-means clustering. Registration is achieved by minimising the K-means variance of the segmentation induced in the other image. Similarity of images of the same modality makes the method attractive for intra-modality registration. A 3D MR image, with voxel dimensions, 2x2x6 mm, was misaligned. The registered image shows visually accurate registration. The average displacement of a pixel from its correct location was measured to be 3.3 mm. The algorithm was tested on intra-subject MR images and was found to produce good qualitative results. Using the data available, the algorithm produced promising qualitative results in intra-subject registration. Further work is necessary in its application to intersubject registration, due to large variability in brain structure between subjects. Clinical evaluation of the algorithm for selected applications is required

  8. A novel magnetic resonance imaging-compatible motor control method for image-guided robotic surgery

    International Nuclear Information System (INIS)

    Suzuki, Takashi; Liao, Hongen; Kobayashi, Etsuko; Sakuma, Ichiro


    For robotic surgery assistance systems that use magnetic resonance imaging (MRI) for guidance, the problem of electromagnetic interference is common. Image quality is particularly degraded if motors are running during scanning. We propose a novel MRI-compatible method considering the pulse sequence of imaging. Motors are driven for a short time when the MRI system stops signal acquisition (i.e., awaiting relaxation of the proton), so the image does not contain noise from the actuators. The MRI system and motor are synchronized using a radio frequency pulse signal (8.5 MHz) as the trigger, which is acquired via a special antenna mounted near the scanner. This method can be widely applied because it only receives part of the scanning signal and neither hardware nor software of the MRI system needs to be changed. As a feasibility evaluation test, we compared the images and signal-to-noise ratios between the cases with and without this method, under the condition that a piezoelectric motor was driven during scanning as a noise source, which was generally used as a MRI-compatible actuator. The results showed no deterioration in image quality and the benefit of the new method even though the choice of available scanning sequences is limited. (author)

  9. Neutron imaging integrated circuit and method for detecting neutrons (United States)

    Nagarkar, Vivek V.; More, Mitali J.


    The present disclosure provides a neutron imaging detector and a method for detecting neutrons. In one example, a method includes providing a neutron imaging detector including plurality of memory cells and a conversion layer on the memory cells, setting one or more of the memory cells to a first charge state, positioning the neutron imaging detector in a neutron environment for a predetermined time period, and reading a state change at one of the memory cells, and measuring a charge state change at one of the plurality of memory cells from the first charge state to a second charge state less than the first charge state, where the charge state change indicates detection of neutrons at said one of the memory cells.

  10. Integration of image exposure time into a modified laser speckle imaging method

    Energy Technology Data Exchange (ETDEWEB)

    RamIrez-San-Juan, J C; Salazar-Hermenegildo, N; Ramos-Garcia, R; Munoz-Lopez, J [Optics Department, INAOE, Puebla (Mexico); Huang, Y C [Department of Electrical Engineering and Computer Science, University of California, Irvine, CA (United States); Choi, B, E-mail: jcram@inaoep.m [Beckman Laser Institute and Medical Clinic, University of California, Irvine, CA (United States)


    Speckle-based methods have been developed to characterize tissue blood flow and perfusion. One such method, called modified laser speckle imaging (mLSI), enables computation of blood flow maps with relatively high spatial resolution. Although it is known that the sensitivity and noise in LSI measurements depend on image exposure time, a fundamental disadvantage of mLSI is that it does not take into account this parameter. In this work, we integrate the exposure time into the mLSI method and provide experimental support of our approach with measurements from an in vitro flow phantom.

  11. Integration of image exposure time into a modified laser speckle imaging method

    International Nuclear Information System (INIS)

    RamIrez-San-Juan, J C; Salazar-Hermenegildo, N; Ramos-Garcia, R; Munoz-Lopez, J; Huang, Y C; Choi, B


    Speckle-based methods have been developed to characterize tissue blood flow and perfusion. One such method, called modified laser speckle imaging (mLSI), enables computation of blood flow maps with relatively high spatial resolution. Although it is known that the sensitivity and noise in LSI measurements depend on image exposure time, a fundamental disadvantage of mLSI is that it does not take into account this parameter. In this work, we integrate the exposure time into the mLSI method and provide experimental support of our approach with measurements from an in vitro flow phantom.

  12. Method of image segmentation using a neural network. Application to MR imaging of brain tumors

    International Nuclear Information System (INIS)

    Engler, E.; Gautherie, M.


    An original method of numerical images segmentation has been developed. This method is based on pixel clustering using a formal neural network configurated by supervised learning of pre-classified examples. The method has been applied to series of MR images of brain tumors (gliomas) with a view to proceed with a 3D-extraction of the tumor volume. This study is part of a project on cancer thermotherapy including the development of a scan-focused ultrasound system of tumor heating and a 3D-numerical thermal model

  13. An improved image non-blind image deblurring method based on FoEs (United States)

    Zhu, Qidan; Sun, Lei


    Traditional non-blind image deblurring algorithms always use maximum a posterior(MAP). MAP estimates involving natural image priors can reduce the ripples effectively in contrast to maximum likelihood(ML). However, they have been found lacking in terms of restoration performance. Based on this issue, we utilize MAP with KL penalty to replace traditional MAP. We develop an image reconstruction algorithm that minimizes the KL divergence between the reference distribution and the prior distribution. The approximate KL penalty can restrain over-smooth caused by MAP. We use three groups of images and Harris corner detection to prove our method. The experimental results show that our algorithm of non-blind image restoration can effectively reduce the ringing effect and exhibit the state-of-the-art deblurring results.

  14. Sampling methods for low-frequency electromagnetic imaging

    International Nuclear Information System (INIS)

    Gebauer, Bastian; Hanke, Martin; Schneider, Christoph


    For the detection of hidden objects by low-frequency electromagnetic imaging the linear sampling method works remarkably well despite the fact that the rigorous mathematical justification is still incomplete. In this work, we give an explanation for this good performance by showing that in the low-frequency limit the measurement operator fulfils the assumptions for the fully justified variant of the linear sampling method, the so-called factorization method. We also show how the method has to be modified in the physically relevant case of electromagnetic imaging with divergence-free currents. We present numerical results to illustrate our findings, and to show that similar performance can be expected for the case of conducting objects and layered backgrounds

  15. Optimized optical clearing method for imaging central nervous system (United States)

    Yu, Tingting; Qi, Yisong; Gong, Hui; Luo, Qingming; Zhu, Dan


    The development of various optical clearing methods provides a great potential for imaging entire central nervous system by combining with multiple-labelling and microscopic imaging techniques. These methods had made certain clearing contributions with respective weaknesses, including tissue deformation, fluorescence quenching, execution complexity and antibody penetration limitation that makes immunostaining of tissue blocks difficult. The passive clarity technique (PACT) bypasses those problems and clears the samples with simple implementation, excellent transparency with fine fluorescence retention, but the passive tissue clearing method needs too long time. In this study, we not only accelerate the clearing speed of brain blocks but also preserve GFP fluorescence well by screening an optimal clearing temperature. The selection of proper temperature will make PACT more applicable, which evidently broaden the application range of this method.

  16. Ultrasensitive SERS detection of mercury based on the assembled gold nanochains. (United States)

    Xu, Liguang; Yin, Honghong; Ma, Wei; Kuang, Hua; Wang, Libing; Xu, Chuanlai


    Mercuric ions (Hg(2+)) mediate the transformation of single-stranded DNA to form double helical DNA by T-Hg(2+)-T interaction between base pairs. With this strategy, DNA modified gold nanoparticles (Au NPs) were assembled into chains which were displayed remarkable surface-enhanced Raman scattering (SERS) signal. Under optimized conditions, the length of gold nanochains was directly proportional to the mercuric ions concentrations over 0.001-0.5 ng mL(-1) and the limit of detection (LOD) in drinking water was as low as 0.45 pg mL(-1). With ultrasensitivity and excellent selectivity, this feasible and simple method is potentially as a promising tool for monitoring of mercury ions in food safety and environmental applications. Copyright © 2014 Elsevier B.V. All rights reserved.

  17. Development of ultrasensitive spectroscopic analysis technology -Development of atomic spectroscopy technology-

    International Nuclear Information System (INIS)

    Cha, Hyung Ki; Song, Ky Seok; Rhee, Young Joo; Baik, Dae Hyun; Shin, Jang Soo; Kim, Duck Hyun; Yang, Ki Ho; Yi, Jong Hoon


    In this project, three principal techniques are developed. The laser photoionization spectrometry (LAPIS) is used for the ultrasensitive detection for heavy metals such as Pb and Cd. The Laser atomic fluorescence spectrometry is applied to the quantitative analysis of the lanthanide and actinide elements. And the DIAL remote sensing system is used for monitoring the ozone concentration and the atmospheric pollution. A time-of-flight mass spectrometer and a high efficient atomic beam generator were designed and manufactured. Various spectroscopic parameters and optimum analytical condition were investigated. By using the laser fluorescence technique, U, Eu and Sm in solution were quantitatively analyzed. The basic researches for the direct analytical method of solid samples were also carried out. The DIAL system for ozone remote sensing was developed and ozone concentration above Taejon were obtained. (Author)

  18. Ultra-sensitive detection of nuclear signatures in support of IAEA safeguards

    International Nuclear Information System (INIS)

    Hotchkis, M.; Child, D.; Tuniz, C.; Williams, M.


    The International Atomic Energy Agency (IAEA) applies a range of ultra-sensitive detection techniques to provide assurance that Member States are in compliance with their safeguards agreements. Environmental samples are collected which can contain minute traces of nuclear material or other evidence. Careful analysis of these samples reveals the nature of the activities undertaken in the vicinity of the sampling point. This paper reviews the analytical techniques that are being applied. To ensure that the IAEA has access to the best available methods, samples are distributed to a group of qualified laboratories around the world for analysis. The Accelerator Mass Spectrometry facility at the Australian Nuclear Science and Technology Organisation (ANSTO) is part of this select group of laboratories, and is the only AMS facility currently accredited with the IAEA. AMS provides the highest sensitivity available for detection of particularly useful signature radioisotopes, including 129 I, 236 U and plutonium isotopes

  19. Image segmentation and particles classification using texture analysis method

    Directory of Open Access Journals (Sweden)

    Mayar Aly Atteya

    Full Text Available Introduction: Ingredients of oily fish include a large amount of polyunsaturated fatty acids, which are important elements in various metabolic processes of humans, and have also been used to prevent diseases. However, in an attempt to reduce cost, recent developments are starting a replace the ingredients of fish oil with products of microalgae, that also produce polyunsaturated fatty acids. To do so, it is important to closely monitor morphological changes in algae cells and monitor their age in order to achieve the best results. This paper aims to describe an advanced vision-based system to automatically detect, classify, and track the organic cells using a recently developed SOPAT-System (Smart On-line Particle Analysis Technology, a photo-optical image acquisition device combined with innovative image analysis software. Methods The proposed method includes image de-noising, binarization and Enhancement, as well as object recognition, localization and classification based on the analysis of particles’ size and texture. Results The methods allowed for correctly computing cell’s size for each particle separately. By computing an area histogram for the input images (1h, 18h, and 42h, the variation could be observed showing a clear increase in cell. Conclusion The proposed method allows for algae particles to be correctly identified with accuracies up to 99% and classified correctly with accuracies up to 100%.

  20. Electrodynamics, Differential Forms and the Method of Images (United States)

    Low, Robert J.


    This paper gives a brief description of how Maxwell's equations are expressed in the language of differential forms and use this to provide an elegant demonstration of how the method of images (well known in electrostatics) also works for electrodynamics in the presence of an infinite plane conducting boundary. The paper should be accessible to an…

  1. The iterative shrinkage method for impulsive noise reduction from images

    International Nuclear Information System (INIS)

    Beygi, Sajjad; Kafashan, Mohammadmehdi; Bahrami, Hamid Reza; Mugler, Dale H


    In this paper, we present a novel scheme to compensate impulsive noise from images using the sparse shrinkage method. In this scheme, we assume the remaining noise after using a simple median filtering in place of corrupted pixels, found by boundary discriminative noise detection method, to be Gaussian additive noise. This assumption will later be verified by the means of simulation. Knowing that the pure image in the discrete wavelet transform (DWT) domain is a sparse vector, we define an optimization problem to minimize the l 0 -norm of the estimated image vector from the noisy one in the DWT domain. l 0 -norm makes the optimization problem a combinatorial optimization problem which is NP-hard to solve. To come up with a solution for our optimization problem, we convert the l 0 -norm problem to a continuous optimization problem which is then solved to find the estimated image with reduced noise. In the simulation and discussion part, the performance of our proposed method in reducing impulsive noise is compared to that of existing methods in the literature. We show that our proposed algorithm generally performs better in terms of both subjective and objective evaluations and is less complex. (paper)

  2. a Hyperspectral Image Classification Method Using Isomap and Rvm (United States)

    Chang, H.; Wang, T.; Fang, H.; Su, Y.


    Classification is one of the most significant applications of hyperspectral image processing and even remote sensing. Though various algorithms have been proposed to implement and improve this application, there are still drawbacks in traditional classification methods. Thus further investigations on some aspects, such as dimension reduction, data mining, and rational use of spatial information, should be developed. In this paper, we used a widely utilized global manifold learning approach, isometric feature mapping (ISOMAP), to address the intrinsic nonlinearities of hyperspectral image for dimension reduction. Considering the impropriety of Euclidean distance in spectral measurement, we applied spectral angle (SA) for substitute when constructed the neighbourhood graph. Then, relevance vector machines (RVM) was introduced to implement classification instead of support vector machines (SVM) for simplicity, generalization and sparsity. Therefore, a probability result could be obtained rather than a less convincing binary result. Moreover, taking into account the spatial information of the hyperspectral image, we employ a spatial vector formed by different classes' ratios around the pixel. At last, we combined the probability results and spatial factors with a criterion to decide the final classification result. To verify the proposed method, we have implemented multiple experiments with standard hyperspectral images compared with some other methods. The results and different evaluation indexes illustrated the effectiveness of our method.

  3. Applications of γ-ray image method to astronomy

    International Nuclear Information System (INIS)

    Wuensche, C.A.; Braga, J.; Jayanthi, U.B.; Villela, T.


    The use of codified mask technique in a gamma ray telescope is presented. The image reconstruction method is described showing the mask operation. The signal/noise relation for redundant uniform arrangements which constitute the mask, is discussed. The MASCO telescope is described in detail showing the main characteristics of project. (M.C.K.)

  4. Monitoring a chemical plume remediation via the radio imaging method

    International Nuclear Information System (INIS)

    McCorkle, R.W.; Spence, T.; Linder, K.E.; Betsill, J.D.


    In this paper, the authors present the results of a site characterization, monitoring, and remediation effort at Sandia National Laboratories (SNL). The primary objective of the study is to determine the feasibility of using the Radio Imaging Method (RIM) to solve a near-surface waste site characterization problem. The goals are to demonstrate the method during the site characterization phase, then continue with an in-situ monitoring and analysis of the remediation process

  5. A single-image method of aberration retrieval for imaging systems under partially coherent illumination

    International Nuclear Information System (INIS)

    Xu, Shuang; Liu, Shiyuan; Zhang, Chuanwei; Wei, Haiqing


    We propose a method for retrieving small lens aberrations in optical imaging systems under partially coherent illumination, which only requires to measure one single defocused image of intensity. By deriving a linear theory of imaging systems, we obtain a generalized formulation of aberration sensitivity in a matrix form, which provides a set of analytic kernels that relate the measured intensity distribution directly to the unknown Zernike coefficients. Sensitivity analysis is performed and test patterns are optimized to ensure well-posedness of the inverse problem. Optical lithography simulations have validated the theoretical derivation and confirmed its simplicity and superior performance in retrieving small lens aberrations. (fast track communication)

  6. Development of CCD Imaging System Using Thermoelectric Cooling Method

    Directory of Open Access Journals (Sweden)

    Youngsik Park


    Full Text Available We developed low light CCD imaging system using thermoelectric cooling method collaboration with a company to design a commercial model. It consists of Kodak KAF-0401E (768x512 pixels CCD chip,thermoelectric module manufactured by Thermotek. This TEC system can reach an operative temperature of -25deg. We employed an Uniblitz VS25S shutter and it has capability a minimum exposure time 80ms. The system components are an interface card using a Korea Astronomy Observatory (hereafter KAO ISA bus controller, image acquisition with AD9816 chip, that is 12bit video processor. The performance test with this imaging system showed good operation within the initial specification of our design. It shows a dark current less than 0.4e-/pixel/sec at a temperature of -10deg, a linearity 99.9+/-0.1%, gain 4.24e-adu, and system noise is 25.3e- (rms. For low temperature CCD operation, we designed a TEC, which uses a one-stage peltier module and forced air heat exchanger. This TEC imaging system enables accurate photometry (+/-0.01mag even though the CCD is not at 'conventional' cryogenic temperatures (140K. The system can be a useful instrument for any other imaging applications. Finally, with this system, we obtained several images of astronomical objects for system performance tests.

  7. Influence of image reconstruction methods on statistical parametric mapping of brain PET images

    International Nuclear Information System (INIS)

    Yin Dayi; Chen Yingmao; Yao Shulin; Shao Mingzhe; Yin Ling; Tian Jiahe; Cui Hongyan


    Objective: Statistic parametric mapping (SPM) was widely recognized as an useful tool in brain function study. The aim of this study was to investigate if imaging reconstruction algorithm of PET images could influence SPM of brain. Methods: PET imaging of whole brain was performed in six normal volunteers. Each volunteer had two scans with true and false acupuncturing. The PET scans were reconstructed using ordered subsets expectation maximization (OSEM) and filtered back projection (FBP) with 3 varied parameters respectively. The images were realigned, normalized and smoothed using SPM program. The difference between true and false acupuncture scans was tested using a matched pair t test at every voxel. Results: (1) SPM corrected multiple comparison (P corrected uncorrected <0.001): SPM derived from the images with different reconstruction method were different. The largest difference, in number and position of the activated voxels, was noticed between FBP and OSEM re- construction algorithm. Conclusions: The method of PET image reconstruction could influence the results of SPM uncorrected multiple comparison. Attention should be paid when the conclusion was drawn using SPM uncorrected multiple comparison. (authors)

  8. Fast nonconvex nonsmooth minimization methods for image restoration and reconstruction. (United States)

    Nikolova, Mila; Ng, Michael K; Tam, Chi-Pan


    Nonconvex nonsmooth regularization has advantages over convex regularization for restoring images with neat edges. However, its practical interest used to be limited by the difficulty of the computational stage which requires a nonconvex nonsmooth minimization. In this paper, we deal with nonconvex nonsmooth minimization methods for image restoration and reconstruction. Our theoretical results show that the solution of the nonconvex nonsmooth minimization problem is composed of constant regions surrounded by closed contours and neat edges. The main goal of this paper is to develop fast minimization algorithms to solve the nonconvex nonsmooth minimization problem. Our experimental results show that the effectiveness and efficiency of the proposed algorithms.

  9. Single photon imaging and timing array sensor apparatus and method (United States)

    Smith, R. Clayton


    An apparatus and method are disclosed for generating a three-dimension image of an object or target. The apparatus is comprised of a photon source for emitting a photon at a target. The emitted photons are received by a photon receiver for receiving the photon when reflected from the target. The photon receiver determines a reflection time of the photon and further determines an arrival position of the photon on the photon receiver. An analyzer is communicatively coupled to the photon receiver, wherein the analyzer generates a three-dimensional image of the object based upon the reflection time and the arrival position.

  10. Method of imaging the electrical conductivity distribution of a subsurface (United States)

    Johnson, Timothy C.


    A method of imaging electrical conductivity distribution of a subsurface containing metallic structures with known locations and dimensions is disclosed. Current is injected into the subsurface to measure electrical potentials using multiple sets of electrodes, thus generating electrical resistivity tomography measurements. A numeric code is applied to simulate the measured potentials in the presence of the metallic structures. An inversion code is applied that utilizes the electrical resistivity tomography measurements and the simulated measured potentials to image the subsurface electrical conductivity distribution and remove effects of the subsurface metallic structures with known locations and dimensions.

  11. Interpolation decoding method with variable parameters for fractal image compression

    International Nuclear Information System (INIS)

    He Chuanjiang; Li Gaoping; Shen Xiaona


    The interpolation fractal decoding method, which is introduced by [He C, Yang SX, Huang X. Progressive decoding method for fractal image compression. IEE Proc Vis Image Signal Process 2004;3:207-13], involves generating progressively the decoded image by means of an interpolation iterative procedure with a constant parameter. It is well-known that the majority of image details are added at the first steps of iterations in the conventional fractal decoding; hence the constant parameter for the interpolation decoding method must be set as a smaller value in order to achieve a better progressive decoding. However, it needs to take an extremely large number of iterations to converge. It is thus reasonable for some applications to slow down the iterative process at the first stages of decoding and then to accelerate it afterwards (e.g., at some iteration as we need). To achieve the goal, this paper proposed an interpolation decoding scheme with variable (iteration-dependent) parameters and proved the convergence of the decoding process mathematically. Experimental results demonstrate that the proposed scheme has really achieved the above-mentioned goal

  12. Cardiodiagnostic imaging. MRT, CT, echocardiography and other methods

    International Nuclear Information System (INIS)

    Erbel, R.; Kreitner, K.F.; Barkhausen, J.; Thelen, M.


    The book presents a differentiated approach to cardiac imaging. The focus is n cardio-MR/-CT and echocardiography. These are highly complex methods involving new equipment, new protocols and indications. The techniques are new and difficult to learn for everybody concerned. MR, CT and echocardiography must always be viewed in the context of other diagnostic methods. The interdisciplinary approach of the book addresses both radiologists and cardiologists and relies on the vast experience of the authors. The book offers more than 500 large high-quality reference images reflecting the latest state of the art. It has amethodological section in which the current methods are described (X-ray, echocardiography, nuclear medicine, angiography, CT, MRT etc.) along with their advantages and shortcomings, and a clinical section in which the main indications are described in the common standardized way (anatomy, clinical picture, interpretation, differential diagnosis). (orig.)

  13. Meshfree Local Radial Basis Function Collocation Method with Image Nodes

    Energy Technology Data Exchange (ETDEWEB)

    Baek, Seung Ki; Kim, Minjae [Pukyong National University, Busan (Korea, Republic of)


    We numerically solve two-dimensional heat diffusion problems by using a simple variant of the meshfree local radial-basis function (RBF) collocation method. The main idea is to include an additional set of sample nodes outside the problem domain, similarly to the method of images in electrostatics, to perform collocation on the domain boundaries. We can thereby take into account the temperature profile as well as its gradients specified by boundary conditions at the same time, which holds true even for a node where two or more boundaries meet with different boundary conditions. We argue that the image method is computationally efficient when combined with the local RBF collocation method, whereas the addition of image nodes becomes very costly in case of the global collocation. We apply our modified method to a benchmark test of a boundary value problem, and find that this simple modification reduces the maximum error from the analytic solution significantly. The reduction is small for an initial value problem with simpler boundary conditions. We observe increased numerical instability, which has to be compensated for by a sufficient number of sample nodes and/or more careful parameter choices for time integration.

  14. The gridding method for image reconstruction by Fourier transformation

    International Nuclear Information System (INIS)

    Schomberg, H.; Timmer, J.


    This paper explores a computational method for reconstructing an n-dimensional signal f from a sampled version of its Fourier transform f. The method involves a window function w and proceeds in three steps. First, the convolution g = w * f is computed numerically on a Cartesian grid, using the available samples of f. Then, g = wf is computed via the inverse discrete Fourier transform, and finally f is obtained as g/w. Due to the smoothing effect of the convolution, evaluating w * f is much less error prone than merely interpolating f. The method was originally devised for image reconstruction in radio astronomy, but is actually applicable to a broad range of reconstructive imaging methods, including magnetic resonance imaging and computed tomography. In particular, it provides a fast and accurate alternative to the filtered backprojection. The basic method has several variants with other applications, such as the equidistant resampling of arbitrarily sampled signals or the fast computation of the Radon (Hough) transform

  15. Total variation superiorized conjugate gradient method for image reconstruction (United States)

    Zibetti, Marcelo V. W.; Lin, Chuan; Herman, Gabor T.


    The conjugate gradient (CG) method is commonly used for the relatively-rapid solution of least squares problems. In image reconstruction, the problem can be ill-posed and also contaminated by noise; due to this, approaches such as regularization should be utilized. Total variation (TV) is a useful regularization penalty, frequently utilized in image reconstruction for generating images with sharp edges. When a non-quadratic norm is selected for regularization, as is the case for TV, then it is no longer possible to use CG. Non-linear CG is an alternative, but it does not share the efficiency that CG shows with least squares and methods such as fast iterative shrinkage-thresholding algorithms (FISTA) are preferred for problems with TV norm. A different approach to including prior information is superiorization. In this paper it is shown that the conjugate gradient method can be superiorized. Five different CG variants are proposed, including preconditioned CG. The CG methods superiorized by the total variation norm are presented and their performance in image reconstruction is demonstrated. It is illustrated that some of the proposed variants of the superiorized CG method can produce reconstructions of superior quality to those produced by FISTA and in less computational time, due to the speed of the original CG for least squares problems. In the Appendix we examine the behavior of one of the superiorized CG methods (we call it S-CG); one of its input parameters is a positive number ɛ. It is proved that, for any given ɛ that is greater than the half-squared-residual for the least squares solution, S-CG terminates in a finite number of steps with an output for which the half-squared-residual is less than or equal to ɛ. Importantly, it is also the case that the output will have a lower value of TV than what would be provided by unsuperiorized CG for the same value ɛ of the half-squared residual.

  16. The method of images and Green's function for spherical domains

    International Nuclear Information System (INIS)

    Gutkin, Eugene; Newton, Paul K


    Motivated by problems in electrostatics and vortex dynamics, we develop two general methods for constructing Green's function for simply connected domains on the surface of the unit sphere. We prove a Riemann mapping theorem showing that such domains can be conformally mapped to the upper hemisphere. We then categorize all domains on the sphere for which Green's function can be constructed by an extension of the classical method of images. We illustrate our methods by several examples, such as the upper hemisphere, geodesic triangles, and latitudinal rectangles. We describe the point vortex motion in these domains, which is governed by a Hamiltonian determined by the Dirichlet Green's function

  17. The best printing methods to print satellite images

    Directory of Open Access Journals (Sweden)

    G.A. Yousif


    In this paper different printing systems were used to print an image of SPOT-4 satellite, caver part of Sharm Elshekh area, Sinai, Egypt, on the same type of paper as much as possible, especially in the photography. This step is followed by measuring the experimental data, and analyzed colors to determine the best printing systems for satellite image printing data. The laser system is the more printing system where produce a wider range of color and highest densities of ink and access much color detail. Followed by the offset system which it recorded the best dot gain. Moreover, the study shows that it can use the advantages of each method according to the satellite image color and quantity to be produced.


    Directory of Open Access Journals (Sweden)

    D. S. Korshunov


    Full Text Available The paper deals with an approach for image quality improvement of the space objects in the visible range of electromagnetic wave spectrum. The proposed method is based on the joint taking into account of both the motion velocity of the space supervisory apparatus and a space object observed in the near-earth space when the time of photo-detector exposure is chosen. The timing of exposure is carried out by light-signal characteristics, which determines the optimal value of the charge package formed in the charge-coupled device being irradiated. Thus, the parameters of onboard observation equipment can be selected, which provides space images suitable for interpretation. The linear resolving capacity is used as quality indicator for space images, giving a complete picture for the image contrast and geometric properties of the object on the photo. Observation scenario modeling of the space object, done by sputnik-inspector, has shown the possibility of increasing the linear resolution up to10% - 20% or up to 40% - 50% depending on the non-complanarity angle at the movement along orbits. The proposed approach to the increase of photographs quality provides getting sharp and highcontrast images of space objects by the optical-electronic equipment of the space-based remote sensing. The usage of these images makes it possible to detect in time the space technology failures, which are the result of its exploitation in the nearearth space. The proposed method can be also applied at the stage of space systems design for optical-electronic surveillance in computer models used for facilities assessment of the shooting equipment information tract.

  19. Optical image encryption method based on incoherent imaging and polarized light encoding (United States)

    Wang, Q.; Xiong, D.; Alfalou, A.; Brosseau, C.


    We propose an incoherent encoding system for image encryption based on a polarized encoding method combined with an incoherent imaging. Incoherent imaging is the core component of this proposal, in which the incoherent point-spread function (PSF) of the imaging system serves as the main key to encode the input intensity distribution thanks to a convolution operation. An array of retarders and polarizers is placed on the input plane of the imaging structure to encrypt the polarized state of light based on Mueller polarization calculus. The proposal makes full use of randomness of polarization parameters and incoherent PSF so that a multidimensional key space is generated to deal with illegal attacks. Mueller polarization calculus and incoherent illumination of imaging structure ensure that only intensity information is manipulated. Another key advantage is that complicated processing and recording related to a complex-valued signal are avoided. The encoded information is just an intensity distribution, which is advantageous for data storage and transition because information expansion accompanying conventional encryption methods is also avoided. The decryption procedure can be performed digitally or using optoelectronic devices. Numerical simulation tests demonstrate the validity of the proposed scheme.

  20. An automated image processing method for classification of diabetic retinopathy stages from conjunctival microvasculature images (United States)

    Khansari, Maziyar M.; O'Neill, William; Penn, Richard; Blair, Norman P.; Chau, Felix; Shahidi, Mahnaz


    The conjunctiva is a densely vascularized tissue of the eye that provides an opportunity for imaging of human microcirculation. In the current study, automated fine structure analysis of conjunctival microvasculature images was performed to discriminate stages of diabetic retinopathy (DR). The study population consisted of one group of nondiabetic control subjects (NC) and 3 groups of diabetic subjects, with no clinical DR (NDR), non-proliferative DR (NPDR), or proliferative DR (PDR). Ordinary least square regression and Fisher linear discriminant analyses were performed to automatically discriminate images between group pairs of subjects. Human observers who were masked to the grouping of subjects performed image discrimination between group pairs. Over 80% and 70% of images of subjects with clinical and non-clinical DR were correctly discriminated by the automated method, respectively. The discrimination rates of the automated method were higher than human observers. The fine structure analysis of conjunctival microvasculature images provided discrimination of DR stages and can be potentially useful for DR screening and monitoring.

  1. a Comparative Case Study of Reflection Seismic Imaging Method (United States)

    Alamooti, M.; Aydin, A.


    Seismic imaging is the most common means of gathering information about subsurface structural features. The accuracy of seismic images may be highly variable depending on the complexity of the subsurface and on how seismic data is processed. One of the crucial steps in this process, especially in layered sequences with complicated structure, is the time and/or depth migration of seismic data.The primary purpose of the migration is to increase the spatial resolution of seismic images by repositioning the recorded seismic signal back to its original point of reflection in time/space, which enhances information about complex structure. In this study, our objective is to process a seismic data set (courtesy of the University of South Carolina) to generate an image on which the Magruder fault near Allendale SC can be clearly distinguished and its attitude can be accurately depicted. The data was gathered by common mid-point method with 60 geophones equally spaced along an about 550 m long traverse over a nearly flat ground. The results obtained from the application of different migration algorithms (including finite-difference and Kirchhoff) are compared in time and depth domains to investigate the efficiency of each algorithm in reducing the processing time and improving the accuracy of seismic images in reflecting the correct position of the Magruder fault.

  2. The evolving role of new imaging methods in breast screening. (United States)

    Houssami, Nehmat; Ciatto, Stefano


    The potential to avert breast cancer deaths through screening means that efforts continue to identify methods which may enhance early detection. While the role of most new imaging technologies remains in adjunct screening or in the work-up of mammography-detected abnormalities, some of the new breast imaging tests (such as MRI) have roles in screening groups of women defined by increased cancer risk. This paper highlights the evidence and the current role of new breast imaging technologies in screening, focusing on those that have broader application in population screening, including digital mammography, breast ultrasound in women with dense breasts, and computer-aided detection. It highlights that evidence on new imaging in screening comes mostly from non-randomised studies that have quantified test detection capability as adjunct to mammography, or have compared measures of screening performance for new technologies with that of conventional mammography. Two RCTs have provided high-quality evidence on the equivalence of digital and conventional mammography and on outcomes of screen-reading complemented by CAD. Many of these imaging technologies enhance cancer detection but also increase recall and false positives in screening. Copyright © 2011 Elsevier Inc. All rights reserved.

  3. Ectomography - a tomographic method for gamma camera imaging

    International Nuclear Information System (INIS)

    Dale, S.; Edholm, P.E.; Hellstroem, L.G.; Larsson, S.


    In computerised gamma camera imaging the projections are readily obtained in digital form, and the number of picture elements may be relatively few. This condition makes emission techniques suitable for ectomography - a tomographic technique for directly visualising arbitrary sections of the human body. The camera rotates around the patient to acquire different projections in a way similar to SPECT. This method differs from SPECT, however, in that the camera is placed at an angle to the rotational axis, and receives two-dimensional, rather than one-dimensional, projections. Images of body sections are reconstructed by digital filtration and combination of the acquired projections. The main advantages of ectomography - a high and uniform resolution, a low and uniform attenuation and a high signal-to-noise ratio - are obtained when imaging sections close and parallel to a body surface. The filtration eliminates signals representing details outside the section and gives the section a certain thickness. Ectomographic transverse images of a line source and of a human brain have been reconstructed. Details within the sections are correctly visualised and details outside are effectively eliminated. For comparison, the same sections have been imaged with SPECT. (author)

  4. Quantitative magnetic resonance micro-imaging methods for pharmaceutical research. (United States)

    Mantle, M D


    The use of magnetic resonance imaging (MRI) as a tool in pharmaceutical research is now well established and the current literature covers a multitude of different pharmaceutically relevant research areas. This review focuses on the use of quantitative magnetic resonance micro-imaging techniques and how they have been exploited to extract information that is of direct relevance to the pharmaceutical industry. The article is divided into two main areas. The first half outlines the theoretical aspects of magnetic resonance and deals with basic magnetic resonance theory, the effects of nuclear spin-lattice (T(1)), spin-spin (T(2)) relaxation and molecular diffusion upon image quantitation, and discusses the applications of rapid magnetic resonance imaging techniques. In addition to the theory, the review aims to provide some practical guidelines for the pharmaceutical researcher with an interest in MRI as to which MRI pulse sequences/protocols should be used and when. The second half of the article reviews the recent advances and developments that have appeared in the literature concerning the use of quantitative micro-imaging methods to pharmaceutically relevant research. Copyright © 2010 Elsevier B.V. All rights reserved.

  5. A simple method for multiday imaging of slice cultures. (United States)

    Seidl, Armin H; Rubel, Edwin W


    The organotypic slice culture (Stoppini et al. A simple method for organotypic cultures of nervous tissue. 1991;37:173-182) has become the method of choice to answer a variety of questions in neuroscience. For many experiments, however, it would be beneficial to image or manipulate a slice culture repeatedly, for example, over the course of many days. We prepared organotypic slice cultures of the auditory brainstem of P3 and P4 mice and kept them in vitro for up to 4 weeks. Single cells in the auditory brainstem were transfected with plasmids expressing fluorescent proteins by way of electroporation (Haas et al. Single-cell electroporation for gene transfer in vivo. 2001;29:583-591). The culture was then placed in a chamber perfused with oxygenated ACSF and the labeled cell imaged with an inverted wide-field microscope repeatedly for multiple days, recording several time-points per day, before returning the slice to the incubator. We describe a simple method to image a slice culture preparation during the course of multiple days and over many continuous hours, without noticeable damage to the tissue or photobleaching. Our method uses a simple, inexpensive custom-built insulator constructed around the microscope to maintain controlled temperature and uses a perfusion chamber as used for in vitro slice recordings. (c) 2009 Wiley-Liss, Inc.

  6. A method for dynamic subtraction MR imaging of the liver

    Directory of Open Access Journals (Sweden)

    Setti Ernesto


    Full Text Available Abstract Background Subtraction of Dynamic Contrast-Enhanced 3D Magnetic Resonance (DCE-MR volumes can result in images that depict and accurately characterize a variety of liver lesions. However, the diagnostic utility of subtraction images depends on the extent of co-registration between non-enhanced and enhanced volumes. Movement of liver structures during acquisition must be corrected prior to subtraction. Currently available methods are computer intensive. We report a new method for the dynamic subtraction of MR liver images that does not require excessive computer time. Methods Nineteen consecutive patients (median age 45 years; range 37–67 were evaluated by VIBE T1-weighted sequences (TR 5.2 ms, TE 2.6 ms, flip angle 20°, slice thickness 1.5 mm acquired before and 45s after contrast injection. Acquisition parameters were optimized for best portal system enhancement. Pre and post-contrast liver volumes were realigned using our 3D registration method which combines: (a rigid 3D translation using maximization of normalized mutual information (NMI, and (b fast 2D non-rigid registration which employs a complex discrete wavelet transform algorithm to maximize pixel phase correlation and perform multiresolution analysis. Registration performance was assessed quantitatively by NMI. Results The new registration procedure was able to realign liver structures in all 19 patients. NMI increased by about 8% after rigid registration (native vs. rigid registration 0.073 ± 0.031 vs. 0.078 ± 0.031, n.s., paired t-test and by a further 23% (0.096 ± 0.035 vs. 0.078 ± 0.031, p t-test after non-rigid realignment. The overall average NMI increase was 31%. Conclusion This new method for realigning dynamic contrast-enhanced 3D MR volumes of liver leads to subtraction images that enhance diagnostic possibilities for liver lesions.

  7. Brief review of image reconstruction methods for imaging in nuclear medicine

    International Nuclear Information System (INIS)

    Murayama, Hideo


    Emission computed tomography (ECT) has as its major emphasis the quantitative determination of the moment to moment changes in the chemistry and flow physiology of injected or inhaled compounds labeled with radioactive atoms in a human body. The major difference lies in the fact that ECT seeks to describe the location and intensity of sources of emitted photons in an attenuating medium whereas transmission X-ray computed tomography (TCT) seeks to determine the distribution of the attenuating medium. A second important difference between ECT and TCT is that of available statistics. ECT statistics are low because each photon without control in emitting direction must be detected and analyzed, not as in TCT. The following sections review the historical development of image reconstruction methods for imaging in nuclear medicine, relevant intrinsic concepts for image reconstruction on ECT, and current status of volume imaging as well as a unique approach on iterative techniques for ECT. (author). 130 refs

  8. Imaging Method Based on Time Reversal Channel Compensation

    Directory of Open Access Journals (Sweden)

    Bing Li


    Full Text Available The conventional time reversal imaging (TRI method builds imaging function by using the maximal value of signal amplitude. In this circumstance, some remote targets are missed (near-far problem or low resolution is obtained in lossy and/or dispersive media, and too many transceivers are employed to locate targets, which increases the complexity and cost of system. To solve these problems, a novel TRI algorithm is presented in this paper. In order to achieve a high resolution, the signal amplitude corresponding to focal time observed at target position is used to reconstruct the target image. For disposing near-far problem and suppressing spurious images, combining with cross-correlation property and amplitude compensation, channel compensation function (CCF is introduced. Moreover, the complexity and cost of system are reduced by employing only five transceivers to detect four targets whose number is close to that of transceivers. For the sake of demonstrating the practicability of the proposed analytical framework, the numerical experiments are actualized in both nondispersive-lossless (NDL media and dispersive-conductive (DPC media. Results show that the performance of the proposed method is superior to that of conventional TRI algorithm even under few echo signals.

  9. Method for imaging with low frequency electromagnetic fields (United States)

    Lee, Ki H.; Xie, Gan Q.


    A method for imaging with low frequency electromagnetic fields, and for interpreting the electromagnetic data using ray tomography, in order to determine the earth conductivity with high accuracy and resolution. The imaging method includes the steps of placing one or more transmitters, at various positions in a plurality of transmitter holes, and placing a plurality of receivers in a plurality of receiver holes. The transmitters generate electromagnetic signals which diffuse through a medium, such as earth, toward the receivers. The measured diffusion field data H is then transformed into wavefield data U. The traveltimes corresponding to the wavefield data U, are then obtained, by charting the wavefield data U, using a different regularization parameter .alpha. for each transform. The desired property of the medium, such as conductivity, is then derived from the velocity, which in turn is constructed from the wavefield data U using ray tomography.

  10. Method and apparatus for animal positioning in imaging systems (United States)

    Hadjioannou, Arion-Xenofon; Stout, David B.; Silverman, Robert W.


    An apparatus for imaging an animal includes a first mounting surface, a bed sized to support the animal and releasably secured to or integral with the first mounting surface. The apparatus also includes a plurality of straps, each having a first end in a fixed position relative to the bed and a second end for tightening around a limb of the animal. A method for in-vivo imaging of an animal includes providing an animal that has limbs, providing a first mounting surface, and providing a bed removably secured to or integral with the mounting surface and sized to support the animal as well as being coupled to a plurality of straps. The method also includes placing the animal on the bed between the plurality of straps and tightening at least two of the plurality of straps around at least two of the limbs such that the animal is substantially secured in place relative to the bed.

  11. A novel attack method about double-random-phase-encoding-based image hiding method (United States)

    Xu, Hongsheng; Xiao, Zhijun; Zhu, Xianchen


    By using optical image processing techniques, a novel text encryption and hiding method applied by double-random phase-encoding technique is proposed in the paper. The first step is that the secret message is transformed into a 2-dimension array. The higher bits of the elements in the array are used to fill with the bit stream of the secret text, while the lower bits are stored specific values. Then, the transformed array is encoded by double random phase encoding technique. Last, the encoded array is embedded on a public host image to obtain the image embedded with hidden text. The performance of the proposed technique is tested via analytical modeling and test data stream. Experimental results show that the secret text can be recovered either accurately or almost accurately, while maintaining the quality of the host image embedded with hidden data by properly selecting the method of transforming the secret text into an array and the superimposition coefficient.

  12. A deep level set method for image segmentation


    Tang, Min; Valipour, Sepehr; Zhang, Zichen Vincent; Cobzas, Dana; MartinJagersand


    This paper proposes a novel image segmentation approachthat integrates fully convolutional networks (FCNs) with a level setmodel. Compared with a FCN, the integrated method can incorporatesmoothing and prior information to achieve an accurate segmentation.Furthermore, different than using the level set model as a post-processingtool, we integrate it into the training phase to fine-tune the FCN. Thisallows the use of unlabeled data during training in a semi-supervisedsetting. Using two types o...

  13. A Method of Poisson's Ration Imaging Within a Material Part (United States)

    Roth, Don J. (Inventor)


    The present invention is directed to a method of displaying the Poisson's ratio image of a material part. In the present invention, longitudinal data is produced using a longitudinal wave transducer and shear wave data is produced using a shear wave transducer. The respective data is then used to calculate the Poisson's ratio for the entire material part. The Poisson's ratio approximations are then used to display the data.

  14. Soft-tissues Image Processing: Comparison of Traditional Segmentation Methods with 2D active Contour Methods

    Czech Academy of Sciences Publication Activity Database

    Mikulka, J.; Gescheidtová, E.; Bartušek, Karel


    Roč. 12, č. 4 (2012), s. 153-161 ISSN 1335-8871 R&D Projects: GA ČR GAP102/11/0318; GA ČR GAP102/12/1104; GA MŠk ED0017/01/01 Institutional support: RVO:68081731 Keywords : Medical image processing * image segmentation * liver tumor * temporomandibular joint disc * watershed method Subject RIV: JA - Electronics ; Optoelectronics, Electrical Engineering Impact factor: 1.233, year: 2012

  15. Innovative method for ultra-sensitive measurement of krypton isotopes

    International Nuclear Information System (INIS)

    Lavielle, B.; Gilabert, E.; Thomas, B.; Rebeix, R.; Canchel, G.; Moulin, C.; Topin, S.; Pointurier, F.


    Kr 81 (T 1/2 = 2.29*10 5 y), that is produced in the atmosphere by cosmic rays and Kr 85 (T 1/2 = 10.77 y), that is produced in the fission of nuclear fuels, are considered as the right chronometer elements for the dating of underground waters, polar ice, of for tracking civil and military nuclear activities. The use of Kr 81 and Kr 85 implies the development of extracting lines and detection devices with so high accuracy that only a few thousands of krypton atoms can be detected. The approach developed aims at detecting Kr 81 and Kr 85 in 1 liter of air and 20 liters of water. 3 steps are necessary. The first step consists in separating and purifying the krypton. The extraction of gases from water is made by helium bubbling. Chemically active gases like N 2 , CO 2 , CH 4 and O 2 are eliminated by chemical traps while Ar si separated by cryogenics. The second step involves a double focus mass spectrometer designed to perform an important enrichment in Kr 81 and Kr 85 . The last step is the high-accuracy measurement of krypton isotopes performed with the new tool named FAKIR (Facility for Analyzing Krypton Isotopic Ratios) that is based on UV laser ionization and on the mass-discrimination of the ions through their time of flight

  16. Design of a practical model-observer-based image quality assessment method for CT imaging systems (United States)

    Tseng, Hsin-Wu; Fan, Jiahua; Cao, Guangzhi; Kupinski, Matthew A.; Sainath, Paavana


    The channelized Hotelling observer (CHO) is a powerful method for quantitative image quality evaluations of CT systems and their image reconstruction algorithms. It has recently been used to validate the dose reduction capability of iterative image-reconstruction algorithms implemented on CT imaging systems. The use of the CHO for routine and frequent system evaluations is desirable both for quality assurance evaluations as well as further system optimizations. The use of channels substantially reduces the amount of data required to achieve accurate estimates of observer performance. However, the number of scans required is still large even with the use of channels. This work explores different data reduction schemes and designs a new approach that requires only a few CT scans of a phantom. For this work, the leave-one-out likelihood (LOOL) method developed by Hoffbeck and Landgrebe is studied as an efficient method of estimating the covariance matrices needed to compute CHO performance. Three different kinds of approaches are included in the study: a conventional CHO estimation technique with a large sample size, a conventional technique with fewer samples, and the new LOOL-based approach with fewer samples. The mean value and standard deviation of area under ROC curve (AUC) is estimated by shuffle method. Both simulation and real data results indicate that an 80% data reduction can be achieved without loss of accuracy. This data reduction makes the proposed approach a practical tool for routine CT system assessment.

  17. Adult Moyamoya disease angiographic images evolutive characters and treatment methods

    International Nuclear Information System (INIS)

    Qian Jiangnan; Ling Feng


    Objective: To discuss the angiographic images with evolutional characters and the treatment methods of the Moyamoya disease. Methods: The clinical manifestations, the radiographic changes and the comparative analysis between medicine treatment and surgery treatment, together with the laboratory tests findings were analyzed in one cases adult Moyamoya disease during six years. Conclusions: The angiographic characteristics of MMD show the supplied artery trunk stenosis, and followed by occlusion, with later appearance of vascular smoking sign. Medical treatment proved to be of null. Direct or indirect intra or extra cranial vascular anastomosis are effective for treatment

  18. Apparatus and method for motion tracking in brain imaging

    DEFF Research Database (Denmark)


    Disclosed is apparatus and method for motion tracking of a subject in medical brain imaging. The method comprises providing a light projector and a first camera; projecting a first pattern sequence (S1) onto a surface region of the subject with the light projector, wherein the subject is positioned......2,1) based on the detected first pattern sequence (S1'); projecting the second pattern sequence (S2) onto a surface region of the subject with the light projector; detecting the projected second pattern sequence (S2') with the first camera; and determining motion tracking parameters based...

  19. Ultrasensitive surveillance of sensors and processes

    International Nuclear Information System (INIS)

    Wegerich, S.W.; Jarman, K.K.; Gross, K.C.


    A method and apparatus for monitoring a source of data for determining an operating state of a working system are disclosed. The method includes determining a sensor (or source of data) arrangement associated with monitoring the source of data for a system, activating a method for performing a sequential probability ratio test if the data source includes a single data (sensor) source, activating a second method for performing a regression sequential possibility ratio testing procedure if the arrangement includes a pair of sensors (data sources) with signals which are linearly or non-linearly related; activating a third method for performing a bounded angle ratio test procedure if the sensor arrangement includes multiple sensors and utilizing at least one of the first, second and third methods to accumulate sensor signals and determining the operating state of the system

  20. Methods for modeling and quantification in functional imaging by positron emissions tomography and magnetic resonance imaging

    International Nuclear Information System (INIS)

    Costes, Nicolas


    This report presents experiences and researches in the field of in vivo medical imaging by positron emission tomography (PET) and magnetic resonance imaging (MRI). In particular, advances in terms of reconstruction, quantification and modeling in PET are described. The validation of processing and analysis methods is supported by the creation of data by simulation of the imaging process in PET. The recent advances of combined PET/MRI clinical cameras, allowing simultaneous acquisition of molecular/metabolic PET information, and functional/structural MRI information opens the door to unique methodological innovations, exploiting spatial alignment and simultaneity of the PET and MRI signals. It will lead to an increase in accuracy and sensitivity in the measurement of biological phenomena. In this context, the developed projects address new methodological issues related to quantification, and to the respective contributions of MRI or PET information for a reciprocal improvement of the signals of the two modalities. They open perspectives for combined analysis of the two imaging techniques, allowing optimal use of synchronous, anatomical, molecular and functional information for brain imaging. These innovative concepts, as well as data correction and analysis methods, will be easily translated into other areas of investigation using combined PET/MRI. (author) [fr

  1. Does thorax EIT image analysis depend on the image reconstruction method? (United States)

    Zhao, Zhanqi; Frerichs, Inéz; Pulletz, Sven; Müller-Lisse, Ullrich; Möller, Knut


    Different methods were proposed to analyze the resulting images of electrical impedance tomography (EIT) measurements during ventilation. The aim of our study was to examine if the analysis methods based on back-projection deliver the same results when applied on images based on other reconstruction algorithms. Seven mechanically ventilated patients with ARDS were examined by EIT. The thorax contours were determined from the routine CT images. EIT raw data was reconstructed offline with (1) filtered back-projection with circular forward model (BPC); (2) GREIT reconstruction method with circular forward model (GREITC) and (3) GREIT with individual thorax geometry (GREITT). Three parameters were calculated on the resulting images: linearity, global ventilation distribution and regional ventilation distribution. The results of linearity test are 5.03±2.45, 4.66±2.25 and 5.32±2.30 for BPC, GREITC and GREITT, respectively (median ±interquartile range). The differences among the three methods are not significant (p = 0.93, Kruskal-Wallis test). The proportions of ventilation in the right lung are 0.58±0.17, 0.59±0.20 and 0.59±0.25 for BPC, GREITC and GREITT, respectively (p = 0.98). The differences of the GI index based on different reconstruction methods (0.53±0.16, 0.51±0.25 and 0.54±0.16 for BPC, GREITC and GREITT, respectively) are also not significant (p = 0.93). We conclude that the parameters developed for images generated with GREITT are comparable with filtered back-projection and GREITC.

  2. Plasmon-Based Colorimetric Nanosensors for Ultrasensitive Molecular Diagnostics. (United States)

    Tang, Longhua; Li, Jinghong


    Colorimetric detection of target analytes with high specificity and sensitivity is of fundamental importance to clinical and personalized point-of-care diagnostics. Because of their extraordinary optical properties, plasmonic nanomaterials have been introduced into colorimetric sensing systems, which provide significantly improved sensitivity in various biosensing applications. Here we review the recent progress on these plasmonic nanoparticles-based colorimetric nanosensors for ultrasensitive molecular diagnostics. According to their different colorimetric signal generation mechanisms, these plasmonic nanosensors are classified into two categories: (1) interparticle distance-dependent colorimetric assay based on target-induced forming cross-linking assembly/aggregate of plasmonic nanoparticles; and (2) size/morphology-dependent colorimetric assay by target-controlled growth/etching of the plasmonic nanoparticles. The sensing fundamentals and cutting-edge applications will be provided for each of them, particularly focusing on signal generation and/or amplification mechanisms that realize ultrasensitive molecular detection. Finally, we also discuss the challenge and give our future perspective in this emerging field.

  3. Ultra-sensitive detection of leukemia by graphene (United States)

    Akhavan, Omid; Ghaderi, Elham; Hashemi, Ehsan; Rahighi, Reza


    Graphene oxide nanoplatelets (GONPs) with extremely sharp edges (lateral dimensions ~20-200 nm and thicknesses leukemia cells. The blood serums containing the extracted guanine were used in differential pulse voltammetry (DPV) with reduced graphene oxide nanowall (rGONW) electrodes to develop fast and ultra-sensitive electrochemical detection of leukemia cells at leukemia fractions (LFs) of ~10-11 (as the lower detection limit). The stability of the DPV signals obtained by oxidation of the extracted guanine on the rGONWs was studied after 20 cycles. Without the guanine extraction, the DPV peaks relating to guanine oxidation of normal and abnormal cells overlapped at LFs diagnosis.Graphene oxide nanoplatelets (GONPs) with extremely sharp edges (lateral dimensions ~20-200 nm and thicknesses leukemia cells. The blood serums containing the extracted guanine were used in differential pulse voltammetry (DPV) with reduced graphene oxide nanowall (rGONW) electrodes to develop fast and ultra-sensitive electrochemical detection of leukemia cells at leukemia fractions (LFs) of ~10-11 (as the lower detection limit). The stability of the DPV signals obtained by oxidation of the extracted guanine on the rGONWs was studied after 20 cycles. Without the guanine extraction, the DPV peaks relating to guanine oxidation of normal and abnormal cells overlapped at LFs diagnosis. Electronic supplementary information (ESI) available. See DOI: 10.1039/C4NR04589K

  4. Ultrasensitive Inertial and Force Sensors with Diamagnetically Levitated Magnets (United States)

    Prat-Camps, J.; Teo, C.; Rusconi, C. C.; Wieczorek, W.; Romero-Isart, O.


    We theoretically show that a magnet can be stably levitated on top of a punctured superconductor sheet in the Meissner state without applying any external field. The trapping potential created by such induced-only superconducting currents is characterized for magnetic spheres ranging from tens of nanometers to tens of millimeters. Such a diamagnetically levitated magnet is predicted to be extremely well isolated from the environment. We propose to use it as an ultrasensitive force and inertial sensor. A magnetomechanical readout of its displacement can be performed by using superconducting quantum interference devices. An analysis using current technology shows that force and acceleration sensitivities on the order of 10-23 N /√{Hz } (for a 100-nm magnet) and 10-14 g /√{Hz } (for a 10-mm magnet) might be within reach in a cryogenic environment. Such remarkable sensitivities, both in force and acceleration, can be used for a variety of purposes, from designing ultrasensitive inertial sensors for technological applications (e.g., gravimetry, avionics, and space industry), to scientific investigations on measuring Casimir forces of magnetic origin and gravitational physics.

  5. High-resolution imaging methods in array signal processing

    DEFF Research Database (Denmark)

    Xenaki, Angeliki

    in active sonar signal processing for detection and imaging of submerged oil contamination in sea water from a deep-water oil leak. The submerged oil _eld is modeled as a uid medium exhibiting spatial perturbations in the acoustic parameters from their mean ambient values which cause weak scattering...... of the incident acoustic energy. A highfrequency active sonar is selected to insonify the medium and receive the backscattered waves. High-frequency acoustic methods can both overcome the optical opacity of water (unlike methods based on electromagnetic waves) and resolve the small-scale structure...... of the submerged oil field (unlike low-frequency acoustic methods). The study shows that high-frequency acoustic methods are suitable not only for large-scale localization of the oil contamination in the water column but also for statistical characterization of the submerged oil field through inference...

  6. Image processing methods and architectures in diagnostic pathology.

    Directory of Open Access Journals (Sweden)

    Oscar DĂŠniz


    Full Text Available Grid technology has enabled the clustering and the efficient and secure access to and interaction among a wide variety of geographically distributed resources such as: supercomputers, storage systems, data sources, instruments and special devices and services. Their main applications include large-scale computational and data intensive problems in science and engineering. General grid structures and methodologies for both software and hardware in image analysis for virtual tissue-based diagnosis has been considered in this paper. This methods are focus on the user level middleware. The article describes the distributed programming system developed by the authors for virtual slide analysis in diagnostic pathology. The system supports different image analysis operations commonly done in anatomical pathology and it takes into account secured aspects and specialized infrastructures with high level services designed to meet application requirements. Grids are likely to have a deep impact on health related applications, and therefore they seem to be suitable for tissue-based diagnosis too. The implemented system is a joint application that mixes both Web and Grid Service Architecture around a distributed architecture for image processing. It has shown to be a successful solution to analyze a big and heterogeneous group of histological images under architecture of massively parallel processors using message passing and non-shared memory.

  7. Color management systems: methods and technologies for increased image quality (United States)

    Caretti, Maria


    All the steps in the imaging chain -- from handling the originals in the prepress to outputting them on any device - - have to be well calibrated and adjusted to each other, in order to reproduce color images in a desktop environment as accurate as possible according to the original. Today most of the steps in the prepress production are digital and therefore it is realistic to believe that the color reproduction can be well controlled. This is true thanks to the last years development of fast, cost effective scanners, digital sources and digital proofing devices not the least. It is likely to believe that well defined tools and methods to control this imaging flow will lead to large cost and time savings as well as increased overall image quality. Until now, there has been a lack of good, reliable, easy-to- use systems (e.g. hardware, software, documentation, training and support) in an extent that has made them accessible to the large group of users of graphic arts production systems. This paper provides an overview of the existing solutions to manage colors in a digital pre-press environment. Their benefits and limitations are discussed as well as how they affect the production workflow and organization. The difference between a color controlled environment and one that is not is explained.

  8. Information theoretic methods for image processing algorithm optimization (United States)

    Prokushkin, Sergey F.; Galil, Erez


    Modern image processing pipelines (e.g., those used in digital cameras) are full of advanced, highly adaptive filters that often have a large number of tunable parameters (sometimes > 100). This makes the calibration procedure for these filters very complex, and the optimal results barely achievable in the manual calibration; thus an automated approach is a must. We will discuss an information theory based metric for evaluation of algorithm adaptive characteristics ("adaptivity criterion") using noise reduction algorithms as an example. The method allows finding an "orthogonal decomposition" of the filter parameter space into the "filter adaptivity" and "filter strength" directions. This metric can be used as a cost function in automatic filter optimization. Since it is a measure of a physical "information restoration" rather than perceived image quality, it helps to reduce the set of the filter parameters to a smaller subset that is easier for a human operator to tune and achieve a better subjective image quality. With appropriate adjustments, the criterion can be used for assessment of the whole imaging system (sensor plus post-processing).

  9. Statistical methods of evaluating and comparing imaging techniques

    International Nuclear Information System (INIS)

    Freedman, L.S.


    Over the past 20 years several new methods of generating images of internal organs and the anatomy of the body have been developed and used to enhance the accuracy of diagnosis and treatment. These include ultrasonic scanning, radioisotope scanning, computerised X-ray tomography (CT) and magnetic resonance imaging (MRI). The new techniques have made a considerable impact on radiological practice in hospital departments, not least on the investigational process for patients suspected or known to have malignant disease. As a consequence of the increased range of imaging techniques now available, there has developed a need to evaluate and compare their usefulness. Over the past 10 years formal studies of the application of imaging technology have been conducted and many reports have appeared in the literature. These studies cover a range of clinical situations. Likewise, the methodologies employed for evaluating and comparing the techniques in question have differed widely. While not attempting an exhaustive review of the clinical studies which have been reported, this paper aims to examine the statistical designs and analyses which have been used. First a brief review of the different types of study is given. Examples of each type are then chosen to illustrate statistical issues related to their design and analysis. In the final sections it is argued that a form of classification for these different types of study might be helpful in clarifying relationships between them and bringing a perspective to the field. A classification based upon a limited analogy with clinical trials is suggested

  10. A Method for Determining Skeletal Lengths from DXA Images

    Directory of Open Access Journals (Sweden)

    Fogelman Ignac


    Full Text Available Abstract Background Skeletal ratios and bone lengths are widely used in anthropology and forensic pathology and hip axis length is a useful predictor of fracture. The aim of this study was to show that skeletal ratios, such as length of femur to height, could be accurately measured from a DXA (dual energy X-ray absorptiometry image. Methods 90 normal Caucasian females, 18–80 years old, with whole body DXA data were used as subjects. Two methods, linear pixel count (LPC and reticule and ruler (RET were used to measure skeletal sizes on DXA images and compared with real clinical measures from 20 subjects and 20 x-rays of the femur and tibia taken in 2003. Results Although both methods were highly correlated, the LPC inter- and intra-observer error was lower at 1.6% compared to that of RET at 2.3%. Both methods correlated positively with real clinical measures, with LPC having a marginally stronger correlation coefficient (r2 = 0.94; r2 = 0.84; average r2 = 0.89 than RET (r2 = 0.86; r2 = 0.84; average r2 = 0.85 with X-rays and real measures respectively. Also, the time taken to use LPC was half that of RET at 5 minutes per scan. Conclusion Skeletal ratios can be accurately and precisely measured from DXA total body scan images. The LPC method is easy to use and relatively rapid. This new phenotype will be useful for osteoporosis research for individuals or large-scale epidemiological or genetic studies.

  11. Reusable nanosilver-coated magnetic particles for ultrasensitive SERS-based detection of malachite green in water samples (United States)

    Song, Dan; Yang, Rong; Wang, Chongwen; Xiao, Rui; Long, Feng


    A novel nanosilver-deposited silica-coated Fe3O4 magnetic particle (Fe3O4@SiO2@Ag) with uniform size, good SERS activity and magnetic responsiveness was synthesized using amination polymer. The Fe3O4@SiO2@Ag magnetic particles have been successfully applied for ultrasensitive SERS detection of malachite green (MG) in water samples. The mechanism is that MG can be adsorbed on the silver surface of nanosilver-coated magnetic particles via one nitrogen atom, and the Raman signal intensity of MG is significantly enhanced by the nanosilver layer formed on the magnetic particles. The developed sensing system exhibited a sensitive response to MG in the range of 10 fM to 100 μM with a low limit of detection (LOD) 2 fM under optimal conditions. The LOD was several orders of magnitude lower than those of other methods. This SERS-based sensor showed good reproducibility and stability for MG detection. The silver-coated magnetic particles could easily be regenerated as SERS substrates only using low pH solution for multiple sensing events. The recovery of MG added to several water samples at different concentrations ranged from 90% to 110%. The proposed method facilitates the ultrasensitive analysis of dyes to satisfy the high demand for ensuring the safety of water sources. PMID:26964502

  12. Internal scanning method as unique imaging method of optical vortex scanning microscope (United States)

    Popiołek-Masajada, Agnieszka; Masajada, Jan; Szatkowski, Mateusz


    The internal scanning method is specific for the optical vortex microscope. It allows to move the vortex point inside the focused vortex beam with nanometer resolution while the whole beam stays in place. Thus the sample illuminated by the focused vortex beam can be scanned just by the vortex point. We show that this method enables high resolution imaging. The paper presents the preliminary experimental results obtained with the first basic image recovery procedure. A prospect of developing more powerful tools for topography recovery with the optical vortex scanning microscope is discussed shortly.

  13. Simultaneous collection method of on-peak window image and off-peak window image in Tl-201 imaging

    International Nuclear Information System (INIS)

    Murakami, Tomonori; Noguchi, Yasushi; Kojima, Akihiro; Takagi, Akihiro; Matsumoto, Masanori


    Tl-201 imaging detects the photopeak (71 keV, in on-peak window) of characteristic X-rays of Hg-201 formed from Tl-201 decay. The peak is derived from 4 rays of different energy and emission intensity and does not follow in Gaussian distribution. In the present study, authors made an idea for the method in the title to attain the more effective single imaging, which was examined for its accuracy and reliability with phantoms and applied clinically to Tl-201 scintigraphy in a patient. The authors applied the triple energy window method for data acquisition: the energy window setting was made on Hg-201 X-rays photopeak in three of the lower (3%, L), main (72 keV, M) and upper (14%, U) windows with the gamma camera with 2-gated detector (Toshiba E. CAM/ICON). L, M and U images obtained simultaneously were then constructed to images of on-peak (L+M, Mock on-peak) and off-peak (M+U) window settings for evaluation. Phantoms for line source with Tl-201-containing swab and for multi-defect with acrylic plate containing Tl-201 solution were imaged in water. The female patient with thyroid cancer was subjected to preoperative scintigraphy under the defined conditions. Mock on-, off-peak images were found to be equivalent to the true (ordinary, clinical) on-, off-peak ones, and the present method was thought usable for evaluation of usefulness of off-peak window data. (R.T.)

  14. Lung function imaging methods in Cystic Fibrosis pulmonary disease. (United States)

    Kołodziej, Magdalena; de Veer, Michael J; Cholewa, Marian; Egan, Gary F; Thompson, Bruce R


    Monitoring of pulmonary physiology is fundamental to the clinical management of patients with Cystic Fibrosis. The current standard clinical practise uses spirometry to assess lung function which delivers a clinically relevant functional readout of total lung function, however does not supply any visible or localised information. High Resolution Computed Tomography (HRCT) is a well-established current 'gold standard' method for monitoring lung anatomical changes in Cystic Fibrosis patients. HRCT provides excellent morphological information, however, the X-ray radiation dose can become significant if multiple scans are required to monitor chronic diseases such as cystic fibrosis. X-ray phase-contrast imaging is another emerging X-ray based methodology for Cystic Fibrosis lung assessment which provides dynamic morphological and functional information, albeit with even higher X-ray doses than HRCT. Magnetic Resonance Imaging (MRI) is a non-ionising radiation imaging method that is garnering growing interest among researchers and clinicians working with Cystic Fibrosis patients. Recent advances in MRI have opened up the possibilities to observe lung function in real time to potentially allow sensitive and accurate assessment of disease progression. The use of hyperpolarized gas or non-contrast enhanced MRI can be tailored to clinical needs. While MRI offers significant promise it still suffers from poor spatial resolution and the development of an objective scoring system especially for ventilation assessment.

  15. Image reconstruction in computerized tomography using the convolution method

    International Nuclear Information System (INIS)

    Oliveira Rebelo, A.M. de.


    In the present work an algoritin was derived, using the analytical convolution method (filtered back-projection) for two-dimensional or three-dimensional image reconstruction in computerized tomography applied to non-destructive testing and to the medical use. This mathematical model is based on the analytical Fourier transform method for image reconstruction. This model consists of a discontinuous system formed by an NxN array of cells (pixels). The attenuation in the object under study of a colimated gamma ray beam has been determined for various positions and incidence angles (projections) in terms of the interaction of the beam with the intercepted pixels. The contribution of each pixel to beam attenuation was determined using the weight function W ij which was used for simulated tests. Simulated tests using standard objects with attenuation coefficients in the range of 0,2 to 0,7 cm -1 were carried out using cell arrays of up to 25x25. One application was carried out in the medical area simulating image reconstruction of an arm phantom with attenuation coefficients in the range of 0,2 to 0,5 cm -1 using cell arrays of 41x41. The simulated results show that, in objects with a great number of interfaces and great variations of attenuation coefficients at these interfaces, a good reconstruction is obtained with the number of projections equal to the reconstruction matrix dimension. A good reconstruction is otherwise obtained with fewer projections. (author) [pt

  16. Image reconstruction methods for the PBX-M pinhole camera

    International Nuclear Information System (INIS)

    Holland, A.; Powell, E.T.; Fonck, R.J.


    This paper describes two methods which have been used to reconstruct the soft x-ray emission profile of the PBX-M tokamak from the projected images recorded by the PBX-M pinhole camera. Both methods must accurately represent the shape of the reconstructed profile while also providing a degree of immunity to noise in the data. The first method is a simple least squares fit to the data. This has the advantage of being fast and small, and thus easily implemented on the PDP-11 computer used to control the video digitizer for the pinhole camera. The second method involves the application of a maximum entropy algorithm to an overdetermined system. This has the advantage of allowing the use of a default profile. This profile contains additional knowledge about the plasma shape which can be obtained from equilibrium fits to the external magnetic measurements. Additionally the reconstruction is guaranteed positive, and the fit to the data can be relaxed by specifying both the amount and distribution of noise in the image. The algorithm described has the advantage of being considerably faster, for an overdetermined system, than the usual Lagrange multiplier approach to finding the maximum entropy solution. 13 refs., 24 figs

  17. Fringe image analysis based on the amplitude modulation method. (United States)

    Gai, Shaoyan; Da, Feipeng


    A novel phase-analysis method is proposed. To get the fringe order of a fringe image, the amplitude-modulation fringe pattern is carried out, which is combined with the phase-shift method. The primary phase value is obtained by a phase-shift algorithm, and the fringe-order information is encoded in the amplitude-modulation fringe pattern. Different from other methods, the amplitude-modulation fringe identifies the fringe order by the amplitude of the fringe pattern. In an amplitude-modulation fringe pattern, each fringe has its own amplitude; thus, the order information is integrated in one fringe pattern, and the absolute fringe phase can be calculated correctly and quickly with the amplitude-modulation fringe image. The detailed algorithm is given, and the error analysis of this method is also discussed. Experimental results are presented by a full-field shape measurement system where the data has been processed using the proposed algorithm. (c) 2010 Optical Society of America.

  18. Determining wood chip size: image analysis and clustering methods

    Directory of Open Access Journals (Sweden)

    Paolo Febbi


    Full Text Available One of the standard methods for the determination of the size distribution of wood chips is the oscillating screen method (EN 15149- 1:2010. Recent literature demonstrated how image analysis could return highly accurate measure of the dimensions defined for each individual particle, and could promote a new method depending on the geometrical shape to determine the chip size in a more accurate way. A sample of wood chips (8 litres was sieved through horizontally oscillating sieves, using five different screen hole diameters (3.15, 8, 16, 45, 63 mm; the wood chips were sorted in decreasing size classes and the mass of all fractions was used to determine the size distribution of the particles. Since the chip shape and size influence the sieving results, Wang’s theory, which concerns the geometric forms, was considered. A cluster analysis on the shape descriptors (Fourier descriptors and size descriptors (area, perimeter, Feret diameters, eccentricity was applied to observe the chips distribution. The UPGMA algorithm was applied on Euclidean distance. The obtained dendrogram shows a group separation according with the original three sieving fractions. A comparison has been made between the traditional sieve and clustering results. This preliminary result shows how the image analysis-based method has a high potential for the characterization of wood chip size distribution and could be further investigated. Moreover, this method could be implemented in an online detection machine for chips size characterization. An improvement of the results is expected by using supervised multivariate methods that utilize known class memberships. The main objective of the future activities will be to shift the analysis from a 2-dimensional method to a 3- dimensional acquisition process.

  19. Methods and applications in high flux neutron imaging

    International Nuclear Information System (INIS)

    Ballhausen, H.


    This treatise develops new methods for high flux neutron radiography and high flux neutron tomography and describes some of their applications in actual experiments. Instead of single images, time series can be acquired with short exposure times due to the available high intensity. To best use the increased amount of information, new estimators are proposed, which extract accurate results from the recorded ensembles, even if the individual piece of data is very noisy and in addition severely affected by systematic errors such as an influence of gamma background radiation. The spatial resolution of neutron radiographies, usually limited by beam divergence and inherent resolution of the scintillator, can be significantly increased by scanning the sample with a pinhole-micro-collimator. This technique circumvents any limitations in present detector design and, due to the available high intensity, could be successfully tested. Imaging with scattered neutrons as opposed to conventional total attenuation based imaging determines separately the absorption and scattering cross sections within the sample. For the first time even coherent angle dependent scattering could be visualized space-resolved. New applications of high flux neutron imaging are presented, such as materials engineering experiments on innovative metal joints, time-resolved tomography on multilayer stacks of fuel cells under operation, and others. A new implementation of an algorithm for the algebraic reconstruction of tomography data executes even in case of missing information, such as limited angle tomography, and returns quantitative reconstructions. The setup of the world-leading high flux radiography and tomography facility at the Institut Laue-Langevin is presented. A comprehensive appendix covers the physical and technical foundations of neutron imaging. (orig.)

  20. Error of image saturation in the structured-light method. (United States)

    Qi, Zhaoshuai; Wang, Zhao; Huang, Junhui; Xing, Chao; Gao, Jianmin


    In the phase-measuring structured-light method, image saturation will induce large phase errors. Usually, by selecting proper system parameters (such as the phase-shift number, exposure time, projection intensity, etc.), the phase error can be reduced. However, due to lack of a complete theory of phase error, there is no rational principle or basis for the selection of the optimal system parameters. For this reason, the phase error due to image saturation is analyzed completely, and the effects of the two main factors, including the phase-shift number and saturation degree, on the phase error are studied in depth. In addition, the selection of optimal system parameters is discussed, including the proper range and the selection principle of the system parameters. The error analysis and the conclusion are verified by simulation and experiment results, and the conclusion can be used for optimal parameter selection in practice.

  1. DAFS measurements using the image-plate Weissenberg method

    International Nuclear Information System (INIS)

    Sugioka, N.; Matsumoto, K.; Sasaki, S.; Tanaka, M.; Mori, T.


    An instrumental technique for DAFS measurements which can provide site-specific information is proposed. The approach uses (i) focusing optics with parabolic mirrors and a double-crystal monochromator, (ii) the Laue and Bragg settings and (iii) data collection by the image-plate Weissenberg method. Six image exposures are recorded per plate at five intrinsic energies and one reference energy. The single-crystal measurements were performed at the Co K-absorption edge, and the 200, 220 and 311 reflections of CoO and 511 and 911 reflections of Co 3 O 4 were used for analysis. The regression analysis of χ(k), Fourier transforms of k 3 χ(k) and back-Fourier filtering have been performed

  2. Expanded image database of pistachio x-ray images and classification by conventional methods (United States)

    Keagy, Pamela M.; Schatzki, Thomas F.; Le, Lan Chau; Casasent, David P.; Weber, David


    In order to develop sorting methods for insect damaged pistachio nuts, a large data set of pistachio x-ray images (6,759 nuts) was created. Both film and linescan sensor images were acquired, nuts dissected and internal conditions coded using the U.S. Grade standards and definitions for pistachios. A subset of 1199 good and 686 insect damaged nuts was used to calculate and test discriminant functions. Statistical parameters of image histograms were evaluated for inclusion by forward stepwise discrimination. Using three variables in the discriminant function, 89% of test set nuts were correctly identified. Comparable data for 6 human subjects ranged from 67 to 92%. If the loss of good nuts is held to 1% by requiring a high probability to discard a nut as insect damaged, approximately half of the insect damage present in clean pistachio nuts may be detected and removed by x-ray inspection.

  3. Impact of upstream and downstream constraints on a signaling module’s ultrasensitivity

    International Nuclear Information System (INIS)

    Altszyler, Edgar; Chernomoretz, Ariel; Ventura, Alejandra; Colman-Lerner, Alejandro


    Much work has been done on the study of the biochemical mechanisms that result in ultrasensitive behavior of simple biochemical modules. However, in a living cell, such modules are embedded in a bigger network that constrains the range of inputs that the module will receive as well as the range of the module’s outputs that network will be able to detect. Here, we studied how the effective ultrasensitivity of a modular system is affected by these restrictions. We use a simple setup to explore to what extent the dynamic range spanned by upstream and downstream components of an ultrasensitive module impact on the effective sensitivity of the system. Interestingly, we found for some ultrasensitive motifs that dynamic range limitations imposed by downstream components can produce effective sensitivities much larger than that of the original module when considered in isolation. (paper)

  4. General filtering method for electronic speckle pattern interferometry fringe images with various densities based on variational image decomposition. (United States)

    Li, Biyuan; Tang, Chen; Gao, Guannan; Chen, Mingming; Tang, Shuwei; Lei, Zhenkun


    Filtering off speckle noise from a fringe image is one of the key tasks in electronic speckle pattern interferometry (ESPI). In general, ESPI fringe images can be divided into three categories: low-density fringe images, high-density fringe images, and variable-density fringe images. In this paper, we first present a general filtering method based on variational image decomposition that can filter speckle noise for ESPI fringe images with various densities. In our method, a variable-density ESPI fringe image is decomposed into low-density fringes, high-density fringes, and noise. A low-density fringe image is decomposed into low-density fringes and noise. A high-density fringe image is decomposed into high-density fringes and noise. We give some suitable function spaces to describe low-density fringes, high-density fringes, and noise, respectively. Then we construct several models and numerical algorithms for ESPI fringe images with various densities. And we investigate the performance of these models via our extensive experiments. Finally, we compare our proposed models with the windowed Fourier transform method and coherence enhancing diffusion partial differential equation filter. These two methods may be the most effective filtering methods at present. Furthermore, we use the proposed method to filter a collection of the experimentally obtained ESPI fringe images with poor quality. The experimental results demonstrate the performance of our proposed method.

  5. Early Detection of Diabetic Retinopathy in Fluorescent Angiography Retinal Images Using Image Processing Methods

    Directory of Open Access Journals (Sweden)

    Meysam Tavakoli


    Full Text Available Introduction: Diabetic retinopathy (DR is the single largest cause of sight loss and blindness in the working age population of Western countries; it is the most common cause of blindness in adults between 20 and 60 years of age. Early diagnosis of DR is critical for preventing vision loss so early detection of microaneurysms (MAs as the first signs of DR is important. This paper addresses the automatic detection of MAs in fluorescein angiography fundus images, which plays a key role in computer assisted diagnosis of DR, a serious and frequent eye disease. Material and Methods: The algorithm can be divided into three main steps. The first step or pre-processing was for background normalization and contrast enhancement of the image. The second step aimed at detecting landmarks, i.e., all patterns possibly corresponding to vessels and the optic nerve head, which was achieved using a local radon transform. Then, MAs were extracted, which were used in the final step to automatically classify candidates into real MA and other objects. A database of 120 fluorescein angiography fundus images was used to train and test the algorithm. The algorithm was compared to manually obtained gradings of those images. Results: Sensitivity of diagnosis for DR was 94%, with specificity of 75%, and sensitivity of precise microaneurysm localization was 92%, at an average number of 8 false positives per image. Discussion and Conclusion: Sensitivity and specificity of this algorithm make it one of the best methods in this field. Using local radon transform in this algorithm eliminates the noise sensitivity for microaneurysm detection in retinal image analysis.

  6. Cavity opto-electromechanical system combining strong electrical actuation with ultrasensitive transduction


    McRae, Terry G.; Lee, Kwan H.; Harris, Glen I.; Knittel, Joachim; Bowen, Warwick P.


    A cavity opto-electromechanical system is reported which combines the ultrasensitive transduction of cavity optomechanical systems with the electrical actuation of nanoelectromechanical systems. Ultrasensitive mechanical transduction is achieved via opto-mechanical coupling. Electrical gradient forces as large as 0.40 $\\mu$N are realized, facilitating strong actuation with ultralow dissipation. A scanning probe microscope is implemented, capable of characterizing the mechanical modes. The int...

  7. Definition imaging of an orebody with the radio imaging method (RIM)

    International Nuclear Information System (INIS)

    Stolarczyk, L.G.


    Waste rock dilution is an economic concern in the planning and design of mining methods for mineralized ore zones. Diamond core drilling and mineralogical examination of core are routinely used to determine the general shape of the ore body. Drilling on closer centers enhances the definition of the mineralization and oregrade across the orebody. In a practical sense, drilling time and cost limit definition. Crosshole scanning between drillholes with the radio imaging method (RIM) has been used to map changes in mineralization in the rock mass. The magnitude and phase of the RIM radio wave depend on the electrical conductivity of the rock mass. The conductivity strongly depends on the percent mineralization. Since the attenuation rate and phase constants of the radio wave are proportional to the one half power of conductivity, the measured crosshole radio wave data can be processed in a tomography algorithm to reconstruct images (map the change in conductivity (mineralization)). The tomography image enhances definition in the orebody while reducing the number of drillholes. This paper compares reconstructed images of the radio wave propagation constants to percent mineralization in the ore body

  8. An Optimized Method for Terrain Reconstruction Based on Descent Images

    Directory of Open Access Journals (Sweden)

    Xu Xinchao


    Full Text Available An optimization method is proposed to perform high-accuracy terrain reconstruction of the landing area of Chang’e III. First, feature matching is conducted using geometric model constraints. Then, the initial terrain is obtained and the initial normal vector of each point is solved on the basis of the initial terrain. By changing the vector around the initial normal vector in small steps a set of new vectors is obtained. By combining these vectors with the direction of light and camera, the functions are set up on the basis of a surface reflection model. Then, a series of gray values is derived by solving the equations. The new optimized vector is recorded when the obtained gray value is closest to the corresponding pixel. Finally, the optimized terrain is obtained after iteration of the vector field. Experiments were conducted using the laboratory images and descent images of Chang’e III. The results showed that the performance of the proposed method was better than that of the classical feature matching method. It can provide a reference for terrain reconstruction of the landing area in subsequent moon exploration missions.

  9. Various imaging methods in the detection of small hepatomas

    International Nuclear Information System (INIS)

    Nakatsuka, Haruki; Kaminou, Toshio; Takemoto, Kazumasa; Takashima, Sumio; Kobayashi, Nobuyuki; Nakamura, Kenji; Onoyama, Yasuto; Kurioka, Naruto


    Fifty-one patients with small hepatomas under 5 cm in diameter were studied to compare the detectability of various imaging methods. Positive finding was obtained in 50 % of the patients by scintigraphy, in 74 % by ultrasonography and in 79 % by CT during screening tests. Rate of detection in retrospective analysis, after the site of the tumor had been known, were 73 %, 93 % and 87 % respectively. Rate of detection was 92 % by celiac arteriography and 98 % by selective hepatic arteriography. In 21 patients, who had the tumor under 3 cm, the rate was 32 % for scintigraphy, 74 % for ultrasonography and 65 % for CT during screening, whereas it was 58 %, 84 % and 75 % retrospectively. By celiac arteriography, it was 85 %, and by hepatic arteriography, 95 %. Rate of detection of small hepatomas in screening tests differed remarkably from that in retrospective analysis. No single method of imaging can disclose reliably the presense of small hepatoma, therefore more than one method should be used in screening. (author)

  10. Parallel MR image reconstruction using augmented Lagrangian methods. (United States)

    Ramani, Sathish; Fessler, Jeffrey A


    Magnetic resonance image (MRI) reconstruction using SENSitivity Encoding (SENSE) requires regularization to suppress noise and aliasing effects. Edge-preserving and sparsity-based regularization criteria can improve image quality, but they demand computation-intensive nonlinear optimization. In this paper, we present novel methods for regularized MRI reconstruction from undersampled sensitivity encoded data--SENSE-reconstruction--using the augmented Lagrangian (AL) framework for solving large-scale constrained optimization problems. We first formulate regularized SENSE-reconstruction as an unconstrained optimization task and then convert it to a set of (equivalent) constrained problems using variable splitting. We then attack these constrained versions in an AL framework using an alternating minimization method, leading to algorithms that can be implemented easily. The proposed methods are applicable to a general class of regularizers that includes popular edge-preserving (e.g., total-variation) and sparsity-promoting (e.g., l(1)-norm of wavelet coefficients) criteria and combinations thereof. Numerical experiments with synthetic and in vivo human data illustrate that the proposed AL algorithms converge faster than both general-purpose optimization algorithms such as nonlinear conjugate gradient (NCG) and state-of-the-art MFISTA.

  11. Imaging method of minute injured area at achilles tendon from multiple MR Images

    International Nuclear Information System (INIS)

    Tokui, Takahiro; Imura, Masataka; Kuroda, Yoshihiro; Oshiro, Osamu; Oguchi, Makoto; Fujiwara, Kazuhisa; Tabata, Yoshito; Ishigaki, Rikuta


    Ruptures of Achilles tendon frequently occur while doing sports. Since two-thirds of the people who suffered from the rupture of Achilles tendon feel the pain at Achilles tendon before rupture, to detect the predictor of the rupture is possible. Achilles tendon is soft tissue consisting of unidirectionally-aligned collagen fibers. Therefore, ordinary MRI scanner, ultrasonic instrument or X-ray scanner cannot acquire medical images of Achilles tendon. However, because MR signal intensity changes according to the angle between static magnetic field direction and fiber orientation, MR device can detect strong signal when the angle is 55 deg. In this research, the authors propose the imaging method to detect injured area at Achilles tendon. The method calculates and visualizes the value representing fiber tropism from the matching between MR signal intensity and the model of signal intensity of angle dependence. (author)

  12. Cardiovascular dysfunction in obesity and new diagnostic imaging techniques: the role of noninvasive image methods. (United States)

    Barbosa, José Augusto A; Rodrigues, Alexandre B; Mota, Cleonice Carvalho C; Barbosa, Márcia M; Simões e Silva, Ana C


    Obesity is a major public health problem affecting adults and children in both developed and developing countries. This condition often leads to metabolic syndrome, which increases the risk of cardiovascular disease. A large number of studies have been carried out to understand the pathogenesis of cardiovascular dysfunction in obese patients. Endothelial dysfunction plays a key role in the progression of atherosclerosis and the development of coronary artery disease, hypertension and congestive heart failure. Noninvasive methods in the field of cardiovascular imaging, such as measuring intima-media thickness, flow-mediated dilatation, tissue Doppler, and strain, and strain rate, constitute new tools for the early detection of cardiac and vascular dysfunction. These techniques will certainly enable a better evaluation of initial cardiovascular injury and allow the correct, timely management of obese patients. The present review summarizes the main aspects of cardiovascular dysfunction in obesity and discusses the application of recent noninvasive imaging methods for the early detection of cardiovascular alterations.

  13. On the convergence of nonconvex minimization methods for image recovery. (United States)

    Xiao, Jin; Ng, Michael Kwok-Po; Yang, Yu-Fei


    Nonconvex nonsmooth regularization method has been shown to be effective for restoring images with neat edges. Fast alternating minimization schemes have also been proposed and developed to solve the nonconvex nonsmooth minimization problem. The main contribution of this paper is to show the convergence of these alternating minimization schemes, based on the Kurdyka-Łojasiewicz property. In particular, we show that the iterates generated by the alternating minimization scheme, converges to a critical point of this nonconvex nonsmooth objective function. We also extend the analysis to nonconvex nonsmooth regularization model with box constraints, and obtain similar convergence results of the related minimization algorithm. Numerical examples are given to illustrate our convergence analysis.

  14. Image Signal Transfer Method in Artificial Retina using Laser

    Energy Technology Data Exchange (ETDEWEB)

    Yoon, I.Y.; Lee, B.H.; Kim, S.J. [Seoul National University, Seoul (Korea)


    Recently, the research on artificial retina for the blind is active. In this paper a new optical link method for the retinal prosthesis is proposed. Laser diode system was chosen to transfer image into the eye in this project and the new optical system was designed and evaluated. The use of laser diode array in artificial retina system makes system simple for lack of signal processing part inside of the eyeball. Designed optical system is enough to focus laser diode array on photodiode array in 20X20 application. (author). 11 refs., 7 figs., 2 tabs.

  15. Application of image processing methods to industrial radiography

    International Nuclear Information System (INIS)

    Goutte, R.; Odet, C.; Tuncer, T.; Bodson, F.; Varcin, E.


    This study was carried out with the financial support of the Commission of the European Communities as part of the CECA research program comprising of IRSID, INSA de Lyon and the Framatome and Creusot Loire companies. Its purpose was to evaluate the possibility of using digital enhancement of radiographic images to improve defect visibility in industrial radiography, thereby providing assistance in defect detection and a method for automatic analysis of radiographs. This paper provides full results obtained from work on digital processing of radiographs showing real and artificial defects. Furthermore, work on simulated automatic defect detection is also presented. 2 refs

  16. A novel ultrasensitive carboxymethyl chitosan-quantum dot-based fluorescence "turn on-off" nanosensor for lysozyme detection. (United States)

    Song, Yu; Li, Yang; Liu, Ziping; Liu, Linlin; Wang, Xinyan; Su, Xingguang; Ma, Qiang


    In this work, we developed an ultrasensitive "turn on-off" fluorescence nanosensor for lysozyme (Lyz) detection. The novel nanosensor was constructed with the carboxymethyl chitosan modified CdTe quantum dots (CMCS-QDs). Firstly, the CMCS-QDs were fabricated via the electrostatic interaction between amino groups in CMCS polymeric chains and carboxyl groups on the surface of QDs. In the fluorescence "turn-on" step, the strong binding ability between Zn(2+) and CMCS on the surface of QDs can enhance the photoluminescence intensity (PL) of QDs. In the following fluorescence "turn-off" step, the N-acetyl-glucosamine (NAG) section along the CMCS chains was hydrolyzed by Lyz. As a result, Zn(2+) was released from the surface of QDs, and the Lyz-QDs complexes were formed to quench the QDs PL. Under the optimal conditions, there was a good linear relationship between the PL of QDs and the Lyz concentration (0.1-1.2 ng/mL) with the detection limit of 0.031 ng/mL. The developed method was ultrasensitive, highly selective and fast. It has been successfully employed in the detection of Lyz in the serum with satisfactory results. Copyright © 2014 Elsevier B.V. All rights reserved.

  17. Self-Assembled Core-Satellite Gold Nanoparticle Networks for Ultrasensitive Detection of Chiral Molecules by Recognition Tunneling Current. (United States)

    Zhang, Yuanchao; Liu, Jingquan; Li, Da; Dai, Xing; Yan, Fuhua; Conlan, Xavier A; Zhou, Ruhong; Barrow, Colin J; He, Jin; Wang, Xin; Yang, Wenrong


    Chirality sensing is a very challenging task. Here, we report a method for ultrasensitive detection of chiral molecule l/d-carnitine based on changes in the recognition tunneling current across self-assembled core-satellite gold nanoparticle (GNP) networks. The recognition tunneling technique has been demonstrated to work at the single molecule level where the binding between the reader molecules and the analytes in a nanojunction. This process was observed to generate a unique and sensitive change in tunneling current, which can be used to identify the analytes of interest. The molecular recognition mechanism between amino acid l-cysteine and l/d-carnitine has been studied with the aid of SERS. The different binding strength between homo- or heterochiral pairs can be effectively probed by the copper ion replacement fracture. The device resistance was measured before and after the sequential exposures to l/d-carnitine and copper ions. The normalized resistance change was found to be extremely sensitive to the chirality of carnitine molecule. The results suggested that a GNP networks device optimized for recognition tunneling was successfully built and that such a device can be used for ultrasensitive detection of chiral molecules.

  18. Research on Methods of Infrared and Color Image Fusion Based on Wavelet Transform

    Directory of Open Access Journals (Sweden)

    Zhao Rentao


    Full Text Available There is significant difference in the imaging features of infrared image and color image, but their fusion images also have very good complementary information. In this paper, based on the characteristics of infrared image and color image, first of all, wavelet transform is applied to the luminance component of the infrared image and color image. In multi resolution the relevant regional variance is regarded as the activity measure, relevant regional variance ratio as the matching measure, and the fusion image is enhanced in the process of integration, thus getting the fused images by final synthesis module and multi-resolution inverse transform. The experimental results show that the fusion image obtained by the method proposed in this paper is better than the other methods in keeping the useful information of the original infrared image and the color information of the original color image. In addition, the fusion image has stronger adaptability and better visual effect.

  19. Method to acquire regions of fruit, branch and leaf from image of red apple in orchard (United States)

    Lv, Jidong; Xu, Liming


    This work proposed a method to acquire regions of fruit, branch and leaf from red apple image in orchard. To acquire fruit image, R-G image was extracted from the RGB image for corrosive working, hole filling, subregion removal, expansive working and opening operation in order. Finally, fruit image was acquired by threshold segmentation. To acquire leaf image, fruit image was subtracted from RGB image before extracting 2G-R-B image. Then, leaf image was acquired by subregion removal and threshold segmentation. To acquire branch image, dynamic threshold segmentation was conducted in the R-G image. Then, the segmented image was added to fruit image to acquire adding fruit image which was subtracted from RGB image with leaf image. Finally, branch image was acquired by opening operation, subregion removal and threshold segmentation after extracting the R-G image from the subtracting image. Compared with previous methods, more complete image of fruit, leaf and branch can be acquired from red apple image with this method.

  20. A method for volumetric retinal tissue oxygen tension imaging. (United States)

    Felder, Anthony E; Wanek, Justin; Teng, Pang-Yu; Blair, Norman P; Shahidi, Mahnaz


    Inadequate retinal oxygenation occurs in many vision-threatening retinal diseases, including diabetic retinopathy, retinal vascular occlusions, and age-related macular degeneration. Therefore, techniques that assess retinal oxygenation are necessary to understand retinal physiology in health and disease. The purpose of the current study is to report a method for the three-dimensional (3D) imaging of retinal tissue oxygen tension (tPO 2 ) in rats. Imaging was performed in Long Evans pigmented rats under systemic normoxia (N = 6) or hypoxia (N = 3). A vertical laser line was horizontally scanned on the retina and a series of optical section phase-delayed phosphorescence images were acquired. From these images, phosphorescence volumes at each phase delay were constructed and a 3D retinal tPO 2 volume was generated. Retinal tPO 2 volumes were quantitatively analyzed by generating retinal depth profiles of mean tPO 2 (M tPO2 ) and the spatial variation of tPO 2 (SV tPO2 ). The effects of systemic condition (normoxia/hypoxia) and retinal depth on M tPO2 and SV tPO2 were determined by mixed linear model. Each 3D retinal tPO 2 volume was approximately 500 × 750 × 200 μm (horizontal × vertical × depth) and consisted of 45 en face tPO 2 images through the retinal depth. M tPO2 at the chorioretinal interface was significantly correlated with systemic arterial oxygen tension (P = 0.007; N = 9). There were significant effects of both systemic condition and retinal depth on M tPO2 and SV tPO2 , such that both were lower under hypoxia than normoxia and higher in the outer retina than inner retina (P < 0.001). For the first time, 3D imaging of retinal tPO 2 was demonstrated, with potential future application for assessment of physiological alterations in animal models of retinal diseases.

  1. X-ray imaging using amorphous selenium: a photoinduced discharge readout method for digital mammography. (United States)

    Rowlands, J A; Hunter, D M; Araj, N


    A new digital image readout method for electrostatic charge images on photoconductive plates is described. The method can be used to read out images on selenium plates similar to those used in xeromammography. The readout method, called the air-gap photoinduced discharge method (PID), discharges the latent image pixel by pixel and measures the charge. The PID readout method, like electrometer methods, is linear. However, the PID method permits much better resolution than scanning electrometers while maintaining quantum limited performance at high radiation exposure levels. Thus the air-gap PID method appears to be uniquely superior for high-resolution digital imaging tasks such as mammography.

  2. THz-SAR Vibrating Target Imaging via the Bayesian Method

    Directory of Open Access Journals (Sweden)

    Bin Deng


    Full Text Available Target vibration bears important information for target recognition, and terahertz, due to significant micro-Doppler effects, has strong advantages for remotely sensing vibrations. In this paper, the imaging characteristics of vibrating targets with THz-SAR are at first analyzed. An improved algorithm based on an excellent Bayesian approach, that is, the expansion-compression variance-component (ExCoV method, has been proposed for reconstructing scattering coefficients of vibrating targets, which provides more robust and efficient initialization and overcomes the deficiencies of sidelobes as well as artifacts arising from the traditional correlation method. A real vibration measurement experiment of idle cars was performed to validate the range model. Simulated SAR data of vibrating targets and a tank model in a real background in 220 GHz show good performance at low SNR. Rapidly evolving high-power terahertz devices will offer viable THz-SAR application at a distance of several kilometers.

  3. Methods for calculating the electrode position Jacobian for impedance imaging. (United States)

    Boyle, A; Crabb, M G; Jehl, M; Lionheart, W R B; Adler, A


    Electrical impedance tomography (EIT) or electrical resistivity tomography (ERT) current and measure voltages at the boundary of a domain through electrodes. The movement or incorrect placement of electrodes may lead to modelling errors that result in significant reconstructed image artifacts. These errors may be accounted for by allowing for electrode position estimates in the model. Movement may be reconstructed through a first-order approximation, the electrode position Jacobian. A reconstruction that incorporates electrode position estimates and conductivity can significantly reduce image artifacts. Conversely, if electrode position is ignored it can be difficult to distinguish true conductivity changes from reconstruction artifacts which may increase the risk of a flawed interpretation. In this work, we aim to determine the fastest, most accurate approach for estimating the electrode position Jacobian. Four methods of calculating the electrode position Jacobian were evaluated on a homogeneous halfspace. Results show that Fréchet derivative and rank-one update methods are competitive in computational efficiency but achieve different solutions for certain values of contact impedance and mesh density.

  4. Application of radiological imaging methods to radioactive waste characterization

    Energy Technology Data Exchange (ETDEWEB)

    Tessaro, Ana Paula Gimenes; Souza, Daiane Cristini B. de; Vicente, Roberto, E-mail: [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)


    Radiological imaging technologies are most frequently used for medical diagnostic purposes but are also useful in materials characterization and other non-medical applications in research and industry. The characterization of radioactive waste packages or waste samples can also benefit from these techniques. In this paper, the application of some imaging methods is examined for the physical characterization of radioactive wastes constituted by spent ion-exchange resins and activated charcoal beds stored at the Radioactive Waste Management Department of IPEN. These wastes are generated when the filter media of the water polishing system of the IEA-R1 Nuclear Research Reactor is no longer able to maintain the required water quality and are replaced. The IEA-R1 is a 5MW pool-type reactor, moderated and cooled by light water, and fission and activation products released from the reactor core must be continuously removed to prevent activity buildup in the water. The replacement of the sorbents is carried out by pumping from the filter tanks into several 200 L drums, each drum getting a variable amount of water. Considering that the results of radioanalytical methods to determine the concentrations of radionuclides are usually expressed on dry basis,the amount of water must be known to calculate the total activity of each package. At first sight this is a trivial problem that demanded, however some effort to be solved. The findings on this subject are reported in this paper. (author)

  5. Ultra-sensitive high performance liquid chromatography-laser-induced fluorescence based proteomics for clinical applications. (United States)

    Patil, Ajeetkumar; Bhat, Sujatha; Pai, Keerthilatha M; Rai, Lavanya; Kartha, V B; Chidangil, Santhosh


    An ultra-sensitive high performance liquid chromatography-laser induced fluorescence (HPLC-LIF) based technique has been developed by our group at Manipal, for screening, early detection, and staging for various cancers, using protein profiling of clinical samples like, body fluids, cellular specimens, and biopsy-tissue. More than 300 protein profiles of different clinical samples (serum, saliva, cellular samples and tissue homogenates) from volunteers (normal, and different pre-malignant/malignant conditions) were recorded using this set-up. The protein profiles were analyzed using principal component analysis (PCA) to achieve objective detection and classification of malignant, premalignant and healthy conditions with high sensitivity and specificity. The HPLC-LIF protein profiling combined with PCA, as a routine method for screening, diagnosis, and staging of cervical cancer and oral cancer, is discussed in this paper. In recent years, proteomics techniques have advanced tremendously in life sciences and medical sciences for the detection and identification of proteins in body fluids, tissue homogenates and cellular samples to understand biochemical mechanisms leading to different diseases. Some of the methods include techniques like high performance liquid chromatography, 2D-gel electrophoresis, MALDI-TOF-MS, SELDI-TOF-MS, CE-MS and LC-MS techniques. We have developed an ultra-sensitive high performance liquid chromatography-laser induced fluorescence (HPLC-LIF) based technique, for screening, early detection, and staging for various cancers, using protein profiling of clinical samples like, body fluids, cellular specimens, and biopsy-tissue. More than 300 protein profiles of different clinical samples (serum, saliva, cellular samples and tissue homogenates) from healthy and volunteers with different malignant conditions were recorded by using this set-up. The protein profile data were analyzed using principal component analysis (PCA) for objective

  6. Quantifying Uncertainty in Near Surface Electromagnetic Imaging Using Bayesian Methods (United States)

    Blatter, D. B.; Ray, A.; Key, K.


    Geoscientists commonly use electromagnetic methods to image the Earth's near surface. Field measurements of EM fields are made (often with the aid an artificial EM source) and then used to infer near surface electrical conductivity via a process known as inversion. In geophysics, the standard inversion tool kit is robust and can provide an estimate of the Earth's near surface conductivity that is both geologically reasonable and compatible with the measured field data. However, standard inverse methods struggle to provide a sense of the uncertainty in the estimate they provide. This is because the task of finding an Earth model that explains the data to within measurement error is non-unique - that is, there are many, many such models; but the standard methods provide only one "answer." An alternative method, known as Bayesian inversion, seeks to explore the full range of Earth model parameters that can adequately explain the measured data, rather than attempting to find a single, "ideal" model. Bayesian inverse methods can therefore provide a quantitative assessment of the uncertainty inherent in trying to infer near surface conductivity from noisy, measured field data. This study applies a Bayesian inverse method (called trans-dimensional Markov chain Monte Carlo) to transient airborne EM data previously collected over Taylor Valley - one of the McMurdo Dry Valleys in Antarctica. Our results confirm the reasonableness of previous estimates (made using standard methods) of near surface conductivity beneath Taylor Valley. In addition, we demonstrate quantitatively the uncertainty associated with those estimates. We demonstrate that Bayesian inverse methods can provide quantitative uncertainty to estimates of near surface conductivity.

  7. Correcting saturation of detectors for particle/droplet imaging methods

    International Nuclear Information System (INIS)

    Kalt, Peter A M


    Laser-based diagnostic methods are being applied to more and more flows of theoretical and practical interest and are revealing interesting new flow features. Imaging particles or droplets in nephelometry and laser sheet dropsizing methods requires a trade-off of maximized signal-to-noise ratio without over-saturating the detector. Droplet and particle imaging results in lognormal distribution of pixel intensities. It is possible to fit a derived lognormal distribution to the histogram of measured pixel intensities. If pixel intensities are clipped at a saturated value, it is possible to estimate a presumed probability density function (pdf) shape without the effects of saturation from the lognormal fit to the unsaturated histogram. Information about presumed shapes of the pixel intensity pdf is used to generate corrections that can be applied to data to account for saturation. The effects of even slight saturation are shown to be a significant source of error on the derived average. The influence of saturation on the derived root mean square (rms) is even more pronounced. It is found that errors on the determined average exceed 5% when the number of saturated samples exceeds 3% of the total. Errors on the rms are 20% for a similar saturation level. This study also attempts to delineate limits, within which the detector saturation can be accurately corrected. It is demonstrated that a simple method for reshaping the clipped part of the pixel intensity histogram makes accurate corrections to account for saturated pixels. These outcomes can be used to correct a saturated signal, quantify the effect of saturation on a derived average and offer a method to correct the derived average in the case of slight to moderate saturation of pixels


    Energy Technology Data Exchange (ETDEWEB)

    Druckmueller, M., E-mail: [Institute of Mathematics, Faculty of Mechanical Engineering, Brno University of Technology, Technicka 2, 616 69 Brno (Czech Republic)


    A new image enhancement tool ideally suited for the visualization of fine structures in extreme ultraviolet images of the corona is presented in this paper. The Noise Adaptive Fuzzy Equalization method is particularly suited for the exceptionally high dynamic range images from the Atmospheric Imaging Assembly instrument on the Solar Dynamics Observatory. This method produces artifact-free images and gives significantly better results than methods based on convolution or Fourier transform which are often used for that purpose.

  9. Development of a practical image-based scatter correction method for brain perfusion SPECT: comparison with the TEW method

    International Nuclear Information System (INIS)

    Shidahara, Miho; Kato, Takashi; Kawatsu, Shoji; Yoshimura, Kumiko; Ito, Kengo; Watabe, Hiroshi; Kim, Kyeong Min; Iida, Hidehiro; Kato, Rikio


    An image-based scatter correction (IBSC) method was developed to convert scatter-uncorrected into scatter-corrected SPECT images. The purpose of this study was to validate this method by means of phantom simulations and human studies with 99m Tc-labeled tracers, based on comparison with the conventional triple energy window (TEW) method. The IBSC method corrects scatter on the reconstructed image I AC μb with Chang's attenuation correction factor. The scatter component image is estimated by convolving I AC μb with a scatter function followed by multiplication with an image-based scatter fraction function. The IBSC method was evaluated with Monte Carlo simulations and 99m Tc-ethyl cysteinate dimer SPECT human brain perfusion studies obtained from five volunteers. The image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were compared. Using data obtained from the simulations, the image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were found to be nearly identical for both gray and white matter. In human brain images, no significant differences in image contrast were observed between the IBSC and TEW methods. The IBSC method is a simple scatter correction technique feasible for use in clinical routine. (orig.)

  10. Development of a practical image-based scatter correction method for brain perfusion SPECT: comparison with the TEW method

    Energy Technology Data Exchange (ETDEWEB)

    Shidahara, Miho; Kato, Takashi; Kawatsu, Shoji; Yoshimura, Kumiko; Ito, Kengo [National Center for Geriatrics and Gerontology Research Institute, Department of Brain Science and Molecular Imaging, Obu, Aichi (Japan); Watabe, Hiroshi; Kim, Kyeong Min; Iida, Hidehiro [National Cardiovascular Center Research Institute, Department of Investigative Radiology, Suita (Japan); Kato, Rikio [National Center for Geriatrics and Gerontology, Department of Radiology, Obu (Japan)


    An image-based scatter correction (IBSC) method was developed to convert scatter-uncorrected into scatter-corrected SPECT images. The purpose of this study was to validate this method by means of phantom simulations and human studies with {sup 99m}Tc-labeled tracers, based on comparison with the conventional triple energy window (TEW) method. The IBSC method corrects scatter on the reconstructed image I{sub AC}{sup {mu}}{sup b} with Chang's attenuation correction factor. The scatter component image is estimated by convolving I{sub AC}{sup {mu}}{sup b} with a scatter function followed by multiplication with an image-based scatter fraction function. The IBSC method was evaluated with Monte Carlo simulations and {sup 99m}Tc-ethyl cysteinate dimer SPECT human brain perfusion studies obtained from five volunteers. The image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were compared. Using data obtained from the simulations, the image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were found to be nearly identical for both gray and white matter. In human brain images, no significant differences in image contrast were observed between the IBSC and TEW methods. The IBSC method is a simple scatter correction technique feasible for use in clinical routine. (orig.)

  11. Development of a practical image-based scatter correction method for brain perfusion SPECT: comparison with the TEW method. (United States)

    Shidahara, Miho; Watabe, Hiroshi; Kim, Kyeong Min; Kato, Takashi; Kawatsu, Shoji; Kato, Rikio; Yoshimura, Kumiko; Iida, Hidehiro; Ito, Kengo


    An image-based scatter correction (IBSC) method was developed to convert scatter-uncorrected into scatter-corrected SPECT images. The purpose of this study was to validate this method by means of phantom simulations and human studies with 99mTc-labeled tracers, based on comparison with the conventional triple energy window (TEW) method. The IBSC method corrects scatter on the reconstructed image I(mub)AC with Chang's attenuation correction factor. The scatter component image is estimated by convolving I(mub)AC with a scatter function followed by multiplication with an image-based scatter fraction function. The IBSC method was evaluated with Monte Carlo simulations and 99mTc-ethyl cysteinate dimer SPECT human brain perfusion studies obtained from five volunteers. The image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were compared. Using data obtained from the simulations, the image counts and contrast of the scatter-corrected images obtained by the IBSC and TEW methods were found to be nearly identical for both gray and white matter. In human brain images, no significant differences in image contrast were observed between the IBSC and TEW methods. The IBSC method is a simple scatter correction technique feasible for use in clinical routine.

  12. Ultra-sensitive detection of plutonium by accelerator mass spectrometry

    Energy Technology Data Exchange (ETDEWEB)

    Fifield, L.K.; Cresswell, R.G.; Ophel, T.R.; Ditada, M. [Australian National Univ., Canberra, ACT (Australia). Dept. of Nuclear Physics; Day, J.P.; Clacher, A. [Manchester Univ. (United Kingdom). Dept. of Chemistry; Priest, N.D. [AEA Technology, Harwell (United Kingdom)


    On the bases of the measurements performed to date, a sensitivity of 10{sup 6} atoms is achievable with accelerator mass spectroscopy (AMS) for each of the plutonium isotopes. Not only does this open the way to the sort of study outlined, but it also makes possible other novel applications, of which two examples are given: (i)the ration of {sup 240}Pu to {sup 239}Pu as a sensitive indicator of the source of the plutonium; (ii) the biochemistry of plutonium in humans. The ultra-sensitive atom counting capability of AMS will make it possible to use the very long-lived {sup 244}Pu (8x10{sup 7}a) in human volunteer studies without any significant increase in radiation body burden. This paper will describe the AMS technique as applied to plutonium using the ANU`s 14UD accelerator, will present the results obtained to date, and will discuss the prospects for the future.

  13. Development of an Ultrasensitive Immunoassay for Detecting Tartrazine

    Directory of Open Access Journals (Sweden)

    Chuanlai Xu


    Full Text Available We have developed an ultrasensitive indirect competitive enzyme-linked immunosorbent assay for the determination of tartrazine. Two carboxylated analogues of tartrazine with different spacer lengths, and one derivative from commercial tartrazine after a little chemical modification, were synthesized as haptens in order to produce antibodies specific to tartrazine. The effect of sulfonic acid groups on the hapten structure of tartrazine was also studied carefully for the first time. A most specific monoclonal antibody against tartrazine was created and exhibited an IC50 value of 0.105 ng/mL and a limit of detection of 0.014 ng/mL, with no cross-reactivity to other structurally-related pigments. The established immunoassay was applied to the determination of tartrazine in fortified samples of orange juice and in real positive samples of carbonated beverages.

  14. Ultra-sensitive detection of plutonium by accelerator mass spectrometry

    Energy Technology Data Exchange (ETDEWEB)

    Fifield, L K; Cresswell, R G; Ophel, T R; Ditada, M [Australian National Univ., Canberra, ACT (Australia). Dept. of Nuclear Physics; Day, J P; Clacher, A [Manchester Univ. (United Kingdom). Dept. of Chemistry; Priest, N D [AEA Technology, Harwell (United Kingdom)


    On the bases of the measurements performed to date, a sensitivity of 10{sup 6} atoms is achievable with accelerator mass spectroscopy (AMS) for each of the plutonium isotopes. Not only does this open the way to the sort of study outlined, but it also makes possible other novel applications, of which two examples are given: (i)the ration of {sup 240}Pu to {sup 239}Pu as a sensitive indicator of the source of the plutonium; (ii) the biochemistry of plutonium in humans. The ultra-sensitive atom counting capability of AMS will make it possible to use the very long-lived {sup 244}Pu (8x10{sup 7}a) in human volunteer studies without any significant increase in radiation body burden. This paper will describe the AMS technique as applied to plutonium using the ANU`s 14UD accelerator, will present the results obtained to date, and will discuss the prospects for the future.

  15. The BetaCage: Ultrasensitive Screener for Radioactive Backgrounds (United States)

    Thompson, Michael; BetaCage Collaboration


    Rare event searches, such as dark matter detection and neutrinoless double beta decay, require screening of materials for backgrounds such as beta emission and alpha decaying isotopes. The BetaCage is a proposed ultra-sensitive time-projection chamber to screen for alpha-emitting and low energy beta-emitting (10-200 keV) contaminants. The expected sensitivity is 0.1 beta particles (perkeV -m2 - day) and 0.1 alpha particles (perm2 - day) , where the former will be limited by Compton scattering of external photons in the screening samples and the latter is expected to be signal-limited. The prototype BetaCage under commissioning at South Dakota School of Mines & Technology is filled with P10 gas (10% methane, 90% argon) in place of neon and is 40×40×20 cm in size. Details on design, construction and characterization will be presented.

  16. Development of an ultrasensitive immunoassay for detecting tartrazine. (United States)

    Li, Zhuokun; Song, Shanshan; Xu, Liguang; Kuang, Hua; Guo, Shidong; Xu, Chuanlai


    We have developed an ultrasensitive indirect competitive enzyme-linked immunosorbent assay for the determination of tartrazine. Two carboxylated analogues of tartrazine with different spacer lengths, and one derivative from commercial tartrazine after a little chemical modification, were synthesized as haptens in order to produce antibodies specific to tartrazine. The effect of sulfonic acid groups on the hapten structure of tartrazine was also studied carefully for the first time. A most specific monoclonal antibody against tartrazine was created and exhibited an IC50 value of 0.105 ng/mL and a limit of detection of 0.014 ng/mL, with no cross-reactivity to other structurally-related pigments. The established immunoassay was applied to the determination of tartrazine in fortified samples of orange juice and in real positive samples of carbonated beverages.

  17. Tunable nanogap devices for ultra-sensitive electrochemical impedance biosensing

    Energy Technology Data Exchange (ETDEWEB)

    Lu, Yong [Department of Chemistry, Wannan Medical College, Wuhu 241002 (China); Guo, Zheng [Nanomaterials and Environmental Detection Laboratory, Institute of Intelligent Machines, Chinese Academy of Sciences, Hefei 230031 (China); Song, Jing-Jing; Huang, Qin-An; Zhu, Si-Wei [Department of Chemistry, Wannan Medical College, Wuhu 241002 (China); Huang, Xing-Jiu [Nanomaterials and Environmental Detection Laboratory, Institute of Intelligent Machines, Chinese Academy of Sciences, Hefei 230031 (China); Wei, Yan, E-mail: [Department of Chemistry, Wannan Medical College, Wuhu 241002 (China)


    A wealth of research has been available discussing nanogap devices for detecting very small quantities of biomolecules by observing their electrical behavior generally performed in dry conditions. We report that a gold nanogapped electrode with tunable gap length for ultra-sensitive detection of streptavidin based on electrochemical impedance technique. The gold nanogap is fabricated using simple monolayer film deposition and in-situ growth of gold nanoparticles in a traditional interdigitated array (IDA) microelectrode. The electrochemical impedance biosensor with a 25-nm nanogap is found to be ultra-sensitive to the specific binding of streptavidin to biotin. The binding of the streptavidin hinder the electron transfer between two electrodes, resulting in a large increase in electron-transfer resistance (R{sub et}) for operating the impedance. A linear relation between the relative R{sub et} and the logarithmic value of streptavidin concentration is observed in the concentration range from 1 pM (picomolar) to 100 nM (nanomolar). The lowest detectable concentration actually measured reaches 1 pM. We believe that such an electrochemical impedance nanogap biosensor provides a useful approach towards biomolecular detection that could be extended to a number of other systems. - Highlights: • A tunable gold nanogap device was used as to electrochemical impedance biosensor. • Linear range from 1 pM to 100 nM with LOD of 1 pM for streptavidin detection was obtained. • The nanogap devices exhibit a satisfactory precision, stability, and reproducibility. • The combination of electrochemical impedance technique and nanogap devices was achieved.

  18. [Imaging of pleural diseases: evaluation of imaging methods based on chest radiography]. (United States)

    Poyraz, Necdet; Kalkan, Havva; Ödev, Kemal; Ceran, Sami


    The most commonly employed radiologic method in diagnosis of pleural diseases is conventional chest radiograph. The commonest chest- X-Ray findings are the presence of pleural effusion and thickening. Small pleural effusions are not readily identified on posteroanterior chest radiograph. However, lateral decubitus chest radiograph and chest ultrasonography may show small pleural effusions. These are more efficient methods than posteroanterior chest radiograph in the erect position for demonstrating small amounts of free pleural effusions. Chest ultrasonograph may be able to help in distinguishing the pleural pathologies from parenchymal lesions. On chest radiograph pleural effusions or pleural thickening may obscure the visibility of the underlying disease or parenchymal abnormality. Thus, computed tomography (CT) may provide additional information of determining the extent and severity of pleural disease and may help to differentiate malign pleural lesions from the benign ones. Moreover, CT may provide the differentiation of parenchmal abnormalities from pleural pathologies. CT (coronal and sagittal reformatted images) that also show invasion of chest wall, mediastinum and diaphragm, as well as enlarged hilar or mediastinal lymph nodes. Standart non-invasive imaging techniques may be supplemented with magnetic resonans imaging (MRI).

  19. An improved method for polarimetric image restoration in interferometry (United States)

    Pratley, Luke; Johnston-Hollitt, Melanie


    Interferometric radio astronomy data require the effects of limited coverage in the Fourier plane to be accounted for via a deconvolution process. For the last 40 years this process, known as `cleaning', has been performed almost exclusively on all Stokes parameters individually as if they were independent scalar images. However, here we demonstrate for the case of the linear polarization P, this approach fails to properly account for the complex vector nature resulting in a process which is dependent on the axes under which the deconvolution is performed. We present here an improved method, `Generalized Complex CLEAN', which properly accounts for the complex vector nature of polarized emission and is invariant under rotations of the deconvolution axes. We use two Australia Telescope Compact Array data sets to test standard and complex CLEAN versions of the Högbom and SDI (Steer-Dwedney-Ito) CLEAN algorithms. We show that in general the complex CLEAN version of each algorithm produces more accurate clean components with fewer spurious detections and lower computation cost due to reduced iterations than the current methods. In particular, we find that the complex SDI CLEAN produces the best results for diffuse polarized sources as compared with standard CLEAN algorithms and other complex CLEAN algorithms. Given the move to wide-field, high-resolution polarimetric imaging with future telescopes such as the Square Kilometre Array, we suggest that Generalized Complex CLEAN should be adopted as the deconvolution method for all future polarimetric surveys and in particular that the complex version of an SDI CLEAN should be used.

  20. The narrow-band imaging examination method in otorhinolaryngology

    Directory of Open Access Journals (Sweden)

    Robert Šifrer


    Full Text Available Early diagnostics could improve the prognosis of patients with squamous-cell carcinomas of the head and neck. Narrow-Band Imaging (NBI is the latest examination method in the group of biologic endoscopies. NBI improves the distinction between malignant and benign mucosal lesions. Early suspect oncologic lesions that may otherwise be missed by normal white light illumination can also be diagnosed. The biggest benefit of NBI technology is achieved by using it together with a HDTV camera that enables better contrast and higher resolution. NBI is based on better imaging of superficial mucosal vasculature. The biologic potential of mucosal lesions could be predicted from vascular changes. The colour of normal mucosa under NBI is blue and green and the vessels show no pathological features. Well-demarcated brownish areas and scattered thick dark spots and abnormal winding and branching out of vessels on the mucosa are all oncologically suspicious features. Authors report the experience from literature on the use of NBI to identify carcinomas of the oral cavity, epipharynx, oropharynx, hypopharynx and larynx and evaluation of unknown primaries. In addition, the literature reports the benefit of NBI in identifying early stage carcinomas in previously irradiated patients. Persistence and recurrence of carcinoma and the development of new primary tumour could easily be missed by using only standard white-light illumination. The method proved to be highly sensitive and specific for predicting malignant changes in the above-mentioned circumstances. Authors report their own experience with NBI technology as well. For further improvement of the method, new technologic development is expected to enable the connection of NBI and HDTV with flexible endoscopes.

  1. Evaluation of the Accuracy of the Dark Frame Subtraction Method in CCD Image Processing

    National Research Council Canada - National Science Library

    Levesque, Martin P; Lelievre, Mario


    .... This method is frequently used for removing the image background gradient (a thermal artefact) in CCD images. This report demonstrates that this method may not be suitable for the detection of objects with very low signal-to-noise ratio...

  2. An improved computing method for the image edge detection

    Institute of Scientific and Technical Information of China (English)

    Gang Wang; Liang Xiao; Anzhi He


    The framework of detecting the image edge based on the sub-pixel multi-fractal measure (SPMM) is presented. The measure is defined, which gives the sub-pixel local distribution of the image gradient. The more precise singularity exponent of every pixel can be obtained by performing the SPMM analysis on the image. Using the singularity exponents and the multi-fractal spectrum of the image, the image can be segmented into a series of sets with different singularity exponents, thus the image edge can be detected automatically and easily. The simulation results show that the SPMM has higher quality factor in the image edge detection.

  3. Equipment and methods for synthetic aperture anatomic and flow imaging

    DEFF Research Database (Denmark)

    Jensen, Jørgen Arendt; Nikolov, Svetoslav; Misaridis, Thanassis


    Conventional ultrasound imaging is done by sequentially probing in each image direction. The frame rate is, thus, limited by the speed of sound and the number of lines necessary to form an image. This is especially limiting in flow imaging, since multiple lines are used for flow estimation. Another...... problem is that each receiving transducer element must be connected to a receiver, which makes the expansion of the number of receive channels expensive. Synthetic aperture (SA) imaging is a radical change from the sequential image formation. Here ultrasound is emitted in all directions and the image...... is formed in all directions simultaneously over a number of acquisitions. SA images can therefore be perfectly focused in both transmit and receive for all depths, thus significantly improving image quality. A further advantage is that very fast imaging can be done, since only a few emissions are needed...

  4. A method for generating high resolution satellite image time series (United States)

    Guo, Tao


    There is an increasing demand for satellite remote sensing data with both high spatial and temporal resolution in many applications. But it still is a challenge to simultaneously improve spatial resolution and temporal frequency due to the technical limits of current satellite observation systems. To this end, much R&D efforts have been ongoing for years and lead to some successes roughly in two aspects, one includes super resolution, pan-sharpen etc. methods which can effectively enhance the spatial resolution and generate good visual effects, but hardly preserve spectral signatures and result in inadequate analytical value, on the other hand, time interpolation is a straight forward method to increase temporal frequency, however it increase little informative contents in fact. In this paper we presented a novel method to simulate high resolution time series data by combing low resolution time series data and a very small number of high resolution data only. Our method starts with a pair of high and low resolution data set, and then a spatial registration is done by introducing LDA model to map high and low resolution pixels correspondingly. Afterwards, temporal change information is captured through a comparison of low resolution time series data, and then projected onto the high resolution data plane and assigned to each high resolution pixel according to the predefined temporal change patterns of each type of ground objects. Finally the simulated high resolution data is generated. A preliminary experiment shows that our method can simulate a high resolution data with a reasonable accuracy. The contribution of our method is to enable timely monitoring of temporal changes through analysis of time sequence of low resolution images only, and usage of costly high resolution data can be reduces as much as possible, and it presents a highly effective way to build up an economically operational monitoring solution for agriculture, forest, land use investigation

  5. Quantitative Clinical Imaging Methods for Monitoring Intratumoral Evolution. (United States)

    Kim, Joo Yeun; Gatenby, Robert A


    regional dynamics so that the internal diversity of tumors is the net result of complex multiscale somatic Darwinian interactions.Methods in landscape ecology harness Darwinian dynamics to link the environmental properties of a given region to the local populations which are assumed to represent maximally fit phenotypes within those conditions. Consider a common task of a landscape ecologist: defining the spatial distribution of species in a large region, e.g., in a satellite image. Clearly the most accurate approach requires a meter by meter survey of the multiple square kilometers in the region of interest. However, this is both impractical and potentially destructive. Instead, landscape ecology breaks the task into component parts relying on the Darwinian interdependence of environmental properties and fitness of specific species' phenotypic and genotypic properties. First, the satellite map is carefully analyzed to define the number and distribution of habitats. Then the species distribution in a representative sampling of each habitat is empirically determined. Ultimately, this permits sufficient bridging of spatial scales to accurately predict spatial distribution of plant and animal species within large regions.Currently, identifying intratumoral subpopulations requires detailed histological and molecular studies that are expensive and time consuming. Furthermore, this method is subject to sampling bias, is invasive for vital organs such as the brain, and inherently destructive precluding repeated assessments for monitoring post-treatment response and proteogenomic evolution. In contrast, modern cross-sectional imaging can interrogate the entire tumor noninvasively, allowing repeated analysis without disrupting the region of interest. In particular, magnetic resonance imaging (MRI) provides exceptional spatial resolution and generates signals that are unique to the molecular constituents of tissue. Here we propose that MRI scans may be the equivalent of satellite

  6. A Frequency Matching Method for Generation of a Priori Sample Models from Training Images

    DEFF Research Database (Denmark)

    Lange, Katrine; Cordua, Knud Skou; Frydendall, Jan


    This paper presents a Frequency Matching Method (FMM) for generation of a priori sample models based on training images and illustrates its use by an example. In geostatistics, training images are used to represent a priori knowledge or expectations of models, and the FMM can be used to generate...... new images that share the same multi-point statistics as a given training image. The FMM proceeds by iteratively updating voxel values of an image until the frequency of patterns in the image matches the frequency of patterns in the training image; making the resulting image statistically...... indistinguishable from the training image....

  7. Methods of evaluating SPECT images. The usefulness of the Matsuda`s method by the Patlak plot method in children

    Energy Technology Data Exchange (ETDEWEB)

    Takaishi, Yasuko [Nippon Medical School, Tokyo (Japan); Hashimoto, Kiyoshi; Fujino, Osamu [and others


    Single photon emission computed tomography (SPECT) is a tool to study cerebral blood flow (CBF) kinetics. There are three methods of evaluating SPECT images: visual, semi-quantitative (evaluation of the radioactivity ratio of the cerebral region to the cerebellum (R/CE) or to the thalamus (R/TH)) and quantitative (Matsuda`s method by Patlak plot method using {sup 99m}Tc-hexamethylpropylene amine oxime radionuclide angiography). We evaluated SPECT images by the quantitative method in 14 patients with neurological disorders and examined the correlation of the results to those obtained by the semi-quantitative method. There was no significant correlation between the R/CE or R/TH ratio and regional CBF except two regions. The evaluation by the semi-quantitative method may have been inappropriate, probably because the cerebellar or thalamic blood flow was not constant in each case. Evaluation by the quantitative method, on the other hand, seemed to be useful not only for the comparison of CBF among normal subjects, but also in the demonstration of progressive changes of CBF in the same case. The Matsuda`s method by the Patlak plot method is suitable for examination of children, since it dose not require aortic blood sampling. (author)

  8. Cardiac imaging systems and methods employing computerized tomographic scanning

    International Nuclear Information System (INIS)

    Richey, J.B.; Wake, R.H.; Walters, R.G.; Hunt, W.F.; Cool, S.L.


    The invention relates to cardiac imaging systems and methods employing computerised tomographic scanning. Apparatus is described which allows an image of the radiation attenuation of the heart at a desired phase of the cardiac cycle. The patients ECG signal can be used in a transverse-and-rotate type CT scanner as a time base, so that the beam reaches the heart at a desired phase of the cardiac cycle, or, in a purely rotational-type CT scanner continuously generated scan data is only stored for corresponding phases of successive cardiac cycles. Alternatively, gating of the beams themselves by shuttering or switching the power supply can be controlled by the ECG signal. A pacemaker is used to stabilize the cardiac period. Also used is a system for recognising unacceptable variations in the cardiac period and discarding corresponding scan data. In a transverse-and-rotate type fan-beam CT scanner, the effective beam width is narrowed to reduce the duration of the traverse of the heart. (U.K.)

  9. Methods of Hematoxylin and Erosin Image Information Acquisition and Optimization in Confocal Microscopy. (United States)

    Yoon, Woong Bae; Kim, Hyunjin; Kim, Kwang Gi; Choi, Yongdoo; Chang, Hee Jin; Sohn, Dae Kyung


    We produced hematoxylin and eosin (H&E) staining-like color images by using confocal laser scanning microscopy (CLSM), which can obtain the same or more information in comparison to conventional tissue staining. We improved images by using several image converting techniques, including morphological methods, color space conversion methods, and segmentation methods. An image obtained after image processing showed coloring very similar to that in images produced by H&E staining, and it is advantageous to conduct analysis through fluorescent dye imaging and microscopy rather than analysis based on single microscopic imaging. The colors used in CLSM are different from those seen in H&E staining, which is the method most widely used for pathologic diagnosis and is familiar to pathologists. Computer technology can facilitate the conversion of images by CLSM to be very similar to H&E staining images. We believe that the technique used in this study has great potential for application in clinical tissue analysis.

  10. Advanced methods and algorithm for high precision astronomical imaging

    International Nuclear Information System (INIS)

    Ngole-Mboula, Fred-Maurice


    One of the biggest challenges of modern cosmology is to gain a more precise knowledge of the dark energy and the dark matter nature. Fortunately, the dark matter can be traced directly through its gravitational effect on galaxies shapes. The European Spatial Agency Euclid mission will precisely provide data for such a purpose. A critical step is analyzing these data will be to accurately model the instrument Point Spread Function (PSF), which the focus of this thesis.We developed non parametric methods to reliably estimate the PSFs across an instrument field-of-view, based on unresolved stars images and accounting for noise, under sampling and PSFs spatial variability. At the core of these contributions, modern mathematical tools and concepts such as sparsity. An important extension of this work will be to account for the PSFs wavelength dependency. (author) [fr

  11. In vivo cellular imaging using fluorescent proteins - Methods and Protocols

    Directory of Open Access Journals (Sweden)

    M. Monti


    Full Text Available The discovery and genetic engineering of fluorescent proteins has revolutionized cell biology. What was previously invisible to the cell often can be made visible with the use of fluorescent proteins. With this words, Robert M. Hoffman introduces In vivo Cellular Imaging Using Fluorescent proteins, the eighteen chapters book dedicated to the description of how fluorescence proteins have changed the way to analyze cellular processes in vivo. Modern researches aim to study new and less invasive methods able to follow the behavior of different cell types in different biological contexts: for example, how cancer cells migrate or how they respond to different therapies. Also, in vivo systems can help researchers to better understand animal embryonic development so as how fluorescence proteins may be used to monitor different processes in living organisms at the molecular and cellular level.

  12. Microfluidic method for measuring viscosity using images from smartphone (United States)

    Kim, Sooyeong; Kim, Kyung Chun; Yeom, Eunseop


    The viscosity of a fluid is the most important characteristic in fluid rheology. Many microfluidic devices have been proposed for easily measuring the fluid viscosity of small samples. A hybrid system consisting of a smartphone and microfluidic device can offer a mobile laboratory for performing a wide range of detection and analysis functions related to healthcare. In this study, a new mobile sensing method based on a microfluidic device was proposed for fluid viscosity measurements. By separately delivering sample and reference fluids into the two inlets of a Y-shaped microfluidic device, an interfacial line is induced at downstream of the device. Because the interfacial width (W) between the sample and reference fluid flows was determined by their pressure ratio, the viscosity (μ) of the sample could be estimated by measuring the interfacial width. To distinguish the interfacial width of a sample, optical images of the flows at downstream of the Y-shaped microfluidic device were acquired using a smartphone. To check the measurement accuracy of the proposed method, the viscosities of glycerol mixtures were compared with those measured by a conventional viscometer. The proposed technique was applied to monitor the variations in blood and oil samples depending on storage or rancidity. We expect that this mobile sensing method based on a microfluidic device could be utilized as a viscometer with significant advantages in terms of mobility, ease-of-operation, and data management.

  13. Unconventional Imaging Methods to Capture Transient Structures during Actomyosin Interaction

    Directory of Open Access Journals (Sweden)

    Eisaku Katayama


    Full Text Available Half a century has passed since the cross-bridge structure was recognized as the molecular machine that generates muscle tension. Despite various approaches by a number of scientists, information on the structural changes in the myosin heads, particularly its transient configurations, remains scant even now, in part because of their small size and rapid stochastic movements during the power stroke. Though progress in cryo-electron microscopy is eagerly awaited as the ultimate means to elucidate structural details, the introduction of some unconventional methods that provide high-contrast raw images of the target protein assemblies is quite useful, if available, to break the current impasse. Quick-freeze deep–etch–replica electron microscopy coupled with dedicated image analysis procedures, and high-speed atomic-force microscopy are two such candidates. We have applied the former to visualize actin-associated myosin heads under in vitro motility assay conditions, and found that they take novel configurations similar to the SH1–SH2-crosslinked myosin that we characterized recently. By incorporating biochemical and biophysical results, we have revised the cross-bridge mechanism to involve the new conformer as an important main player. The latter “microscopy” is unique and advantageous enabling continuous observation of various protein assemblies as they function. Direct observation of myosin-V’s movement along actin filaments revealed several unexpected behaviors such as foot-stomping of the leading head and unwinding of the coiled-coil tail. The potential contribution of these methods with intermediate spatial resolution is discussed.

  14. Local coding based matching kernel method for image classification.

    Directory of Open Access Journals (Sweden)

    Yan Song

    Full Text Available This paper mainly focuses on how to effectively and efficiently measure visual similarity for local feature based representation. Among existing methods, metrics based on Bag of Visual Word (BoV techniques are efficient and conceptually simple, at the expense of effectiveness. By contrast, kernel based metrics are more effective, but at the cost of greater computational complexity and increased storage requirements. We show that a unified visual matching framework can be developed to encompass both BoV and kernel based metrics, in which local kernel plays an important role between feature pairs or between features and their reconstruction. Generally, local kernels are defined using Euclidean distance or its derivatives, based either explicitly or implicitly on an assumption of Gaussian noise. However, local features such as SIFT and HoG often follow a heavy-tailed distribution which tends to undermine the motivation behind Euclidean metrics. Motivated by recent advances in feature coding techniques, a novel efficient local coding based matching kernel (LCMK method is proposed. This exploits the manifold structures in Hilbert space derived from local kernels. The proposed method combines advantages of both BoV and kernel based metrics, and achieves a linear computational complexity. This enables efficient and scalable visual matching to be performed on large scale image sets. To evaluate the effectiveness of the proposed LCMK method, we conduct extensive experiments with widely used benchmark datasets, including 15-Scenes, Caltech101/256, PASCAL VOC 2007 and 2011 datasets. Experimental results confirm the effectiveness of the relatively efficient LCMK method.

  15. A fast and automatic mosaic method for high-resolution satellite images (United States)

    Chen, Hongshun; He, Hui; Xiao, Hongyu; Huang, Jing


    We proposed a fast and fully automatic mosaic method for high-resolution satellite images. First, the overlapped rectangle is computed according to geographical locations of the reference and mosaic images and feature points on both the reference and mosaic images are extracted by a scale-invariant feature transform (SIFT) algorithm only from the overlapped region. Then, the RANSAC method is used to match feature points of both images. Finally, the two images are fused into a seamlessly panoramic image by the simple linear weighted fusion method or other method. The proposed method is implemented in C++ language based on OpenCV and GDAL, and tested by Worldview-2 multispectral images with a spatial resolution of 2 meters. Results show that the proposed method can detect feature points efficiently and mosaic images automatically.

  16. High-performance method of morphological medical image processing

    Directory of Open Access Journals (Sweden)

    Ryabykh M. S.


    Full Text Available the article shows the implementation of grayscale morphology vHGW algorithm for selection borders in the medical image. Image processing is executed using OpenMP and NVIDIA CUDA technology for images with different resolution and different size of the structuring element.

  17. Novel axolotl cardiac function analysis method using magnetic resonance imaging

    NARCIS (Netherlands)

    Sanches, Pedro Gomes; Op 't Veld, Roel C.; de Graaf, Wolter; Strijkers, Gustav J.; Grüll, Holger


    The salamander axolotl is capable of complete regeneration of amputated heart tissue. However, non-invasive imaging tools for assessing its cardiac function were so far not employed. In this study, cardiac magnetic resonance imaging is introduced as a non-invasive technique to image heart function

  18. Novel axolotl cardiac function analysis method using magnetic resonance imaging

    NARCIS (Netherlands)

    Sanches, P.G.; Op ‘t Veld, R.C.; de Graaf, W.; Strijkers, G.J.; Grüll, H.


    The salamander axolotl is capable of complete regeneration of amputated heart tissue. However, non-invasive imaging tools for assessing its cardiac function were so far not employed. In this study, cardiac magnetic resonance imaging is introduced as a noninvasive technique to image heart function of

  19. Cardiovascular dysfunction in obesity and new diagnostic imaging techniques: the role of noninvasive image methods

    Directory of Open Access Journals (Sweden)

    Barbosa JA


    Full Text Available José Augusto A Barbosa¹, Alexandre B Rodrigues¹, Cleonice Carvalho C Mota¹, Márcia M Barbosa², Ana C Simões e Silva¹¹Department of Pediatrics, Faculty of Medicine, Federal University of Minas Gerais (UFMG, Belo Horizonte, Minas Gerais, Brazil; ²Ecocenter, Socor Hospital, Belo Horizonte, Minas Gerais, BrazilAbstract: Obesity is a major public health problem affecting adults and children in both developed and developing countries. This condition often leads to metabolic syndrome, which increases the risk of cardiovascular disease. A large number of studies have been carried out to understand the pathogenesis of cardiovascular dysfunction in obese patients. Endothelial dysfunction plays a key role in the progression of atherosclerosis and the development of coronary artery disease, hypertension and congestive heart failure. Noninvasive methods in the field of cardiovascular imaging, such as measuring intima-media thickness, flow-mediated dilatation, tissue Doppler, and strain, and strain rate, constitute new tools for the early detection of cardiac and vascular dysfunction. These techniques will certainly enable a better evaluation of initial cardiovascular injury and allow the correct, timely management of obese patients. The present review summarizes the main aspects of cardiovascular dysfunction in obesity and discusses the application of recent noninvasive imaging methods for the early detection of cardiovascular alterations.Keywords: cardiovascular risk, endothelium dysfunction, obesity, strain and strain rate, tissue Doppler

  20. A preconditioned inexact newton method for nonlinear sparse electromagnetic imaging

    KAUST Repository

    Desmal, Abdulla


    A nonlinear inversion scheme for the electromagnetic microwave imaging of domains with sparse content is proposed. Scattering equations are constructed using a contrast-source (CS) formulation. The proposed method uses an inexact Newton (IN) scheme to tackle the nonlinearity of these equations. At every IN iteration, a system of equations, which involves the Frechet derivative (FD) matrix of the CS operator, is solved for the IN step. A sparsity constraint is enforced on the solution via thresholded Landweber iterations, and the convergence is significantly increased using a preconditioner that levels the FD matrix\\'s singular values associated with contrast and equivalent currents. To increase the accuracy, the weight of the regularization\\'s penalty term is reduced during the IN iterations consistently with the scheme\\'s quadratic convergence. At the end of each IN iteration, an additional thresholding, which removes small \\'ripples\\' that are produced by the IN step, is applied to maintain the solution\\'s sparsity. Numerical results demonstrate the applicability of the proposed method in recovering sparse and discontinuous dielectric profiles with high contrast values.


    Directory of Open Access Journals (Sweden)

    Patricia Donnadieu


    Full Text Available Owing the phase image method (Hytch, 1998, strain fields can be derived from HREM images. The method is here applied to the nanoscale precipitates responsible for hardening in Aluminum alloys. Since the method is a very sensitive one, we have examined the impact of several aspects of the image quality (noise, fluctuations, distortion. The strain field information derived from the HREM image analysis is further introduced in a simulation of the dislocation motion in the matrix.

  2. A simple method for detecting tumor in T2-weighted MRI brain images. An image-based analysis

    International Nuclear Information System (INIS)

    Lau, Phooi-Yee; Ozawa, Shinji


    The objective of this paper is to present a decision support system which uses a computer-based procedure to detect tumor blocks or lesions in digitized medical images. The authors developed a simple method with a low computation effort to detect tumors on T2-weighted Magnetic Resonance Imaging (MRI) brain images, focusing on the connection between the spatial pixel value and tumor properties from four different perspectives: cases having minuscule differences between two images using a fixed block-based method, tumor shape and size using the edge and binary images, tumor properties based on texture values using spatial pixel intensity distribution controlled by a global discriminate value, and the occurrence of content-specific tumor pixel for threshold images. Measurements of the following medical datasets were performed: different time interval images, and different brain disease images on single and multiple slice images. Experimental results have revealed that our proposed technique incurred an overall error smaller than those in other proposed methods. In particular, the proposed method allowed decrements of false alarm and missed alarm errors, which demonstrate the effectiveness of our proposed technique. In this paper, we also present a prototype system, known as PCB, to evaluate the performance of the proposed methods by actual experiments, comparing the detection accuracy and system performance. (author)

  3. [Testing method research for key performance indicator of imaging acousto-optic tunable filter (AOTF)]. (United States)

    Hu, Shan-Zhou; Chen, Fen-Fei; Zeng, Li-Bo; Wu, Qiong-Shui


    Imaging AOTF is an important optical filter component for new spectral imaging instruments developed in recent years. The principle of imaging AOTF component was demonstrated, and a set of testing methods for some key performances were studied, such as diffraction efficiency, wavelength shift with temperature, homogeneity in space for diffraction efficiency, imaging shift, etc.

  4. Digital images segmentation: a state of art of the different methods ...

    African Journals Online (AJOL)

    An image is a planar representation of a scene or a 3 D object. The primary information associated to each point of the image is transcribed in grey level or in colour. Image analysis is the set of methods which permits the extraction of pertinent information from the image according to the concerned application, to treat them ...

  5. An ultrasensitive NanoLuc-based luminescence system for monitoring Plasmodium berghei throughout its life cycle. (United States)

    De Niz, Mariana; Stanway, Rebecca R; Wacker, Rahel; Keller, Derya; Heussler, Volker T


    Bioluminescence imaging is widely used for cell-based assays and animal imaging studies, both in biomedical research and drug development. Its main advantages include its high-throughput applicability, affordability, high sensitivity, operational simplicity, and quantitative outputs. In malaria research, bioluminescence has been used for drug discovery in vivo and in vitro, exploring host-pathogen interactions, and studying multiple aspects of Plasmodium biology. While the number of fluorescent proteins available for imaging has undergone a great expansion over the last two decades, enabling simultaneous visualization of multiple molecular and cellular events, expansion of available luciferases has lagged behind. The most widely used bioluminescent probe in malaria research is the Photinus pyralis firefly luciferase, followed by the more recently introduced Click-beetle and Renilla luciferases. Ultra-sensitive imaging of Plasmodium at low parasite densities has not been previously achieved. With the purpose of overcoming these challenges, a Plasmodium berghei line expressing the novel ultra-bright luciferase enzyme NanoLuc, called PbNLuc has been generated, and is presented in this work. NanoLuc shows at least 150 times brighter signal than firefly luciferase in vitro, allowing single parasite detection in mosquito, liver, and sexual and asexual blood stages. As a proof-of-concept, the PbNLuc parasites were used to image parasite development in the mosquito, liver and blood stages of infection, and to specifically explore parasite liver stage egress, and pre-patency period in vivo. PbNLuc is a suitable parasite line for sensitive imaging of the entire Plasmodium life cycle. Its sensitivity makes it a promising line to be used as a reference for drug candidate testing, as well as the characterization of mutant parasites to explore the function of parasite proteins, host-parasite interactions, and the better understanding of Plasmodium biology. Since the substrate

  6. The Study of Image Processing Method for AIDS PA Test

    International Nuclear Information System (INIS)

    Zhang, H J; Wang, Q G


    At present, the main test technique of AIDS is PA in China. Because the judgment of PA test image is still depending on operator, the error ration is high. To resolve this problem, we present a new technique of image processing, which first process many samples and get the data including coordinate of center and the rang of kinds images; then we can segment the image with the data; at last, the result is exported after data was judgment. This technique is simple and veracious; and it also turns out to be suitable for the processing and analyzing of other infectious diseases' PA test image

  7. Optical Methods and Instrumentation in Brain Imaging and Therapy

    CERN Document Server


    This book provides a comprehensive up-to-date review of optical approaches used in brain imaging and therapy. It covers a variety of imaging techniques including diffuse optical imaging, laser speckle imaging, photoacoustic imaging and optical coherence tomography. A number of laser-based therapeutic approaches are reviewed, including photodynamic therapy, fluorescence guided resection and photothermal therapy. Fundamental principles and instrumentation are discussed for each imaging and therapeutic technique. Represents the first publication dedicated solely to optical diagnostics and therapeutics in the brain Provides a comprehensive review of the principles of each imaging/therapeutic modality Reviews the latest advances in instrumentation for optical diagnostics in the brain Discusses new optical-based therapeutic approaches for brain diseases

  8. A Review of Imaging Methods for Prostate Cancer Detection

    Directory of Open Access Journals (Sweden)

    Saradwata Sarkar


    Full Text Available Imaging is playing an increasingly important role in the detection of prostate cancer (PCa. This review summarizes the key imaging modalities–multiparametric ultrasound (US, multiparametric magnetic resonance imaging (MRI, MRI-US fusion imaging, and positron emission tomography (PET imaging–-used in the diagnosis and localization of PCa. Emphasis is laid on the biological and functional characteristics of tumors that rationalize the use of a specific imaging technique. Changes to anatomical architecture of tissue can be detected by anatomical grayscale US and T2-weighted MRI. Tumors are known to progress through angiogenesis–-a fact exploited by Doppler and contrast-enhanced US and dynamic contrast-enhanced MRI. The increased cellular density of tumors is targeted by elastography and diffusion-weighted MRI. PET imaging employs several different radionuclides to target the metabolic and cellular activities during tumor growth. Results from studies using these various imaging techniques are discussed and compared.

  9. Extended morphological processing: a practical method for automatic spot detection of biological markers from microscopic images. (United States)

    Kimori, Yoshitaka; Baba, Norio; Morone, Nobuhiro


    A reliable extraction technique for resolving multiple spots in light or electron microscopic images is essential in investigations of the spatial distribution and dynamics of specific proteins inside cells and tissues. Currently, automatic spot extraction and characterization in complex microscopic images poses many challenges to conventional image processing methods. A new method to extract closely located, small target spots from biological images is proposed. This method starts with a simple but practical operation based on the extended morphological top-hat transformation to subtract an uneven background. The core of our novel approach is the following: first, the original image is rotated in an arbitrary direction and each rotated image is opened with a single straight line-segment structuring element. Second, the opened images are unified and then subtracted from the original image. To evaluate these procedures, model images of simulated spots with closely located targets were created and the efficacy of our method was compared to that of conventional morphological filtering methods. The results showed the better performance of our method. The spots of real microscope images can be quantified to confirm that the method is applicable in a given practice. Our method achieved effective spot extraction under various image conditions, including aggregated target spots, poor signal-to-noise ratio, and large variations in the background intensity. Furthermore, it has no restrictions with respect to the shape of the extracted spots. The features of our method allow its broad application in biological and biomedical image information analysis.

  10. A New Pixels Flipping Method for Huge Watermarking Capacity of the Invoice Font Image

    Directory of Open Access Journals (Sweden)

    Li Li


    Full Text Available Invoice printing just has two-color printing, so invoice font image can be seen as binary image. To embed watermarks into invoice image, the pixels need to be flipped. The more huge the watermark is, the more the pixels need to be flipped. We proposed a new pixels flipping method in invoice image for huge watermarking capacity. The pixels flipping method includes one novel interpolation method for binary image, one flippable pixels evaluation mechanism, and one denoising method based on gravity center and chaos degree. The proposed interpolation method ensures that the invoice image keeps features well after scaling. The flippable pixels evaluation mechanism ensures that the pixels keep better connectivity and smoothness and the pattern has highest structural similarity after flipping. The proposed denoising method makes invoice font image smoother and fiter for human vision. Experiments show that the proposed flipping method not only keeps the invoice font structure well but also improves watermarking capacity.

  11. A new pixels flipping method for huge watermarking capacity of the invoice font image. (United States)

    Li, Li; Hou, Qingzheng; Lu, Jianfeng; Xu, Qishuai; Dai, Junping; Mao, Xiaoyang; Chang, Chin-Chen


    Invoice printing just has two-color printing, so invoice font image can be seen as binary image. To embed watermarks into invoice image, the pixels need to be flipped. The more huge the watermark is, the more the pixels need to be flipped. We proposed a new pixels flipping method in invoice image for huge watermarking capacity. The pixels flipping method includes one novel interpolation method for binary image, one flippable pixels evaluation mechanism, and one denoising method based on gravity center and chaos degree. The proposed interpolation method ensures that the invoice image keeps features well after scaling. The flippable pixels evaluation mechanism ensures that the pixels keep better connectivity and smoothness and the pattern has highest structural similarity after flipping. The proposed denoising method makes invoice font image smoother and fiter for human vision. Experiments show that the proposed flipping method not only keeps the invoice font structure well but also improves watermarking capacity.

  12. New Methods for Imaging Evaluation of Chest Wall Deformities

    Directory of Open Access Journals (Sweden)

    Ana Lain


    Full Text Available AimThe purpose of this study is to describe the development of an external 3-dimensional (3D scanner as a noninvasive method for imaging chest wall deformities. It allows objective assessment, reconstruction of the area of interest, and evaluation of the severity of the deformity by using external indexes.External 3D scanning systemThe OrtenBodyOne scanner (Orten, Lyon, France uses depth sensors to scan the entire 3D external body surface of a patient. The depth sensors combine structured light with two classic computer vision techniques: depth from focus and depth from stereo. The data acquired are processed and analyzed using the Orten-Clinic software.Materials and methodsTo investigate the performance of the device, a preliminary prospective study (January 2015–March 2016 was carried out in patients attending our hospital chest wall deformities unit. In total, 100 patients (children and young adults with pectus excavatum or pectus carinatum, treated by surgery or non-operative methods were included. In patients undergoing non-operative treatment, external 3D scanning was performed monthly until complete correction was achieved. In surgically treated patients, scanning was done before and after surgical correction. In 42 patients, computed tomography (CT was additionally performed and correlations between the Haller index calculated by CT and the external Haller index using external scanning were investigated using a Student’s test (r = 0.83.ConclusionExternal scanning is an effective, objective, radiation-free means to diagnose and follow-up patients with chest wall deformities. Externally measured indexes can be used to evaluate the severity of these conditions and the treatment outcomes.

  13. Patient positioning method based on binary image correlation between two edge images for proton-beam radiation therapy

    International Nuclear Information System (INIS)

    Sawada, Akira; Yoda, Kiyoshi; Numano, Masumi; Futami, Yasuyuki; Yamashita, Haruo; Murayama, Shigeyuki; Tsugami, Hironobu


    A new technique based on normalized binary image correlation between two edge images has been proposed for positioning proton-beam radiotherapy patients. A Canny edge detector was used to extract two edge images from a reference x-ray image and a test x-ray image of a patient before positioning. While translating and rotating the edged test image, the absolute value of the normalized binary image correlation between the two edge images is iteratively maximized. Each time before rotation, dilation is applied to the edged test image to avoid a steep reduction of the image correlation. To evaluate robustness of the proposed method, a simulation has been carried out using 240 simulated edged head front-view images extracted from a reference image by varying parameters of the Canny algorithm with a given range of rotation angles and translation amounts in x and y directions. It was shown that resulting registration errors have an accuracy of one pixel in x and y directions and zero degrees in rotation, even when the number of edge pixels significantly differs between the edged reference image and the edged simulation image. Subsequently, positioning experiments using several sets of head, lung, and hip data have been performed. We have observed that the differences of translation and rotation between manual positioning and the proposed method were within one pixel in translation and one degree in rotation. From the results of the validation study, it can be concluded that a significant reduction in workload for the physicians and technicians can be achieved with this method

  14. Unconventional methods of imaging: computational microscopy and compact implementations (United States)

    McLeod, Euan; Ozcan, Aydogan


    In the past two decades or so, there has been a renaissance of optical microscopy research and development. Much work has been done in an effort to improve the resolution and sensitivity of microscopes, while at the same time to introduce new imaging modalities, and make existing imaging systems more efficient and more accessible. In this review, we look at two particular aspects of this renaissance: computational imaging techniques and compact imaging platforms. In many cases, these aspects go hand-in-hand because the use of computational techniques can simplify the demands placed on optical hardware in obtaining a desired imaging performance. In the first main section, we cover lens-based computational imaging, in particular, light-field microscopy, structured illumination, synthetic aperture, Fourier ptychography, and compressive imaging. In the second main section, we review lensfree holographic on-chip imaging, including how images are reconstructed, phase recovery techniques, and integration with smart substrates for more advanced imaging tasks. In the third main section we describe how these and other microscopy modalities have been implemented in compact and field-portable devices, often based around smartphones. Finally, we conclude with some comments about opportunities and demand for better results, and where we believe the field is heading.

  15. Box-Counting Method of 2D Neuronal Image: Method Modification and Quantitative Analysis Demonstrated on Images from the Monkey and Human Brain

    Directory of Open Access Journals (Sweden)

    Nemanja Rajković


    Full Text Available This study calls attention to the difference between traditional box-counting method and its modification. The appropriate scaling factor, influence on image size and resolution, and image rotation, as well as different image presentation, are showed on the sample of asymmetrical neurons from the monkey dentate nucleus. The standard BC method and its modification were evaluated on the sample of 2D neuronal images from the human neostriatum. In addition, three box dimensions (which estimate the space-filling property, the shape, complexity, and the irregularity of dendritic tree were used to evaluate differences in the morphology of type III aspiny neurons between two parts of the neostriatum.

  16. Box-Counting Method of 2D Neuronal Image: Method Modification and Quantitative Analysis Demonstrated on Images from the Monkey and Human Brain. (United States)

    Rajković, Nemanja; Krstonošić, Bojana; Milošević, Nebojša


    This study calls attention to the difference between traditional box-counting method and its modification. The appropriate scaling factor, influence on image size and resolution, and image rotation, as well as different image presentation, are showed on the sample of asymmetrical neurons from the monkey dentate nucleus. The standard BC method and its modification were evaluated on the sample of 2D neuronal images from the human neostriatum. In addition, three box dimensions (which estimate the space-filling property, the shape, complexity, and the irregularity of dendritic tree) were used to evaluate differences in the morphology of type III aspiny neurons between two parts of the neostriatum.

  17. Luminol, horseradish peroxidase, and glucose oxidase ternary functionalized graphene oxide for ultrasensitive glucose sensing. (United States)

    Li, Fang; Ma, Wenjing; Liu, Jiachang; Wu, Xiang; Wang, Yan; He, Jianbo


    Luminol, horseradish peroxidase (HRP), and glucose oxidase (GOx) ternary functionalized graphene oxide (HRP/GOx-luminol-GO) with excellent chemiluminescence (CL) activity and specific enzymatic property was prepared via a simple and general strategy for the first time. In this approach, luminol functionalized GO (luminol-GO) was prepared by gently stirring GO with luminol. Then HRP and GOx were further co-immobilized onto the surface of luminol-GO by storing HRP and GOx with luminol-GO at 4 °C overnight, to form HRP/GOx-luminol-GO bionanocomposites. The synthesized HRP/GOx-luminol-GO could react with H 2 O 2 generated from GOx catalyzed glucose oxidization reaction, to produce strong CL emission in the presence of co-immobilized HRP. Thus, we developed an ultrasensitive, homogeneous, reagentless, selective, and simple CL sensing system for glucose detection. The resulting biosensors exhibited ultra-wide linear range from 5.0 nM to 5.0 mM, and an ultra-low detection limit of 1.2 nM, which was more than 3 orders of magnitude lower than previously reported methods. Furthermore, the sensing system was successfully applied for the detection of glucose in human blood samples.

  18. Zinc oxide nanoparticle-enhanced ultrasensitive chemiluminescence immunoassay for the carcinoma embryonic antigen

    International Nuclear Information System (INIS)

    Pal, Souvik; Bhand, Sunil


    An ultrasensitive enzyme-linked immunosorbent assay (ELISA) is reported for the determination of carcinoma embryonic antigen (CEA) in human serum. It was realized using a microplate reader using a 384-well plate. Monoclonal antibody (Ab) against CEA (1° Ab) acting as the capture probe was immobilized on zinc oxide nanoparticles (ZnO-NPs) in the form of self-assembled monolayers (SAMs). CEA captured by 1° Ab was quantified using a sandwich ELISA wherein a polyclonal second antibody against CEA (2° Ab) was used for detection and quantified using an HRP-labeled secondary antibody (3° Ab). The ZnO-NPs-CEA capture probe was deposited on the bottom of the wells in order to enhance capture of CEA. A 3-fold enhancement in the chemiluminescence (CL) signal of luminol is found (compared to a conventional ELISA). CEA can be quantified by this method in concentrations as low as 1 pg · mL −1 . The upper limit of detection is 20 ng · mL −1 . The use of ZnO-NPs also imparts improved thermal stability. When stored at 4 °C in phosphate-buffered saline of pH 7.4, the probe displays stability of up to 30 days. (author)

  19. Proximity hybridization-mediated isothermal exponential amplification for ultrasensitive electrochemical protein detection

    Directory of Open Access Journals (Sweden)

    Yu Y


    Full Text Available Yanyan Yu, Gaoxing Su, Hongyan Zhu, Qing Zhu, Yong Chen, Bohui Xu, Yuqin Li, Wei Zhang School of Pharmacy, Nantong University, Nantong, People’s Republic of China Abstract: In this study, we fabricated a novel electrochemical biosensing platform on the basis of target-triggered proximity hybridization-mediated isothermal exponential amplification reaction (EXPAR for ultrasensitive protein analysis. Through rational design, the aptamers for protein recognition were integrated within two DNA probes. Via proximity hybridization principle, the affinity protein-binding event was converted into DNA assembly process. The recognition of protein by aptamers can trigger the strand displacement through the increase of the local concentrations of the involved probes. As a consequence, the output DNA was displaced, which can hybridize with the duplex probes immobilized on the electrode surface subsequently, leading to the initiation of the EXPAR as well as the cleavage of duplex probes. Each cleavage will release the gold nanoparticles (AuNPs binding sequence. With the modification of G-quadruplex sequence, electrochemical signals were yielded by the AuNPs through oxidizing 3,3',5,5'-tetramethylbenzidine in the presence of H2O2. The study we proposed exhibited high sensitivity toward platelet-derived growth factor BB (PDGF-BB with the detection limit of 52 fM. And, this method also showed great selectivity among the PDGF isoforms and performed well in spiked human serum samples. Keywords: electrochemical biosensor, proximity hybridization, PDGF-BB, isothermal exponential amplification, G-quadruplex 

  20. A Room Temperature Ultrasensitive Magnetoelectric Susceptometer for Quantitative Tissue Iron Detection (United States)

    Xi, Hao; Qian, Xiaoshi; Lu, Meng-Chien; Mei, Lei; Rupprecht, Sebastian; Yang, Qing X.; Zhang, Q. M.


    Iron is a trace mineral that plays a vital role in the human body. However, absorbing and accumulating excessive iron in body organs (iron overload) can damage or even destroy an organ. Even after many decades of research, progress on the development of noninvasive and low-cost tissue iron detection methods is very limited. Here we report a recent advance in a room-temperature ultrasensitive biomagnetic susceptometer for quantitative tissue iron detection. The biomagnetic susceptometer exploits recent advances in the magnetoelectric (ME) composite sensors that exhibit an ultrahigh AC magnetic sensitivity under the presence of a strong DC magnetic field. The first order gradiometer based on piezoelectric and magnetostrictive laminate (ME composite) structure shows an equivalent magnetic noise of 0.99 nT/rt Hz at 1 Hz in the presence of a DC magnetic field of 0.1 Tesla and a great common mode noise rejection ability. A prototype magnetoelectric liver susceptometry has been demonstrated with liver phantoms. The results indicate its output signals to be linearly responsive to iron concentrations from normal iron dose (0.05 mg Fe/g liver phantom) to 5 mg Fe/g liver phantom iron overload (100X overdose). The results here open up many innovative possibilities for compact-size, portable, cost-affordable, and room-temperature operated medical systems for quantitative determinations of tissue iron.

  1. A Switchable Linker-Based Immunoassay for Ultrasensitive Visible Detection of Salmonella in Tomatoes. (United States)

    Hahn, Jungwoo; Kim, Eunghee; You, Young Sang; Gunasekaran, Sundaram; Lim, Seokwon; Choi, Young Jin


    On-site detection for sensitive identification of foodborne pathogens on fresh produce with minimal use of specialized instrumentation is crucial to the food industry. A switchable linker (SL)-based immunoassay was designed for ultrasensitive on-site detection of Salmonella in tomato samples. The assay is based on large-scale aggregation of gold nanoparticles (GNPs), induced by a quantitative relationship among the biotinylated Salmonella polyclonal antibody (b-Ab) used as the SL, the functionalized GNPs, and Salmonella. Important factors such as the concentration of SLs, time required for large-scale aggregation, and selectivity of b-Ab were optimized to minimize the detection time (within 45 min with gentle agitation) and achieve the lowest limit of detection (LOD; 10 CFU/g in tomato samples) possible. This SL-based immunoassay with its relatively low LOD and short detection time may meet the need for rapid, simple, on-site analysis of pathogens in fresh produce. The novel switchable linker-based immunoassay is a rapid, specific, and sensitive method that has potential applications for routine diagnostics of Salmonella in tomato products. These advantages make it a practical approach for general use in the processing industry to detect Salmonella rapidly and to implement appropriate regulatory procedures. Furthermore, it could be applied to other fresh products including cantaloupe, strawberry, and cucumbers. © 2017 Institute of Food Technologists®.

  2. Ultrasensitive Detection of Ebola Virus Oligonucleotide Based on Upconversion Nanoprobe/Nanoporous Membrane System. (United States)

    Tsang, Ming-Kiu; Ye, WeiWei; Wang, Guojing; Li, Jingming; Yang, Mo; Hao, Jianhua


    Ebola outbreaks are currently of great concern, and therefore, development of effective diagnosis methods is urgently needed. The key for lethal virus detection is high sensitivity, since early-stage detection of virus may increase the probability of survival. Here, we propose a luminescence scheme of assay consisting of BaGdF5:Yb/Er upconversion nanoparticles (UCNPs) conjugated with oligonucleotide probe and gold nanoparticles (AuNPs) linked with target Ebola virus oligonucleotide. As a proof of concept, a homogeneous assay was fabricated and tested, yielding a detection limit at picomolar level. The luminescence resonance energy transfer is ascribed to the spectral overlapping of upconversion luminescence and the absorption characteristics of AuNPs. Moreover, we anchored the UCNPs and AuNPs on a nanoporous alumina (NAAO) membrane to form a heterogeneous assay. Importantly, the detection limit was greatly improved, exhibiting a remarkable value at the femtomolar level. The enhancement is attributed to the increased light-matter interaction throughout the nanopore walls of the NAAO membrane. The specificity test suggested that the nanoprobes were specific to Ebola virus oligonucleotides. The strategy combining UCNPs, AuNPs, and NAAO membrane provides new insight into low-cost, rapid, and ultrasensitive detection of different diseases. Furthermore, we explored the feasibility of clinical application by using inactivated Ebola virus samples. The detection results showed great potential of our heterogeneous design for practical application.

  3. Image Quality Assessment of High-Resolution Satellite Images with Mtf-Based Fuzzy Comprehensive Evaluation Method (United States)

    Wu, Z.; Luo, Z.; Zhang, Y.; Guo, F.; He, L.


    A Modulation Transfer Function (MTF)-based fuzzy comprehensive evaluation method was proposed in this paper for the purpose of evaluating high-resolution satellite image quality. To establish the factor set, two MTF features and seven radiant features were extracted from the knife-edge region of image patch, which included Nyquist, MTF0.5, entropy, peak signal to noise ratio (PSNR), average difference, edge intensity, average gradient, contrast and ground spatial distance (GSD). After analyzing the statistical distribution of above features, a fuzzy evaluation threshold table and fuzzy evaluation membership functions was established. The experiments for comprehensive quality assessment of different natural and artificial objects was done with GF2 image patches. The results showed that the calibration field image has the highest quality scores. The water image has closest image quality to the calibration field, quality of building image is a little poor than water image, but much higher than farmland image. In order to test the influence of different features on quality evaluation, the experiment with different weights were tested on GF2 and SPOT7 images. The results showed that different weights correspond different evaluating effectiveness. In the case of setting up the weights of edge features and GSD, the image quality of GF2 is better than SPOT7. However, when setting MTF and PSNR as main factor, the image quality of SPOT7 is better than GF2.

  4. EIT Imaging of admittivities with a D-bar method and spatial prior: experimental results for absolute and difference imaging. (United States)

    Hamilton, S J


    Electrical impedance tomography (EIT) is an emerging imaging modality that uses harmless electrical measurements taken on electrodes at a body's surface to recover information about the internal electrical conductivity and or permittivity. The image reconstruction task of EIT is a highly nonlinear inverse problem that is sensitive to noise and modeling errors making the image reconstruction task challenging. D-bar methods solve the nonlinear problem directly, bypassing the need for detailed and time-intensive forward models, to provide absolute (static) as well as time-difference EIT images. Coupling the D-bar methodology with the inclusion of high confidence a priori data results in a noise-robust regularized image reconstruction method. In this work, the a priori D-bar method for complex admittivities is demonstrated effective on experimental tank data for absolute imaging for the first time. Additionally, the method is adjusted for, and tested on, time-difference imaging scenarios. The ability of the method to be used for conductivity, permittivity, absolute as well as time-difference imaging provides the user with great flexibility without a high computational cost.

  5. Automated method and system for the alignment and correlation of images from two different modalities (United States)

    Giger, Maryellen L.; Chen, Chin-Tu; Armato, Samuel; Doi, Kunio


    A method and system for the computerized registration of radionuclide images with radiographic images, including generating image data from radiographic and radionuclide images of the thorax. Techniques include contouring the lung regions in each type of chest image, scaling and registration of the contours based on location of lung apices, and superimposition after appropriate shifting of the images. Specific applications are given for the automated registration of radionuclide lungs scans with chest radiographs. The method in the example given yields a system that spatially registers and correlates digitized chest radiographs with V/Q scans in order to correlate V/Q functional information with the greater structural detail of chest radiographs. Final output could be the computer-determined contours from each type of image superimposed on any of the original images, or superimposition of the radionuclide image data, which contains high activity, onto the radiographic chest image.

  6. An Improved Method to Watermark Images Sensitive to Blocking Artifacts


    Afzel Noore


    A new digital watermarking technique for images that are sensitive to blocking artifacts is presented. Experimental results show that the proposed MDCT based approach produces highly imperceptible watermarked images and is robust to attacks such as compression, noise, filtering and geometric transformations. The proposed MDCT watermarking technique is applied to fingerprints for ensuring security. The face image and demographic text data of an individual are used as multi...

  7. The feasibility of images reconstructed with the method of sieves

    International Nuclear Information System (INIS)

    Veklerov, E.; Llacer, J.


    The concept of sieves has been applied with the maximum likelihood estimator (MLE) to image reconstruction. While it makes it possible to recover smooth images consistent with the data, the degree of smoothness provided by it is arbitrary. It is shown that the concept of feasibility is able to resolve this arbitrariness. By varying the values of parameters determining the degree of smoothness, one can generate images on both sides of the feasibility region, as well as within the region. Feasible images recovered by using different sieve parameters are compared with feasible results of other procedures. One- and two-dimensional examples using both simulated and real data sets are considered

  8. Strategically functionalized carbon nanotubes as the ultrasensitive electrochemical probe for picomolar detection of sildenafil citrate (Viagra). (United States)

    Gopalan, Anantha Iyengar; Lee, Kwang Pill; Komathi, Shanmugasundaram


    The present work demonstrates the utility of the functionalized carbon nanotubes, poly(4-aminobenzene sulfonic acid) (PABS) grafted multiwalled carbon nanotubes, MWNT-g-PABS, as an electrode modifier towards achieving ultrasensitive detection of a model drug, sildenafil citrate (SC). PABS units in MWNT-g-PABS interact with SC, pre-concentrate and accumulate at the surface. The electron transduction from SC to electrode is augmented via MWNT-g-PABS. As a result, the MWNT-g-PABS modified electrode exhibited ultrasensitive (57.7 μA/nM) and selective detection of SC with a detection limit of 4.7 pM. The present work provides scope towards targeting ultrasensitivity for the detection of biomolecules/drug through rational design and incorporation of appropriate chemical components to carbon nanotubes. Copyright © 2010 Elsevier B.V. All rights reserved.

  9. A Label-Free and Ultrasensitive Immunosensor for Detection of Human Chorionic Gonadotrophin Based on Graphene FETs. (United States)

    Islam, Kamrul; Suhail, Ahmed; Pan, Genhua


    We report on a label-free immunosensor based on graphene field effect transistors (G-FETs) for the ultrasensitive detection of Human Chorionic Gonadotrophin (hCG), as an indicator of pregnancy and related disorders, such as actopic pregnancy, choriocarcinoma and orchic teratoma. Pyrene based bioactive ester was non-covalently anchored onto the graphene channel in order to retain the sp² lattice. The G-FET transfer characteristics showed repeatable and reliable responses in all surface modifying steps using a direct current (DC) readout system. The hCG concentration gradient showed a detection limit of ~1 pg·mL -1 . The proposed method facilitates the cost-effective and viable production of graphene point-of-care devices for clinical diagnosis.

  10. Demonstration of an ultrasensitive refractive-index plasmonic sensor by enabling its quadrupole resonance in phase interrogation. (United States)

    Lee, Hsin-Cheng; Li, Chung-Tien; Chen, How-Foo; Yen, Ta-Jen


    We present an ultrasensitive plasmonic sensing system by introducing a nanostructured X-shaped plasmonic sensor (XPS) and measuring its localized optical properties in phase interrogation. Our tailored XPS exhibits two major resonant modes of a low-order dipole and a high-order quadrupole, between which the quadrupole resonance allows an ultrahigh sensitivity, due to its higher quality factor. Furthermore, we design an in-house common-path phase-interrogation system, in contrast to conventional wavelength-interrogation methods, to achieve greater sensing capability. The experimental measurement shows that the sensing resolution of the XPS reaches 1.15×10(-6) RIU, not only two orders of magnitude greater than the result of the controlled extinction measurement (i.e., 9.90×10(-5) RIU), but also superior than current reported plasmonic sensors.

  11. A Label-Free and Ultrasensitive Immunosensor for Detection of Human Chorionic Gonadotrophin Based on Graphene FETs

    Directory of Open Access Journals (Sweden)

    Kamrul Islam


    Full Text Available We report on a label-free immunosensor based on graphene field effect transistors (G-FETs for the ultrasensitive detection of Human Chorionic Gonadotrophin (hCG, as an indicator of pregnancy and related disorders, such as actopic pregnancy, choriocarcinoma and orchic teratoma. Pyrene based bioactive ester was non-covalently anchored onto the graphene channel in order to retain the sp2 lattice. The G-FET transfer characteristics showed repeatable and reliable responses in all surface modifying steps using a direct current (DC readout system. The hCG concentration gradient showed a detection limit of ~1 pg·mL−1. The proposed method facilitates the cost-effective and viable production of graphene point-of-care devices for clinical diagnosis.

  12. Moving-Article X-Ray Imaging System and Method for 3-D Image Generation (United States)

    Fernandez, Kenneth R. (Inventor)


    An x-ray imaging system and method for a moving article are provided for an article moved along a linear direction of travel while the article is exposed to non-overlapping x-ray beams. A plurality of parallel linear sensor arrays are disposed in the x-ray beams after they pass through the article. More specifically, a first half of the plurality are disposed in a first of the x-ray beams while a second half of the plurality are disposed in a second of the x-ray beams. Each of the parallel linear sensor arrays is oriented perpendicular to the linear direction of travel. Each of the parallel linear sensor arrays in the first half is matched to a corresponding one of the parallel linear sensor arrays in the second half in terms of an angular position in the first of the x-ray beams and the second of the x-ray beams, respectively.

  13. W-transform method for feature-oriented multiresolution image retrieval

    Energy Technology Data Exchange (ETDEWEB)

    Kwong, M.K.; Lin, B. [Argonne National Lab., IL (United States). Mathematics and Computer Science Div.


    Image database management is important in the development of multimedia technology. Since an enormous amount of digital images is likely to be generated within the next few decades in order to integrate computers, television, VCR, cables, telephone and various imaging devices. Effective image indexing and retrieval systems are urgently needed so that images can be easily organized, searched, transmitted, and presented. Here, the authors present a local-feature-oriented image indexing and retrieval method based on Kwong, and Tang`s W-transform. Multiresolution histogram comparison is an effective method for content-based image indexing and retrieval. However, most recent approaches perform multiresolution analysis for whole images but do not exploit the local features present in the images. Since W-transform is featured by its ability to handle images of arbitrary size, with no periodicity assumptions, it provides a natural tool for analyzing local image features and building indexing systems based on such features. In this approach, the histograms of the local features of images are used in the indexing, system. The system not only can retrieve images that are similar or identical to the query images but also can retrieve images that contain features specified in the query images, even if the retrieved images as a whole might be very different from the query images. The local-feature-oriented method also provides a speed advantage over the global multiresolution histogram comparison method. The feature-oriented approach is expected to be applicable in managing large-scale image systems such as video databases and medical image databases.

  14. Signal to noise comparison of metabolic imaging methods on a clinical 3T MRI

    DEFF Research Database (Denmark)

    Müller, C. A.; Hansen, Rie Beck; Skinner, J. G.

    MRI with hyperpolarized tracers has enabled new diagnostic applications, e.g. metabolic imaging in cancer research. However, the acquisition of the transient, hyperpolarized signal with spatial and frequency resolution requires dedicated imaging methods. Here, we compare three promising candidate...... for 2D MR spectroscopic imaging (MRSI): (i) multi-echo balanced steady-state free precession (me-bSSFP), 1,2 (ii) echo planar spectroscopic imaging (EPSI) sequence and (iii) phase-encoded, pulseacquisition chemical-shift imaging (CSI)...

  15. Wavelet-Based Bayesian Methods for Image Analysis and Automatic Target Recognition

    National Research Council Canada - National Science Library

    Nowak, Robert


    .... We have developed two new techniques. First, we have develop a wavelet-based approach to image restoration and deconvolution problems using Bayesian image models and an alternating-maximation method...

  16. A Novel Image Tag Completion Method Based on Convolutional Neural Transformation

    KAUST Repository

    Geng, Yanyan; Zhang, Guohui; Li, Weizhi; Gu, Yi; Liang, Ru-Ze; Liang, Gaoyuan; Wang, Jingbin; Wu, Yanbin; Patil, Nitin; Wang, Jing-Yan


    In the problems of image retrieval and annotation, complete textual tag lists of images play critical roles. However, in real-world applications, the image tags are usually incomplete, thus it is important to learn the complete tags for images. In this paper, we study the problem of image tag complete and proposed a novel method for this problem based on a popular image representation method, convolutional neural network (CNN). The method estimates the complete tags from the convolutional filtering outputs of images based on a linear predictor. The CNN parameters, linear predictor, and the complete tags are learned jointly by our method. We build a minimization problem to encourage the consistency between the complete tags and the available incomplete tags, reduce the estimation error, and reduce the model complexity. An iterative algorithm is developed to solve the minimization problem. Experiments over benchmark image data sets show its effectiveness.

  17. A Novel Image Tag Completion Method Based on Convolutional Neural Transformation

    KAUST Repository

    Geng, Yanyan


    In the problems of image retrieval and annotation, complete textual tag lists of images play critical roles. However, in real-world applications, the image tags are usually incomplete, thus it is important to learn the complete tags for images. In this paper, we study the problem of image tag complete and proposed a novel method for this problem based on a popular image representation method, convolutional neural network (CNN). The method estimates the complete tags from the convolutional filtering outputs of images based on a linear predictor. The CNN parameters, linear predictor, and the complete tags are learned jointly by our method. We build a minimization problem to encourage the consistency between the complete tags and the available incomplete tags, reduce the estimation error, and reduce the model complexity. An iterative algorithm is developed to solve the minimization problem. Experiments over benchmark image data sets show its effectiveness.

  18. A Geometric Correction Method of Plane Image Based on OpenCV

    Directory of Open Access Journals (Sweden)

    Li Xiaopeng


    Full Text Available Using OpenCV, a geometric correction method of plane image from single grid image in a state of unknown camera position is presented. The method can remove the perspective and lens distortions from an image. The method is simple and easy to implement, and the efficiency is high. Experiments indicate that this method has high precision, and can be used in some domains such as plane measurement.

  19. The establishment of the method of three dimension volumetric fusion of emission and transmission images for PET imaging

    International Nuclear Information System (INIS)

    Zhang Xiangsong; He Zuoxiang


    Objective: To establish the method of three dimension volumetric fusion of emission and transmission images for PET imaging. Methods: The volume data of emission and transmission images acquired with Siemens ECAT HR + PET scanner were transferred to PC computer by local area network. The PET volume data were converted into 8 bit byte type, and scaled to the range of 0-255. The data coordinates of emission and transmission images were normalized by three-dimensional coordinate conversion in the same way. The images were fused with the mode of alpha-blending. The accuracy of image fusion was confirmed by its clinical application in 13 cases. Results: The three dimension volumetric fusion of emission and transmission images clearly displayed the silhouette and anatomic configuration in chest, including chest wall, lung, heart, mediastinum, et al. Forty-eight lesions in chest in 13 cases were accurately located by the image fusion. Conclusions: The volume data of emission and transmission images acquired with Siemens ECAT HR + PET scanner have the same data coordinate. The three dimension fusion software can conveniently used for the three dimension volumetric fusion of emission and transmission images, and also can correctly locate the lesions in chest

  20. Image combination enhancement method for X-ray compton back-scattering security inspection body scanner

    International Nuclear Information System (INIS)

    Wang Huaiying; Zhang Yujin; Yang Lirui; Li Dong


    As for X-ray Compton Back-Scattering (CBS) body scanner, image clearness is very important for the performance of detecting the contraband hidden on the body. A new image combination enhancement method is provided based on characteristics of CBS body images and points of human vision. After processed by this method, the CBS image will be obviously improved with clear levels, distinct outline and uniform background. (authors)

  1. Basic artefacts of diagnostic imaging by the magnetic resonance method

    International Nuclear Information System (INIS)

    Vitak, T.; Seidl, Z.; Obenberger, J.; Vaneckova, M.; Danes, J.; Krasensky, J.; Peterkova, V


    Artefacts in diagnostic imaging are defined as a geometric or anatomic misrepresentation of the reality by the image formed. The article deals with artefacts due to field and frequency shifts, in particular due to the water-fat chemical shift and due to magnetic susceptibility. The physical nature of the artefacts is explained and their diagnostic significance is discussed. (P.A.)

  2. Solubility investigations in support of ultrasensitive noble gas detector development

    International Nuclear Information System (INIS)

    Gross, K.C.; Andersen, A.; Russ, W.R.; Stuenkel, D.; Valentine, J.D.


    Argonne National Laboratory (ANL) and the University of Cincinnati (UC) have been developing a new class of ultrasensitive noble gas detectors that are based upon the ANL discovery that corn oil has a high affinity for heavy noble gas absorption at room temperature but releases the noble gases with warming or by other low-energy-input means. Environmental applications for this new class of fluid-based detectors include ultrahigh sensitivity radioxenon detectors for comprehensive test ban treaty surveillance, improved fission gas detectors for enhanced environmental surveillance in the vicinity of US Department of Energy, US Department of Defense, and US Nuclear Regulatory Commission licensed facilities, and improved integrating Rn detectors for earthquake prediction. They present the results of theoretical and experimental investigations into the solubility phenomena of heavy noble gases (Rn, Xe, and Kr) in triglyceride oils. They intend for the findings presented herein to be used to guide future selection, development, and refinement of vegetable and other hydrocarbon oils to bring further enhancements to noble gas detection efficiencies

  3. Specific and ultrasensitive ciprofloxacin detection by responsive photonic crystal sensor

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Rong; Wang, Yong [Department of Chemistry, School of Science, Tianjin University, Tianjin 300072 (China); Yu, Li-Ping, E-mail: [Department of Chemistry, School of Science, Tianjin University, Tianjin 300072 (China); State Key Laboratory of Medicinal Chemical Biology, Nankai University, Tianjin 300071 (China)


    Highlights: • Sensor was designed by integrating complexes into responsive photonic crystal. • Ternary tryptophan–zinc(II)–ciprofloxacin complexes were chosen for sensing. • Excellent sensing of ciprofloxacin was achieved in aqueous media. - Abstract: A new approach for specific and ultrasensitive measurement of ciprofloxacin has been developed by integrating ternary complexes into responsive photonic crystal (RPC). Tryptophan was first immobilized within the polyacrylamide hydrogel substrates of RPC. The determination of ciprofloxacin was via the existence of zinc(II) ions that function as a ‘bridge’ to form specific tryptophan–zinc(II)–ciprofloxacin complexes step by step, which resulted in a stepwise red-shift of the diffraction wavelength. A maximum wavelength shift from 798 to 870 nm for ciprofloxacin was observed when the RPC film was immersed in 10{sup −4} M ciprofloxacin. A linear relationship has been obtained between the Δλ of diffraction peak and logarithm of ciprofloxacin concentration at pH 5.0 in the range of 10{sup −10} to 10{sup −4} M. And the least detectable concentration in present work is about 5 × 10{sup −11} M. The results demonstrated that the as-designed ternary complexes-based RPC sensor exhibited high sensitivity, satisfactory specificity and excellent recoverability for sensing of ciprofloxacin in aqueous media and were validated by detecting ciprofloxacin in the eye-drop sample.

  4. Ultra-sensitive radionuclide spectrometry. Radiometrics and mass spectrometry synergy

    International Nuclear Information System (INIS)

    Povinec, P.P.


    Recent developments in radiometrics and mass spectrometry techniques for ultra-sensitive analysis of radionuclides in the marine environment are reviewed. In the radiometrics sector the dominant development has been the utilization of large HPGe detectors in underground laboratories with anti-cosmic or anti-Compton shielding for the analysis of short and medium-lived radionuclides in the environment. In the mass spectrometry sector, applications of inductively coupled plasma mass spectrometry (ICP-MS) and accelerator mass spectrometry (AMS) for the analysis of long-lived radionuclides in the environment are the most important recent achievements. The recent developments do not only considerably decrease the detection limits for several radionuclides (up to several orders of magnitude), but they also enable to decrease sample volumes so that sampling, e.g., of the water column can be much easier and more effective. A comparison of radiometrics and mass spectrometry results for the analysis of radionuclides in the marine environment shows a reasonable agreement - within quoted uncertainties, for wide range of activities and different sample matrices analyzed. (author)

  5. Fabrication of ultra-sensitive leak detection standards

    International Nuclear Information System (INIS)

    Winkelman, C.R.


    The primary difficulty with flow rate measurements below 10 -10 standard cubic centimeters per second (std. cc/sec) is that there are no commercially available standards. The requirements, however, dictate that the problem of design and construction of a qualifiable standard in the ultra-sensitive range had to be solved. There are a number of leak types which were considered - capillary leaks, orifice leaks, and the pore type leaks, among others. The capillary leak was not used because of the cracking or sorting effects that are common to this type leak. For example, a gas blend flowing through a capillary leak will result in the lighter gases passing through the leak first. The difficulty of fabricating the proper hole size in relation to the flow rate requirements ruled out the orifice type leak. The choice was the pore type leak which utilizes the basic concept of a stainless steel knife edge driven into a fixed section composed of stainless steel with a gold over-lay and maintained under force

  6. Solubility investigations in support of ultrasensitive noble gas detector development

    International Nuclear Information System (INIS)

    Gross, K. C.


    Argonne National Laboratory (ANL) and the University of Cincinnati (UC) have been developing a new class of ultrasensitive noble gas detectors that are based upon the ANL discovery that corn oil has a high affinity for heavy noble gas absorption at room temperature, but releases the noble gases with warming or by other low-energy-input means. Environmental applications for this new class of fluid-based detectors include ultrahigh sensitivity radioxenon detectors for Comprehensive Test Ban Treaty Surveillance, improved fission gas detectors for enhanced environmental surveillance in the vicinity of DOE, DOD, and NRC-licensed facilities, and improved integrating Rn detectors for earthquake prediction. The purpose of the present paper is to present the results of theoretical and experimental investigations into the solubility phenomena of heavy noble gases (Rn, Xe, and Kr) in triglyceride oils. It is the authors' intention that the findings presented herein may be used to guide future selection, development, and refinement of vegetable and other hydrocarbon oils to bring further enhancements to noble gas detection efficiencies

  7. Aptamer-Phage Reporters for Ultrasensitive Lateral Flow Assays. (United States)

    Adhikari, Meena; Strych, Ulrich; Kim, Jinsu; Goux, Heather; Dhamane, Sagar; Poongavanam, Mohan-Vivekanandan; Hagström, Anna E V; Kourentzi, Katerina; Conrad, Jacinta C; Willson, Richard C


    We introduce the modification of bacteriophage particles with aptamers for use as bioanalytical reporters, and demonstrate the use of these particles in ultrasensitive lateral flow assays. M13 phage displaying an in vivo biotinylatable peptide (AviTag) genetically fused to the phage tail protein pIII were used as reporter particle scaffolds, with biotinylated aptamers attached via avidin-biotin linkages, and horseradish peroxidase (HRP) reporter enzymes covalently attached to the pVIII coat protein. These modified viral nanoparticles were used in immunochromatographic sandwich assays for the direct detection of IgE and of the penicillin-binding protein from Staphylococcus aureus (PBP2a). We also developed an additional lateral flow assay for IgE, in which the analyte is sandwiched between immobilized anti-IgE antibodies and aptamer-bearing reporter phage modified with HRP. The limit of detection of this LFA was 0.13 ng/mL IgE, ∼100 times lower than those of previously reported IgE assays.

  8. A method of combining STEM image with parallel beam diffraction and electron-optical conditions for diffractive imaging

    International Nuclear Information System (INIS)

    He Haifeng; Nelson, Chris


    We describe a method of combining STEM imaging functionalities with nanoarea parallel beam electron diffraction on a modern TEM. This facilitates the search for individual particles whose diffraction patterns are needed for diffractive imaging or structural studies of nanoparticles. This also lays out a base for 3D diffraction data collection

  9. Reliable clarity automatic-evaluation method for optical remote sensing images (United States)

    Qin, Bangyong; Shang, Ren; Li, Shengyang; Hei, Baoqin; Liu, Zhiwen


    Image clarity, which reflects the sharpness degree at the edge of objects in images, is an important quality evaluate index for optical remote sensing images. Scholars at home and abroad have done a lot of work on estimation of image clarity. At present, common clarity-estimation methods for digital images mainly include frequency-domain function methods, statistical parametric methods, gradient function methods and edge acutance methods. Frequency-domain function method is an accurate clarity-measure approach. However, its calculation process is complicate and cannot be carried out automatically. Statistical parametric methods and gradient function methods are both sensitive to clarity of images, while their results are easy to be affected by the complex degree of images. Edge acutance method is an effective approach for clarity estimate, while it needs picking out the edges manually. Due to the limits in accuracy, consistent or automation, these existing methods are not applicable to quality evaluation of optical remote sensing images. In this article, a new clarity-evaluation method, which is based on the principle of edge acutance algorithm, is proposed. In the new method, edge detection algorithm and gradient search algorithm are adopted to automatically search the object edges in images. Moreover, The calculation algorithm for edge sharpness has been improved. The new method has been tested with several groups of optical remote sensing images. Compared with the existing automatic evaluation methods, the new method perform better both in accuracy and consistency. Thus, the new method is an effective clarity evaluation method for optical remote sensing images.

  10. Evaluation of an automatic brain segmentation method developed for neonates on adult MR brain images (United States)

    Moeskops, Pim; Viergever, Max A.; Benders, Manon J. N. L.; Išgum, Ivana


    Automatic brain tissue segmentation is of clinical relevance in images acquired at all ages. The literature presents a clear distinction between methods developed for MR images of infants, and methods developed for images of adults. The aim of this work is to evaluate a method developed for neonatal images in the segmentation of adult images. The evaluated method employs supervised voxel classification in subsequent stages, exploiting spatial and intensity information. Evaluation was performed using images available within the MRBrainS13 challenge. The obtained average Dice coefficients were 85.77% for grey matter, 88.66% for white matter, 81.08% for cerebrospinal fluid, 95.65% for cerebrum, and 96.92% for intracranial cavity, currently resulting in the best overall ranking. The possibility of applying the same method to neonatal as well as adult images can be of great value in cross-sectional studies that include a wide age range.

  11. A novel electrochemical sensing strategy for rapid and ultrasensitive detection of Salmonella by rolling circle amplification and DNA–AuNPs probe

    Energy Technology Data Exchange (ETDEWEB)

    Zhu, Dan; Yan, Yurong; Lei, Pinhua; Shen, Bo [Key Laboratory of Clinical Laboratory Diagnostics (Ministry of Education), College of Laboratory Medicine, Chongqing Medical University, Chongqing 400016 (China); Cheng, Wei [Key Laboratory of Clinical Laboratory Diagnostics (Ministry of Education), College of Laboratory Medicine, Chongqing Medical University, Chongqing 400016 (China); The Center for Clinical Molecular Medical detection, The First Affiliated Hospital of Chongqing Medical University, Chongqing 400016 (China); Ju, Huangxian [Key Laboratory of Clinical Laboratory Diagnostics (Ministry of Education), College of Laboratory Medicine, Chongqing Medical University, Chongqing 400016 (China); State Key Laboratory of Analytical Chemistry for Life Science, Department of Chemistry, Nanjing University, Nanjing 210093 (China); Ding, Shijia, E-mail: [Key Laboratory of Clinical Laboratory Diagnostics (Ministry of Education), College of Laboratory Medicine, Chongqing Medical University, Chongqing 400016 (China)


    A novel electrochemical sensing strategy was developed for ultrasensitive and rapid detection of Salmonella by combining the rolling circle amplification with DNA–AuNPs probe. - Highlights: • This paper presented a novel sensing strategy for the rapid and ultrasensitive detection for Salmonella. • Combination of rolling circle amplification and DNA–AuNPs probe is the first time for Salmonella electrochemical detection. • The method displayed excellent sensitivity and specificity for detection of Salmonella. • The fabricated biosensor was successfully applied to detect Salmonella in milk samples. - Abstract: A novel electrochemical sensing strategy was developed for ultrasensitive and rapid detection of Salmonella by combining the rolling circle amplification with DNA–AuNPs probe. The target DNA could be specifically captured by probe 1 on the sensing interface. Then the circularization mixture was added to form a typical sandwich structure. In the presence of dNTPs and phi29 DNA polymerase, the RCA was initiated to produce micrometer-long single-strand DNA. Finally, the detection probe (DNA–AuNPs) could recognize RCA product to produce enzymatic electrochemical signal. Under optimal conditions, the calibration curve of synthetic target DNA had good linearity from 10 aM to 10 pM with a detection limit of 6.76 aM (S/N = 3). The developed method had been successfully applied to detect Salmonella as low as 6 CFU mL{sup −1} in real milk sample. This proposed strategy showed great potential for clinical diagnosis, food safety and environmental monitoring.

  12. A novel electrochemical sensing strategy for rapid and ultrasensitive detection of Salmonella by rolling circle amplification and DNA–AuNPs probe

    International Nuclear Information System (INIS)

    Zhu, Dan; Yan, Yurong; Lei, Pinhua; Shen, Bo; Cheng, Wei; Ju, Huangxian; Ding, Shijia


    A novel electrochemical sensing strategy was developed for ultrasensitive and rapid detection of Salmonella by combining the rolling circle amplification with DNA–AuNPs probe. - Highlights: • This paper presented a novel sensing strategy for the rapid and ultrasensitive detection for Salmonella. • Combination of rolling circle amplification and DNA–AuNPs probe is the first time for Salmonella electrochemical detection. • The method displayed excellent sensitivity and specificity for detection of Salmonella. • The fabricated biosensor was successfully applied to detect Salmonella in milk samples. - Abstract: A novel electrochemical sensing strategy was developed for ultrasensitive and rapid detection of Salmonella by combining the rolling circle amplification with DNA–AuNPs probe. The target DNA could be specifically captured by probe 1 on the sensing interface. Then the circularization mixture was added to form a typical sandwich structure. In the presence of dNTPs and phi29 DNA polymerase, the RCA was initiated to produce micrometer-long single-strand DNA. Finally, the detection probe (DNA–AuNPs) could recognize RCA product to produce enzymatic electrochemical signal. Under optimal conditions, the calibration curve of synthetic target DNA had good linearity from 10 aM to 10 pM with a detection limit of 6.76 aM (S/N = 3). The developed method had been successfully applied to detect Salmonella as low as 6 CFU mL −1 in real milk sample. This proposed strategy showed great potential for clinical diagnosis, food safety and environmental monitoring

  13. Comparison of classification algorithms for various methods of preprocessing radar images of the MSTAR base (United States)

    Borodinov, A. A.; Myasnikov, V. V.


    The present work is devoted to comparing the accuracy of the known qualification algorithms in the task of recognizing local objects on radar images for various image preprocessing methods. Preprocessing involves speckle noise filtering and normalization of the object orientation in the image by the method of image moments and by a method based on the Hough transform. In comparison, the following classification algorithms are used: Decision tree; Support vector machine, AdaBoost, Random forest. The principal component analysis is used to reduce the dimension. The research is carried out on the objects from the base of radar images MSTAR. The paper presents the results of the conducted studies.

  14. Recent advances in computational methods and clinical applications for spine imaging

    CERN Document Server

    Glocker, Ben; Klinder, Tobias; Li, Shuo


    This book contains the full papers presented at the MICCAI 2014 workshop on Computational Methods and Clinical Applications for Spine Imaging. The workshop brought together scientists and clinicians in the field of computational spine imaging. The chapters included in this book present and discuss the new advances and challenges in these fields, using several methods and techniques in order to address more efficiently different and timely applications involving signal and image acquisition, image processing and analysis, image segmentation, image registration and fusion, computer simulation, image based modeling, simulation and surgical planning, image guided robot assisted surgical and image based diagnosis. The book also includes papers and reports from the first challenge on vertebra segmentation held at the workshop.

  15. Statistical methods for segmentation and classification of images

    DEFF Research Database (Denmark)

    Rosholm, Anders


    The central matter of the present thesis is Bayesian statistical inference applied to classification of images. An initial review of Markov Random Fields relates to the modeling aspect of the indicated main subject. In that connection, emphasis is put on the relatively unknown sub-class of Pickard...... with a Pickard Random Field modeling of a considered (categorical) image phenomemon. An extension of the fast PRF based classification technique is presented. The modification introduces auto-correlation into the model of an involved noise process, which previously has been assumed independent. The suitability...... of the extended model is documented by tests on controlled image data containing auto-correlated noise....

  16. Hybrid Geometric Calibration Method for Multi-Platform Spaceborne SAR Image with Sparse Gcps (United States)

    Lv, G.; Tang, X.; Ai, B.; Li, T.; Chen, Q.


    Geometric calibration is able to provide high-accuracy geometric coordinates of spaceborne SAR image through accurate geometric parameters in the Range-Doppler model by ground control points (GCPs). However, it is very difficult to obtain GCPs that covering large-scale areas, especially in the mountainous regions. In addition, the traditional calibration method is only used for single platform SAR images and can't support the hybrid geometric calibration for multi-platform images. To solve the above problems, a hybrid geometric calibration method for multi-platform spaceborne SAR images with sparse GCPs is proposed in this paper. First, we calibrate the master image that contains GCPs. Secondly, the point tracking algorithm is used to obtain the tie points (TPs) between the master and slave images. Finally, we calibrate the slave images using TPs as the GCPs. We take the Beijing-Tianjin- Hebei region as an example to study SAR image hybrid geometric calibration method using 3 TerraSAR-X images, 3 TanDEM-X images and 5 GF-3 images covering more than 235 kilometers in the north-south direction. Geometric calibration of all images is completed using only 5 GCPs. The GPS data extracted from GNSS receiver are used to assess the plane accuracy after calibration. The results after geometric calibration with sparse GCPs show that the geometric positioning accuracy is 3 m for TSX/TDX images and 7.5 m for GF-3 images.

  17. Color image enhancement of medical images using alpha-rooting and zonal alpha-rooting methods on 2D QDFT (United States)

    Grigoryan, Artyom M.; John, Aparna; Agaian, Sos S.


    2-D quaternion discrete Fourier transform (2-D QDFT) is the Fourier transform applied to color images when the color images are considered in the quaternion space. The quaternion numbers are four dimensional hyper-complex numbers. Quaternion representation of color image allows us to see the color of the image as a single unit. In quaternion approach of color image enhancement, each color is seen as a vector. This permits us to see the merging effect of the color due to the combination of the primary colors. The color images are used to be processed by applying the respective algorithm onto each channels separately, and then, composing the color image from the processed channels. In this article, the alpha-rooting and zonal alpha-rooting methods are used with the 2-D QDFT. In the alpha-rooting method, the alpha-root of the transformed frequency values of the 2-D QDFT are determined before taking the inverse transform. In the zonal alpha-rooting method, the frequency spectrum of the 2-D QDFT is divided by different zones and the alpha-rooting is applied with different alpha values for different zones. The optimization of the choice of alpha values is done with the genetic algorithm. The visual perception of 3-D medical images is increased by changing the reference gray line.

  18. Accuracy of signal-to-noise ratio measurement method for magnetic resonance images

    International Nuclear Information System (INIS)

    Ogura, Akio; Miyai, Akira; Maeda, Fumie; Fukutake, Hiroyuki; Kikumoto, Rikiya


    The signal-to-noise ratio (SNR) of a magnetic resonance image is a common measure of imager performance. However, evaluations for the calculation of the SNR use various methods. A problem with measuring SNR is caused by the distortion of noise statistics in commonly used magnitude images. In this study, measurement accuracy was compared among four methods of evaluating SNR according to the size and position of regions of interest (ROIs). The results indicated that the method that used the difference between two images showed the best agreement with the theoretical value. In the method that used a single image, the SNR calculated by using a small size of ROI showed better agreement with the theoretical value because of noise bias and image artifacts. However, in the method that used the difference between two images, a large size of ROI was better in reducing statistical errors. In the same way, the methods that used air noise and air signal were better when applied to a large ROI. In addition, the image subtraction process used to calculate pixel-by-pixel differences in images may reach zero on a minus pixel value when using an image processor with the MRI system and apparatuses associated with it. A revised equation is presented for this case. It is important to understand the characteristics of each method and to choose a suitable method carefully according to the purpose of the study. (author)

  19. Introduction to statistical methods in signal and image processing


    Forbes , Florence


    Doctoral; This is a 3 part lecture starting with basics on Bayesian analysis in particular for image and signal analysis applications. The last part is devoted to an introduction to variational approximations.

  20. Robust Methods for Image Processing in Anthropology and Biomedicine

    Czech Academy of Sciences Publication Activity Database

    Kalina, Jan

    -, č. 86 (2011), s. 53-53 ISSN 0926-4981 Institutional research plan: CEZ:AV0Z10300504 Keywords : image analysis * robust estimation * forensic anthropology Subject RIV: BB - Applied Statistics, Operational Research

  1. A simple approach for ultrasensitive detection of bisphenols by multiplexed surface-enhanced Raman scattering

    Energy Technology Data Exchange (ETDEWEB)

    De Bleye, C., E-mail:; Dumont, E.; Hubert, C.; Sacré, P.-Y.; Netchacovitch, L.; Chavez, P.-F.; Hubert, Ph.; Ziemons, E.


    Bisphenol A (BPA) is well known for its use in plastic manufacture and thermal paper production despite its risk of health toxicity as an endocrine disruptor in humans. Since the publication of new legislation regarding the use of BPA, manufacturers have begun to replace BPA with other phenolic molecules such as bisphenol F (BPF) and bisphenol B (BPB), but there are no guarantees regarding the health safety of these compounds at this time. In this context, a very simple, cheap and fast surface-enhanced Raman scattering (SERS) method was developed for the sensitive detection of these molecules in spiked tap water solutions. Silver nanoparticles were used as SERS substrates. An original strategy was employed to circumvent the issue of the affinity of bisphenols for metallic surfaces and the silver nanoparticles surface was functionalized using pyridine in order to improve again the sensitivity of the detection. Semi-quantitative detections were performed in tap water solutions at a concentrations range from 0.25 to 20 μg L{sup −1} for BPA and BPB and from 5 to 100 μg L{sup −1} for BPF. Moreover, a feasibility study for performing a multiplex-SERS detection of these molecules was also performed before successfully implementing the developed SERS method on real samples. - Highlights: • Development of a simple, fast and ultrasensitive SERS method to detect bisphenols. • Multiplexed-SERS detection of bisphenol A, bisphenol B and bisphenol F. • Implementation of the SERS developed method on real samples to detect bisphenols.

  2. A simple approach for ultrasensitive detection of bisphenols by multiplexed surface-enhanced Raman scattering

    International Nuclear Information System (INIS)

    De Bleye, C.; Dumont, E.; Hubert, C.; Sacré, P.-Y.; Netchacovitch, L.; Chavez, P.-F.; Hubert, Ph.; Ziemons, E.


    Bisphenol A (BPA) is well known for its use in plastic manufacture and thermal paper production despite its risk of health toxicity as an endocrine disruptor in humans. Since the publication of new legislation regarding the use of BPA, manufacturers have begun to replace BPA with other phenolic molecules such as bisphenol F (BPF) and bisphenol B (BPB), but there are no guarantees regarding the health safety of these compounds at this time. In this context, a very simple, cheap and fast surface-enhanced Raman scattering (SERS) method was developed for the sensitive detection of these molecules in spiked tap water solutions. Silver nanoparticles were used as SERS substrates. An original strategy was employed to circumvent the issue of the affinity of bisphenols for metallic surfaces and the silver nanoparticles surface was functionalized using pyridine in order to improve again the sensitivity of the detection. Semi-quantitative detections were performed in tap water solutions at a concentrations range from 0.25 to 20 μg L −1 for BPA and BPB and from 5 to 100 μg L −1 for BPF. Moreover, a feasibility study for performing a multiplex-SERS detection of these molecules was also performed before successfully implementing the developed SERS method on real samples. - Highlights: • Development of a simple, fast and ultrasensitive SERS method to detect bisphenols. • Multiplexed-SERS detection of bisphenol A, bisphenol B and bisphenol F. • Implementation of the SERS developed method on real samples to detect bisphenols

  3. Nuclear medicine and imaging research (instrumentation and quantitative methods of evaluation)

    International Nuclear Information System (INIS)

    Beck, R.N.; Cooper, M.; Chen, C.T.


    This document is the annual progress report for project entitled ''Instrumentation and Quantitative Methods of Evaluation.'' Progress is reported in separate sections individually abstracted and indexed for the database. Subject areas reported include theoretical studies of imaging systems and methods, hardware developments, quantitative methods of evaluation, and knowledge transfer: education in quantitative nuclear medicine imaging

  4. An ultrasensitive hollow-silica-based biosensor for pathogenic Escherichia coli DNA detection. (United States)

    Ariffin, Eda Yuhana; Lee, Yook Heng; Futra, Dedi; Tan, Ling Ling; Karim, Nurul Huda Abd; Ibrahim, Nik Nuraznida Nik; Ahmad, Asmat


    A novel electrochemical DNA biosensor for ultrasensitive and selective quantitation of Escherichia coli DNA based on aminated hollow silica spheres (HSiSs) has been successfully developed. The HSiSs were synthesized with facile sonication and heating techniques. The HSiSs have an inner and an outer surface for DNA immobilization sites after they have been functionalized with 3-aminopropyltriethoxysilane. From field emission scanning electron microscopy images, the presence of pores was confirmed in the functionalized HSiSs. Furthermore, Brunauer-Emmett-Teller (BET) analysis indicated that the HSiSs have four times more surface area than silica spheres that have no pores. These aminated HSiSs were deposited onto a screen-printed carbon paste electrode containing a layer of gold nanoparticles (AuNPs) to form a AuNP/HSiS hybrid sensor membrane matrix. Aminated DNA probes were grafted onto the AuNP/HSiS-modified screen-printed electrode via imine covalent bonds with use of glutaraldehyde cross-linker. The DNA hybridization reaction was studied by differential pulse voltammetry using an anthraquinone redox intercalator as the electroactive DNA hybridization label. The DNA biosensor demonstrated a linear response over a wide target sequence concentration range of 1.0×10 -12 -1.0×10 -2 μM, with a low detection limit of 8.17×10 -14 μM (R 2 = 0.99). The improved performance of the DNA biosensor appeared to be due to the hollow structure and rough surface morphology of the hollow silica particles, which greatly increased the total binding surface area for high DNA loading capacity. The HSiSs also facilitated molecule diffusion through the silica hollow structure, and substantially improved the overall DNA hybridization assay. Graphical abstract Step-by-step DNA biosensor fabrication based on aminated hollow silica spheres.

  5. An atlas-based multimodal registration method for 2D images with discrepancy structures. (United States)

    Lv, Wenchao; Chen, Houjin; Peng, Yahui; Li, Yanfeng; Li, Jupeng


    An atlas-based multimodal registration method for 2-dimension images with discrepancy structures was proposed in this paper. Atlas was utilized for complementing the discrepancy structure information in multimodal medical images. The scheme includes three steps: floating image to atlas registration, atlas to reference image registration, and field-based deformation. To evaluate the performance, a frame model, a brain model, and clinical images were employed in registration experiments. We measured the registration performance by the squared sum of intensity differences. Results indicate that this method is robust and performs better than the direct registration for multimodal images with discrepancy structures. We conclude that the proposed method is suitable for multimodal images with discrepancy structures. Graphical Abstract An Atlas-based multimodal registration method schematic diagram.

  6. Spiking cortical model-based nonlocal means method for speckle reduction in optical coherence tomography images (United States)

    Zhang, Xuming; Li, Liu; Zhu, Fei; Hou, Wenguang; Chen, Xinjian


    Optical coherence tomography (OCT) images are usually degraded by significant speckle noise, which will strongly hamper their quantitative analysis. However, speckle noise reduction in OCT images is particularly challenging because of the difficulty in differentiating between noise and the information components of the speckle pattern. To address this problem, the spiking cortical model (SCM)-based nonlocal means method is presented. The proposed method explores self-similarities of OCT images based on rotation-invariant features of image patches extracted by SCM and then restores the speckled images by averaging the similar patches. This method can provide sufficient speckle reduction while preserving image details very well due to its effectiveness in finding reliable similar patches under high speckle noise contamination. When applied to the retinal OCT image, this method provides signal-to-noise ratio improvements of >16 dB with a small 5.4% loss of similarity.

  7. A bibliography on blind methods for identifying image forgery

    Czech Academy of Sciences Publication Activity Database

    Saic, Stanislav; Mahdian, Babak


    Roč. 25, č. 6 (2010), s. 389-399 ISSN 0923-5965 R&D Projects: GA ČR(CZ) GA102/08/0470 Institutional research plan: CEZ:AV0Z10750506 Keywords : Image forensics * Digital forgery * Image tampering * Blind forgery detection * Multimedia security Subject RIV: IN - Informatics, Computer Science Impact factor: 1.186, year: 2010

  8. Image restoration by the method of convex projections: part 2 applications and numerical results. (United States)

    Sezan, M I; Stark, H


    The image restoration theory discussed in a previous paper by Youla and Webb [1] is applied to a simulated image and the results compared with the well-known method known as the Gerchberg-Papoulis algorithm. The results show that the method of image restoration by projection onto convex sets, by providing a convenient technique for utilizing a priori information, performs significantly better than the Gerchberg-Papoulis method.

  9. Ultrasensitive direct competitive FLISA using highly luminescent quantum dot beads for tuning affinity of competing antigens to antibodies

    Energy Technology Data Exchange (ETDEWEB)

    Xiong, Sicheng; Zhou, Yaofeng; Huang, Xiaolin [State Key Laboratory of Food Science and Technology, Nanchang University, Nanchang 330047 (China); Yu, Ruijin [College of Science, Northwest A& F University, Yangling, Shaanxi 712100 (China); Lai, Weihua [State Key Laboratory of Food Science and Technology, Nanchang University, Nanchang 330047 (China); Xiong, Yonghua, E-mail: [State Key Laboratory of Food Science and Technology, Nanchang University, Nanchang 330047 (China)


    Herein, for the first time we report a novel direct competitive fluorescence-linked immunosorbent assay (dcFLISA) for the ultrasensitive detection of ochratoxin A (OTA) by introducing a large size polymer beads loaded with quantum dots (QBs) as carrier of competing antigen for decreasing binding affinity to antibody and enhancing the fluorescent signal intensity. When using 255 nm QBs as carrier of competing antigen, the equilibrium dissociation constant of QB based competing antigen to antibodies can be tuned to 100 times higher than that of the horseradish peroxidase (HRP) based competing antigen by controlling labeled amounts of antigen on the surface of QBs. Various parameters that influenced the sensitivity of dcFLISA were investigated and optimized. Under optimum detection parameters, the dynamic linear range of developed dcFLISA for detecting OTA was established at 0.05 pg/mL to 1.56 pg/mL with a half maximal inhibitory concentration at 0.14 ± 0.04 pg/mL (n = 5), which is three orders of magnitude lower than that of conventional HRP-based dcELISA (0.24 ng/mL). The developed FLISA is also highly accurate, reliable, and shows no cross reaction to other mycotoxins. In summary, the proposed method offers a straightforward approach to improve the sensitivity of direct competitive immunoassay for trace small chemical molecule detection in food quality control, environmental monitoring, and clinical diagnosis. - Highlights: • Highly luminescent QBs were used as a carrier of competing antigen for ultrasensitive detection of OTA. • It is the first time to use a large size QBs as a carrier for tuning affinity of competing antigen to antibodies. • IC{sub 50} value of QB-based dcFLISA is three orders of magnitude lower than that of HRP-based dcELISA.

  10. Ultrasensitive direct competitive FLISA using highly luminescent quantum dot beads for tuning affinity of competing antigens to antibodies

    International Nuclear Information System (INIS)

    Xiong, Sicheng; Zhou, Yaofeng; Huang, Xiaolin; Yu, Ruijin; Lai, Weihua; Xiong, Yonghua


    Herein, for the first time we report a novel direct competitive fluorescence-linked immunosorbent assay (dcFLISA) for the ultrasensitive detection of ochratoxin A (OTA) by introducing a large size polymer beads loaded with quantum dots (QBs) as carrier of competing antigen for decreasing binding affinity to antibody and enhancing the fluorescent signal intensity. When using 255 nm QBs as carrier of competing antigen, the equilibrium dissociation constant of QB based competing antigen to antibodies can be tuned to 100 times higher than that of the horseradish peroxidase (HRP) based competing antigen by controlling labeled amounts of antigen on the surface of QBs. Various parameters that influenced the sensitivity of dcFLISA were investigated and optimized. Under optimum detection parameters, the dynamic linear range of developed dcFLISA for detecting OTA was established at 0.05 pg/mL to 1.56 pg/mL with a half maximal inhibitory concentration at 0.14 ± 0.04 pg/mL (n = 5), which is three orders of magnitude lower than that of conventional HRP-based dcELISA (0.24 ng/mL). The developed FLISA is also highly accurate, reliable, and shows no cross reaction to other mycotoxins. In summary, the proposed method offers a straightforward approach to improve the sensitivity of direct competitive immunoassay for trace small chemical molecule detection in food quality control, environmental monitoring, and clinical diagnosis. - Highlights: • Highly luminescent QBs were used as a carrier of competing antigen for ultrasensitive detection of OTA. • It is the first time to use a large size QBs as a carrier for tuning affinity of competing antigen to antibodies. • IC_5_0 value of QB-based dcFLISA is three orders of magnitude lower than that of HRP-based dcELISA.

  11. Preclinical detection of porcine circovirus type 2 infection using an ultrasensitive nanoparticle DNA probe-based PCR assay.

    Directory of Open Access Journals (Sweden)

    Yong Huang

    Full Text Available Porcine circovirus type 2 (PCV2 has emerged as one of the most important pathogens affecting swine production globally. Preclinical identification of PCV2 is very important for effective prophylaxis of PCV2-associated diseases. In this study, we developed an ultrasensitive nanoparticle DNA probe-based PCR assay (UNDP-PCR for PCV2 detection. Magnetic microparticles coated with PCV2 specific DNA probes were used to enrich PCV2 DNA from samples, then gold nanoparticles coated with PCV2 specific oligonucleotides were added to form a sandwich nucleic acid-complex. After the complex was formed, the oligonucleotides were released and characterized by PCR. This assay exhibited about 500-fold more sensitive than conventional PCR, with a detection limit of 2 copies of purified PCV2 genomic DNA and 10 viral copies of PCV2 in serum. The assay has a wide detection range for all of PCV2 genotypes with reliable reproducibility. No cross-reactivity was observed from the samples of other related viruses including porcine circovirus type 1, porcine parvovirus, porcine pseudorabies virus, porcine reproductive and respiratory syndrome virus and classical swine fever virus. The positive detection rate of PCV2 specific UNDP-PCR in 40 preclinical field samples was 27.5%, which appeared greater than that by conventional and real-time PCR and appeared application potency in evaluation of the viral loads levels of preclinical infection samples. The UNDP-PCR assay reported here can reliably rule out false negative results from antibody-based assays, provide a nucleic acid extraction free, specific, ultrasensitive, economic and rapid diagnosis method for preclinical PCV2 infection in field, which may help prevent large-scale outbreaks.

  12. Boosting the oxidase mimicking activity of nanoceria by fluoride capping: rivaling protein enzymes and ultrasensitive F- detection (United States)

    Liu, Biwu; Huang, Zhicheng; Liu, Juewen


    Nanomaterial-based enzyme mimics (nanozymes) are currently a new forefront of chemical research. However, the application of nanozymes is limited by their low catalytic activity and low turnover numbers. Cerium dioxide nanoparticles (nanoceria) are among the few with oxidase activity. Herein, we report an interesting finding addressing their limitations. The oxidase activity of nanoceria is improved by over 100-fold by fluoride capping, making it more close to real oxidases. The turnover number reached 700 in 15 min, drastically improved from ~15 turnovers for the naked particles. The mechanism is attributed to surface charge modulation and facilitated electron transfer by F- capping based on ζ-potential and free radical measurements. Ultrasensitive sensing of fluoride was achieved with a detection limit of 0.64 μM F- in water and in toothpastes, while no other tested anions can achieve the activity enhancement.Nanomaterial-based enzyme mimics (nanozymes) are currently a new forefront of chemical research. However, the application of nanozymes is limited by their low catalytic activity and low turnover numbers. Cerium dioxide nanoparticles (nanoceria) are among the few with oxidase activity. Herein, we report an interesting finding addressing their limitations. The oxidase activity of nanoceria is improved by over 100-fold by fluoride capping, making it more close to real oxidases. The turnover number reached 700 in 15 min, drastically improved from ~15 turnovers for the naked particles. The mechanism is attributed to surface charge modulation and facilitated electron transfer by F- capping based on ζ-potential and free radical measurements. Ultrasensitive sensing of fluoride was achieved with a detection limit of 0.64 μM F- in water and in toothpastes, while no other tested anions can achieve the activity enhancement. Electronic supplementary information (ESI) available: Methods, TMB oxidation kinetics and control experiments. See DOI: 10.1039/c6nr02730j

  13. An automated 3D reconstruction method of UAV images (United States)

    Liu, Jun; Wang, He; Liu, Xiaoyang; Li, Feng; Sun, Guangtong; Song, Ping


    In this paper a novel fully automated 3D reconstruction approach based on low-altitude unmanned aerial vehicle system (UAVs) images will be presented, which does not require previous camera calibration or any other external prior knowledge. Dense 3D point clouds are generated by integrating orderly feature extraction, image matching, structure from motion (SfM) and multi-view stereo (MVS) algorithms, overcoming many of the cost, time limitations of rigorous photogrammetry techniques. An image topology analysis strategy is introduced to speed up large scene reconstruction by taking advantage of the flight-control data acquired by UAV. Image topology map can significantly reduce the running time of feature matching by limiting the combination of images. A high-resolution digital surface model of the study area is produced base on UAV point clouds by constructing the triangular irregular network. Experimental results show that the proposed approach is robust and feasible for automatic 3D reconstruction of low-altitude UAV images, and has great potential for the acquisition of spatial information at large scales mapping, especially suitable for rapid response and precise modelling in disaster emergency.

  14. Standardization and performance evaluation of "modified" and "ultrasensitive" versions of the Abbott RealTime HIV-1 assay, adapted to quantify minimal residual viremia. (United States)

    Amendola, Alessandra; Bloisi, Maria; Marsella, Patrizia; Sabatini, Rosella; Bibbò, Angela; Angeletti, Claudio; Capobianchi, Maria Rosaria


    Numerous studies investigating clinical significance of HIV-1 minimal residual viremia (MRV) suggest potential utility of assays more sensitive than those routinely used to monitor viral suppression. However currently available methods, based on different technologies, show great variation in detection limit and input plasma volume, and generally suffer from lack of standardization. In order to establish new tools suitable for routine quantification of minimal residual viremia in patients under virological suppression, some modifications were introduced into standard procedure of the Abbott RealTime HIV-1 assay leading to a "modified" and an "ultrasensitive" protocols. The following modifications were introduced: calibration curve extended towards low HIV-1 RNA concentration; 4 fold increased sample volume by concentrating starting material; reduced volume of internal control; adoption of "open-mode" software for quantification. Analytical performances were evaluated using the HIV-1 RNA Working Reagent 1 for NAT assays (NIBSC). Both tests were applied to clinical samples from virologically suppressed patients. The "modified" and the "ultrasensitive" configurations of the assay reached a limit of detection of 18.8 (95% CI: 11.1-51.0 cp/mL) and 4.8 cp/mL (95% CI: 2.6-9.1 cp/mL), respectively, with high precision and accuracy. In clinical samples from virologically suppressed patients, "modified" and "ultrasensitive" protocols allowed to detect and quantify HIV RNA in 12.7% and 46.6%, respectively, of samples resulted "not-detectable", and in 70.0% and 69.5%, respectively, of samples "detected laboratories for measuring MRV. Copyright © 2011 Elsevier B.V. All rights reserved.

  15. Depth extraction method with high accuracy in integral imaging based on moving array lenslet technique (United States)

    Wang, Yao-yao; Zhang, Juan; Zhao, Xue-wei; Song, Li-pei; Zhang, Bo; Zhao, Xing


    In order to improve depth extraction accuracy, a method using moving array lenslet technique (MALT) in pickup stage is proposed, which can decrease the depth interval caused by pixelation. In this method, the lenslet array is moved along the horizontal and vertical directions simultaneously for N times in a pitch to get N sets of elemental images. Computational integral imaging reconstruction method for MALT is taken to obtain the slice images of the 3D scene, and the sum modulus (SMD) blur metric is taken on these slice images to achieve the depth information of the 3D scene. Simulation and optical experiments are carried out to verify the feasibility of this method.

  16. Multiresolution, Geometric, and Learning Methods in Statistical Image Processing, Object Recognition, and Sensor Fusion

    National Research Council Canada - National Science Library

    Willsky, Alan


    .... Our research blends methods from several fields-statistics and probability, signal and image processing, mathematical physics, scientific computing, statistical learning theory, and differential...

  17. Synthesis and assessment methods for an edge-alignment-free hybrid image (United States)

    Sripian, Peeraya; Yamaguchi, Yasushi


    A hybrid image allows multiple image interpretations to be modulated by the viewing distance. It can be constructed on the basis of the multiscale perceptual mechanisms of the human visual system by combining the low and high spatial frequencies of two different images. The hybrid image was introduced as an experimental tool for visual recognition study in terms of spatial frequency perception. To produce a compelling hybrid image, the original hybrid image synthesis method could only use similar shapes of source images that were aligned in the edges. If any two different images can be hybrid, it would be beneficial as a new experimental tool. In addition, there is no measure for the actual perception of spatial frequency, whether a single spatial frequency or both spatial frequencies are perceived from the hybrid stimulus. This paper describes two methods for synthesizing a hybrid image from dissimilar shape images or unaligned images; this hybrid image is known as an "edge-alignment-free hybrid image." A noise-inserted method can be done by intentionally inserting and enhancing noises into the high-frequency image. With this method, the low-frequency blobs are covered with high-frequency noises when viewed up close. A color-inserted method uses complementary color gratings in the background of the high-frequency image to emphasize the high-frequency image when viewed up close, whereas the gratings disappear when viewed from far away. To ascertain that our approach successfully separates the spatial frequency at each viewing distance, we measured this property using our proposed assessment method. Our proposed method allows the experimenter to quantify the probability of perceiving both spatial frequencies and a single spatial frequency in a hybrid image. The experimental results confirmed that our proposed synthesis methods successfully hid the low-frequency image and emphasized the high-frequency image at a close viewing distance. At the same time, the

  18. A high-speed computerized tomography image reconstruction using direct two-dimensional Fourier transform method

    International Nuclear Information System (INIS)

    Niki, Noboru; Mizutani, Toshio; Takahashi, Yoshizo; Inouye, Tamon.


    The nescessity for developing real-time computerized tomography (CT) aiming at the dynamic observation of organs such as hearts has lately been advocated. It is necessary for its realization to reconstruct the images which are markedly faster than present CTs. Although various reconstructing methods have been proposed so far, the method practically employed at present is the filtered backprojection (FBP) method only, which can give high quality image reconstruction, but takes much computing time. In the past, the two-dimensional Fourier transform (TFT) method was regarded as unsuitable to practical use because the quality of images obtained was not good, in spite of the promising method for high speed reconstruction because of its less computing time. However, since it was revealed that the image quality by TFT method depended greatly on interpolation accuracy in two-dimensional Fourier space, the authors have developed a high-speed calculation algorithm that can obtain high quality images by pursuing the relationship between the image quality and the interpolation method. In this case, radial data sampling points in Fourier space are increased to β-th power of 2 times, and the linear or spline interpolation is used. Comparison of this method with the present FBP method resulted in the conclusion that the image quality is almost the same in practical image matrix, the computational time by TFT method becomes about 1/10 of FBP method, and the memory capacity also reduces by about 20 %. (Wakatsuki, Y.)


    Directory of Open Access Journals (Sweden)

    S. Hosseinian


    Full Text Available Nowadays, various medical X-ray imaging methods such as digital radiography, computed tomography and fluoroscopy are used as important tools in diagnostic and operative processes especially in the computer and robotic assisted surgeries. The procedures of extracting information from these images require appropriate deblurring and denoising processes on the pre- and intra-operative images in order to obtain more accurate information. This issue becomes more considerable when the X-ray images are planned to be employed in the photogrammetric processes for 3D reconstruction from multi-view X-ray images since, accurate data should be extracted from images for 3D modelling and the quality of X-ray images affects directly on the results of the algorithms. For restoration of X-ray images, it is essential to consider the nature and characteristics of these kinds of images. X-ray images exhibit severe quantum noise due to limited X-ray photons involved. The assumptions of Gaussian modelling are not appropriate for photon-limited images such as X-ray images, because of the nature of signal-dependant quantum noise. These images are generally modelled by Poisson distribution which is the most common model for low-intensity imaging. In this paper, existing methods are evaluated. For this purpose, after demonstrating the properties of medical X-ray images, the more efficient and recommended methods for restoration of X-ray images would be described and assessed. After explaining these approaches, they are implemented on samples from different kinds of X-ray images. By considering the results, it is concluded that using PURE-LET, provides more effective and efficient denoising than other examined methods in this research.

  20. A non-iterative twin image elimination method with two in-line digital holograms (United States)

    Kim, Jongwu; Lee, Heejung; Jeon, Philjun; Kim, Dug Young


    We propose a simple non-iterative in-line holographic measurement method which can effectively eliminate a twin image in digital holographic 3D imaging. It is shown that a twin image can be effectively eliminated with only two measured holograms by using a simple numerical propagation algorithm and arithmetic calculations.