WorldWideScience

Sample records for based proton monte

  1. Technical Note: Defining cyclotron-based clinical scanning proton machines in a FLUKA Monte Carlo system.

    Science.gov (United States)

    Fiorini, Francesca; Schreuder, Niek; Van den Heuvel, Frank

    2018-02-01

    Cyclotron-based pencil beam scanning (PBS) proton machines represent nowadays the majority and most affordable choice for proton therapy facilities, however, their representation in Monte Carlo (MC) codes is more complex than passively scattered proton system- or synchrotron-based PBS machines. This is because degraders are used to decrease the energy from the cyclotron maximum energy to the desired energy, resulting in a unique spot size, divergence, and energy spread depending on the amount of degradation. This manuscript outlines a generalized methodology to characterize a cyclotron-based PBS machine in a general-purpose MC code. The code can then be used to generate clinically relevant plans starting from commercial TPS plans. The described beam is produced at the Provision Proton Therapy Center (Knoxville, TN, USA) using a cyclotron-based IBA Proteus Plus equipment. We characterized the Provision beam in the MC FLUKA using the experimental commissioning data. The code was then validated using experimental data in water phantoms for single pencil beams and larger irregular fields. Comparisons with RayStation TPS plans are also presented. Comparisons of experimental, simulated, and planned dose depositions in water plans show that same doses are calculated by both programs inside the target areas, while penumbrae differences are found at the field edges. These differences are lower for the MC, with a γ(3%-3 mm) index never below 95%. Extensive explanations on how MC codes can be adapted to simulate cyclotron-based scanning proton machines are given with the aim of using the MC as a TPS verification tool to check and improve clinical plans. For all the tested cases, we showed that dose differences with experimental data are lower for the MC than TPS, implying that the created FLUKA beam model is better able to describe the experimental beam. © 2017 The Authors. Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists

  2. A fast GPU-based Monte Carlo simulation of proton transport with detailed modeling of nonelastic interactions.

    Science.gov (United States)

    Wan Chan Tseung, H; Ma, J; Beltran, C

    2015-06-01

    Very fast Monte Carlo (MC) simulations of proton transport have been implemented recently on graphics processing units (GPUs). However, these MCs usually use simplified models for nonelastic proton-nucleus interactions. Our primary goal is to build a GPU-based proton transport MC with detailed modeling of elastic and nonelastic proton-nucleus collisions. Using the cuda framework, the authors implemented GPU kernels for the following tasks: (1) simulation of beam spots from our possible scanning nozzle configurations, (2) proton propagation through CT geometry, taking into account nuclear elastic scattering, multiple scattering, and energy loss straggling, (3) modeling of the intranuclear cascade stage of nonelastic interactions when they occur, (4) simulation of nuclear evaporation, and (5) statistical error estimates on the dose. To validate our MC, the authors performed (1) secondary particle yield calculations in proton collisions with therapeutically relevant nuclei, (2) dose calculations in homogeneous phantoms, (3) recalculations of complex head and neck treatment plans from a commercially available treatment planning system, and compared with (GEANT)4.9.6p2/TOPAS. Yields, energy, and angular distributions of secondaries from nonelastic collisions on various nuclei are in good agreement with the (GEANT)4.9.6p2 Bertini and Binary cascade models. The 3D-gamma pass rate at 2%-2 mm for treatment plan simulations is typically 98%. The net computational time on a NVIDIA GTX680 card, including all CPU-GPU data transfers, is ∼ 20 s for 1 × 10(7) proton histories. Our GPU-based MC is the first of its kind to include a detailed nuclear model to handle nonelastic interactions of protons with any nucleus. Dosimetric calculations are in very good agreement with (GEANT)4.9.6p2/TOPAS. Our MC is being integrated into a framework to perform fast routine clinical QA of pencil-beam based treatment plans, and is being used as the dose calculation engine in a clinically

  3. SU-F-T-156: Monte Carlo Simulation Using TOPAS for Synchrotron Based Proton Discrete Spot Scanning System

    Energy Technology Data Exchange (ETDEWEB)

    Moskvin, V; Pirlepesov, F; Tsiamas, P; Axente, M; Lukose, R; Zhao, L; Farr, J [St. Jude Children’s Hospital, Memphis, TN (United States); Shin, J [Massachusetts General Hospital, Brookline, MA (United States)

    2016-06-15

    Purpose: This study provides an overview of the design and commissioning of the Monte Carlo (MC) model of the spot-scanning proton therapy nozzle and its implementation for the patient plan simulation. Methods: The Hitachi PROBEAT V scanning nozzle was simulated based on vendor specifications using the TOPAS extension of Geant4 code. FLUKA MC simulation was also utilized to provide supporting data for the main simulation. Validation of the MC model was performed using vendor provided data and measurements collected during acceptance/commissioning of the proton therapy machine. Actual patient plans using CT based treatment geometry were simulated and compared to the dose distributions produced by the treatment planning system (Varian Eclipse 13.6), and patient quality assurance measurements. In-house MATLAB scripts are used for converting DICOM data into TOPAS input files. Results: Comparison analysis of integrated depth doses (IDDs), therapeutic ranges (R90), and spot shape/sizes at different distances from the isocenter, indicate good agreement between MC and measurements. R90 agreement is within 0.15 mm across all energy tunes. IDDs and spot shapes/sizes differences are within statistical error of simulation (less than 1.5%). The MC simulated data, validated with physical measurements, were used for the commissioning of the treatment planning system. Patient geometry simulations were conducted based on the Eclipse produced DICOM plans. Conclusion: The treatment nozzle and standard option beam model were implemented in the TOPAS framework to simulate a highly conformal discrete spot-scanning proton beam system.

  4. A new approach to integrate GPU-based Monte Carlo simulation into inverse treatment plan optimization for proton therapy.

    Science.gov (United States)

    Li, Yongbao; Tian, Zhen; Song, Ting; Wu, Zhaoxia; Liu, Yaqiang; Jiang, Steve; Jia, Xun

    2017-01-07

    Monte Carlo (MC)-based spot dose calculation is highly desired for inverse treatment planning in proton therapy because of its accuracy. Recent studies on biological optimization have also indicated the use of MC methods to compute relevant quantities of interest, e.g. linear energy transfer. Although GPU-based MC engines have been developed to address inverse optimization problems, their efficiency still needs to be improved. Also, the use of a large number of GPUs in MC calculation is not favorable for clinical applications. The previously proposed adaptive particle sampling (APS) method can improve the efficiency of MC-based inverse optimization by using the computationally expensive MC simulation more effectively. This method is more efficient than the conventional approach that performs spot dose calculation and optimization in two sequential steps. In this paper, we propose a computational library to perform MC-based spot dose calculation on GPU with the APS scheme. The implemented APS method performs a non-uniform sampling of the particles from pencil beam spots during the optimization process, favoring those from the high intensity spots. The library also conducts two computationally intensive matrix-vector operations frequently used when solving an optimization problem. This library design allows a streamlined integration of the MC-based spot dose calculation into an existing proton therapy inverse planning process. We tested the developed library in a typical inverse optimization system with four patient cases. The library achieved the targeted functions by supporting inverse planning in various proton therapy schemes, e.g. single field uniform dose, 3D intensity modulated proton therapy, and distal edge tracking. The efficiency was 41.6  ±  15.3% higher than the use of a GPU-based MC package in a conventional calculation scheme. The total computation time ranged between 2 and 50 min on a single GPU card depending on the problem size.

  5. Towards offline PET monitoring at a cyclotron-based proton therapy facility. Experiments and Monte Carlo simulations

    Energy Technology Data Exchange (ETDEWEB)

    Wuerl, Matthias

    2016-08-01

    Matthias Wuerl presents two essential steps to implement offline PET monitoring of proton dose delivery at a clinical facility, namely the setting up of an accurate Monte Carlo model of the clinical beamline and the experimental validation of positron emitter production cross-sections. In the first part, the field size dependence of the dose output is described for scanned proton beams. Both the Monte Carlo and an analytical computational beam model were able to accurately predict target dose, while the latter tends to overestimate dose in normal tissue. In the second part, the author presents PET measurements of different phantom materials, which were activated by the proton beam. The results indicate that for an irradiation with a high number of protons for the sake of good statistics, dead time losses of the PET scanner may become important and lead to an underestimation of positron-emitter production yields.

  6. MO-A-BRD-10: A Fast and Accurate GPU-Based Proton Transport Monte Carlo Simulation for Validating Proton Therapy Treatment Plans

    Energy Technology Data Exchange (ETDEWEB)

    Wan Chan Tseung, H; Ma, J; Beltran, C [Mayo Clinic, Rochester, MN (United States)

    2014-06-15

    Purpose: To build a GPU-based Monte Carlo (MC) simulation of proton transport with detailed modeling of elastic and non-elastic (NE) protonnucleus interactions, for use in a very fast and cost-effective proton therapy treatment plan verification system. Methods: Using the CUDA framework, we implemented kernels for the following tasks: (1) Simulation of beam spots from our possible scanning nozzle configurations, (2) Proton propagation through CT geometry, taking into account nuclear elastic and multiple scattering, as well as energy straggling, (3) Bertini-style modeling of the intranuclear cascade stage of NE interactions, and (4) Simulation of nuclear evaporation. To validate our MC, we performed: (1) Secondary particle yield calculations in NE collisions with therapeutically-relevant nuclei, (2) Pencil-beam dose calculations in homogeneous phantoms, (3) A large number of treatment plan dose recalculations, and compared with Geant4.9.6p2/TOPAS. A workflow was devised for calculating plans from a commercially available treatment planning system, with scripts for reading DICOM files and generating inputs for our MC. Results: Yields, energy and angular distributions of secondaries from NE collisions on various nuclei are in good agreement with the Geant4.9.6p2 Bertini and Binary cascade models. The 3D-gamma pass rate at 2%–2mm for 70–230 MeV pencil-beam dose distributions in water, soft tissue, bone and Ti phantoms is 100%. The pass rate at 2%–2mm for treatment plan calculations is typically above 98%. The net computational time on a NVIDIA GTX680 card, including all CPU-GPU data transfers, is around 20s for 1×10{sup 7} proton histories. Conclusion: Our GPU-based proton transport MC is the first of its kind to include a detailed nuclear model to handle NE interactions on any nucleus. Dosimetric calculations demonstrate very good agreement with Geant4.9.6p2/TOPAS. Our MC is being integrated into a framework to perform fast routine clinical QA of pencil

  7. Clinical implementation of a GPU-based simplified Monte Carlo method for a treatment planning system of proton beam therapy.

    Science.gov (United States)

    Kohno, R; Hotta, K; Nishioka, S; Matsubara, K; Tansho, R; Suzuki, T

    2011-11-21

    We implemented the simplified Monte Carlo (SMC) method on graphics processing unit (GPU) architecture under the computer-unified device architecture platform developed by NVIDIA. The GPU-based SMC was clinically applied for four patients with head and neck, lung, or prostate cancer. The results were compared to those obtained by a traditional CPU-based SMC with respect to the computation time and discrepancy. In the CPU- and GPU-based SMC calculations, the estimated mean statistical errors of the calculated doses in the planning target volume region were within 0.5% rms. The dose distributions calculated by the GPU- and CPU-based SMCs were similar, within statistical errors. The GPU-based SMC showed 12.30-16.00 times faster performance than the CPU-based SMC. The computation time per beam arrangement using the GPU-based SMC for the clinical cases ranged 9-67 s. The results demonstrate the successful application of the GPU-based SMC to a clinical proton treatment planning.

  8. TU-EF-304-07: Monte Carlo-Based Inverse Treatment Plan Optimization for Intensity Modulated Proton Therapy

    Energy Technology Data Exchange (ETDEWEB)

    Li, Y [Tsinghua University, Beijing, Beijing (China); UT Southwestern Medical Center, Dallas, TX (United States); Tian, Z; Jiang, S; Jia, X [UT Southwestern Medical Center, Dallas, TX (United States); Song, T [Southern Medical University, Guangzhou, Guangdong (China); UT Southwestern Medical Center, Dallas, TX (United States); Wu, Z; Liu, Y [Tsinghua University, Beijing, Beijing (China)

    2015-06-15

    Purpose: Intensity-modulated proton therapy (IMPT) is increasingly used in proton therapy. For IMPT optimization, Monte Carlo (MC) is desired for spots dose calculations because of its high accuracy, especially in cases with a high level of heterogeneity. It is also preferred in biological optimization problems due to the capability of computing quantities related to biological effects. However, MC simulation is typically too slow to be used for this purpose. Although GPU-based MC engines have become available, the achieved efficiency is still not ideal. The purpose of this work is to develop a new optimization scheme to include GPU-based MC into IMPT. Methods: A conventional approach using MC in IMPT simply calls the MC dose engine repeatedly for each spot dose calculations. However, this is not the optimal approach, because of the unnecessary computations on some spots that turned out to have very small weights after solving the optimization problem. GPU-memory writing conflict occurring at a small beam size also reduces computational efficiency. To solve these problems, we developed a new framework that iteratively performs MC dose calculations and plan optimizations. At each dose calculation step, the particles were sampled from different spots altogether with Metropolis algorithm, such that the particle number is proportional to the latest optimized spot intensity. Simultaneously transporting particles from multiple spots also mitigated the memory writing conflict problem. Results: We have validated the proposed MC-based optimization schemes in one prostate case. The total computation time of our method was ∼5–6 min on one NVIDIA GPU card, including both spot dose calculation and plan optimization, whereas a conventional method naively using the same GPU-based MC engine were ∼3 times slower. Conclusion: A fast GPU-based MC dose calculation method along with a novel optimization workflow is developed. The high efficiency makes it attractive for clinical

  9. TU-EF-304-07: Monte Carlo-Based Inverse Treatment Plan Optimization for Intensity Modulated Proton Therapy

    International Nuclear Information System (INIS)

    Li, Y; Tian, Z; Jiang, S; Jia, X; Song, T; Wu, Z; Liu, Y

    2015-01-01

    Purpose: Intensity-modulated proton therapy (IMPT) is increasingly used in proton therapy. For IMPT optimization, Monte Carlo (MC) is desired for spots dose calculations because of its high accuracy, especially in cases with a high level of heterogeneity. It is also preferred in biological optimization problems due to the capability of computing quantities related to biological effects. However, MC simulation is typically too slow to be used for this purpose. Although GPU-based MC engines have become available, the achieved efficiency is still not ideal. The purpose of this work is to develop a new optimization scheme to include GPU-based MC into IMPT. Methods: A conventional approach using MC in IMPT simply calls the MC dose engine repeatedly for each spot dose calculations. However, this is not the optimal approach, because of the unnecessary computations on some spots that turned out to have very small weights after solving the optimization problem. GPU-memory writing conflict occurring at a small beam size also reduces computational efficiency. To solve these problems, we developed a new framework that iteratively performs MC dose calculations and plan optimizations. At each dose calculation step, the particles were sampled from different spots altogether with Metropolis algorithm, such that the particle number is proportional to the latest optimized spot intensity. Simultaneously transporting particles from multiple spots also mitigated the memory writing conflict problem. Results: We have validated the proposed MC-based optimization schemes in one prostate case. The total computation time of our method was ∼5–6 min on one NVIDIA GPU card, including both spot dose calculation and plan optimization, whereas a conventional method naively using the same GPU-based MC engine were ∼3 times slower. Conclusion: A fast GPU-based MC dose calculation method along with a novel optimization workflow is developed. The high efficiency makes it attractive for clinical

  10. SU-F-T-193: Evaluation of a GPU-Based Fast Monte Carlo Code for Proton Therapy Biological Optimization

    Energy Technology Data Exchange (ETDEWEB)

    Taleei, R; Qin, N; Jiang, S [UT Southwestern Medical Center, Dallas, TX (United States); Peeler, C [UT MD Anderson Cancer Center, Houston, TX (United States); Jia, X [The University of Texas Southwestern Medical Ctr, Dallas, TX (United States)

    2016-06-15

    Purpose: Biological treatment plan optimization is of great interest for proton therapy. It requires extensive Monte Carlo (MC) simulations to compute physical dose and biological quantities. Recently, a gPMC package was developed for rapid MC dose calculations on a GPU platform. This work investigated its suitability for proton therapy biological optimization in terms of accuracy and efficiency. Methods: We performed simulations of a proton pencil beam with energies of 75, 150 and 225 MeV in a homogeneous water phantom using gPMC and FLUKA. Physical dose and energy spectra for each ion type on the central beam axis were scored. Relative Biological Effectiveness (RBE) was calculated using repair-misrepair-fixation model. Microdosimetry calculations were performed using Monte Carlo Damage Simulation (MCDS). Results: Ranges computed by the two codes agreed within 1 mm. Physical dose difference was less than 2.5 % at the Bragg peak. RBE-weighted dose agreed within 5 % at the Bragg peak. Differences in microdosimetric quantities such as dose average lineal energy transfer and specific energy were < 10%. The simulation time per source particle with FLUKA was 0.0018 sec, while gPMC was ∼ 600 times faster. Conclusion: Physical dose computed by FLUKA and gPMC were in a good agreement. The RBE differences along the central axis were small, and RBE-weighted dose difference was found to be acceptable. The combined accuracy and efficiency makes gPMC suitable for proton therapy biological optimization.

  11. Reconstruction for proton computed tomography by tracing proton trajectories: A Monte Carlo study

    International Nuclear Information System (INIS)

    Li Tianfang; Liang Zhengrong; Singanallur, Jayalakshmi V.; Satogata, Todd J.; Williams, David C.; Schulte, Reinhard W.

    2006-01-01

    Proton computed tomography (pCT) has been explored in the past decades because of its unique imaging characteristics, low radiation dose, and its possible use for treatment planning and on-line target localization in proton therapy. However, reconstruction of pCT images is challenging because the proton path within the object to be imaged is statistically affected by multiple Coulomb scattering. In this paper, we employ GEANT4-based Monte Carlo simulations of the two-dimensional pCT reconstruction of an elliptical phantom to investigate the possible use of the algebraic reconstruction technique (ART) with three different path-estimation methods for pCT reconstruction. The first method assumes a straight-line path (SLP) connecting the proton entry and exit positions, the second method adapts the most-likely path (MLP) theoretically determined for a uniform medium, and the third method employs a cubic spline path (CSP). The ART reconstructions showed progressive improvement of spatial resolution when going from the SLP [2 line pairs (lp) cm -1 ] to the curved CSP and MLP path estimates (5 lp cm -1 ). The MLP-based ART algorithm had the fastest convergence and smallest residual error of all three estimates. This work demonstrates the advantage of tracking curved proton paths in conjunction with the ART algorithm and curved path estimates

  12. Neutron shielding calculations in a proton therapy facility based on Monte Carlo simulations and analytical models: Criterion for selecting the method of choice

    International Nuclear Information System (INIS)

    Titt, U.; Newhauser, W. D.

    2005-01-01

    Proton therapy facilities are shielded to limit the amount of secondary radiation to which patients, occupational workers and members of the general public are exposed. The most commonly applied shielding design methods for proton therapy facilities comprise semi-empirical and analytical methods to estimate the neutron dose equivalent. This study compares the results of these methods with a detailed simulation of a proton therapy facility by using the Monte Carlo technique. A comparison of neutron dose equivalent values predicted by the various methods reveals the superior accuracy of the Monte Carlo predictions in locations where the calculations converge. However, the reliability of the overall shielding design increases if simulation results, for which solutions have not converged, e.g. owing to too few particle histories, can be excluded, and deterministic models are being used at these locations. Criteria to accept or reject Monte Carlo calculations in such complex structures are not well understood. An optimum rejection criterion would allow all converging solutions of Monte Carlo simulation to be taken into account, and reject all solutions with uncertainties larger than the design safety margins. In this study, the optimum rejection criterion of 10% was found. The mean ratio was 26, 62% of all receptor locations showed a ratio between 0.9 and 10, and 92% were between 1 and 100. (authors)

  13. A Fano cavity test for Monte Carlo proton transport algorithms

    International Nuclear Information System (INIS)

    Sterpin, Edmond; Sorriaux, Jefferson; Souris, Kevin; Vynckier, Stefaan; Bouchard, Hugo

    2014-01-01

    Purpose: In the scope of reference dosimetry of radiotherapy beams, Monte Carlo (MC) simulations are widely used to compute ionization chamber dose response accurately. Uncertainties related to the transport algorithm can be verified performing self-consistency tests, i.e., the so-called “Fano cavity test.” The Fano cavity test is based on the Fano theorem, which states that under charged particle equilibrium conditions, the charged particle fluence is independent of the mass density of the media as long as the cross-sections are uniform. Such tests have not been performed yet for MC codes simulating proton transport. The objectives of this study are to design a new Fano cavity test for proton MC and to implement the methodology in two MC codes: Geant4 and PENELOPE extended to protons (PENH). Methods: The new Fano test is designed to evaluate the accuracy of proton transport. Virtual particles with an energy ofE 0 and a mass macroscopic cross section of (Σ)/(ρ) are transported, having the ability to generate protons with kinetic energy E 0 and to be restored after each interaction, thus providing proton equilibrium. To perform the test, the authors use a simplified simulation model and rigorously demonstrate that the computed cavity dose per incident fluence must equal (ΣE 0 )/(ρ) , as expected in classic Fano tests. The implementation of the test is performed in Geant4 and PENH. The geometry used for testing is a 10 × 10 cm 2 parallel virtual field and a cavity (2 × 2 × 0.2 cm 3 size) in a water phantom with dimensions large enough to ensure proton equilibrium. Results: For conservative user-defined simulation parameters (leading to small step sizes), both Geant4 and PENH pass the Fano cavity test within 0.1%. However, differences of 0.6% and 0.7% were observed for PENH and Geant4, respectively, using larger step sizes. For PENH, the difference is attributed to the random-hinge method that introduces an artificial energy straggling if step size is not

  14. Clinically Applicable Monte Carlo–based Biological Dose Optimization for the Treatment of Head and Neck Cancers With Spot-Scanning Proton Therapy

    Energy Technology Data Exchange (ETDEWEB)

    Wan Chan Tseung, Hok Seum, E-mail: wanchantseung.hok@mayo.edu; Ma, Jiasen; Kreofsky, Cole R.; Ma, Daniel J.; Beltran, Chris

    2016-08-01

    Purpose: Our aim is to demonstrate the feasibility of fast Monte Carlo (MC)–based inverse biological planning for the treatment of head and neck tumors in spot-scanning proton therapy. Methods and Materials: Recently, a fast and accurate graphics processor unit (GPU)–based MC simulation of proton transport was developed and used as the dose-calculation engine in a GPU-accelerated intensity modulated proton therapy (IMPT) optimizer. Besides dose, the MC can simultaneously score the dose-averaged linear energy transfer (LET{sub d}), which makes biological dose (BD) optimization possible. To convert from LET{sub d} to BD, a simple linear relation was assumed. By use of this novel optimizer, inverse biological planning was applied to 4 patients, including 2 small and 1 large thyroid tumor targets, as well as 1 glioma case. To create these plans, constraints were placed to maintain the physical dose (PD) within 1.25 times the prescription while maximizing target BD. For comparison, conventional intensity modulated radiation therapy (IMRT) and IMPT plans were also created using Eclipse (Varian Medical Systems) in each case. The same critical-structure PD constraints were used for the IMRT, IMPT, and biologically optimized plans. The BD distributions for the IMPT plans were obtained through MC recalculations. Results: Compared with standard IMPT, the biologically optimal plans for patients with small tumor targets displayed a BD escalation that was around twice the PD increase. Dose sparing to critical structures was improved compared with both IMRT and IMPT. No significant BD increase could be achieved for the large thyroid tumor case and when the presence of critical structures mitigated the contribution of additional fields. The calculation of the biologically optimized plans can be completed in a clinically viable time (<30 minutes) on a small 24-GPU system. Conclusions: By exploiting GPU acceleration, MC-based, biologically optimized plans were created for

  15. Proton therapy Monte Carlo SRNA-VOX code

    Directory of Open Access Journals (Sweden)

    Ilić Radovan D.

    2012-01-01

    Full Text Available The most powerful feature of the Monte Carlo method is the possibility of simulating all individual particle interactions in three dimensions and performing numerical experiments with a preset error. These facts were the motivation behind the development of a general-purpose Monte Carlo SRNA program for proton transport simulation in technical systems described by standard geometrical forms (plane, sphere, cone, cylinder, cube. Some of the possible applications of the SRNA program are: (a a general code for proton transport modeling, (b design of accelerator-driven systems, (c simulation of proton scattering and degrading shapes and composition, (d research on proton detectors; and (e radiation protection at accelerator installations. This wide range of possible applications of the program demands the development of various versions of SRNA-VOX codes for proton transport modeling in voxelized geometries and has, finally, resulted in the ISTAR package for the calculation of deposited energy distribution in patients on the basis of CT data in radiotherapy. All of the said codes are capable of using 3-D proton sources with an arbitrary energy spectrum in an interval of 100 keV to 250 MeV.

  16. Optimization of GATE and PHITS Monte Carlo code parameters for spot scanning proton beam based on simulation with FLUKA general-purpose code

    International Nuclear Information System (INIS)

    Kurosu, Keita; Das, Indra J.; Moskvin, Vadim P.

    2016-01-01

    Spot scanning, owing to its superior dose-shaping capability, provides unsurpassed dose conformity, in particular for complex targets. However, the robustness of the delivered dose distribution and prescription has to be verified. Monte Carlo (MC) simulation has the potential to generate significant advantages for high-precise particle therapy, especially for medium containing inhomogeneities. However, the inherent choice of computational parameters in MC simulation codes of GATE, PHITS and FLUKA that is observed for uniform scanning proton beam needs to be evaluated. This means that the relationship between the effect of input parameters and the calculation results should be carefully scrutinized. The objective of this study was, therefore, to determine the optimal parameters for the spot scanning proton beam for both GATE and PHITS codes by using data from FLUKA simulation as a reference. The proton beam scanning system of the Indiana University Health Proton Therapy Center was modeled in FLUKA, and the geometry was subsequently and identically transferred to GATE and PHITS. Although the beam transport is managed by spot scanning system, the spot location is always set at the center of a water phantom of 600 × 600 × 300 mm 3 , which is placed after the treatment nozzle. The percentage depth dose (PDD) is computed along the central axis using 0.5 × 0.5 × 0.5 mm 3 voxels in the water phantom. The PDDs and the proton ranges obtained with several computational parameters are then compared to those of FLUKA, and optimal parameters are determined from the accuracy of the proton range, suppressed dose deviation, and computational time minimization. Our results indicate that the optimized parameters are different from those for uniform scanning, suggesting that the gold standard for setting computational parameters for any proton therapy application cannot be determined consistently since the impact of setting parameters depends on the proton irradiation technique

  17. SU-E-T-254: Optimization of GATE and PHITS Monte Carlo Code Parameters for Uniform Scanning Proton Beam Based On Simulation with FLUKA General-Purpose Code

    International Nuclear Information System (INIS)

    Kurosu, K; Takashina, M; Koizumi, M; Das, I; Moskvin, V

    2014-01-01

    Purpose: Monte Carlo codes are becoming important tools for proton beam dosimetry. However, the relationships between the customizing parameters and percentage depth dose (PDD) of GATE and PHITS codes have not been reported which are studied for PDD and proton range compared to the FLUKA code and the experimental data. Methods: The beam delivery system of the Indiana University Health Proton Therapy Center was modeled for the uniform scanning beam in FLUKA and transferred identically into GATE and PHITS. This computational model was built from the blue print and validated with the commissioning data. Three parameters evaluated are the maximum step size, cut off energy and physical and transport model. The dependence of the PDDs on the customizing parameters was compared with the published results of previous studies. Results: The optimal parameters for the simulation of the whole beam delivery system were defined by referring to the calculation results obtained with each parameter. Although the PDDs from FLUKA and the experimental data show a good agreement, those of GATE and PHITS obtained with our optimal parameters show a minor discrepancy. The measured proton range R90 was 269.37 mm, compared to the calculated range of 269.63 mm, 268.96 mm, and 270.85 mm with FLUKA, GATE and PHITS, respectively. Conclusion: We evaluated the dependence of the results for PDDs obtained with GATE and PHITS Monte Carlo generalpurpose codes on the customizing parameters by using the whole computational model of the treatment nozzle. The optimal parameters for the simulation were then defined by referring to the calculation results. The physical model, particle transport mechanics and the different geometrybased descriptions need accurate customization in three simulation codes to agree with experimental data for artifact-free Monte Carlo simulation. This study was supported by Grants-in Aid for Cancer Research (H22-3rd Term Cancer Control-General-043) from the Ministry of Health

  18. Monte Carlo event generator for black hole production and decay in proton-proton collisions - QBH version 1.02

    Science.gov (United States)

    Gingrich, Douglas M.

    2010-11-01

    We describe the Monte Carlo event generator for black hole production and decay in proton-proton collisions - QBH version 1.02. The generator implements a model for quantum black hole production and decay based on the conservation of local gauge symmetries and democratic decays. The code in written entirely in C++ and interfaces to the PYTHIA 8 Monte Carlo code for fragmentation and decays. Program summaryProgram title: QBH Catalogue identifier: AEGU_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEGU_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 10 048 No. of bytes in distributed program, including test data, etc.: 118 420 Distribution format: tar.gz Programming language: C++ Computer: x86 Operating system: Scientific Linux, Mac OS X RAM: 1 GB Classification: 11.6 External routines: PYTHIA 8130 ( http://home.thep.lu.se/~torbjorn/pythiaaux/present.html) and LHAPDF ( http://projects.hepforge.org/lhapdf/) Nature of problem: Simulate black hole production and decay in proton-proton collision. Solution method: Monte Carlo simulation using importance sampling. Running time: Eight events per second.

  19. Clinical implementation of full Monte Carlo dose calculation in proton beam therapy

    International Nuclear Information System (INIS)

    Paganetti, Harald; Jiang, Hongyu; Parodi, Katia; Slopsema, Roelf; Engelsman, Martijn

    2008-01-01

    The goal of this work was to facilitate the clinical use of Monte Carlo proton dose calculation to support routine treatment planning and delivery. The Monte Carlo code Geant4 was used to simulate the treatment head setup, including a time-dependent simulation of modulator wheels (for broad beam modulation) and magnetic field settings (for beam scanning). Any patient-field-specific setup can be modeled according to the treatment control system of the facility. The code was benchmarked against phantom measurements. Using a simulation of the ionization chamber reading in the treatment head allows the Monte Carlo dose to be specified in absolute units (Gy per ionization chamber reading). Next, the capability of reading CT data information was implemented into the Monte Carlo code to model patient anatomy. To allow time-efficient dose calculation, the standard Geant4 tracking algorithm was modified. Finally, a software link of the Monte Carlo dose engine to the patient database and the commercial planning system was established to allow data exchange, thus completing the implementation of the proton Monte Carlo dose calculation engine ('DoC++'). Monte Carlo re-calculated plans are a valuable tool to revisit decisions in the planning process. Identification of clinically significant differences between Monte Carlo and pencil-beam-based dose calculations may also drive improvements of current pencil-beam methods. As an example, four patients (29 fields in total) with tumors in the head and neck regions were analyzed. Differences between the pencil-beam algorithm and Monte Carlo were identified in particular near the end of range, both due to dose degradation and overall differences in range prediction due to bony anatomy in the beam path. Further, the Monte Carlo reports dose-to-tissue as compared to dose-to-water by the planning system. Our implementation is tailored to a specific Monte Carlo code and the treatment planning system XiO (Computerized Medical Systems Inc

  20. Srna-Monte Carlo codes for proton transport simulation in combined and voxelized geometries

    International Nuclear Information System (INIS)

    Ilic, R.D.; Lalic, D.; Stankovic, S.J.

    2002-01-01

    This paper describes new Monte Carlo codes for proton transport simulations in complex geometrical forms and in materials of different composition. The SRNA codes were developed for three dimensional (3D) dose distribution calculation in proton therapy and dosimetry. The model of these codes is based on the theory of proton multiple scattering and a simple model of compound nucleus decay. The developed package consists of two codes: SRNA-2KG and SRNA-VOX. The first code simulates proton transport in combined geometry that can be described by planes and second order surfaces. The second one uses the voxelized geometry of material zones and is specifically adopted for the application of patient computer tomography data. Transition probabilities for both codes are given by the SRNADAT program. In this paper, we will present the models and algorithms of our programs, as well as the results of the numerical experiments we have carried out applying them, along with the results of proton transport simulation obtained through the PETRA and GEANT programs. The simulation of the proton beam characterization by means of the Multi-Layer Faraday Cup and spatial distribution of positron emitters obtained by our program indicate the imminent application of Monte Carlo techniques in clinical practice. (author)

  1. Srna - Monte Carlo codes for proton transport simulation in combined and voxelized geometries

    Directory of Open Access Journals (Sweden)

    Ilić Radovan D.

    2002-01-01

    Full Text Available This paper describes new Monte Carlo codes for proton transport simulations in complex geometrical forms and in materials of different composition. The SRNA codes were developed for three dimensional (3D dose distribution calculation in proton therapy and dosimetry. The model of these codes is based on the theory of proton multiple scattering and a simple model of compound nucleus decay. The developed package consists of two codes: SRNA-2KG and SRNA-VOX. The first code simulates proton transport in combined geometry that can be described by planes and second order surfaces. The second one uses the voxelized geometry of material zones and is specifically adopted for the application of patient computer tomography data. Transition probabilities for both codes are given by the SRNADAT program. In this paper, we will present the models and algorithms of our programs, as well as the results of the numerical experiments we have carried out applying them, along with the results of proton transport simulation obtained through the PETRA and GEANT programs. The simulation of the proton beam characterization by means of the Multi-Layer Faraday Cup and spatial distribution of positron emitters obtained by our program indicate the imminent application of Monte Carlo techniques in clinical practice.

  2. Monte Carlo simulations of a low energy proton beamline for radiobiological experiments.

    Science.gov (United States)

    Dahle, Tordis J; Rykkelid, Anne Marit; Stokkevåg, Camilla H; Mairani, Andrea; Görgen, Andreas; Edin, Nina J; Rørvik, Eivind; Fjæra, Lars Fredrik; Malinen, Eirik; Ytre-Hauge, Kristian S

    2017-06-01

    In order to determine the relative biological effectiveness (RBE) of protons with high accuracy, radiobiological experiments with detailed knowledge of the linear energy transfer (LET) are needed. Cell survival data from high LET protons are sparse and experiments with low energy protons to achieve high LET values are therefore required. The aim of this study was to quantify LET distributions from a low energy proton beam by using Monte Carlo (MC) simulations, and to further compare to a proton beam representing a typical minimum energy available at clinical facilities. A Markus ionization chamber and Gafchromic films were employed in dose measurements in the proton beam at Oslo Cyclotron Laboratory. Dose profiles were also calculated using the FLUKA MC code, with the MC beam parameters optimized based on comparisons with the measurements. LET spectra and dose-averaged LET (LET d ) were then estimated in FLUKA, and compared with LET calculated from an 80 MeV proton beam. The initial proton energy was determined to be 15.5 MeV, with a Gaussian energy distribution of 0.2% full width at half maximum (FWHM) and a Gaussian lateral spread of 2 mm FWHM. The LET d increased with depth, from approximately 5 keV/μm in the entrance to approximately 40 keV/μm in the distal dose fall-off. The LET d values were considerably higher and the LET spectra were much narrower than the corresponding spectra from the 80 MeV beam. MC simulations accurately modeled the dose distribution from the proton beam and could be used to estimate the LET at any position in the setup. The setup can be used to study the RBE for protons at high LET d , which is not achievable in clinical proton therapy facilities.

  3. SU-G-TeP3-03: Dose Enhancement of Gold Nanoparticle in Proton Therapy: A Monte Carlo Study Based On the Transmission Electron Microscopy Imaging

    Energy Technology Data Exchange (ETDEWEB)

    Ma, Y; Beaulieu, L [Centre Hospitalier University de Quebec, Quebec, QC (Canada); Laprise-Pelletier, M; Lagueux, J; Cote, M; Fordin, M [Dept of Mining, Metallurgy and Materials Engineering, University Laval, Quebec, QC, CA (Canada)

    2016-06-15

    Purpose: Gold nanoparticle (GNP) is a promising radiosensitizer that selectively boosts tumor dose in radiotherapy. Transmission electron microscopy (TEM) imaging observations recently revealed for the first time that GNP exists in vivo in the form of highly localized vesicles, instead of hypothetical uniform distribution. This work investigates the corresponding difference of energy deposition in proton therapy. Methods: First, single vesicles of various radii were constructed by packing GNPs (as Φ50 nm gold spheres) in spheres and were simulated, as well as a single GNP. The radial energy depositions (REDs) were scored using 100 concentric spherical shells from 0.1µm to 10µm, 0.1µm thickness each, for both vesicles and GNP, and compared. TEM images, 8 days after injection in a PC3 prostate cancer murine model, were used to extract position/dimension of vesicles, as well as contours of cytoplasmic and nucleus membranes. Vesicles were then constructed based on the TEM images. A 100 MeV proton beam was studied by using the Geant4-DNA code, which simulates all energy deposition events. Results: The vesicle REDs, normalized to the same proton energy loss as in a single GNP, are larger (smaller) than that of a single GNP when radius >2µm (<2µm). The peak increase (at about 3µm radius) is about 10% and 18% for Φ1µm and Φ1.6µm vesicles respectively, relative to a single GNP. The TEM-based simulation resulted in a larger energy deposition (by about one order of magnitude) that follows completely different pattern from that of hypothetical GNP distributions (regular dotted pattern in extracellular and/or extranucleus regions). Conclusion: The in vivo energy deposition, both in pattern and magnitude, of proton therapy is greatly affected by the true distribution of the GNP, as illustrated by the presence of GNP vesicles compared to hypothetical scenarios. Work supported by NSERC Discovery Grant #435510, Canada.

  4. Monte Carlo evaluation of magnetically focused proton beams for radiosurgery

    Science.gov (United States)

    McAuley, Grant A.; Heczko, Sarah L.; Nguyen, Theodore T.; Slater, James M.; Slater, Jerry D.; Wroe, Andrew J.

    2018-03-01

    The purpose of this project is to investigate the advantages in dose distribution and delivery of proton beams focused by a triplet of quadrupole magnets in the context of potential radiosurgery treatments. Monte Carlo simulations were performed using various configurations of three quadrupole magnets located immediately upstream of a water phantom. Magnet parameters were selected to match what can be commercially manufactured as assemblies of rare-earth permanent magnetic materials. Focused unmodulated proton beams with a range of ~10 cm in water were target matched with passive collimated beams (the current beam delivery method for proton radiosurgery) and properties of transverse dose, depth dose and volumetric dose distributions were compared. Magnetically focused beams delivered beam spots of low eccentricity to Bragg peak depth with full widths at the 90% reference dose contour from ~2.5 to 5 mm. When focused initial beam diameters were larger than matching unfocused beams (10 of 11 cases) the focused beams showed 16%–83% larger peak-to-entrance dose ratios and 1.3 to 3.4-fold increases in dose delivery efficiency. Peak-to-entrance and efficiency benefits tended to increase with larger magnet gradients and larger initial diameter focused beams. Finally, it was observed that focusing tended to shift dose in the water phantom volume from the 80%–20% dose range to below 20% of reference dose, compared to unfocused beams. We conclude that focusing proton beams immediately upstream from tissue entry using permanent magnet assemblies can produce beams with larger peak-to-entrance dose ratios and increased dose delivery efficiencies. Such beams could potentially be used in the clinic to irradiate small-field radiosurgical targets with fewer beams, lower entrance dose and shorter treatment times.

  5. GMC ['gimik]: a one-variable Monte Carlo dose algorithm for proton therapy

    Science.gov (United States)

    Depauw, N.; Clasie, B.; Madden, T.; Rosenfeld, A.; Kooy, H.

    2014-03-01

    This work presents the CPU implementation of GMC ['gimik]: a fast yet accurate one-variable Monte Carlo dose algorithm for proton therapy to be incorporated into our in-house treatment planning system, Astroid. GMC is based on a simple mathematical model using the formulated proton scattering power and tabulated data of empirical depth-dose distributions. These Bragg peaks determine the energy deposited along the particle's track. The polar scattering angle is based on the particle's local energy and the voxel's density, while the azimuthal component of that scattering angle is the single variable in GMC, uniformly distributed from 0 to 2π. The halo effect of the beam, currently not implemented, will consider large scattering angles and secondary protons for a small percentage of the incident histories. GMC shows strong agreement with both the empirical data and GEANT4-based simulations. Its current CPU implementation runs at ~300 m.s--1, approximately ten times faster than GEANT4. Significant speed improvement is expected with the upcoming implementation of multi-threading and the portage to the GPU architecture. In conclusion, a one-variable Monte Carlo dose algorithm was produced for proton therapy dose computations. Its simplicity allows for fast dose computation while conserving accuracy against heterogeneities, hence drastically improving the current algorithms used in treatment planning systems.

  6. SU-E-T-289: Scintillating Fiber Based In-Vivo Dose Monitoring System to the Rectum in Proton Therapy of Prostate Cancer: A Geant4 Monte Carlo Simulation

    International Nuclear Information System (INIS)

    Tesfamicael, B; Gueye, P; Lyons, D; Mahesh, M; Avery, S

    2014-01-01

    Purpose: To construct a dose monitoring system based on an endorectal balloon coupled to thin scintillating fibers to study the dose delivered to the rectum during prostate cancer proton therapy Methods: The Geant4 Monte Carlo toolkit version 9.6p02 was used to simulate prostate cancer proton therapy treatments of an endorectal balloon (for immobilization of a 2.9 cm diameter prostate gland) and a set of 34 scintillating fibers symmetrically placed around the balloon and perpendicular to the proton beam direction (for dosimetry measurements) Results: A linear response of the fibers to the dose delivered was observed within <2%, a property that makes them good candidates for real time dosimetry. Results obtained show that the closest fiber recorded about 1/3 of the dose to the target with a 1/r 2 decrease in the dose distribution as one goes toward the frontal and distal top fibers. Very low dose was recorded by the bottom fibers (about 45 times comparatively), which is a clear indication that the overall volume of the rectal wall that is exposed to a higher dose is relatively minimized. Further analysis indicated a simple scaling relationship between the dose to the prostate and the dose to the top fibers (a linear fit gave a slope of −0.07±0.07 MeV per treatment Gy) Conclusion: Thin (1 mm × 1 mm × 100 cm) long scintillating fibers were found to be ideal for real time in-vivo dose measurement to the rectum for prostate cancer proton therapy. The linear response of the fibers to the dose delivered makes them good candidates of dosimeters. With thorough calibration and the ability to define a good correlation between the dose to the target and the dose to the fibers, such dosimeters can be used for real time dose verification to the target

  7. Characterizing a Proton Beam Scanning System for Monte Carlo Dose Calculation in Patients

    Science.gov (United States)

    Grassberger, C; Lomax, Tony; Paganetti, H

    2015-01-01

    The presented work has two goals. First, to demonstrate the feasibility of accurately characterizing a proton radiation field at treatment head exit for Monte Carlo dose calculation of active scanning patient treatments. Second, to show that this characterization can be done based on measured depth dose curves and spot size alone, without consideration of the exact treatment head delivery system. This is demonstrated through calibration of a Monte Carlo code to the specific beam lines of two institutions, Massachusetts General Hospital (MGH) and Paul Scherrer Institute (PSI). Comparison of simulations modeling the full treatment head at MGH to ones employing a parameterized phase space of protons at treatment head exit reveals the adequacy of the method for patient simulations. The secondary particle production in the treatment head is typically below 0.2% of primary fluence, except for low–energy electrons (protons), whose contribution to skin dose is negligible. However, there is significant difference between the two methods in the low-dose penumbra, making full treatment head simulations necessary to study out-of field effects such as secondary cancer induction. To calibrate the Monte Carlo code to measurements in a water phantom, we use an analytical Bragg peak model to extract the range-dependent energy spread at the two institutions, as this quantity is usually not available through measurements. Comparison of the measured with the simulated depth dose curves demonstrates agreement within 0.5mm over the entire energy range. Subsequently, we simulate three patient treatments with varying anatomical complexity (liver, head and neck and lung) to give an example how this approach can be employed to investigate site-specific discrepancies between treatment planning system and Monte Carlo simulations. PMID:25549079

  8. Monte Carlo calculations supporting patient plan verification in proton therapy

    Directory of Open Access Journals (Sweden)

    Thiago Viana Miranda Lima

    2016-03-01

    Full Text Available Patient’s treatment plan verification covers substantial amount of the quality assurance (QA resources, this is especially true for Intensity Modulated Proton Therapy (IMPT. The use of Monte Carlo (MC simulations in supporting QA has been widely discussed and several methods have been proposed. In this paper we studied an alternative approach from the one being currently applied clinically at Centro Nazionale di Adroterapia Oncologica (CNAO. We reanalysed the previously published data (Molinelli et al. 2013, where 9 patient plans were investigated in which the warning QA threshold of 3% mean dose deviation was crossed. The possibility that these differences between measurement and calculated dose were related to dose modelling (Treatment Planning Systems (TPS vs MC, limitations on dose delivery system or detectors mispositioning was originally explored but other factors such as the geometric description of the detectors were not ruled out. For the purpose of this work we compared ionisation-chambers measurements with different MC simulations results. It was also studied some physical effects introduced by this new approach for example inter detector interference and the delta ray thresholds. The simulations accounting for a detailed geometry typically are superior (statistical difference - p-value around 0.01 to most of the MC simulations used at CNAO (only inferior to the shift approach used. No real improvement were observed in reducing the current delta-ray threshold used (100 keV and no significant interference between ion chambers in the phantom were detected (p-value 0.81. In conclusion, it was observed that the detailed geometrical description improves the agreement between measurement and MC calculations in some cases. But in other cases position uncertainty represents the dominant uncertainty. The inter chamber disturbance was not detected for the therapeutic protons energies and the results from the current delta threshold are

  9. Monte Carlo simulation of proton boron fusion reaction for radiation therapy

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Sun Mi; Yoon, Do Kun; Suh, Tae Suk [Catholic University of Korea, Seoul (Korea, Republic of)

    2016-05-15

    The principle of the proton boron fusion therapy (PBFT) is based on this reaction as the radiation therapy technique. First, because three alpha particles can contribute to the death of the tumor cell by the use of one proton, high therapy efficiency can be achieved by using smaller flux than conventional proton therapy or the boron neutron capture therapy (BNCT), after the thermal neutron was captured by the labeled boron in the tumor region, an alpha particle is emitted from the capture reaction point. An alpha particle induces the death of the tumor cell by the one capture reaction. However, three alpha particles are emitted from the point of the proton boron fusion reaction. If this reaction is applied to the radiation therapy, the therapy results could be more effective in inducing the death of tumor cells using a smaller flux. In addition, the proton's energy loss during its propagation through matter is described by the Bragg-peak. After the boron-labeled compound is accumulated in the tumor region, if the portion of the proton's maximum dose (Bragg-peak) is included at the tumor region, which is the boron uptake region (BUR), a dramatic therapy effect with less damage to normal tissue can be expected. This study was performed to introduce a therapy method using the proton boron fusion reaction and verify the theoretical validity of PBFT using Monte Carlo simulations. In this study, there are two parts of the simulation to confirm the validity of PBFT. First, the variation of the Bragg-peak of the proton depending on the location of the BUR was examined. The other simulation was performed to confirm the existence of the prompt gamma ray peak of 719 keV from energy spectrum simulation. PBFT method is still at the conceptual stage, the verification of its effectiveness is required for the use of a physical approach.

  10. SU-E-T-239: Monte Carlo Modelling of SMC Proton Nozzles Using TOPAS

    International Nuclear Information System (INIS)

    Chung, K; Kim, J; Shin, J; Han, Y; Ju, S; Hong, C; Kim, D; Kim, H; Shin, E; Ahn, S; Chung, S; Choi, D

    2014-01-01

    Purpose: To expedite and cross-check the commissioning of the proton therapy nozzles at Samsung Medical Center using TOPAS. Methods: We have two different types of nozzles at Samsung Medical Center (SMC), a multi-purpose nozzle and a pencil beam scanning dedicated nozzle. Both nozzles have been modelled in Monte Carlo simulation by using TOPAS based on the vendor-provided geometry. The multi-purpose nozzle is mainly composed of wobbling magnets, scatterers, ridge filters and multi-leaf collimators (MLC). Including patient specific apertures and compensators, all the parts of the nozzle have been implemented in TOPAS following the geometry information from the vendor.The dedicated scanning nozzle has a simpler structure than the multi-purpose nozzle with a vacuum pipe at the down stream of the nozzle.A simple water tank volume has been implemented to measure the dosimetric characteristics of proton beams from the nozzles. Results: We have simulated the two proton beam nozzles at SMC. Two different ridge filters have been tested for the spread-out Bragg peak (SOBP) generation of wobbling mode in the multi-purpose nozzle. The spot sizes and lateral penumbra in two nozzles have been simulated and analyzed using a double Gaussian model. Using parallel geometry, both the depth dose curve and dose profile have been measured simultaneously. Conclusion: The proton therapy nozzles at SMC have been successfully modelled in Monte Carlo simulation using TOPAS. We will perform a validation with measured base data and then use the MC simulation to interpolate/extrapolate the measured data. We believe it will expedite the commissioning process of the proton therapy nozzles at SMC

  11. Srna-Monte Carlo codes for proton transport simulation in combined and voxelized geometries

    CERN Document Server

    Ilic, R D; Stankovic, S J

    2002-01-01

    This paper describes new Monte Carlo codes for proton transport simulations in complex geometrical forms and in materials of different composition. The SRNA codes were developed for three dimensional (3D) dose distribution calculation in proton therapy and dosimetry. The model of these codes is based on the theory of proton multiple scattering and a simple model of compound nucleus decay. The developed package consists of two codes: SRNA-2KG and SRNA-VOX. The first code simulates proton transport in combined geometry that can be described by planes and second order surfaces. The second one uses the voxelized geometry of material zones and is specifically adopted for the application of patient computer tomography data. Transition probabilities for both codes are given by the SRNADAT program. In this paper, we will present the models and algorithms of our programs, as well as the results of the numerical experiments we have carried out applying them, along with the results of proton transport simulation obtaine...

  12. Monte Carlo simulations for angular and spatial distributions in therapeutic-energy proton beams

    Science.gov (United States)

    Lin, Yi-Chun; Pan, C. Y.; Chiang, K. J.; Yuan, M. C.; Chu, C. H.; Tsai, Y. W.; Teng, P. K.; Lin, C. H.; Chao, T. C.; Lee, C. C.; Tung, C. J.; Chen, A. E.

    2017-11-01

    The purpose of this study is to compare the angular and spatial distributions of therapeutic-energy proton beams obtained from the FLUKA, GEANT4 and MCNP6 Monte Carlo codes. The Monte Carlo simulations of proton beams passing through two thin targets and a water phantom were investigated to compare the primary and secondary proton fluence distributions and dosimetric differences among these codes. The angular fluence distributions, central axis depth-dose profiles, and lateral distributions of the Bragg peak cross-field were calculated to compare the proton angular and spatial distributions and energy deposition. Benchmark verifications from three different Monte Carlo simulations could be used to evaluate the residual proton fluence for the mean range and to estimate the depth and lateral dose distributions and the characteristic depths and lengths along the central axis as the physical indices corresponding to the evaluation of treatment effectiveness. The results showed a general agreement among codes, except that some deviations were found in the penumbra region. These calculated results are also particularly helpful for understanding primary and secondary proton components for stray radiation calculation and reference proton standard determination, as well as for determining lateral dose distribution performance in proton small-field dosimetry. By demonstrating these calculations, this work could serve as a guide to the recent field of Monte Carlo methods for therapeutic-energy protons.

  13. The Monte Carlo SRNA-VOX code for 3D proton dose distribution in voxelized geometry using CT data

    International Nuclear Information System (INIS)

    Ilic, Radovan D; Spasic-Jokic, Vesna; Belicev, Petar; Dragovic, Milos

    2005-01-01

    This paper describes the application of the SRNA Monte Carlo package for proton transport simulations in complex geometry and different material compositions. The SRNA package was developed for 3D dose distribution calculation in proton therapy and dosimetry and it was based on the theory of multiple scattering. The decay of proton induced compound nuclei was simulated by the Russian MSDM model and our own using ICRU 63 data. The developed package consists of two codes: the SRNA-2KG, which simulates proton transport in combinatorial geometry and the SRNA-VOX, which uses the voxelized geometry using the CT data and conversion of the Hounsfield's data to tissue elemental composition. Transition probabilities for both codes are prepared by the SRNADAT code. The simulation of the proton beam characterization by multi-layer Faraday cup, spatial distribution of positron emitters obtained by the SRNA-2KG code and intercomparison of computational codes in radiation dosimetry, indicate immediate application of the Monte Carlo techniques in clinical practice. In this paper, we briefly present the physical model implemented in the SRNA package, the ISTAR proton dose planning software, as well as the results of the numerical experiments with proton beams to obtain 3D dose distribution in the eye and breast tumour

  14. The Monte Carlo srna code as the engine in istar proton dose planning software for the tesla accelerator installation

    Directory of Open Access Journals (Sweden)

    Ilić Radovan D.

    2004-01-01

    Full Text Available This paper describes the application of SRNA Monte Carlo package for proton transport simulations in complex geometry and different material composition. SRNA package was developed for 3D dose distribution calculation in proton therapy and dosimetry and it was based on the theory of multiple scattering. The compound nuclei decay was simulated by our own and the Russian MSDM models using ICRU 63 data. The developed package consists of two codes SRNA-2KG, which simulates proton transport in the combinatorial geometry and SRNA-VOX, which uses the voxelized geometry using the CT data and conversion of the Hounsfield’s data to tissue elemental composition. Transition probabilities for both codes are prepared by the SRNADAT code. The simulation of proton beam characterization by Multi-Layer Faraday Cup, spatial distribution of positron emitters obtained by SRNA-2KG code, and intercomparison of computational codes in radiation dosimetry, indicate the immediate application of the Monte Carlo techniques in clinical practice. In this paper, we briefly present the physical model implemented in SRNA pack age, the ISTAR proton dose planning software, as well as the results of the numerical experiments with proton beams to obtain 3D dose distribution in the eye and breast tumor.

  15. Range uncertainties in proton therapy and the role of Monte Carlo simulations

    International Nuclear Information System (INIS)

    Paganetti, Harald

    2012-01-01

    The main advantages of proton therapy are the reduced total energy deposited in the patient as compared to photon techniques and the finite range of the proton beam. The latter adds an additional degree of freedom to treatment planning. The range in tissue is associated with considerable uncertainties caused by imaging, patient setup, beam delivery and dose calculation. Reducing the uncertainties would allow a reduction of the treatment volume and thus allow a better utilization of the advantages of protons. This paper summarizes the role of Monte Carlo simulations when aiming at a reduction of range uncertainties in proton therapy. Differences in dose calculation when comparing Monte Carlo with analytical algorithms are analyzed as well as range uncertainties due to material constants and CT conversion. Range uncertainties due to biological effects and the role of Monte Carlo for in vivo range verification are discussed. Furthermore, the current range uncertainty recipes used at several proton therapy facilities are revisited. We conclude that a significant impact of Monte Carlo dose calculation can be expected in complex geometries where local range uncertainties due to multiple Coulomb scattering will reduce the accuracy of analytical algorithms. In these cases Monte Carlo techniques might reduce the range uncertainty by several mm. (topical review)

  16. Transmission calculation by empirical numerical model and Monte Carlo simulation in high energy proton radiography of thick objects

    Science.gov (United States)

    Zheng, Na; Xu, Hai-Bo

    2015-10-01

    An empirical numerical model that includes nuclear absorption, multiple Coulomb scattering and energy loss is presented for the calculation of transmission through thick objects in high energy proton radiography. In this numerical model the angular distributions are treated as Gaussians in the laboratory frame. A Monte Carlo program based on the Geant4 toolkit was developed and used for high energy proton radiography experiment simulations and verification of the empirical numerical model. The two models are used to calculate the transmission fraction of carbon and lead step-wedges in proton radiography at 24 GeV/c, and to calculate radial transmission of the French Test Object in proton radiography at 24 GeV/c with different angular cuts. It is shown that the results of the two models agree with each other, and an analysis of the slight differences is given. Supported by NSAF (11176001) and Science and Technology Developing Foundation of China Academy of Engineering Physics (2012A0202006)

  17. Monte-Carlo simulation of proton radiotherapy for human eye

    International Nuclear Information System (INIS)

    Liu Yunpeng; Tang Xiaobin; Xie Qin; Chen Feida; Geng Changran; Chen Da

    2010-01-01

    The 62 MeV proton beam was selected to develop a MCNPX model of the human eye to approximate dose delivered from proton therapy by. In the course of proton therapy, two treatment simulations were considered. The first simulation was an ideal treatment scenario. In this case, the dose of tumor was 50.03 Gy, which was at the level of effective treatment, while other organizations were in the range of acceptable dose. The second case was a worst case scenario to simulate a patient gazing directly into the treatment beam during therapy. The bulk of dose deposited in the cornea, lens, and anterior chamber region. However, the dose of tumor area was zero. The calculated results show an agreement accordance with the relative reference, which confirmed that the MCNPX code can simulate proton radiotherapy perfectly, and is a capable platform for patient planning. The data from the worst case can be used for dose reconstruction of the clinical accident. (authors)

  18. Monte Carlo simulation of prompt γ-ray emission in proton therapy using a specific track length estimator

    International Nuclear Information System (INIS)

    El Kanawati, W; Létang, J M; Sarrut, D; Freud, N; Dauvergne, D; Pinto, M; Testa, É

    2015-01-01

    A Monte Carlo (MC) variance reduction technique is developed for prompt-γ emitters calculations in proton therapy. Prompt-γ emitted through nuclear fragmentation reactions and exiting the patient during proton therapy could play an important role to help monitoring the treatment. However, the estimation of the number and the energy of emitted prompt-γ per primary proton with MC simulations is a slow process. In order to estimate the local distribution of prompt-γ emission in a volume of interest for a given proton beam of the treatment plan, a MC variance reduction technique based on a specific track length estimator (TLE) has been developed. First an elemental database of prompt-γ emission spectra is established in the clinical energy range of incident protons for all elements in the composition of human tissues. This database of the prompt-γ spectra is built offline with high statistics. Regarding the implementation of the prompt-γ TLE MC tally, each proton deposits along its track the expectation of the prompt-γ spectra from the database according to the proton kinetic energy and the local material composition. A detailed statistical study shows that the relative efficiency mainly depends on the geometrical distribution of the track length. Benchmarking of the proposed prompt-γ TLE MC technique with respect to an analogous MC technique is carried out. A large relative efficiency gain is reported, ca. 10 5 . (paper)

  19. Proton energy and scattering angle radiographs to improve proton treatment planning : a Monte Carlo study

    NARCIS (Netherlands)

    Biegun, Aleksandra; Takatsu, Jun; Nakaji, Taku; van Goethem, Marc-Jan; van der Graaf, Emiel; Koffeman, E.; Visser, Jan; Brandenburg, Sijtze

    2016-01-01

    The novel proton radiography imaging technique has a large potential to be used in direct measurement of the proton energy loss (proton stopping power, PSP) in various tissues in the patient. The uncertainty of PSPs, currently obtained from translation of X-ray Computed Tomography (xCT) images,

  20. The proton therapy nozzles at Samsung Medical Center: A Monte Carlo simulation study using TOPAS

    Science.gov (United States)

    Chung, Kwangzoo; Kim, Jinsung; Kim, Dae-Hyun; Ahn, Sunghwan; Han, Youngyih

    2015-07-01

    To expedite the commissioning process of the proton therapy system at Samsung Medical Center (SMC), we have developed a Monte Carlo simulation model of the proton therapy nozzles by using TOol for PArticle Simulation (TOPAS). At SMC proton therapy center, we have two gantry rooms with different types of nozzles: a multi-purpose nozzle and a dedicated scanning nozzle. Each nozzle has been modeled in detail following the geometry information provided by the manufacturer, Sumitomo Heavy Industries, Ltd. For this purpose, the novel features of TOPAS, such as the time feature or the ridge filter class, have been used, and the appropriate physics models for proton nozzle simulation have been defined. Dosimetric properties, like percent depth dose curve, spreadout Bragg peak (SOBP), and beam spot size, have been simulated and verified against measured beam data. Beyond the Monte Carlo nozzle modeling, we have developed an interface between TOPAS and the treatment planning system (TPS), RayStation. An exported radiotherapy (RT) plan from the TPS is interpreted by using an interface and is then translated into the TOPAS input text. The developed Monte Carlo nozzle model can be used to estimate the non-beam performance, such as the neutron background, of the nozzles. Furthermore, the nozzle model can be used to study the mechanical optimization of the design of the nozzle.

  1. Fast multipurpose Monte Carlo simulation for proton therapy using multi- and many-core CPU architectures

    International Nuclear Information System (INIS)

    Souris, Kevin; Lee, John Aldo; Sterpin, Edmond

    2016-01-01

    Purpose: Accuracy in proton therapy treatment planning can be improved using Monte Carlo (MC) simulations. However the long computation time of such methods hinders their use in clinical routine. This work aims to develop a fast multipurpose Monte Carlo simulation tool for proton therapy using massively parallel central processing unit (CPU) architectures. Methods: A new Monte Carlo, called MCsquare (many-core Monte Carlo), has been designed and optimized for the last generation of Intel Xeon processors and Intel Xeon Phi coprocessors. These massively parallel architectures offer the flexibility and the computational power suitable to MC methods. The class-II condensed history algorithm of MCsquare provides a fast and yet accurate method of simulating heavy charged particles such as protons, deuterons, and alphas inside voxelized geometries. Hard ionizations, with energy losses above a user-specified threshold, are simulated individually while soft events are regrouped in a multiple scattering theory. Elastic and inelastic nuclear interactions are sampled from ICRU 63 differential cross sections, thereby allowing for the computation of prompt gamma emission profiles. MCsquare has been benchmarked with the GATE/GEANT4 Monte Carlo application for homogeneous and heterogeneous geometries. Results: Comparisons with GATE/GEANT4 for various geometries show deviations within 2%–1 mm. In spite of the limited memory bandwidth of the coprocessor simulation time is below 25 s for 10 7 primary 200 MeV protons in average soft tissues using all Xeon Phi and CPU resources embedded in a single desktop unit. Conclusions: MCsquare exploits the flexibility of CPU architectures to provide a multipurpose MC simulation tool. Optimized code enables the use of accurate MC calculation within a reasonable computation time, adequate for clinical practice. MCsquare also simulates prompt gamma emission and can thus be used also for in vivo range verification.

  2. Monte-Carlo Geant4 numerical simulation of experiments at 247-MeV proton microscope

    Science.gov (United States)

    Kantsyrev, A. V.; Skoblyakov, A. V.; Bogdanov, A. V.; Golubev, A. A.; Shilkin, N. S.; Yuriev, D. S.; Mintsev, V. B.

    2018-01-01

    A radiographic facility for an investigation of fast dynamic processes with areal density of targets up to 5 g/cm2 is under development on the basis of high-current proton linear accelerator at the Institute for Nuclear Research (Troitsk, Russia). A virtual model of the proton microscope developed in a software toolkit Geant4 is presented in the article. Fullscale Monte-Carlo numerical simulation of static radiographic experiments at energy of a proton beam 247 MeV was performed. The results of simulation of proton radiography experiments with static model of shock-compressed xenon are presented. The results of visualization of copper and polymethyl methacrylate step wedges static targets also described.

  3. Comparison of some popular Monte Carlo solution for proton transportation within pCT problem

    International Nuclear Information System (INIS)

    Evseev, Ivan; Assis, Joaquim T. de; Yevseyeva, Olga; Hormaza, Joel M.

    2007-01-01

    The proton transport in matter is described by the Boltzmann kinetic equation for the proton flux density. This equation, however, does not have a general analytical solution. Some approximate analytical solutions have been developed within a number of significant simplifications. Alternatively, the Monte Carlo simulations are widely used. Current work is devoted to the discussion of the proton energy spectra obtained by simulation with SRIM2006, GEANT4 and MCNPX packages. The simulations have been performed considering some further applications of the obtained results in computed tomography with proton beam (pCT). Thus the initial and outgoing proton energies (3 / 300 MeV) as well as the thickness of irradiated target (water and aluminum phantoms within 90% of the full range for a given proton beam energy) were considered in the interval of values typical for pCT applications. One from the most interesting results of this comparison is that while the MCNPX spectra are in a good agreement with analytical description within Fokker-Plank approximation and the GEANT4 simulated spectra are slightly shifted from them the SRIM2006 simulations predict a notably higher mean energy loss for protons. (author)

  4. DEVELOPMENT OF A MULTIMODAL MONTE CARLO BASED TREATMENT PLANNING SYSTEM.

    Science.gov (United States)

    Kumada, Hiroaki; Takada, Kenta; Sakurai, Yoshinori; Suzuki, Minoru; Takata, Takushi; Sakurai, Hideyuki; Matsumura, Akira; Sakae, Takeji

    2017-10-26

    To establish boron neutron capture therapy (BNCT), the University of Tsukuba is developing a treatment device and peripheral devices required in BNCT, such as a treatment planning system. We are developing a new multimodal Monte Carlo based treatment planning system (developing code: Tsukuba Plan). Tsukuba Plan allows for dose estimation in proton therapy, X-ray therapy and heavy ion therapy in addition to BNCT because the system employs PHITS as the Monte Carlo dose calculation engine. Regarding BNCT, several verifications of the system are being carried out for its practical usage. The verification results demonstrate that Tsukuba Plan allows for accurate estimation of thermal neutron flux and gamma-ray dose as fundamental radiations of dosimetry in BNCT. In addition to the practical use of Tsukuba Plan in BNCT, we are investigating its application to other radiation therapies. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  5. Monte Carlo simulation of secondary neutron dose for scanning proton therapy using FLUKA.

    Directory of Open Access Journals (Sweden)

    Chaeyeong Lee

    Full Text Available Proton therapy is a rapidly progressing field for cancer treatment. Globally, many proton therapy facilities are being commissioned or under construction. Secondary neutrons are an important issue during the commissioning process of a proton therapy facility. The purpose of this study is to model and validate scanning nozzles of proton therapy at Samsung Medical Center (SMC by Monte Carlo simulation for beam commissioning. After the commissioning, a secondary neutron ambient dose from proton scanning nozzle (Gantry 1 was simulated and measured. This simulation was performed to evaluate beam properties such as percent depth dose curve, Bragg peak, and distal fall-off, so that they could be verified with measured data. Using the validated beam nozzle, the secondary neutron ambient dose was simulated and then compared with the measured ambient dose from Gantry 1. We calculated secondary neutron dose at several different points. We demonstrated the validity modeling a proton scanning nozzle system to evaluate various parameters using FLUKA. The measured secondary neutron ambient dose showed a similar tendency with the simulation result. This work will increase the knowledge necessary for the development of radiation safety technology in medical particle accelerators.

  6. Patient-specific bolus for range shifter air gap reduction in intensity-modulated proton therapy of head-and-neck cancer studied with Monte Carlo based plan optimization.

    Science.gov (United States)

    Michiels, Steven; Barragán, Ana Maria; Souris, Kevin; Poels, Kenneth; Crijns, Wouter; Lee, John A; Sterpin, Edmond; Nuyts, Sandra; Haustermans, Karin; Depuydt, Tom

    2017-09-23

    Intensity-modulated proton therapy (IMPT) of superficial lesions requires pre-absorbing range shifter (RS) to deliver the more shallow spots. RS air gap minimization is important to avoid spot size degradation, but remains challenging in complex geometries such as in head-and-neck cancer (HNC). In this study, clinical endpoints were investigated for patient-specific bolus and for conventional RS solutions, making use of a Monte Carlo (MC) dose engine for IMPT optimization. For 5 oropharyngeal cancer patients, IMPT spot maps were generated using beamlets calculated with MC. The plans were optimized for three different RS configurations: 3D printed on-skin bolus, snout- and nozzle-mounted RS. Organ-at-risk (OAR) doses and late toxicity probabilities were compared between all configuration-specific optimized plans. The use of bolus reduced the mean dose to all OARs compared to snout and nozzle-mounted RS. The contralateral parotid gland and supraglottic larynx received on average 2.9Gy and 4.2Gy less dose compared to the snout RS. Bolus reduced the average probability for xerostomia by 3.0%. For dysphagia, bolus reduced the probability by 2.7%. Quantification of the dosimetric advantage of patient-specific bolus shows significant reductions compared to conventional RS solutions for xerostomia and dysphagia probability. These results motivate the development of a patient-specific bolus solution in IMPT for HNC. Copyright © 2017 Elsevier B.V. All rights reserved.

  7. A Monte Carlo pencil beam scanning model for proton treatment plan simulation using GATE/GEANT4

    Energy Technology Data Exchange (ETDEWEB)

    Grevillot, L; Freud, N; Sarrut, D [Universite de Lyon, CREATIS, CNRS UMR5220, Inserm U1044, INSA-Lyon, Universite Lyon 1, Centre Leon Berard, Lyon (France); Bertrand, D; Dessy, F, E-mail: loic.grevillot@creatis.insa-lyon.fr [IBA, B-1348, Louvain-la Neuve (Belgium)

    2011-08-21

    This work proposes a generic method for modeling scanned ion beam delivery systems, without simulation of the treatment nozzle and based exclusively on beam data library (BDL) measurements required for treatment planning systems (TPS). To this aim, new tools dedicated to treatment plan simulation were implemented in the Gate Monte Carlo platform. The method was applied to a dedicated nozzle from IBA for proton pencil beam scanning delivery. Optical and energy parameters of the system were modeled using a set of proton depth-dose profiles and spot sizes measured at 27 therapeutic energies. For further validation of the beam model, specific 2D and 3D plans were produced and then measured with appropriate dosimetric tools. Dose contributions from secondary particles produced by nuclear interactions were also investigated using field size factor experiments. Pristine Bragg peaks were reproduced with 0.7 mm range and 0.2 mm spot size accuracy. A 32 cm range spread-out Bragg peak with 10 cm modulation was reproduced with 0.8 mm range accuracy and a maximum point-to-point dose difference of less than 2%. A 2D test pattern consisting of a combination of homogeneous and high-gradient dose regions passed a 2%/2 mm gamma index comparison for 97% of the points. In conclusion, the generic modeling method proposed for scanned ion beam delivery systems was applicable to an IBA proton therapy system. The key advantage of the method is that it only requires BDL measurements of the system. The validation tests performed so far demonstrated that the beam model achieves clinical performance, paving the way for further studies toward TPS benchmarking. The method involves new sources that are available in the new Gate release V6.1 and could be further applied to other particle therapy systems delivering protons or other types of ions like carbon.

  8. A Monte Carlo pencil beam scanning model for proton treatment plan simulation using GATE/GEANT4.

    Science.gov (United States)

    Grevillot, L; Bertrand, D; Dessy, F; Freud, N; Sarrut, D

    2011-08-21

    This work proposes a generic method for modeling scanned ion beam delivery systems, without simulation of the treatment nozzle and based exclusively on beam data library (BDL) measurements required for treatment planning systems (TPS). To this aim, new tools dedicated to treatment plan simulation were implemented in the Gate Monte Carlo platform. The method was applied to a dedicated nozzle from IBA for proton pencil beam scanning delivery. Optical and energy parameters of the system were modeled using a set of proton depth-dose profiles and spot sizes measured at 27 therapeutic energies. For further validation of the beam model, specific 2D and 3D plans were produced and then measured with appropriate dosimetric tools. Dose contributions from secondary particles produced by nuclear interactions were also investigated using field size factor experiments. Pristine Bragg peaks were reproduced with 0.7 mm range and 0.2 mm spot size accuracy. A 32 cm range spread-out Bragg peak with 10 cm modulation was reproduced with 0.8 mm range accuracy and a maximum point-to-point dose difference of less than 2%. A 2D test pattern consisting of a combination of homogeneous and high-gradient dose regions passed a 2%/2 mm gamma index comparison for 97% of the points. In conclusion, the generic modeling method proposed for scanned ion beam delivery systems was applicable to an IBA proton therapy system. The key advantage of the method is that it only requires BDL measurements of the system. The validation tests performed so far demonstrated that the beam model achieves clinical performance, paving the way for further studies toward TPS benchmarking. The method involves new sources that are available in the new Gate release V6.1 and could be further applied to other particle therapy systems delivering protons or other types of ions like carbon.

  9. Monte Carlo calculations and measurements of absorbed dose per monitor unit for the treatment of uveal melanoma with proton therapy

    Energy Technology Data Exchange (ETDEWEB)

    Koch, Nicholas; Newhauser, Wayne D; Titt, Uwe; Starkschall, George [Department of Radiation Physics, University of Texas M. D. Anderson Cancer Center, 1515 Holcombe Blvd, Houston, TX 77030 (United States); Gombos, Dan [Section of Ophthalmology, Department of Head and Neck Surgery MDACC Unit 441 (United States); Coombes, Kevin [Graduate School of Biomedical Sciences, University of Texas Health Science Center, 6767 Bertner Avenue, Houston, TX 77030 (United States)], E-mail: kochn@musc.edu

    2008-03-21

    The treatment of uveal melanoma with proton radiotherapy has provided excellent clinical outcomes. However, contemporary treatment planning systems use simplistic dose algorithms that limit the accuracy of relative dose distributions. Further, absolute predictions of absorbed dose per monitor unit are not yet available in these systems. The purpose of this study was to determine if Monte Carlo methods could predict dose per monitor unit (D/MU) value at the center of a proton spread-out Bragg peak (SOBP) to within 1% on measured values for a variety of treatment fields relevant to ocular proton therapy. The MCNPX Monte Carlo transport code, in combination with realistic models for the ocular beam delivery apparatus and a water phantom, was used to calculate dose distributions and D/MU values, which were verified by the measurements. Measured proton beam data included central-axis depth dose profiles, relative cross-field profiles and absolute D/MU measurements under several combinations of beam penetration ranges and range-modulation widths. The Monte Carlo method predicted D/MU values that agreed with measurement to within 1% and dose profiles that agreed with measurement to within 3% of peak dose or within 0.5 mm distance-to-agreement. Lastly, a demonstration of the clinical utility of this technique included calculations of dose distributions and D/MU values in a realistic model of the human eye. It is possible to predict D/MU values accurately for clinical relevant range-modulated proton beams for ocular therapy using the Monte Carlo method. It is thus feasible to use the Monte Carlo method as a routine absolute dose algorithm for ocular proton therapy.

  10. The optimal balance between quality and efficiency in proton radiography imaging technique at various proton beam energies: A Monte Carlo study.

    Science.gov (United States)

    Biegun, A K; van Goethem, M-J; van der Graaf, E R; van Beuzekom, M; Koffeman, E N; Nakaji, T; Takatsu, J; Visser, J; Brandenburg, S

    2017-09-01

    Proton radiography is a novel imaging modality that allows direct measurement of the proton energy loss in various tissues. Currently, due to the conversion of so-called Hounsfield units from X-ray Computed Tomography (CT) into relative proton stopping powers (RPSP), the uncertainties of RPSP are 3-5% or higher, which need to be minimized down to 1% to make the proton treatment plans more accurate. In this work, we simulated a proton radiography system, with position-sensitive detectors (PSDs) and a residual energy detector (RED). The simulations were built using Geant4, a Monte Carlo simulation toolkit. A phantom, consisting of several materials was placed between the PSDs of various Water Equivalent Thicknesses (WET), corresponding to an ideal detector, a gaseous detector, silicon and plastic scintillator detectors. The energy loss radiograph and the scattering angle distributions of the protons were studied for proton beam energies of 150MeV, 190MeV and 230MeV. To improve the image quality deteriorated by the multiple Coulomb scattering (MCS), protons with small angles were selected. Two ways of calculating a scattering angle were considered using the proton's direction and position. A scattering angle cut of 8.7mrad was applied giving an optimal balance between quality and efficiency of the radiographic image. For the three proton beam energies, the number of protons used in image reconstruction with the direction method was half the number of protons kept using the position method. Copyright © 2017 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  11. SU-E-T-314: The Application of Cloud Computing in Pencil Beam Scanning Proton Therapy Monte Carlo Simulation

    International Nuclear Information System (INIS)

    Wang, Z; Gao, M

    2014-01-01

    Purpose: Monte Carlo simulation plays an important role for proton Pencil Beam Scanning (PBS) technique. However, MC simulation demands high computing power and is limited to few large proton centers that can afford a computer cluster. We study the feasibility of utilizing cloud computing in the MC simulation of PBS beams. Methods: A GATE/GEANT4 based MC simulation software was installed on a commercial cloud computing virtual machine (Linux 64-bits, Amazon EC2). Single spot Integral Depth Dose (IDD) curves and in-air transverse profiles were used to tune the source parameters to simulate an IBA machine. With the use of StarCluster software developed at MIT, a Linux cluster with 2–100 nodes can be conveniently launched in the cloud. A proton PBS plan was then exported to the cloud where the MC simulation was run. Results: The simulated PBS plan has a field size of 10×10cm 2 , 20cm range, 10cm modulation, and contains over 10,000 beam spots. EC2 instance type m1.medium was selected considering the CPU/memory requirement and 40 instances were used to form a Linux cluster. To minimize cost, master node was created with on-demand instance and worker nodes were created with spot-instance. The hourly cost for the 40-node cluster was $0.63 and the projected cost for a 100-node cluster was $1.41. Ten million events were simulated to plot PDD and profile, with each job containing 500k events. The simulation completed within 1 hour and an overall statistical uncertainty of < 2% was achieved. Good agreement between MC simulation and measurement was observed. Conclusion: Cloud computing is a cost-effective and easy to maintain platform to run proton PBS MC simulation. When proton MC packages such as GATE and TOPAS are combined with cloud computing, it will greatly facilitate the pursuing of PBS MC studies, especially for newly established proton centers or individual researchers

  12. Experimental validation of the TOPAS Monte Carlo system for passive scattering proton therapy

    International Nuclear Information System (INIS)

    Testa, M.; Schümann, J.; Lu, H.-M.; Paganetti, H.; Shin, J.; Faddegon, B.; Perl, J.

    2013-01-01

    Purpose: TOPAS (TOol for PArticle Simulation) is a particle simulation code recently developed with the specific aim of making Monte Carlo simulations user-friendly for research and clinical physicists in the particle therapy community. The authors present a thorough and extensive experimental validation of Monte Carlo simulations performed with TOPAS in a variety of setups relevant for proton therapy applications. The set of validation measurements performed in this work represents an overall end-to-end testing strategy recommended for all clinical centers planning to rely on TOPAS for quality assurance or patient dose calculation and, more generally, for all the institutions using passive-scattering proton therapy systems. Methods: The authors systematically compared TOPAS simulations with measurements that are performed routinely within the quality assurance (QA) program in our institution as well as experiments specifically designed for this validation study. First, the authors compared TOPAS simulations with measurements of depth-dose curves for spread-out Bragg peak (SOBP) fields. Second, absolute dosimetry simulations were benchmarked against measured machine output factors (OFs). Third, the authors simulated and measured 2D dose profiles and analyzed the differences in terms of field flatness and symmetry and usable field size. Fourth, the authors designed a simple experiment using a half-beam shifter to assess the effects of multiple Coulomb scattering, beam divergence, and inverse square attenuation on lateral and longitudinal dose profiles measured and simulated in a water phantom. Fifth, TOPAS’ capabilities to simulate time dependent beam delivery was benchmarked against dose rate functions (i.e., dose per unit time vs time) measured at different depths inside an SOBP field. Sixth, simulations of the charge deposited by protons fully stopping in two different types of multilayer Faraday cups (MLFCs) were compared with measurements to benchmark the

  13. The impact of different Monte Carlo models on the cross section measurement of top-pair production at 7 TeV proton-proton collisions

    Energy Technology Data Exchange (ETDEWEB)

    Krause, Claudius

    2012-04-15

    High energy proton-proton collisions lead to a large amount of secondary particles to be measured in a detector. A final state containing top quarks is of particular interest. But top quarks are only produced in a small fraction of the collisions. Hence, criteria must be defined to separate events containing top quarks from the background. From detectors, we record signals, for example hits in the tracker system or deposits in the calorimeters. In order to obtain the momentum of the particles, we apply algorithms to reconstruct tracks in space. More sophisticated algorithms are needed to identify the flavour of quarks, such as b-tagging. Several steps are needed to test these algorithms. Collision products of proton-proton events are generated using Monte Carlo techniques and their passage through the detector is simulated. After that, the algorithms are applied and the signal efficiency and the mistagging rate can be obtained. There are, however, many different approaches and algorithms realized in programs, so the question arises if the choice of the Monte Carlo generator influences the measured quantities. In this thesis, two commonly used Monte Carlo generators, SHERPA and MadGraph/MadEvent, are compared and the differences in the selection efficiency of semimuonic tt events are estimated. In addition, the distributions of kinematic variables are shown. A special chapter about the matching of matrix elements with parton showers is included. The main algorithms, CKKW for SHERPA and MLM for MadGraph/MadEvent, are introduced.

  14. Proton radiography and fluoroscopy of lung tumors: A Monte Carlo study using patient-specific 4DCT phantoms

    Science.gov (United States)

    Han, Bin; Xu, X. George; Chen, George T. Y.

    2011-01-01

    Purpose: Monte Carlo methods are used to simulate and optimize a time-resolved proton range telescope (TRRT) in localization of intrafractional and interfractional motions of lung tumor and in quantification of proton range variations. Methods: The Monte Carlo N-Particle eXtended (MCNPX) code with a particle tracking feature was employed to evaluate the TRRT performance, especially in visualizing and quantifying proton range variations during respiration. Protons of 230 MeV were tracked one by one as they pass through position detectors, patient 4DCT phantom, and finally scintillator detectors that measured residual ranges. The energy response of the scintillator telescope was investigated. Mass density and elemental composition of tissues were defined for 4DCT data. Results: Proton water equivalent length (WEL) was deduced by a reconstruction algorithm that incorporates linear proton track and lateral spatial discrimination to improve the image quality. 4DCT data for three patients were used to visualize and measure tumor motion and WEL variations. The tumor trajectories extracted from the WEL map were found to be within ∼1 mm agreement with direct 4DCT measurement. Quantitative WEL variation studies showed that the proton radiograph is a good representation of WEL changes from entrance to distal of the target. Conclusions:MCNPX simulation results showed that TRRT can accurately track the motion of the tumor and detect the WEL variations. Image quality was optimized by choosing proton energy, testing parameters of image reconstruction algorithm, and comparing to ground truth 4DCT. The future study will demonstrate the feasibility of using the time resolved proton radiography as an imaging tool for proton treatments of lung tumors. PMID:21626923

  15. A fast - Monte Carlo toolkit on GPU for treatment plan dose recalculation in proton therapy

    Science.gov (United States)

    Senzacqua, M.; Schiavi, A.; Patera, V.; Pioli, S.; Battistoni, G.; Ciocca, M.; Mairani, A.; Magro, G.; Molinelli, S.

    2017-10-01

    In the context of the particle therapy a crucial role is played by Treatment Planning Systems (TPSs), tools aimed to compute and optimize the tratment plan. Nowadays one of the major issues related to the TPS in particle therapy is the large CPU time needed. We developed a software toolkit (FRED) for reducing dose recalculation time by exploiting Graphics Processing Units (GPU) hardware. Thanks to their high parallelization capability, GPUs significantly reduce the computation time, up to factor 100 respect to a standard CPU running software. The transport of proton beams in the patient is accurately described through Monte Carlo methods. Physical processes reproduced are: Multiple Coulomb Scattering, energy straggling and nuclear interactions of protons with the main nuclei composing the biological tissues. FRED toolkit does not rely on the water equivalent translation of tissues, but exploits the Computed Tomography anatomical information by reconstructing and simulating the atomic composition of each crossed tissue. FRED can be used as an efficient tool for dose recalculation, on the day of the treatment. In fact it can provide in about one minute on standard hardware the dose map obtained combining the treatment plan, earlier computed by the TPS, and the current patient anatomic arrangement.

  16. Geant4 Monte Carlo simulation of absorbed dose and radiolysis yields enhancement from a gold nanoparticle under MeV proton irradiation

    Energy Technology Data Exchange (ETDEWEB)

    Tran, H.N., E-mail: tranngochoang@tdt.edu.vn [Division of Nuclear Physics, Ton Duc Thang University, Tan Phong Ward, District 7, Ho Chi Minh City (Viet Nam); Faculty of Applied Sciences, Ton Duc Thang University, Tan Phong Ward, District 7, Ho Chi Minh City (Viet Nam); Karamitros, M. [Notre Dame Radiation Laboratory, University of Notre-Dame, IN 46556 (United States); Ivanchenko, V.N. [Geant4 Associates International Ltd, Hebden Bridge (United Kingdom); Guatelli, S.; McKinnon, S. [Centre For Medical Radiation Physics, University of Wollongong (Australia); Illawarra Health and Medical Research, University of Wollongong, NSW (Australia); Murakami, K.; Sasaki, T.; Okada, S. [Computing Research Center, High Energy Accelerator Organization, KEK, Tsukuba City (Japan); Bordage, M.C. [INSERM, UMR 1037, CRCT, F-31000 Toulouse (France); Univ. Toulouse III-Paul Sabatier, UMR 1037, CRCT, F-31000 Toulouse (France); Francis, Z. [Saint Joseph University, Faculty of Sciences, Department of Physics, Beirut (Lebanon); El Bitar, Z. [Institut Pluridisciplinaire Hubert Curien/IN2P3/CNRS, Strasbourg (France); Bernal, M.A. [Instituto de Física Gleb Wataghin, Universidade Estadual de Campinas, SP (Brazil); Shin, J.I. [Division of Heavy Ion Clinical Research, Korea Institute of Radiological and Medical Science, 75, Nowon-ro, Nowon-gu, Seoul (Korea, Republic of); Lee, S.B. [Proton Therapy Center, National Cancer Center, 323, Ilsan-ro, Ilsandong-gu, Goyang-si, Gyeonggi-do (Korea, Republic of); Barberet, Ph. [Univ. Bordeaux, CENBG, UMR 5797, F-33170 Gradignan (France); CNRS, IN2P3, CENBG, UMR 5797, F-33170 Gradignan (France); Tran, T.T. [VNUHCM-University of Science (Viet Nam); Brown, J.M.C. [School of Mathematics and Physics, Queen’s University Belfast, Belfast, Northern Ireland (United Kingdom); and others

    2016-04-15

    Gold nanoparticles have been reported as a possible radio-sensitizer agent in radiation therapy due to their ability to increase energy deposition and subsequent direct damage to cells and DNA within their local vicinity. Moreover, this increase in energy deposition also results in an increase of the radiochemical yields. In this work we present, for the first time, an in silico investigation, based on the general purpose Monte Carlo simulation toolkit Geant4, into energy deposition and radical species production around a spherical gold nanoparticle 50 nm in diameter via proton irradiation. Simulations were preformed for incident proton energies ranging from 2 to 170 MeV, which are of interest for clinical proton therapy.

  17. Distributions of secondary particles in proton and carbon-ion therapy: a comparison between GATE/Geant4 and FLUKA Monte Carlo codes.

    Science.gov (United States)

    Robert, C; Dedes, G; Battistoni, G; Böhlen, T T; Buvat, I; Cerutti, F; Chin, M P W; Ferrari, A; Gueth, P; Kurz, C; Lestand, L; Mairani, A; Montarou, G; Nicolini, R; Ortega, P G; Parodi, K; Prezado, Y; Sala, P R; Sarrut, D; Testa, E

    2013-05-07

    Monte Carlo simulations play a crucial role for in-vivo treatment monitoring based on PET and prompt gamma imaging in proton and carbon-ion therapies. The accuracy of the nuclear fragmentation models implemented in these codes might affect the quality of the treatment verification. In this paper, we investigate the nuclear models implemented in GATE/Geant4 and FLUKA by comparing the angular and energy distributions of secondary particles exiting a homogeneous target of PMMA. Comparison results were restricted to fragmentation of (16)O and (12)C. Despite the very simple target and set-up, substantial discrepancies were observed between the two codes. For instance, the number of high energy (>1 MeV) prompt gammas exiting the target was about twice as large with GATE/Geant4 than with FLUKA both for proton and carbon ion beams. Such differences were not observed for the predicted annihilation photon production yields, for which ratios of 1.09 and 1.20 were obtained between GATE and FLUKA for the proton beam and the carbon ion beam, respectively. For neutrons and protons, discrepancies from 14% (exiting protons-carbon ion beam) to 57% (exiting neutrons-proton beam) have been identified in production yields as well as in the energy spectra for neutrons.

  18. Accelerated GPU based SPECT Monte Carlo simulations.

    Science.gov (United States)

    Garcia, Marie-Paule; Bert, Julien; Benoit, Didier; Bardiès, Manuel; Visvikis, Dimitris

    2016-06-07

    Monte Carlo (MC) modelling is widely used in the field of single photon emission computed tomography (SPECT) as it is a reliable technique to simulate very high quality scans. This technique provides very accurate modelling of the radiation transport and particle interactions in a heterogeneous medium. Various MC codes exist for nuclear medicine imaging simulations. Recently, new strategies exploiting the computing capabilities of graphical processing units (GPU) have been proposed. This work aims at evaluating the accuracy of such GPU implementation strategies in comparison to standard MC codes in the context of SPECT imaging. GATE was considered the reference MC toolkit and used to evaluate the performance of newly developed GPU Geant4-based Monte Carlo simulation (GGEMS) modules for SPECT imaging. Radioisotopes with different photon energies were used with these various CPU and GPU Geant4-based MC codes in order to assess the best strategy for each configuration. Three different isotopes were considered: (99m) Tc, (111)In and (131)I, using a low energy high resolution (LEHR) collimator, a medium energy general purpose (MEGP) collimator and a high energy general purpose (HEGP) collimator respectively. Point source, uniform source, cylindrical phantom and anthropomorphic phantom acquisitions were simulated using a model of the GE infinia II 3/8" gamma camera. Both simulation platforms yielded a similar system sensitivity and image statistical quality for the various combinations. The overall acceleration factor between GATE and GGEMS platform derived from the same cylindrical phantom acquisition was between 18 and 27 for the different radioisotopes. Besides, a full MC simulation using an anthropomorphic phantom showed the full potential of the GGEMS platform, with a resulting acceleration factor up to 71. The good agreement with reference codes and the acceleration factors obtained support the use of GPU implementation strategies for improving computational

  19. Accelerated GPU based SPECT Monte Carlo simulations

    Science.gov (United States)

    Garcia, Marie-Paule; Bert, Julien; Benoit, Didier; Bardiès, Manuel; Visvikis, Dimitris

    2016-06-01

    Monte Carlo (MC) modelling is widely used in the field of single photon emission computed tomography (SPECT) as it is a reliable technique to simulate very high quality scans. This technique provides very accurate modelling of the radiation transport and particle interactions in a heterogeneous medium. Various MC codes exist for nuclear medicine imaging simulations. Recently, new strategies exploiting the computing capabilities of graphical processing units (GPU) have been proposed. This work aims at evaluating the accuracy of such GPU implementation strategies in comparison to standard MC codes in the context of SPECT imaging. GATE was considered the reference MC toolkit and used to evaluate the performance of newly developed GPU Geant4-based Monte Carlo simulation (GGEMS) modules for SPECT imaging. Radioisotopes with different photon energies were used with these various CPU and GPU Geant4-based MC codes in order to assess the best strategy for each configuration. Three different isotopes were considered: 99m Tc, 111In and 131I, using a low energy high resolution (LEHR) collimator, a medium energy general purpose (MEGP) collimator and a high energy general purpose (HEGP) collimator respectively. Point source, uniform source, cylindrical phantom and anthropomorphic phantom acquisitions were simulated using a model of the GE infinia II 3/8" gamma camera. Both simulation platforms yielded a similar system sensitivity and image statistical quality for the various combinations. The overall acceleration factor between GATE and GGEMS platform derived from the same cylindrical phantom acquisition was between 18 and 27 for the different radioisotopes. Besides, a full MC simulation using an anthropomorphic phantom showed the full potential of the GGEMS platform, with a resulting acceleration factor up to 71. The good agreement with reference codes and the acceleration factors obtained support the use of GPU implementation strategies for improving computational efficiency

  20. Monte Carlo Based Framework to Support HAZOP Study

    DEFF Research Database (Denmark)

    Danko, Matej; Frutiger, Jerome; Jelemenský, Ľudovít

    2017-01-01

    This study combines Monte Carlo based process simulation features with classical hazard identification techniques for consequences of deviations from normal operating conditions investigation and process safety examination. A Monte Carlo based method has been used to sample and evaluate different...... deviations in process parameters simultaneously, thereby bringing an improvement to the Hazard and Operability study (HAZOP), which normally considers only one at a time deviation in process parameters. Furthermore, Monte Carlo filtering was then used to identify operability and hazard issues including...

  1. Monte Carlo simulations of the dosimetric impact of radiopaque fiducial markers for proton radiotherapy of the prostate

    Science.gov (United States)

    Newhauser, Wayne; Fontenot, Jonas; Koch, Nicholas; Dong, Lei; Lee, Andrew; Zheng, Yuanshui; Waters, Laurie; Mohan, Radhe

    2007-06-01

    Many clinical studies have demonstrated that implanted radiopaque fiducial markers improve targeting accuracy in external-beam radiotherapy, but little is known about the dose perturbations these markers may cause in patients receiving proton radiotherapy. The objective of this study was to determine what types of implantable markers are visible in setup radiographs and, at the same time, perturb the therapeutic proton dose to the prostate by less than 10%. The radiographic visibility of the markers was assessed by visual inspection of lateral setup radiographs of a pelvic phantom using a kilovoltage x-ray imaging system. The fiducial-induced perturbations in the proton dose were estimated with Monte Carlo simulations. The influence of marker material, size, placement depth and orientation within the pelvis was examined. The radiographic tests confirmed that gold and stainless steel markers were clearly visible and that titanium markers were not. The Monte Carlo simulations revealed that titanium and stainless steel markers minimally perturbed the proton beam, but gold markers cast unacceptably large dose shadows. A 0.9 mm diameter, 3.1 mm long cylindrical stainless steel marker provides good radiographic visibility yet perturbs the proton dose distribution in the prostate by less than 8% when using a parallel opposed lateral beam arrangement.

  2. SU-E-T-241: Monte Carlo Simulation Study About the Prediction of Proton-Induced DNA Strand Breakage On the Double Helix Structure

    Energy Technology Data Exchange (ETDEWEB)

    Shin, J; Park, S; Jeong, J; Jeong, C [National Cancer Center, Goyang, Gyeonggi-do (Korea, Republic of); Lim, Y; Lee, S [National Cancer Center in Korea, Goyang, Gyeonggi-do (Korea, Republic of); SHIN, D [National Cancer Center, Goyangsi, Gyeonggi-do (Korea, Republic of); Incerti, S [Universite Bordeaux 1, CNRS.IN2P3, Centres d’Etudes Nucleaires de Bordeau, Gradignan, Gradignan (France)

    2014-06-01

    Purpose: In particle therapy and radiobiology, the investigation of mechanisms leading to the death of target cancer cells induced by ionising radiation is an active field of research. Recently, several studies based on Monte Carlo simulation codes have been initiated in order to simulate physical interactions of ionising particles at cellular scale and in DNA. Geant4-DNA is the one of them; it is an extension of the general purpose Geant4 Monte Carlo simulation toolkit for the simulation of physical interactions at sub-micrometre scale. In this study, we present Geant4-DNA Monte Carlo simulations for the prediction of DNA strand breakage using a geometrical modelling of DNA structure. Methods: For the simulation of DNA strand breakage, we developed a specific DNA geometrical structure. This structure consists of DNA components, such as the deoxynucleotide pairs, the DNA double helix, the nucleosomes and the chromatin fibre. Each component is made of water because the cross sections models currently available in Geant4-DNA for protons apply to liquid water only. Also, at the macroscopic-scale, protons were generated with various energies available for proton therapy at the National Cancer Center, obtained using validated proton beam simulations developed in previous studies. These multi-scale simulations were combined for the validation of Geant4-DNA in radiobiology. Results: In the double helix structure, the deposited energy in a strand allowed to determine direct DNA damage from physical interaction. In other words, the amount of dose and frequency of damage in microscopic geometries was related to direct radiobiological effect. Conclusion: In this report, we calculated the frequency of DNA strand breakage using Geant4- DNA physics processes for liquid water. This study is now on-going in order to develop geometries which use realistic DNA material, instead of liquid water. This will be tested as soon as cross sections for DNA material become available in Geant4

  3. Prediction of production of {sup 22}Na in a gas-cell target irradiated by protons using Monte Carlo tracking

    Energy Technology Data Exchange (ETDEWEB)

    Eslami, M., E-mail: mohammad.eslami25@yahoo.com [Department of Physics, Faculty of Science, University of Zanjan, Zengan (Zanjan) (Iran, Islamic Republic of); Kakavand, T. [Department of Physics, Faculty of Science, University of Zanjan, Zengan (Zanjan) (Iran, Islamic Republic of); Department of Physics, Faculty of Science, Imam Khomeini International University, Qazvin (Iran, Islamic Republic of); Mirzaii, M.; Rajabifar, S. [Agricultural, Medical and Industrial Research School, Nuclear Science and Technology Research Institute, AEOI, Karaj (Iran, Islamic Republic of)

    2015-01-01

    Highlights: • Angular distribution of the proton beam in a gaseous environment. • Particle energy distribution profile and proton flux within gas-cell target with MCNPX. • Detection of the residual nuclei during the nuclear reactions. • Estimation of production yield for {sup 22,nat}Ne(p,x){sup 22}Na reactions. - Abstract: The {sup 22}Ne(p,n){sup 22}Na is an optimal reaction for the cyclotron production of {sup 22}Na. This work tends to monitor the proton induced production of {sup 22}Na in a gas-cell target, containing natural and enriched neon gas, using Monte Carlo method. The excitation functions of reactions are calculated by both TALYS-1.6 and ALICE/ASH codes and then the optimum energy range of projectile for the high yield production is selected. A free gaseous environment of neon at a particular pressure and temperature is prearranged and the proton beam is transported within it using Monte Carlo codes MCNPX and SRIM. The beam monitoring performed by each of these codes indicates that the gas-cell has to be designed as conical frustum to reach desired interactions. The MCNPX is also employed to calculate the energy distribution of proton in the designed target and estimation of the residual nuclei during irradiation. The production yield of {sup 22}Na in {sup 22}Ne(p,n){sup 22}Na and {sup nat}Ne(p,x){sup 22}Na reactions are estimated and it shows a good agreement with the experimental results. The results demonstrate that Monte Carlo makes available a beneficial manner to design and optimize the gas targets as well as calibration of detectors, which can be used for the radionuclide production purposes.

  4. SU-F-T-217: A Comprehensive Monte-Carlo Study of Out-Of-Field Secondary Neutron Spectra in a Scanned-Beam Proton Therapy Treatment Room

    Energy Technology Data Exchange (ETDEWEB)

    Englbrecht, F; Parodi, K [LMU Munich, Department of Medical Physics, Garching / Munich, Bavaria (Germany); Trinkl, S; Mares, V; Ruehm, W; Wielunski, M [Helmholtz Zentrum Munich, Institute of Radiation Protection, Neuherberg, Bavaria (Germany); Wilkens, J [Technical University of Munich, Department of Physics, Munich, Germany, Garching, Bavaria (Germany); Klinikum rechts der Isar, Department of Radiation Oncology, Munich (Germany); Hillbrand, M [Rinecker Proton Therapy Center, Munich, Bavaria (Germany)

    2016-06-15

    Purpose: To simulate secondary neutron radiation-fields produced at different positions during phantom irradiation inside a scanning proton therapy gantry treatment room. Further, to identify origin, energy distribution and angular emission as function of proton beam energy. Methods: GEANT4 and FLUKA Monte-Carlo codes were used to model the relevant parts of the treatment room in a gantry-equipped pencil beam scanning proton therapy facility including walls, floor, metallic gantry-components, patient table and the homogeneous PMMA target. The proton beams were modeled based on experimental beam ranges in water and spot shapes in air. Neutron energy spectra were simulated at 0°, 45°, 90° and 135° relative to the beam axis at 2m distance from isocenter, as well as 11×11 cm2 fields for 75MeV, 140MeV, 200MeV and for 118MeV with 5cm PMMA range-shifter. The total neutron energy distribution was recorded for these four positions and proton energies. Additionally, the room-components generating secondary neutrons in the room and their contributions to the total spectrum were identified and quantified. Results: FLUKA and GEANT4 simulated neutron spectra showed good general agreement in the whole energy range of 10{sup −}9 to 10{sup 2} MeV. Comparison of measured spectra with the simulated contributions of the various room components helped to limit the complexity of the room model, by identifying the dominant contributions to the secondary neutron spectrum. The iron of the bending magnet and counterweight were identified as sources of secondary evaporation-neutrons, which were lacking in simplified room models. Conclusion: Thorough Monte-Carlo simulations have been performed to complement Bonner-sphere spectrometry measurements of secondary neutrons in a clinical proton therapy treatment room. Such calculations helped disentangling the origin of secondary neutrons and their dominant contributions to measured spectra, besides providing a useful validation of widely

  5. Research on perturbation based Monte Carlo reactor criticality search

    International Nuclear Information System (INIS)

    Li Zeguang; Wang Kan; Li Yangliu; Deng Jingkang

    2013-01-01

    Criticality search is a very important aspect in reactor physics analysis. Due to the advantages of Monte Carlo method and the development of computer technologies, Monte Carlo criticality search is becoming more and more necessary and feasible. Traditional Monte Carlo criticality search method is suffered from large amount of individual criticality runs and uncertainty and fluctuation of Monte Carlo results. A new Monte Carlo criticality search method based on perturbation calculation is put forward in this paper to overcome the disadvantages of traditional method. By using only one criticality run to get initial k eff and differential coefficients of concerned parameter, the polynomial estimator of k eff changing function is solved to get the critical value of concerned parameter. The feasibility of this method was tested. The results show that the accuracy and efficiency of perturbation based criticality search method are quite inspiring and the method overcomes the disadvantages of traditional one. (authors)

  6. SU-E-T-638: Proton Beam Delivery to a Moving Lung Tumor and Monte Carlo Simulation with TOPAS

    Energy Technology Data Exchange (ETDEWEB)

    Coruh, M; Ewell, L [Hampton University, Hampton, Virginia (United States); Demez, N [Hampton University Proton Therapy Institute, Hampton, VA (United States)

    2015-06-15

    Purpose: To estimate the dose delivered to a moving lung tumor by proton therapy beams of different modulation types, and compare with Monte Carlo predictions. Methods: A radiology support devices (RSD) phantom was irradiated with therapeutic proton radiation beams using two different types of modulation: uniform scanning (US) and double scattered (DS). The Eclipse© dose plan was designed to deliver 1.00Gy to the isocenter of a static ∼3×3×3cm (27cc) tumor in the phantom with 100% coverage. The peak to peak amplitude of tumor motion varied from 0.0 to 2.5cm. The radiation dose was measured with an ion-chamber (CC-13) located within the tumor. The time required to deliver the radiation dose varied from an average of 65s for the DS beams to an average of 95s for the US beams. Results: The amount of radiation dose varied from 100% (both US and DS) to the static tumor down to approximately 92% for the moving tumor. The ratio of US dose to DS dose ranged from approximately 1.01 for the static tumor, down to 0.99 for the 2.5cm moving tumor. A Monte Carlo simulation using TOPAS included a lung tumor with 4.0cm of peak to peak motion. In this simulation, the dose received by the tumor varied by ∼40% as the period of this motion varied from 1s to 4s. Conclusion: The radiation dose deposited to a moving tumor was less than for a static tumor, as expected. At large (2.5cm) amplitudes, the DS proton beams gave a dose closer to the desired dose than the US beams, but equal within experimental uncertainty. TOPAS Monte Carlo simulation can give insight into the moving tumor — dose relationship. This work was supported in part by the Philips corporation.

  7. Evaluation of ion chamber dependent correction factors for ionisation chamber dosimetry in proton beams using a Monte Carlo method

    International Nuclear Information System (INIS)

    Palmans, H.; Verhaegen, F.

    1995-01-01

    In the last decade, several clinical proton beam therapy facilities have been developed. To satisfy the demand for uniformity in clinical (routine) proton beam dosimetry two dosimetry protocols (ECHED and AAPM) have been published. Both protocols neglect the influence of ion chamber dependent parameters on dose determination in proton beams because of the scatter properties of these beams, although the problem has not been studied thoroughly yet. A comparison between water calorimetry and ionisation chamber dosimetry showed a discrepancy of 2.6% between the former method and ionometry following the ECHED protocol. Possibly, a small part of this difference can be attributed to chamber dependent correction factors. Indications for this possibility are found in ionometry measurements. To allow the simulation of complex geometries with different media necessary for the study of those corrections, an existing proton Monte Carlo code (PTRAN, Berger) has been modified. The original code, that applies Mollire's multiple scattering theory and Vavilov's energy straggling theory, calculates depth dose profiles, energy distributions and radial distributions for pencil beams in water. Comparisons with measurements and calculations reported in the literature are done to test the program's accuracy. Preliminary results of the influence of chamber design and chamber materials on dose to water determination are presented

  8. Evaluation of ion chamber dependent correction factors for ionisation chamber dosimetry in proton beams using a Monte Carlo method

    Energy Technology Data Exchange (ETDEWEB)

    Palmans, H. [Ghent Univ. (Belgium). Dept. of Biomedical Physics; Verhaegen, F.

    1995-12-01

    In the last decade, several clinical proton beam therapy facilities have been developed. To satisfy the demand for uniformity in clinical (routine) proton beam dosimetry two dosimetry protocols (ECHED and AAPM) have been published. Both protocols neglect the influence of ion chamber dependent parameters on dose determination in proton beams because of the scatter properties of these beams, although the problem has not been studied thoroughly yet. A comparison between water calorimetry and ionisation chamber dosimetry showed a discrepancy of 2.6% between the former method and ionometry following the ECHED protocol. Possibly, a small part of this difference can be attributed to chamber dependent correction factors. Indications for this possibility are found in ionometry measurements. To allow the simulation of complex geometries with different media necessary for the study of those corrections, an existing proton Monte Carlo code (PTRAN, Berger) has been modified. The original code, that applies Mollire`s multiple scattering theory and Vavilov`s energy straggling theory, calculates depth dose profiles, energy distributions and radial distributions for pencil beams in water. Comparisons with measurements and calculations reported in the literature are done to test the program`s accuracy. Preliminary results of the influence of chamber design and chamber materials on dose to water determination are presented.

  9. Neutron H*(10) inside a proton therapy facility: comparison between Monte Carlo simulations and WENDI-2 measurements

    International Nuclear Information System (INIS)

    De Smet, V.; Stichelbaut, F.; Mathot, G.; Vanaudenhove, T.; De Lentdecker, G.; Dubus, A.; Pauly, N.; Gerardy, I.

    2014-01-01

    Inside an IBA proton therapy centre, secondary neutrons are produced due to nuclear interactions of the proton beam with matter mainly inside the cyclotron, the beam line, the treatment nozzle and the patient. Accurate measurements of the neutron ambient dose equivalent H*(10) in such a facility require the use of a detector that has a good sensitivity for neutrons ranging from thermal energies up to 230 MeV, such as for instance the WENDI-2 detector. WENDI-2 measurements have been performed at the Westdeutsches Protonentherapiezentrum Essen, at several positions around the cyclotron room and around a gantry treatment room operated in two different beam delivery modes: Pencil Beam Scanning and Double Scattering. These measurements are compared with Monte Carlo simulation results for the neutron H*(10) obtained with MCNPX 2.5.0 and GEANT4 9.6. In proton therapy, proton beams with energies up to typically 230 MeV are used to treat cancerous tumours very efficiently while sparing surrounding healthy tissues as much as possible. Due to nuclear interactions of the proton beams with matter, mainly inside the cyclotron, the beam line, the treatment nozzle and the patient, secondary neutrons with energies up to 230 MeV are unfortunately produced, as well as photons up to ∼10 MeV. Behind the thick concrete shielding walls which are necessary to attenuate the stray radiation fields, the total ambient dose equivalent H*(10) is very large due to the neutron component. In shielding studies for proton therapy facilities, the neutron H*(10) component is often evaluated using the Monte Carlo codes MCNPX(5), FLUKA(6) or PHITS(7). Recent benchmark simulations performed with GEANT4 have shown that this code would also be a suitable tool for the shielding studies of proton therapy centres. The experimental validation of such shielding studies requires the use of a detector with a good sensitivity for neutrons ranging from thermal energies up to 230 MeV, such as for example the

  10. Dosimetric verification in water of a Monte Carlo treatment planning tool for proton, helium, carbon and oxygen ion beams at the Heidelberg Ion Beam Therapy Center

    Science.gov (United States)

    Tessonnier, T.; Böhlen, T. T.; Ceruti, F.; Ferrari, A.; Sala, P.; Brons, S.; Haberer, T.; Debus, J.; Parodi, K.; Mairani, A.

    2017-08-01

    The introduction of ‘new’ ion species in particle therapy needs to be supported by a thorough assessment of their dosimetric properties and by treatment planning comparisons with clinically used proton and carbon ion beams. In addition to the latter two ions, helium and oxygen ion beams are foreseen at the Heidelberg Ion Beam Therapy Center (HIT) as potential assets for improving clinical outcomes in the near future. We present in this study a dosimetric validation of a FLUKA-based Monte Carlo treatment planning tool (MCTP) for protons, helium, carbon and oxygen ions for spread-out Bragg peaks in water. The comparisons between the ions show the dosimetric advantages of helium and heavier ion beams in terms of their distal and lateral fall-offs with respect to protons, reducing the lateral size of the region receiving 50% of the planned dose up to 12 mm. However, carbon and oxygen ions showed significant doses beyond the target due to the higher fragmentation tail compared to lighter ions (p and He), up to 25%. The Monte Carlo predictions were found to be in excellent geometrical agreement with the measurements, with deviations below 1 mm for all parameters investigated such as target and lateral size as well as distal fall-offs. Measured and simulated absolute dose values agreed within about 2.5% on the overall dose distributions. The MCTP tool, which supports the usage of multiple state-of-the-art relative biological effectiveness models, will provide a solid engine for treatment planning comparisons at HIT.

  11. Dosimetric verification in water of a Monte Carlo treatment planning tool for proton, helium, carbon and oxygen ion beams at the Heidelberg Ion Beam Therapy Center.

    Science.gov (United States)

    Tessonnier, T; Böhlen, T T; Ceruti, F; Ferrari, A; Sala, P; Brons, S; Haberer, T; Debus, J; Parodi, K; Mairani, A

    2017-07-31

    The introduction of 'new' ion species in particle therapy needs to be supported by a thorough assessment of their dosimetric properties and by treatment planning comparisons with clinically used proton and carbon ion beams. In addition to the latter two ions, helium and oxygen ion beams are foreseen at the Heidelberg Ion Beam Therapy Center (HIT) as potential assets for improving clinical outcomes in the near future. We present in this study a dosimetric validation of a FLUKA-based Monte Carlo treatment planning tool (MCTP) for protons, helium, carbon and oxygen ions for spread-out Bragg peaks in water. The comparisons between the ions show the dosimetric advantages of helium and heavier ion beams in terms of their distal and lateral fall-offs with respect to protons, reducing the lateral size of the region receiving 50% of the planned dose up to 12 mm. However, carbon and oxygen ions showed significant doses beyond the target due to the higher fragmentation tail compared to lighter ions (p and He), up to 25%. The Monte Carlo predictions were found to be in excellent geometrical agreement with the measurements, with deviations below 1 mm for all parameters investigated such as target and lateral size as well as distal fall-offs. Measured and simulated absolute dose values agreed within about 2.5% on the overall dose distributions. The MCTP tool, which supports the usage of multiple state-of-the-art relative biological effectiveness models, will provide a solid engine for treatment planning comparisons at HIT.

  12. Accelerated event-by-event Monte Carlo microdosimetric calculations of electrons and protons tracks on a multi-core CPU and a CUDA-enabled GPU.

    Science.gov (United States)

    Kalantzis, Georgios; Tachibana, Hidenobu

    2014-01-01

    For microdosimetric calculations event-by-event Monte Carlo (MC) methods are considered the most accurate. The main shortcoming of those methods is the extensive requirement for computational time. In this work we present an event-by-event MC code of low projectile energy electron and proton tracks for accelerated microdosimetric MC simulations on a graphic processing unit (GPU). Additionally, a hybrid implementation scheme was realized by employing OpenMP and CUDA in such a way that both GPU and multi-core CPU were utilized simultaneously. The two implementation schemes have been tested and compared with the sequential single threaded MC code on the CPU. Performance comparison was established on the speed-up for a set of benchmarking cases of electron and proton tracks. A maximum speedup of 67.2 was achieved for the GPU-based MC code, while a further improvement of the speedup up to 20% was achieved for the hybrid approach. The results indicate the capability of our CPU-GPU implementation for accelerated MC microdosimetric calculations of both electron and proton tracks without loss of accuracy. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  13. Comparison of Monte Carlo simulations with proton experiment for a thick Au absorber

    International Nuclear Information System (INIS)

    Yevseyeva, Olga; Assis, Joaquim T. de; Diaz, Katherin S.; Hormaza, Joel M.; Lopes, Ricardo T.

    2009-01-01

    Proton therapy applications deal with relatively thick targets like the human head or the trunk. Therefore, relatively small differences in the total proton stopping power given, for example, by the different models provided by GEANT4, could lead to significant disagreement in the final proton energy spectra when integrated along lengthy proton trajectories. This work presents a comparison of proton energy spectra for 49.1 MeV protons passing through a couple of Au absorbers with different thicknesses obtained by GEANT4.8.2 simulations using ICRU49, Ziegler1985 and Ziegler2000 models. The comparison was made with the experimental data of Tschalaer, with TRIM/SRIM2008 and MCNPX2.4.0 simulations, and the Payne analytical solution for the transport equation in the Fokker-Plank approximation. It is shown that the simulations reproduce the experimental spectra with some detectable contradictions. It should be noted that all the spectra lay at the proton energies significantly above 2 MeV, i.e. in the so-called 'Bethe-Bloch region'. Therefore the observed disagreements in GEANT4 results, simulated with different models, are somewhat unexpected. Further studies for a better understanding and to obtain definitive conclusions are necessary. (author)

  14. Calculation of primary and secondary dose in proton therapy of brain tumors using Monte Carlo method

    International Nuclear Information System (INIS)

    Moghbel Esfahani, F.; Alamatsaz, M.; Karimian, A.

    2012-01-01

    High-energy beams of protons offer significant advantages for the treatment of deep-seated local tumors. Their physical depth-dose distribution in tissue is characterized by a small entrance dose and a distinct maximum - Bragg peak - near the end of range with a sharp falloff at the distal edge. Therefore, research must be done to investigate the possible negative and positive effects of using proton therapy as a treatment modality. In proton therapy, protons do account for the vast majority of dose. However, when protons travel through matter, secondary particles are created by the interactions of protons and matter en route to and within the patient. It is believed that secondary dose can lead to secondary cancer, especially in pediatric cases. Therefore, the focus of this work is determining both primary and secondary dose. Dose calculations were performed by MCNPX in tumoral and healthy parts of brain. The brain tumor has a 10 mm diameter and is located 16 cm under the skin surface. The brain was simulated by a cylindrical water phantom with the dimensions of 19 x 19cm 2 (length x diameter), with 0.5 cm thickness of plexiglass (C 4 H 6 O 2 ). Then beam characteristics were investigated to ensure the accuracy of the model. Simulations were initially validated with against packages such as SRIM/TRIM. Dose calculations were performed using different configurations to evaluate depth-dose profiles and dose 2D distributions.The results of the simulation show that the best proton energy interval, to cover completely the brain tumor, is from 152 to 154 MeV. (authors)

  15. A combined molecular dynamics and Monte Carlo simulation of the spatial distribution of energy deposition by proton beams in liquid water

    International Nuclear Information System (INIS)

    Garcia-Molina, Rafael; Abril, Isabel; Heredia-Avalos, Santiago; Kyriakou, Ioanna; Emfietzoglou, Dimitris

    2011-01-01

    We have evaluated the spatial distribution of energy deposition by proton beams in liquid water using the simulation code SEICS (Simulation of Energetic Ions and Clusters through Solids), which combines molecular dynamics and Monte Carlo techniques and includes the main interaction phenomena between the projectile and the target constituents: (i) the electronic stopping force due to energy loss to target electronic excitations, including fluctuations due to the energy-loss straggling, (ii) the elastic scattering with the target nuclei, with their corresponding energy loss and (iii) the dynamical changes in projectile charge state due to electronic capture and loss processes. An important feature of SEICS is the accurate account of the excitation spectrum of liquid water, based on a consistent solid-state description of its energy-loss-function over the whole energy and momentum space. We analyse how the above-mentioned interactions affect the depth distribution of the energy delivered in liquid water by proton beams with incident energies of the order of several MeV. Our simulations show that the position of the Bragg peak is determined mainly by the stopping power, whereas its width can be attributed to the energy-loss straggling. Multiple elastic scattering processes contribute slightly only at the distal part of the Bragg peak. The charge state of the projectiles only changes when approaching the end of their trajectories, i.e. near the Bragg peak. We have also simulated the proton-beam energy distribution at several depths in the liquid water target, and found that it is determined mainly by the fluctuation in the energy loss of the projectile, evaluated through the energy-loss straggling. We conclude that a proper description of the target excitation spectrum as well as the inclusion of the energy-loss straggling is essential in the calculation of the proton beam depth-dose distribution.

  16. SU-E-T-591: Measurement and Monte Carlo Simulation of Stray Neutrons in Passive Scattering Proton Therapy: Needs and Challenges

    Energy Technology Data Exchange (ETDEWEB)

    Farah, J; Bonfrate, A; Donadille, L; Dubourg, N; Lacoste, V; Martinetti, F; Sayah, R; Trompier, F; Clairand, I [IRSN - Institute for Radiological Protection and Nuclear Safety, Fontenay-aux-roses (France); Caresana, M [Politecnico di Milano, Milano (Italy); Delacroix, S; Nauraye, C [Institut Curie - Centre de Protontherapie d Orsay, Orsay (France); Herault, J [Centre Antoine Lacassagne, Nice (France); Piau, S; Vabre, I [Institut de Physique Nucleaire d Orsay, Orsay (France)

    2014-06-01

    Purpose: Measure stray radiation inside a passive scattering proton therapy facility, compare values to Monte Carlo (MC) simulations and identify the actual needs and challenges. Methods: Measurements and MC simulations were considered to acknowledge neutron exposure associated with 75 MeV ocular or 180 MeV intracranial passively scattered proton treatments. First, using a specifically-designed high sensitivity Bonner Sphere system, neutron spectra were measured at different positions inside the treatment rooms. Next, measurement-based mapping of neutron ambient dose equivalent was fulfilled using several TEPCs and rem-meters. Finally, photon and neutron organ doses were measured using TLDs, RPLs and PADCs set inside anthropomorphic phantoms (Rando, 1 and 5-years-old CIRS). All measurements were also simulated with MCNPX to investigate the efficiency of MC models in predicting stray neutrons considering different nuclear cross sections and models. Results: Knowledge of the neutron fluence and energy distribution inside a proton therapy room is critical for stray radiation dosimetry. However, as spectrometry unfolding is initiated using a MC guess spectrum and suffers from algorithmic limits a 20% spectrometry uncertainty is expected. H*(10) mapping with TEPCs and rem-meters showed a good agreement between the detectors. Differences within measurement uncertainty (10–15%) were observed and are inherent to the energy, fluence and directional response of each detector. For a typical ocular and intracranial treatment respectively, neutron doses outside the clinical target volume of 0.4 and 11 mGy were measured inside the Rando phantom. Photon doses were 2–10 times lower depending on organs position. High uncertainties (40%) are inherent to TLDs and PADCs measurements due to the need for neutron spectra at detector position. Finally, stray neutrons prediction with MC simulations proved to be extremely dependent on proton beam energy and the used nuclear models and

  17. A Simulation Study for Radiation Treatment Planning Based on the Atomic Physics of the Proton-Boron Fusion Reaction

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Sunmi; Yoon, Do-Kun; Shin, Han-Back; Jung, Joo-Young; Kim, Moo-Sub; Kim, Kyeong-Hyeon; Jang, Hong-Seok; Suh, Tae Suk [the Catholic University of Korea, Seoul (Korea, Republic of)

    2017-03-15

    The purpose of this research is to demonstrate, based on a Monte Carlo simulation code, the procedure of radiation treatment planning for proton-boron fusion therapy (PBFT). A discrete proton beam (60 - 120 MeV) relevant to the Bragg peak was simulated using a Monte Carlo particle extended (MCNPX, Ver. 2.6.0, National Laboratory, Los Alamos NM, USA) simulation code. After computed tomography (CT) scanning of a virtual water phantom including air cavities, the acquired CT images were converted using the simulation source code. We set the boron uptake regions (BURs) in the simulated water phantom to achieve the proton-boron fusion reaction. Proton sources irradiated the BUR, in the phantom. The acquired dose maps were overlapped with the original CT image of the phantom to analyze the dose volume histogram (DVH). We successfully confirmed amplifications of the proton doses (average: 130%) at the target regions. From the DVH result for each simulation, we acquired a relatively accurate dose map for the treatment. A simulation was conducted to characterize the dose distribution and verify the feasibility of proton boron fusion therapy (PBFT). We observed a variation in proton range and developed a tumor targeting technique for treatment that was more accurate and powerful than both conventional proton therapy and boron-neutron capture therapy.

  18. A simulation study for radiation treatment planning based on the atomic physics of the proton-boron fusion reaction

    Science.gov (United States)

    Kim, Sunmi; Yoon, Do-Kun; Shin, Han-Back; Jung, Joo-Young; Kim, Moo-Sub; Kim, Kyeong-Hyeon; Jang, Hong-Seok; Suh, Tae Suk

    2017-03-01

    The purpose of this research is to demonstrate, based on a Monte Carlo simulation code, the procedure of radiation treatment planning for proton-boron fusion therapy (PBFT). A discrete proton beam (60 - 120 MeV) relevant to the Bragg peak was simulated using a Monte Carlo n-particle extended (MCNPX, Ver. 2.6.0, National Laboratory, Los Alamos NM, USA) simulation code. After computed tomography (CT) scanning of a virtual water phantom including air cavities, the acquired CT images were converted using the simulation source code. We set the boron uptake regions (BURs) in the simulated water phantom to achieve the proton-boron fusion reaction. Proton sources irradiated the BUR, in the phantom. The acquired dose maps were overlapped with the original CT image of the phantom to analyze the dose volume histogram (DVH). We successfully confirmed amplifications of the proton doses (average: 130%) at the target regions. From the DVH result for each simulation, we acquired a relatively accurate dose map for the treatment. A simulation was conducted to characterize the dose distribution and verify the feasibility of proton-boron fusion therapy (PBFT). We observed a variation in proton range and developed a tumor-targeting technique for treatment that was more accurate and powerful than both conventional proton therapy and boron-neutron capture therapy.

  19. Range degradation and distal edge behavior of proton radiotherapy beams using 11C activation and Monte Carlo simulation

    Science.gov (United States)

    Elmekawy, Ahmed Farouk

    The distal edge of therapeutic proton radiation beams was investigated by different methods. Proton beams produced at the Hampton University Proton Therapy Institute (HUPTI) were used to irradiate a Polymethylmethacrylate (PMMA) phantom for three different ranges (13.5, 17.0 and 21.0 cm) to investigate the distal slope dependence of the Bragg peak. The activation of 11 C was studied by scanning the phantom less than 10 minutes post-irradiation with a Philips Big Bore Gemini(c) PET/CT. The DICOM images were imported into the Varian Eclipse(c) Treatment Planning System (TPS) for analysis and then analyzed by ImageJ(c) . The distal slope ranged from ?0.1671 +/- 0.0036 to -0.1986 +/- 0.0052 (pixel intensity/slice number) for ranges 13.5 to 21.0 cm respectively. A realistic description of the setup was modeled using the GATE 7.0 Monte Carlo simulation tool and compared to the experiment data. The results show the distal slope ranged from -0.1158+/-0.0133 to -0.0787+/-0.002 (Gy/mm). Additionally, low activity, 11C were simulated to study the 11C reconstructed half-life dependence versus the initial activity for six ranges chosen around the previous activation study. The results of the expected/nominal half-life vs. activity ranged from -5 x 10-4 +/- 2.8104 x 10-4 to 1.6 x 10-3 +/- 9.44 x 10-4 (%diff./Bq). The comparison between two experiments with proton beams on a PMMA phantom and multi-layer ion chamber, and two GATE simulations of a proton beam incident on a water phantom and 11C PET study show that: (i) the distal fall-off variation of the steepness of the slopes are found to be similar thus validating the sensitivity of the PET technique to the range degradation and (ii) the average of the super-ratios difference between all studies observed is primarily due to the difference in the dose deposited in the media.

  20. Continuous energy Monte Carlo method based lattice homogeinzation

    International Nuclear Information System (INIS)

    Li Mancang; Yao Dong; Wang Kan

    2014-01-01

    Based on the Monte Carlo code MCNP, the continuous energy Monte Carlo multi-group constants generation code MCMC has been developed. The track length scheme has been used as the foundation of cross section generation. The scattering matrix and Legendre components require special techniques, and the scattering event method has been proposed to solve this problem. Three methods have been developed to calculate the diffusion coefficients for diffusion reactor core codes and the Legendre method has been applied in MCMC. To the satisfaction of the equivalence theory, the general equivalence theory (GET) and the superhomogenization method (SPH) have been applied to the Monte Carlo method based group constants. The super equivalence method (SPE) has been proposed to improve the equivalence. GET, SPH and SPE have been implemented into MCMC. The numerical results showed that generating the homogenization multi-group constants via Monte Carlo method overcomes the difficulties in geometry and treats energy in continuum, thus provides more accuracy parameters. Besides, the same code and data library can be used for a wide range of applications due to the versatility. The MCMC scheme can be seen as a potential alternative to the widely used deterministic lattice codes. (authors)

  1. Development of Monte Carlo input code for proton, alpha and heavy ion microdosimetric trac structure simulations

    International Nuclear Information System (INIS)

    Douglass, M.; Bezak, E.

    2010-01-01

    Full text: Radiobiology science is important for cancer treatment as it improves our understanding of radiation induced cell death. Monte Carlo simulations playa crucial role in developing improved knowledge of cellular processes. By model Ii ng the cell response to radiation damage and verifying with experimental data, understanding of cell death through direct radiation hits and bystander effects can be obtained. A Monte Carlo input code was developed using 'Geant4' to simulate cellular level radiation interactions. A physics list which enables physically accurate interactions of heavy ions to energies below 100 e V was implemented. A simple biological cell model was also implemented. Each cell consists of three concentric spheres representing the nucleus, cytoplasm and the membrane. This will enable all critical cell death channels to be investigated (i.e. membrane damage, nucleus/DNA). The current simulation has the ability to predict the positions of ionization events within the individual cell components on I micron scale. We have developed a Geant4 simulation for investigation of radiation damage to cells on sub-cellular scale (∼I micron). This code currently allows the positions of the ionisation events within the individual components of the cell enabling a more complete picture of cell death to be developed. The next stage will include expansion of the code to utilise non-regular cell lattice. (author)

  2. Acceleration of Monte Carlo-based scatter compensation for cardiac SPECT

    Energy Technology Data Exchange (ETDEWEB)

    Sohlberg, A; Watabe, H; Iida, H [National Cardiovascular Center Research Institute, 5-7-1 Fujishiro-dai, Suita City, 565-8565 Osaka (Japan)], E-mail: antti.sohlberg@hermesmedical.com

    2008-07-21

    Single proton emission computed tomography (SPECT) images are degraded by photon scatter making scatter compensation essential for accurate reconstruction. Reconstruction-based scatter compensation with Monte Carlo (MC) modelling of scatter shows promise for accurate scatter correction, but it is normally hampered by long computation times. The aim of this work was to accelerate the MC-based scatter compensation using coarse grid and intermittent scatter modelling. The acceleration methods were compared to un-accelerated implementation using MC-simulated projection data of the mathematical cardiac torso (MCAT) phantom modelling {sup 99m}Tc uptake and clinical myocardial perfusion studies. The results showed that when combined the acceleration methods reduced the reconstruction time for 10 ordered subset expectation maximization (OS-EM) iterations from 56 to 11 min without a significant reduction in image quality indicating that the coarse grid and intermittent scatter modelling are suitable for MC-based scatter compensation in cardiac SPECT. (note)

  3. SU-E-T-553: Monte Carlo Calculation of Proton Bragg Peak Displacements in the Presence of Al2O3:C Dosimeters

    Energy Technology Data Exchange (ETDEWEB)

    Young, L; Yang, F [Univ Washington, Seattle, WA (United States)

    2015-06-15

    Purpose: The application of optically stimulated luminescence dosimeters (OSLDs) may be extended to clinical investigations verifying irradiated doses in small animal models. In proton beams, the accurate positioning of the Bragg peak is essential for tumor targeting. The purpose of this study was to estimate the displacement of a pristine Bragg peak when an Al2O3:C nanodot (Landauer, Inc.) is placed on the surface of a water phantom and to evaluate corresponding changes in dose. Methods: Clinical proton pencil beam simulations were carried out with using TOPAS, a Monte Carlo platform layered on top of GEANT4. Point-shaped beams with no energy spread were modeled for energies 100MV, 150MV, 200MV, and 250MV. Dose scoring for 100,000 particle histories was conducted within a water phantom (20cm × 20cm irradiated area, 40cm depth) with its surface placed 214.5cm away from the source. The modeled nanodot had a 4mm radius and 0.2mm thickness. Results: A comparative analysis of Monte Carlo depth dose profiles modeled for these proton pencil beams did not demonstrate an energy dependent in the Bragg peak shift. The shifts in Bragg Peak depth for water phantoms modeled with a nanodot on the phantom surface ranged between 2.7 to 3.2 mm. In all cases, the Bragg Peaks were shifted closer to the irradiation source. The peak dose in phantoms with an OSLD remained unchanged with percent dose differences less than 0.55% when compared to phantom doses without the nanodot. Conclusion: Monte Carlo calculations show that the presence of OSLD nanodots in proton beam therapy will not change the position of a pristine Bragg Peak by more than 3 mm. Although the 3.0 mm shift will not have a detrimental effect in patients receiving proton therapy, this effect may not be negligible in dose verification measurements for mouse models at lower proton beam energies.

  4. Research on Monte Carlo application based on Hadoop

    Directory of Open Access Journals (Sweden)

    Wu Minglei

    2018-01-01

    Full Text Available Monte Carlo method is also known as random simulation method. The more the number of experiments, the more accurate the results obtained. Therefore, a large number of random simulation is required in order to obtain a higher degree of accuracy, but the traditional stand-alone algorithm has been difficult to meet the needs of a large number of simulation. Hadoop platform is a distributed computing platform built on a large data background and an open source software under Apache. It is easier to write and run applications for processing massive amounts of data as an open source software platform. Therefore, this paper takes π value calculation as an example to realize the Monte Carlo algorithm based on Hadoop platform, and get the exact π value with the advantage of Hadoop platform in distributed processing.

  5. Parallel proton transfer pathways in aqueous acid-base reactions

    NARCIS (Netherlands)

    Cox, M.J.; Bakker, H.J.

    2008-01-01

    We study the mechanism of proton transfer (PT) between the photoacid 8-hydroxy-1,3, 6-pyrenetrisulfonic acid (HPTS) and the base chloroacetate in aqueous solution. We investigate both proton and deuteron transfer reactions in solutions with base concentrations ranging from 0.25M to 4M. Using

  6. SU-F-T-157: Physics Considerations Regarding Dosimetric Accuracy of Analytical Dose Calculations for Small Field Proton Therapy: A Monte Carlo Study

    Energy Technology Data Exchange (ETDEWEB)

    Geng, C [Massachusetts General Hospital, Boston, MA (United States); Nanjing University of Aeronautics and Astronautics, Nanjing (China); Daartz, J; Cheung, K; Bussiere, M; Shih, H; Paganetti, H; Schuemann, J [Massachusetts General Hospital, Boston, MA (United States)

    2016-06-15

    Purpose: To evaluate the accuracy of dose calculations by analytical dose calculation methods (ADC) for small field proton therapy in a gantry based passive scattering facility. Methods: 50 patients with intra-cranial disease were evaluated in the study. Treatment plans followed standard prescription and optimization procedures of proton stereotactic radiosurgery. Dose distributions calculated with the Monte Carlo (MC) toolkit TOPAS were used to represent delivered treatments. The MC dose was first adjusted using the output factor (OF) applied clinically. This factor is determined from the field size and the prescribed range. We then introduced a normalization factor to measure the difference in mean dose between the delivered dose (MC dose with OF) and the dose calculated by ADC for each beam. The normalization was determined by the mean dose of the center voxels of the target area. We compared delivered dose distributions and those calculated by ADC in terms of dose volume histogram parameters and beam range distributions. Results: The mean target dose for a whole treatment is generally within 5% comparing delivered dose (MC dose with OF) and ADC dose. However, the differences can be as great as 11% for shallow and small target treated with a thick range compensator. Applying the normalization factor to the MC dose with OF can reduce the mean dose difference to less than 3%. Considering range uncertainties, the generally applied margins (3.5% of the prescribed range + 1mm) to cover uncertainties in range might not be sufficient to guarantee tumor coverage. The range difference for R90 (90% distal dose falloff) is affected by multiple factors, such as the heterogeneity index. Conclusion: This study indicates insufficient accuracy calculating proton doses using ADC. Our results suggest that uncertainties of target doses are reduced using MC techniques, improving the dosimetric accuracy for proton stereotactic radiosurgery. The work was supported by NIH/NCI under CA

  7. Measurement of LET (linear energy transfer) spectra using CR-39 at different depths of water irradiated by 171 MeV protons: A comparison with Monte Carlo simulation

    Energy Technology Data Exchange (ETDEWEB)

    Sahoo, G.S. [Health Physics Division, Bhabha Atomic Research Centre, Mumbai 400085 (India); Tripathy, S.P., E-mail: sam.tripathy@gmail.com [Health Physics Division, Bhabha Atomic Research Centre, Mumbai 400085 (India); Homi Bhabha National Institute, Mumbai 400094 (India); Molokanov, A.G.; Aleynikov, V.E. [Joint Institute for Nuclear Research, Dubna 141980 (Russian Federation); Sharma, S.D. [Homi Bhabha National Institute, Mumbai 400094 (India); Radiological Physics & Advisory Division, Bhabha Atomic Research Centre, Mumbai 400085 (India); Bandyopadhyay, T. [Health Physics Division, Bhabha Atomic Research Centre, Mumbai 400085 (India); Homi Bhabha National Institute, Mumbai 400094 (India)

    2016-05-11

    In this work, we have used CR-39 detectors to estimate the LET (linear energy transfer) spectrum of secondary particles due to 171 MeV proton beam at different depths of water including the Bragg peak region. The measured LET spectra were compared with those obtained from FLUKA Monte Carlo simulation. The absorbed dose (D{sub LET}), dose equivalent (H{sub LET}) were estimated using the LET spectra. The values of D{sub LET} and H{sub LET} per incident proton fluence were found to increase with the increase in depth of water and were maximum at Bragg peak. - Highlights: • Measurement of LET spectrometry using CR-39 detectors at different depths of water. • Comparison of measured spectra with FLUKA Monte carlo simulation. • Absorbed dose and dose equivalent was found to increase with depth of water.

  8. SU-E-T-794: Validation of the MCNPX Monte Carlo Model of the Ocular Beam Line at the Proton Therapy Center Houston

    Energy Technology Data Exchange (ETDEWEB)

    Titt, U; Suzuki, K [The Univ. of Texas MD Anderson Cancer Center, Houston, TX (United States)

    2015-06-15

    Purpose: The PTCH is preparing the ocular proton beam nozzle for clinical use. Currently commissioning measurements are being performed using films, diodes and ionization chambers. In parallel, a Monte Carlo model of the beam line was created for integration into the automated Monte Carlo treatment plan computation system, MC{sup 2}. This work aims to compare Monte Carlo predictions to measured proton doses in order to validate the Monte Carlo model. Methods: A complete model of the double scattering ocular beam line has been created and is capable of simulating proton beams with a comprehensive set of beam modifying devices, including eleven different range modulator wheels. Simulations of doses in water were scored and compare to ion chamber measurements of depth doses, lateral dose profiles extracted from half beam block exposures of films, and diode measurements of lateral penumbrae at various depths. Results: All comparison resulted in an average relative entrance dose difference of less than 3% and peak dose difference of less than 2%. All range differences were smaller than 0.2 mm. The differences in the lateral beam profiles were smaller than 0.2 mm, and the differences in the penumbrae were all smaller than 0.4%. Conclusion: All available data shows excellent agreement of simulations and measurements. More measurements will have to be performed in order to completely and systematically validate the model. Besides simulating and measuring PDDs and lateral profiles of all remaining range modulator wheels, the absolute dosimetry factors in terms of number of source protons per monitor unit have to be determined.

  9. Proton conduction based on intracrystalline chemical reaction

    International Nuclear Information System (INIS)

    Schuck, G.; Lechner, R.E.; Langer, K.

    2002-01-01

    Proton conductivity in M 3 H(SeO 4 ) 2 crystals (M=K, Rb, Cs) is shown to be due to a dynamic disorder in the form of an intracrystalline chemical equilibrium reaction: alternation between the association of the monomers [HSeO 4 ] 1- and [SeO 4 ] 2- resulting in the dimer [H(SeO 4 ) 2 ] 3- (H-bond formation) and the dissociation of the latter into the two monomers (H-bond breaking). By a combination of quasielastic neutron scattering and FTIR spectroscopy, reaction rates were obtained, as well as rates of proton exchange between selenate ions, leading to diffusion. The results demonstrate that this reaction plays a central role in the mechanism of proton transport in these solid-state protonic conductors. (orig.)

  10. The optimal balance between quality and efficiency in proton radiography imaging technique at various proton beam energies : A Monte Carlo study

    NARCIS (Netherlands)

    Biegun, A K; van Goethem, M-J; van der Graaf, E R; van Beuzekom, M; Koffeman, E N; Nakaji, T; Takatsu, J; Visser, J; Brandenburg, S

    Proton radiography is a novel imaging modality that allows direct measurement of the proton energy loss in various tissues. Currently, due to the conversion of so-called Hounsfield units from X-ray Computed Tomography (CT) into relative proton stopping powers (RPSP), the uncertainties of RPSP are

  11. Determination of protonation sites in bases from topological rules

    Science.gov (United States)

    Fuster, Franck; Silvi, Bernard

    2000-02-01

    The topological analysis of the electron localization function ELF has been carried out on a series of neutral and protonated bases revealing that the determination of the most favorable protonation site can be done with the help of a simple rule. The protonation occurs in the most populated valence basin of the base which yields the least topological change of the localization gradient field. This can be viewed as resulting of a competition between the Pauli repulsion and the stabilization brought by the proton. Structural comparisons of hydrogen bonded complexes and protonated bases are consistent with a picture in which the formation of the complex is the initial step of the proton transfer reaction B+HX→BH ++X -. The angular structure of the complex is such as the proton transfer reaction satisfies the least motion principle of Rice and Teller. This description brings support to the model of Legon and Millen which generalizes the VSEPR concepts to hydrogen bonded complexes B···HX in assuming that the hydrogen bond is directed towards the nucleophilic center of the basis, in other words towards the protonation site.

  12. Proton therapy treatment monitoring with the DoPET system: activity range, positron emitters evaluation and comparison with Monte Carlo predictions

    Science.gov (United States)

    Muraro, S.; Battistoni, G.; Belcari, N.; Bisogni, M. G.; Camarlinghi, N.; Cristoforetti, L.; Del Guerra, A.; Ferrari, A.; Fracchiolla, F.; Morrocchi, M.; Righetto, R.; Sala, P.; Schwarz, M.; Sportelli, G.; Topi, A.; Rosso, V.

    2017-12-01

    Ion beam irradiations can deliver conformal dose distributions minimizing damage to healthy tissues thanks to their characteristic dose profiles. Nevertheless, the location of the Bragg peak can be affected by different sources of range uncertainties: a critical issue is the treatment verification. During the treatment delivery, nuclear interactions between the ions and the irradiated tissues generate β+ emitters: the detection of this activity signal can be used to perform the treatment monitoring if an expected activity distribution is available for comparison. Monte Carlo (MC) codes are widely used in the particle therapy community to evaluate the radiation transport and interaction with matter. In this work, FLUKA MC code was used to simulate the experimental conditions of irradiations performed at the Proton Therapy Center in Trento (IT). Several mono-energetic pencil beams were delivered on phantoms mimicking human tissues. The activity signals were acquired with a PET system (DoPET) based on two planar heads, and designed to be installed along the beam line to acquire data also during the irradiation. Different acquisitions are analyzed and compared with the MC predictions, with a special focus on validating the PET detectors response for activity range verification.

  13. Efficient sampling algorithms for Monte Carlo based treatment planning

    International Nuclear Information System (INIS)

    DeMarco, J.J.; Solberg, T.D.; Chetty, I.; Smathers, J.B.

    1998-01-01

    Efficient sampling algorithms are necessary for producing a fast Monte Carlo based treatment planning code. This study evaluates several aspects of a photon-based tracking scheme and the effect of optimal sampling algorithms on the efficiency of the code. Four areas were tested: pseudo-random number generation, generalized sampling of a discrete distribution, sampling from the exponential distribution, and delta scattering as applied to photon transport through a heterogeneous simulation geometry. Generalized sampling of a discrete distribution using the cutpoint method can produce speedup gains of one order of magnitude versus conventional sequential sampling. Photon transport modifications based upon the delta scattering method were implemented and compared with a conventional boundary and collision checking algorithm. The delta scattering algorithm is faster by a factor of six versus the conventional algorithm for a boundary size of 5 mm within a heterogeneous geometry. A comparison of portable pseudo-random number algorithms and exponential sampling techniques is also discussed

  14. SU-E-T-521: Investigation of the Uncertainties Involved in Secondary Neutron/gamma Production in Geant4/MCNP6 Monte Carlo Codes for Proton Therapy Application

    International Nuclear Information System (INIS)

    Mirzakhanian, L; Enger, S; Giusti, V

    2015-01-01

    Purpose: A major concern in proton therapy is the production of secondary neutrons causing secondary cancers, especially in young adults and children. Most utilized Monte Carlo codes in proton therapy are Geant4 and MCNP. However, the default versions of Geant4 and MCNP6 do not have suitable cross sections or physical models to properly handle secondary particle production in proton energy ranges used for therapy. In this study, default versions of Geant4 and MCNP6 were modified to better handle production of secondaries by adding the TENDL-2012 cross-section library. Methods: In-water proton depth-dose was measured at the “The Svedberg Laboratory” in Uppsala (Sweden). The proton beam was mono-energetic with mean energy of 178.25±0.2 MeV. The measurement set-up was simulated by Geant4 version 10.00 (default and modified version) and MCNP6. Proton depth-dose, primary and secondary particle fluence and neutron equivalent dose were calculated. In case of Geant4, the secondary particle fluence was filtered by all the physics processes to identify the main process responsible for the difference between the default and modified version. Results: The proton depth-dose curves and primary proton fluence show a good agreement between both Geant4 versions and MCNP6. With respect to the modified version, default Geant4 underestimates the production of secondary neutrons while overestimates that of gammas. The “ProtonInElastic” process was identified as the main responsible process for the difference between the two versions. MCNP6 shows higher neutron production and lower gamma production than both Geant4 versions. Conclusion: Despite the good agreement on the proton depth dose curve and primary proton fluence, there is a significant discrepancy on secondary neutron production between MCNP6 and both versions of Geant4. Further studies are thus in order to find the possible cause of this discrepancy or more accurate cross-sections/models to handle the nuclear

  15. Monte-Carlo Modeling of Parameters of a Subcritical Cascade Reactor Based on MSBR and LMFBR Technologies

    CERN Document Server

    Bznuni, S A; Zhamkochyan, V M; Polanski, A; Sosnin, A N; Khudaverdyan, A H

    2001-01-01

    Parameters of a subcritical cascade reactor driven by a proton accelerator and based on a primary lead-bismuth target, main reactor constructed analogously to the molten salt breeder (MSBR) reactor core and a booster-reactor analogous to the core of the BN-350 liquid metal cooled fast breeder reactor (LMFBR). It is shown by means of Monte-Carlo modeling that the reactor under study provides safe operation modes (k_{eff}=0.94-0.98), is apable to transmute effectively radioactive nuclear waste and reduces by an order of magnitude the requirements on the accelerator beam current. Calculations show that the maximal neutron flux in the thermal zone is 10^{14} cm^{12}\\cdot s^_{-1}, in the fast booster zone is 5.12\\cdot10^{15} cm^{12}\\cdot s{-1} at k_{eff}=0.98 and proton beam current I=2.1 mA.

  16. Dosimetric evaluation of a commercial proton spot scanning Monte-Carlo dose algorithm: comparisons against measurements and simulations

    Science.gov (United States)

    Saini, Jatinder; Maes, Dominic; Egan, Alexander; Bowen, Stephen R.; St. James, Sara; Janson, Martin; Wong, Tony; Bloch, Charles

    2017-10-01

    RaySearch Americas Inc. (NY) has introduced a commercial Monte Carlo dose algorithm (RS-MC) for routine clinical use in proton spot scanning. In this report, we provide a validation of this algorithm against phantom measurements and simulations in the GATE software package. We also compared the performance of the RayStation analytical algorithm (RS-PBA) against the RS-MC algorithm. A beam model (G-MC) for a spot scanning gantry at our proton center was implemented in the GATE software package. The model was validated against measurements in a water phantom and was used for benchmarking the RS-MC. Validation of the RS-MC was performed in a water phantom by measuring depth doses and profiles for three spread-out Bragg peak (SOBP) beams with normal incidence, an SOBP with oblique incidence, and an SOBP with a range shifter and large air gap. The RS-MC was also validated against measurements and simulations in heterogeneous phantoms created by placing lung or bone slabs in a water phantom. Lateral dose profiles near the distal end of the beam were measured with a microDiamond detector and compared to the G-MC simulations, RS-MC and RS-PBA. Finally, the RS-MC and RS-PBA were validated against measured dose distributions in an Alderson-Rando (AR) phantom. Measurements were made using Gafchromic film in the AR phantom and compared to doses using the RS-PBA and RS-MC algorithms. For SOBP depth doses in a water phantom, all three algorithms matched the measurements to within  ±3% at all points and a range within 1 mm. The RS-PBA algorithm showed up to a 10% difference in dose at the entrance for the beam with a range shifter and  >30 cm air gap, while the RS-MC and G-MC were always within 3% of the measurement. For an oblique beam incident at 45°, the RS-PBA algorithm showed up to 6% local dose differences and broadening of distal fall-off by 5 mm. Both the RS-MC and G-MC accurately predicted the depth dose to within  ±3% and distal fall-off to within 2

  17. GPU-Monte Carlo based fast IMRT plan optimization

    Directory of Open Access Journals (Sweden)

    Yongbao Li

    2014-03-01

    Full Text Available Purpose: Intensity-modulated radiation treatment (IMRT plan optimization needs pre-calculated beamlet dose distribution. Pencil-beam or superposition/convolution type algorithms are typically used because of high computation speed. However, inaccurate beamlet dose distributions, particularly in cases with high levels of inhomogeneity, may mislead optimization, hindering the resulting plan quality. It is desire to use Monte Carlo (MC methods for beamlet dose calculations. Yet, the long computational time from repeated dose calculations for a number of beamlets prevents this application. It is our objective to integrate a GPU-based MC dose engine in lung IMRT optimization using a novel two-steps workflow.Methods: A GPU-based MC code gDPM is used. Each particle is tagged with an index of a beamlet where the source particle is from. Deposit dose are stored separately for beamlets based on the index. Due to limited GPU memory size, a pyramid space is allocated for each beamlet, and dose outside the space is neglected. A two-steps optimization workflow is proposed for fast MC-based optimization. At first step, a rough dose calculation is conducted with only a few number of particle per beamlet. Plan optimization is followed to get an approximated fluence map. In the second step, more accurate beamlet doses are calculated, where sampled number of particles for a beamlet is proportional to the intensity determined previously. A second-round optimization is conducted, yielding the final result.Results: For a lung case with 5317 beamlets, 105 particles per beamlet in the first round, and 108 particles per beam in the second round are enough to get a good plan quality. The total simulation time is 96.4 sec.Conclusion: A fast GPU-based MC dose calculation method along with a novel two-step optimization workflow are developed. The high efficiency allows the use of MC for IMRT optimizations.--------------------------------Cite this article as: Li Y, Tian Z

  18. Secondary Neutron Doses to Pediatric Patients During Intracranial Proton Therapy: Monte Carlo Simulation of the Neutron Energy Spectrum and its Organ Doses.

    Science.gov (United States)

    Matsumoto, Shinnosuke; Koba, Yusuke; Kohno, Ryosuke; Lee, Choonsik; Bolch, Wesley E; Kai, Michiaki

    2016-04-01

    Proton therapy has the physical advantage of a Bragg peak that can provide a better dose distribution than conventional x-ray therapy. However, radiation exposure of normal tissues cannot be ignored because it is likely to increase the risk of secondary cancer. Evaluating secondary neutrons generated by the interaction of the proton beam with the treatment beam-line structure is necessary; thus, performing the optimization of radiation protection in proton therapy is required. In this research, the organ dose and energy spectrum were calculated from secondary neutrons using Monte Carlo simulations. The Monte Carlo code known as the Particle and Heavy Ion Transport code System (PHITS) was used to simulate the transport proton and its interaction with the treatment beam-line structure that modeled the double scattering body of the treatment nozzle at the National Cancer Center Hospital East. The doses of the organs in a hybrid computational phantom simulating a 5-y-old boy were calculated. In general, secondary neutron doses were found to decrease with increasing distance to the treatment field. Secondary neutron energy spectra were characterized by incident neutrons with three energy peaks: 1×10, 1, and 100 MeV. A block collimator and a patient collimator contributed significantly to organ doses. In particular, the secondary neutrons from the patient collimator were 30 times higher than those from the first scatter. These results suggested that proactive protection will be required in the design of the treatment beam-line structures and that organ doses from secondary neutrons may be able to be reduced.

  19. Markov chain Monte Carlo sampling based terahertz holography image denoising.

    Science.gov (United States)

    Chen, Guanghao; Li, Qi

    2015-05-10

    Terahertz digital holography has attracted much attention in recent years. This technology combines the strong transmittance of terahertz and the unique features of digital holography. Nonetheless, the low clearness of the images captured has hampered the popularization of this imaging technique. In this paper, we perform a digital image denoising technique on our multiframe superposed images. The noise suppression model is concluded as Bayesian least squares estimation and is solved with Markov chain Monte Carlo (MCMC) sampling. In this algorithm, a weighted mean filter with a Gaussian kernel is first applied to the noisy image, and then by nonlinear contrast transform, the contrast of the image is restored to the former level. By randomly walking on the preprocessed image, the MCMC-based filter keeps collecting samples, assigning them weights by similarity assessment, and constructs multiple sample sequences. Finally, these sequences are used to estimate the value of each pixel. Our algorithm shares some good qualities with nonlocal means filtering and the algorithm based on conditional sampling proposed by Wong et al. [Opt. Express18, 8338 (2010)10.1364/OE.18.008338OPEXFF1094-4087], such as good uniformity, and, moreover, reveals better performance in structure preservation, as shown in numerical comparison using the structural similarity index measurement and the peak signal-to-noise ratio.

  20. Anhydrous proton conductor based on composites of PEO and ATMP

    Energy Technology Data Exchange (ETDEWEB)

    Sun Baoying [Key Lab of Organic Optoelectronics and Molecular Engineering, Department of Chemistry, Tsinghua University, Beijing 100084 (China); Qiu Xinping, E-mail: qiuxp@tsinghua.edu.c [Key Lab of Organic Optoelectronics and Molecular Engineering, Department of Chemistry, Tsinghua University, Beijing 100084 (China); Zhu Wentao [Key Lab of Organic Optoelectronics and Molecular Engineering, Department of Chemistry, Tsinghua University, Beijing 100084 (China)

    2011-04-15

    A new type anhydrous PEM material based on Poly (ethylene oxide) (PEO)/Amino Trimethylene Phosphonic Acid (ATMP) composite was prepared. In this study, PEO assumed to 'grab' protons via hydrogen bond between PEO and ATMP. Based on this point, the PEO/ATMP composites were prepared firstly as the preliminary study to verify this proton conducting system. Then, PVDF was added to enhance the membrane's stability. The PVDF/PEO/ATMP composite membranes were thermally stable up to 200 {sup o}C in the studied composition ranges. The membrane had relatively compact structure by SEM images. Proton conductivity of 59% PVDF/29% PEO/12% ATMP was up to 6.71 x 10 {sup -3} S cm{sup -1} at 86 {sup o}C after doping with 7.9 wt% phosphoric acid without extra humidification.

  1. Anhydrous proton conductor based on composites of PEO and ATMP

    International Nuclear Information System (INIS)

    Sun Baoying; Qiu Xinping; Zhu Wentao

    2011-01-01

    A new type anhydrous PEM material based on Poly (ethylene oxide) (PEO)/Amino Trimethylene Phosphonic Acid (ATMP) composite was prepared. In this study, PEO assumed to 'grab' protons via hydrogen bond between PEO and ATMP. Based on this point, the PEO/ATMP composites were prepared firstly as the preliminary study to verify this proton conducting system. Then, PVDF was added to enhance the membrane's stability. The PVDF/PEO/ATMP composite membranes were thermally stable up to 200 o C in the studied composition ranges. The membrane had relatively compact structure by SEM images. Proton conductivity of 59% PVDF/29% PEO/12% ATMP was up to 6.71 x 10 -3 S cm -1 at 86 o C after doping with 7.9 wt% phosphoric acid without extra humidification.

  2. Image based Monte Carlo modeling for computational phantom

    International Nuclear Information System (INIS)

    Cheng, M.; Wang, W.; Zhao, K.; Fan, Y.; Long, P.; Wu, Y.

    2013-01-01

    Full text of the publication follows. The evaluation on the effects of ionizing radiation and the risk of radiation exposure on human body has been becoming one of the most important issues for radiation protection and radiotherapy fields, which is helpful to avoid unnecessary radiation and decrease harm to human body. In order to accurately evaluate the dose on human body, it is necessary to construct more realistic computational phantom. However, manual description and verification of the models for Monte Carlo (MC) simulation are very tedious, error-prone and time-consuming. In addition, it is difficult to locate and fix the geometry error, and difficult to describe material information and assign it to cells. MCAM (CAD/Image-based Automatic Modeling Program for Neutronics and Radiation Transport Simulation) was developed as an interface program to achieve both CAD- and image-based automatic modeling. The advanced version (Version 6) of MCAM can achieve automatic conversion from CT/segmented sectioned images to computational phantoms such as MCNP models. Imaged-based automatic modeling program(MCAM6.0) has been tested by several medical images and sectioned images. And it has been applied in the construction of Rad-HUMAN. Following manual segmentation and 3D reconstruction, a whole-body computational phantom of Chinese adult female called Rad-HUMAN was created by using MCAM6.0 from sectioned images of a Chinese visible human dataset. Rad-HUMAN contains 46 organs/tissues, which faithfully represented the average anatomical characteristics of the Chinese female. The dose conversion coefficients (Dt/Ka) from kerma free-in-air to absorbed dose of Rad-HUMAN were calculated. Rad-HUMAN can be applied to predict and evaluate dose distributions in the Treatment Plan System (TPS), as well as radiation exposure for human body in radiation protection. (authors)

  3. Image based Monte Carlo Modeling for Computational Phantom

    Science.gov (United States)

    Cheng, Mengyun; Wang, Wen; Zhao, Kai; Fan, Yanchang; Long, Pengcheng; Wu, Yican

    2014-06-01

    The evaluation on the effects of ionizing radiation and the risk of radiation exposure on human body has been becoming one of the most important issues for radiation protection and radiotherapy fields, which is helpful to avoid unnecessary radiation and decrease harm to human body. In order to accurately evaluate the dose on human body, it is necessary to construct more realistic computational phantom. However, manual description and verfication of the models for Monte carlo(MC)simulation are very tedious, error-prone and time-consuming. In addiation, it is difficult to locate and fix the geometry error, and difficult to describe material information and assign it to cells. MCAM (CAD/Image-based Automatic Modeling Program for Neutronics and Radiation Transport Simulation) was developed as an interface program to achieve both CAD- and image-based automatic modeling by FDS Team (Advanced Nuclear Energy Research Team, http://www.fds.org.cn). The advanced version (Version 6) of MCAM can achieve automatic conversion from CT/segmented sectioned images to computational phantoms such as MCNP models. Imaged-based automatic modeling program(MCAM6.0) has been tested by several medical images and sectioned images. And it has been applied in the construction of Rad-HUMAN. Following manual segmentation and 3D reconstruction, a whole-body computational phantom of Chinese adult female called Rad-HUMAN was created by using MCAM6.0 from sectioned images of a Chinese visible human dataset. Rad-HUMAN contains 46 organs/tissues, which faithfully represented the average anatomical characteristics of the Chinese female. The dose conversion coefficients(Dt/Ka) from kerma free-in-air to absorbed dose of Rad-HUMAN were calculated. Rad-HUMAN can be applied to predict and evaluate dose distributions in the Treatment Plan System (TPS), as well as radiation exposure for human body in radiation protection.

  4. Monte Carlo study of radial energy deposition from primary and secondary particles for narrow and large proton beamlet source models

    International Nuclear Information System (INIS)

    Peeler, Christopher R; Titt, Uwe

    2012-01-01

    In spot-scanning intensity-modulated proton therapy, numerous unmodulated proton beam spots are delivered over a target volume to produce a prescribed dose distribution. To accurately model field size-dependent output factors for beam spots, the energy deposition at positions radial to the central axis of the beam must be characterized. In this study, we determined the difference in the central axis dose for spot-scanned fields that results from secondary particle doses by investigating energy deposition radial to the proton beam central axis resulting from primary protons and secondary particles for mathematical point source and distributed source models. The largest difference in the central axis dose from secondary particles resulting from the use of a mathematical point source and a distributed source model was approximately 0.43%. Thus, we conclude that the central axis dose for a spot-scanned field is effectively independent of the source model used to calculate the secondary particle dose. (paper)

  5. A critical study of different Monte Carlo scoring methods of dose average linear-energy-transfer maps calculated in voxelized geometries irradiated with clinical proton beams.

    Science.gov (United States)

    Cortés-Giraldo, M A; Carabe, A

    2015-04-07

    We compare unrestricted dose average linear energy transfer (LET) maps calculated with three different Monte Carlo scoring methods in voxelized geometries irradiated with proton therapy beams with three different Monte Carlo scoring methods. Simulations were done with the Geant4 (Geometry ANd Tracking) toolkit. The first method corresponds to a step-by-step computation of LET which has been reported previously in the literature. We found that this scoring strategy is influenced by spurious high LET components, which relative contribution in the dose average LET calculations significantly increases as the voxel size becomes smaller. Dose average LET values calculated for primary protons in water with voxel size of 0.2 mm were a factor ~1.8 higher than those obtained with a size of 2.0 mm at the plateau region for a 160 MeV beam. Such high LET components are a consequence of proton steps in which the condensed-history algorithm determines an energy transfer to an electron of the material close to the maximum value, while the step length remains limited due to voxel boundary crossing. Two alternative methods were derived to overcome this problem. The second scores LET along the entire path described by each proton within the voxel. The third followed the same approach of the first method, but the LET was evaluated at each step from stopping power tables according to the proton kinetic energy value. We carried out microdosimetry calculations with the aim of deriving reference dose average LET values from microdosimetric quantities. Significant differences between the methods were reported either with pristine or spread-out Bragg peaks (SOBPs). The first method reported values systematically higher than the other two at depths proximal to SOBP by about 15% for a 5.9 cm wide SOBP and about 30% for a 11.0 cm one. At distal SOBP, the second method gave values about 15% lower than the others. Overall, we found that the third method gave the most consistent

  6. Implementation and verification of nuclear interactions in a Monte-Carlo code for the Procom-ProGam proton therapy planning system

    International Nuclear Information System (INIS)

    Kostyuchenko, V.I.; Makarova, A.S.; Ryazantsev, O.B.; Samarin, S.I.; Uglov, A.S.

    2013-01-01

    Proton interaction with an exposed object material needs to be modeled with account for three basic processes: electromagnetic stopping of protons in matter, multiple coulomb scattering and nuclear interactions. Just the last type of processes is the topic of this paper. Monte Carlo codes are often used to simulate high-energy particle interaction with matter. However, nuclear interaction models implemented in these codes are rather extensive and their use in treatment planning systems requires huge computational resources. We have selected the IThMC code for its ability to reproduce experiments which measure the distribution of the projected ranges of nuclear secondary particles generated by proton beams in a multi-layer Faraday cup. The multi-layer Faraday cup detectors measure charge rather than dose and allow distinguishing between electromagnetic and nuclear interactions. The event generator used in the IThMC code is faster, but less accurate than any other used in testing. Our model of nuclear reactions demonstrates quite good agreement with experiment in the context of their effect on the Bragg peak in therapeutic applications

  7. Proton Conductivity and Operational Features Of PBI-Based Membranes

    DEFF Research Database (Denmark)

    Qingfeng, Li; Jensen, Jens Oluf; Precht Noyé, Pernille

    2005-01-01

    As an approach to high temperature operation of PEMFCs, acid-doped PBI membranes are under active development. The membrane exhibits high proton conductivity under low water contents at temperatures up to 200°C. Mechanisms of proton conduction for the membranes have been proposed. Based on the me...... on the membranes fuel cell tests have been demonstrated. Operating features of the PBI cell include no humidification, high CO tolerance, better heat utilization and possible integration with fuel processing units. Issues for further development are also discussed....

  8. Proton-beam writing channel based on an electrostatic accelerator

    Science.gov (United States)

    Lapin, A. S.; Rebrov, V. A.; Kolin'ko, S. V.; Salivon, V. F.; Ponomarev, A. G.

    2016-09-01

    We have described the structure of the proton-beam writing channel as a continuation of a nuclear scanning microprobe channel. The problem of the accuracy of positioning a probe by constructing a new high-frequency electrostatic scanning system has been solved. Special attention has been paid to designing the probe-forming system and its various configurations have been considered. The probe-forming system that best corresponds to the conditions of the lithographic process has been found based on solving the problem of optimizing proton beam formation. A system for controlling beam scanning using multifunctional module of integrated programmable logic systems has been developed.

  9. DSC and conductivity studies on PVA based proton conducting gel ...

    Indian Academy of Sciences (India)

    Unknown

    An attempt has been made in the present work to prepare polyvinyl alcohol (PVA) based proton conducting gel electrolytes in ammonium thiocyanate (NH4SCN) solution and characterize ... dependence of ionic conductivity exhibits VTF behaviour. Keywords. Polymer gel electrolytes; ionic conductivity; solvent free polymer ...

  10. A Monte Carlo study for the calculation of the average linear energy transfer (LET) distributions for a clinical proton beam line and a radiobiological carbon ion beam line.

    Science.gov (United States)

    Romano, F; Cirrone, G A P; Cuttone, G; Rosa, F Di; Mazzaglia, S E; Petrovic, I; Fira, A Ristic; Varisano, A

    2014-06-21

    Fluence, depth absorbed dose and linear energy transfer (LET) distributions of proton and carbon ion beams have been investigated using the Monte Carlo code Geant4 (GEometry ANd Tracking). An open source application was developed with the aim to simulate two typical transport beam lines, one used for ocular therapy and cell irradiations with protons and the other for cell irradiations with carbon ions. This tool allows evaluation of the primary and total dose averaged LET and predict their spatial distribution in voxelized or sliced geometries. In order to reproduce the LET distributions in a realistic way, and also the secondary particles' contributions due to nuclear interactions were considered in the computations. Pristine and spread-out Bragg peaks were taken into account both for proton and carbon ion beams, with the maximum energy of 62 MeV/n. Depth dose distributions were compared with experimental data, showing good agreement. Primary and total LET distributions were analysed in order to study the influence of contributions of secondary particles in regions at different depths. A non-negligible influence of high-LET components was found in the entrance channel for proton beams, determining the total dose averaged LET by the factor 3 higher than the primary one. A completely different situation was obtained for carbon ions. In this case, secondary particles mainly contributed in the tail that is after the peak. The results showed how the weight of light and heavy secondary ions can considerably influence the computation of LET depth distributions. This has an important role in the interpretation of results coming from radiobiological experiments and, therefore, in hadron treatment planning procedures.

  11. Neutron production in spallation reactions of 0.9 and 1.5 GeV protons on a thick lead target. Comparison between experimental data and Monte-Carlo simulations

    International Nuclear Information System (INIS)

    Krasa, A.; Krizek, F.; Wagner, V.; Kugler, A.; Henzl, V.; Henzlova, D.; Majerle, M.; Adam, J.; Caloun, P.; Bradnova, V.; Chultem, D.; Kalinnikov, V.G.; Krivopustov, M.I.; Solnyshkin, A.A.; Stegajlov, V.I.; Tsupko-Sitnikov, V.M.; Tumehndehlgehr, Ts.; Vasil'ev, S.I.

    2005-01-01

    This paper reports on two experiments performed at the Synchrophasotron/Nuclotron accelerator complex at JINR. Relativistic protons with energies 885 MeV and 1.5 GeV hit a massive cylindrical lead target. The spatial and energetic distributions of the neutron field produced by the spallation reactions were measured by the activation of Al, Au, Bi, Co, and Cu foils placed on the surface of the target and close to it. The yields of the radioactive nuclei produced by threshold reactions in these foils were determined by the analyses of their γ spectra. The comparison with Monte-Carlo based simulations was performed both with the LAHET+MCNP code and the MCNPX code

  12. Independent dose verification system with Monte Carlo simulations using TOPAS for passive scattering proton therapy at the National Cancer Center in Korea

    Science.gov (United States)

    Shin, Wook-Geun; Testa, Mauro; Kim, Hak Soo; Jeong, Jong Hwi; Byeong Lee, Se; Kim, Yeon-Joo; Min, Chul Hee

    2017-10-01

    For the independent validation of treatment plans, we developed a fully automated Monte Carlo (MC)-based patient dose calculation system with the tool for particle simulation (TOPAS) and proton therapy machine installed at the National Cancer Center in Korea to enable routine and automatic dose recalculation for each patient. The proton beam nozzle was modeled with TOPAS to simulate the therapeutic beam, and MC commissioning was performed by comparing percent depth dose with the measurement. The beam set-up based on the prescribed beam range and modulation width was automated by modifying the vendor-specific method. The CT phantom was modeled based on the DICOM CT files with TOPAS-built-in function, and an in-house-developed C++ code directly imports the CT files for positioning the CT phantom, RT-plan file for simulating the treatment plan, and RT-structure file for applying the Hounsfield unit (HU) assignment, respectively. The developed system was validated by comparing the dose distributions with those calculated by the treatment planning system (TPS) for a lung phantom and two patient cases of abdomen and internal mammary node. The results of the beam commissioning were in good agreement of up to 0.8 mm2 g-1 for B8 option in both of the beam range and the modulation width of the spread-out Bragg peaks. The beam set-up technique can predict the range and modulation width with an accuracy of 0.06% and 0.51%, respectively, with respect to the prescribed range and modulation in arbitrary points of B5 option (128.3, 132.0, and 141.2 mm2 g-1 of range). The dose distributions showed higher than 99% passing rate for the 3D gamma index (3 mm distance to agreement and 3% dose difference) between the MC simulations and the clinical TPS in the target volume. However, in the normal tissues, less favorable agreements were obtained for the radiation treatment planning with the lung phantom and internal mammary node cases. The discrepancies might come from the

  13. Monte Carlo-based simulation of dynamic jaws tomotherapy

    International Nuclear Information System (INIS)

    Sterpin, E.; Chen, Y.; Chen, Q.; Lu, W.; Mackie, T. R.; Vynckier, S.

    2011-01-01

    Purpose: Original TomoTherapy systems may involve a trade-off between conformity and treatment speed, the user being limited to three slice widths (1.0, 2.5, and 5.0 cm). This could be overcome by allowing the jaws to define arbitrary fields, including very small slice widths (<1 cm), which are challenging for a beam model. The aim of this work was to incorporate the dynamic jaws feature into a Monte Carlo (MC) model called TomoPen, based on the MC code PENELOPE, previously validated for the original TomoTherapy system. Methods: To keep the general structure of TomoPen and its efficiency, the simulation strategy introduces several techniques: (1) weight modifiers to account for any jaw settings using only the 5 cm phase-space file; (2) a simplified MC based model called FastStatic to compute the modifiers faster than pure MC; (3) actual simulation of dynamic jaws. Weight modifiers computed with both FastStatic and pure MC were compared. Dynamic jaws simulations were compared with the convolution/superposition (C/S) of TomoTherapy in the ''cheese'' phantom for a plan with two targets longitudinally separated by a gap of 3 cm. Optimization was performed in two modes: asymmetric jaws-constant couch speed (''running start stop,'' RSS) and symmetric jaws-variable couch speed (''symmetric running start stop,'' SRSS). Measurements with EDR2 films were also performed for RSS for the formal validation of TomoPen with dynamic jaws. Results: Weight modifiers computed with FastStatic were equivalent to pure MC within statistical uncertainties (0.5% for three standard deviations). Excellent agreement was achieved between TomoPen and C/S for both asymmetric jaw opening/constant couch speed and symmetric jaw opening/variable couch speed, with deviations well within 2%/2 mm. For RSS procedure, agreement between C/S and measurements was within 2%/2 mm for 95% of the points and 3%/3 mm for 98% of the points, where dose is greater than 30% of the prescription dose (gamma analysis

  14. Specific Features of Intramolecular Proton Transfer Reaction in Schiff Bases

    Directory of Open Access Journals (Sweden)

    Aleksander Koll

    2003-06-01

    Full Text Available Abstract: The differences between the intramolecular proton transfer in Mannich and Schiff bases are discussed. The tautomeric forms being in equilibrium in both types of molecules are seriously different. In Mannich bases there are in equilibrium the forms of phenols and phenolates. In Schiff bases each of tautomers is strongly influenced by resonance between zwitterionic and keto structures. Despite the common opinion that the proton transfer forms in compounds with internal π-electronic coupling are mainly keto forms it is shown in this work, that in Schiff bases the content of keto structure is slightly less than zwitterionic one. Almost equal participation of both forms leads to effective resonance between them and stabilization of intramolecular hydrogen bond in this way.

  15. Monte Carlo Techniques for Calculations of Charge Deposition and Displacement Damage from Protons in Visible and Infrared Sensor Arrays

    Science.gov (United States)

    Marshall, Paul; Reed, Robert; Fodness, Bryan; Jordan, Tom; Pickel, Jim; Xapsos, Michael; Burke, Ed

    2004-01-01

    This slide presentation examines motivation for Monte Carlo methods, charge deposition in sensor arrays, displacement damage calculations, and future work. The discussion of charge deposition sensor arrays includes Si active pixel sensor APS arrays and LWIR HgCdTe FPAs. The discussion of displacement damage calculations includes nonionizing energy loss (NIEL), HgCdTe NIEL calculation results including variance, and implications for damage in HgCdTe detector arrays.

  16. Unfolding an under-determined neutron spectrum using genetic algorithm based Monte Carlo

    International Nuclear Information System (INIS)

    Suman, V.; Sarkar, P.K.

    2011-01-01

    Spallation in addition to the other photon-neutron reactions in target materials and different components in accelerators may result in production of huge amount of energetic protons which further leads to the production of neutron and contributes to the main component of the total dose. For dosimetric purposes in accelerator facilities the detector measurements doesn't provide directly the actual neutron flux values but a cumulative picture. To obtain Neutron spectrum from the measured data, response functions of the measuring instrument together with the measurements are used into many unfolding techniques which are frequently used for unfolding the hidden spectral information. Here we discuss a genetic algorithm based unfolding technique which is in the process of development. Genetic Algorithm is a stochastic method based on natural selection, which mimics Darwinian theory of survival of the best. The above said method has been tested to unfold the neutron spectra obtained from a reaction carried out at an accelerator facility, with energetic carbon ions on thick silver target along with its respective neutron response of BC501A liquid scintillation detector. The problem dealt here is under-determined where the number of measurements is less than the required energy bin information. The results so obtained were compared with those obtained using the established unfolding code FERDOR, which unfolds data for completely determined problems. It is seen that the genetic algorithm based solution has a reasonable match with the results of FERDOR, when smoothening carried out by Monte Carlo is taken into consideration. This method appears to be a promising candidate for unfolding neutron spectrum in cases of under-determined and over-determined, where measurements are more. The method also has advantages of flexibility, computational simplicity and works well without need of any initial guess spectrum. (author)

  17. Sci—Fri PM: Topics — 07: Monte Carlo Simulation of Primary Dose and PET Isotope Production for the TRIUMF Proton Therapy Facility

    Energy Technology Data Exchange (ETDEWEB)

    Lindsay, C; Jirasek, A [University of Victoria (Australia); Blackmore, E; Hoehr, C; Schaffer, P; Trinczek, M [TRIUMF (Canada); Sossi, V [University of British Columbia (Canada)

    2014-08-15

    Uveal melanoma is a rare and deadly tumour of the eye with primary metastases in the liver resulting in an 8% 2-year survival rate upon detection. Large growths, or those in close proximity to the optic nerve, pose a particular challenge to the commonly employed eye-sparing technique of eye-plaque brachytherapy. In these cases external beam charged particle therapy offers improved odds in avoiding catastrophic side effects such as neuropathy or blindness. Since 1995, the British Columbia Cancer Agency in partnership with the TRIUMF national laboratory have offered proton therapy in the treatment of difficult ocular tumors. Having seen 175 patients, yielding 80% globe preservation and 82% metastasis free survival as of 2010, this modality has proven to be highly effective. Despite this success, there have been few studies into the use of the world's largest cyclotron in patient care. Here we describe first efforts of modeling the TRIUMF dose delivery system using the FLUKA Monte Carlo package. Details on geometry, estimating beam parameters, measurement of primary dose and simulation of PET isotope production are discussed. Proton depth dose in both modulated and pristine beams is successfully simulated to sub-millimeter precision in range (within limits of measurement) and 2% agreement to measurement within in a treatment volume. With the goal of using PET signals for in vivo dosimetry (alignment), a first look at PET isotope depth distribution is presented — comparing favourably to a naive method of approximating simulated PET slice activity in a Lucite phantom.

  18. Anhydrous proton conductivity of acid doped vinyltriazole-based polymers

    International Nuclear Information System (INIS)

    Pu, Hongting; Ye, Sheng; Wan, Decheng

    2007-01-01

    Poly(1,2,4-vinyltriazole) (PVTr) and poly(1,2,4-vinyltriazole-co-5-vinyltetrazole-co-acrylonitrile) (P(VTr-VT-AN)) were prepared by normal free radical polymerization and click chemistry, respectively. The structure of the polymers was characterized by FTIR spectra, H NMR spectrum and elemental analysis. Compared with polybenzimidazole (PBI) which is one of the most widely studied anhydrous proton conducting polymers, the solubility of vinyltriazole-based polymers is improved significantly. They are soluble in a lot of polar solvents. The glass-transition temperatures of such kind of polymers are between 70 and 85 o C, thus indirectly indicating the improvement of fabricating properties. In phosphoric acid doped membranes, the higher the basicity of the vinyltriazole-based polymers is, the higher the proton conductivity is. The temperature dependence of the proton conductivity of the acid doped membranes can always be fitted by a simple Arrhenius equation. Transmittance of phosphoric acid doped vinyltriazole-based polymers is above 80% in the range of visual spectra and changes a little with the different structure and basicity of the copolymers

  19. Experimental depth dose curves of a 67.5 MeV proton beam for benchmarking and validation of Monte Carlo simulation

    Energy Technology Data Exchange (ETDEWEB)

    Faddegon, Bruce A., E-mail: bfaddegon@radonc.ucsf.edu; Ramos-Méndez, José; Daftari, Inder K. [Department of Radiation Oncology, University of California San Francisco, 1600 Divisadero Street, Suite H1031, San Francisco, California 94143 (United States); Shin, Jungwook [St. Jude Children’s Research Hospital, 252 Danny Thomas Place, Memphis, Tennessee 38105 (United States); Castenada, Carlos M. [Crocker Nuclear Laboratory, University of California Davis, 1 Shields Avenue, Davis, California 95616 (United States)

    2015-07-15

    Purpose: To measure depth dose curves for a 67.5 ± 0.1 MeV proton beam for benchmarking and validation of Monte Carlo simulation. Methods: Depth dose curves were measured in 2 beam lines. Protons in the raw beam line traversed a Ta scattering foil, 0.1016 or 0.381 mm thick, a secondary emission monitor comprised of thin Al foils, and a thin Kapton exit window. The beam energy and peak width and the composition and density of material traversed by the beam were known with sufficient accuracy to permit benchmark quality measurements. Diodes for charged particle dosimetry from two different manufacturers were used to scan the depth dose curves with 0.003 mm depth reproducibility in a water tank placed 300 mm from the exit window. Depth in water was determined with an uncertainty of 0.15 mm, including the uncertainty in the water equivalent depth of the sensitive volume of the detector. Parallel-plate chambers were used to verify the accuracy of the shape of the Bragg peak and the peak-to-plateau ratio measured with the diodes. The uncertainty in the measured peak-to-plateau ratio was 4%. Depth dose curves were also measured with a diode for a Bragg curve and treatment beam spread out Bragg peak (SOBP) on the beam line used for eye treatment. The measurements were compared to Monte Carlo simulation done with GEANT4 using TOPAS. Results: The 80% dose at the distal side of the Bragg peak for the thinner foil was at 37.47 ± 0.11 mm (average of measurement with diodes from two different manufacturers), compared to the simulated value of 37.20 mm. The 80% dose for the thicker foil was at 35.08 ± 0.15 mm, compared to the simulated value of 34.90 mm. The measured peak-to-plateau ratio was within one standard deviation experimental uncertainty of the simulated result for the thinnest foil and two standard deviations for the thickest foil. It was necessary to include the collimation in the simulation, which had a more pronounced effect on the peak-to-plateau ratio for the

  20. Machine learning-based patient specific prompt-gamma dose monitoring in proton therapy

    International Nuclear Information System (INIS)

    Gueth, P; Freud, N; Létang, J M; Sarrut, D; Dauvergne, D; Ray, C; Testa, E

    2013-01-01

    Online dose monitoring in proton therapy is currently being investigated with prompt-gamma (PG) devices. PG emission was shown to be correlated with dose deposition. This relationship is mostly unknown under real conditions. We propose a machine learning approach based on simulations to create optimized treatment-specific classifiers that detect discrepancies between planned and delivered dose. Simulations were performed with the Monte-Carlo platform Gate/Geant4 for a spot-scanning proton therapy treatment and a PG camera prototype currently under investigation. The method first builds a learning set of perturbed situations corresponding to a range of patient translation. This set is then used to train a combined classifier using distal falloff and registered correlation measures. Classifier performances were evaluated using receiver operating characteristic curves and maximum associated specificity and sensitivity. A leave-one-out study showed that it is possible to detect discrepancies of 5 mm with specificity and sensitivity of 85% whereas using only distal falloff decreases the sensitivity down to 77% on the same data set. The proposed method could help to evaluate performance and to optimize the design of PG monitoring devices. It is generic: other learning sets of deviations, other measures and other types of classifiers could be studied to potentially reach better performance. At the moment, the main limitation lies in the computation time needed to perform the simulations. (paper)

  1. Machine learning-based patient specific prompt-gamma dose monitoring in proton therapy

    Science.gov (United States)

    Gueth, P.; Dauvergne, D.; Freud, N.; Létang, J. M.; Ray, C.; Testa, E.; Sarrut, D.

    2013-07-01

    Online dose monitoring in proton therapy is currently being investigated with prompt-gamma (PG) devices. PG emission was shown to be correlated with dose deposition. This relationship is mostly unknown under real conditions. We propose a machine learning approach based on simulations to create optimized treatment-specific classifiers that detect discrepancies between planned and delivered dose. Simulations were performed with the Monte-Carlo platform Gate/Geant4 for a spot-scanning proton therapy treatment and a PG camera prototype currently under investigation. The method first builds a learning set of perturbed situations corresponding to a range of patient translation. This set is then used to train a combined classifier using distal falloff and registered correlation measures. Classifier performances were evaluated using receiver operating characteristic curves and maximum associated specificity and sensitivity. A leave-one-out study showed that it is possible to detect discrepancies of 5 mm with specificity and sensitivity of 85% whereas using only distal falloff decreases the sensitivity down to 77% on the same data set. The proposed method could help to evaluate performance and to optimize the design of PG monitoring devices. It is generic: other learning sets of deviations, other measures and other types of classifiers could be studied to potentially reach better performance. At the moment, the main limitation lies in the computation time needed to perform the simulations.

  2. Reoptimization of Intensity Modulated Proton Therapy Plans Based on Linear Energy Transfer

    Energy Technology Data Exchange (ETDEWEB)

    Unkelbach, Jan, E-mail: junkelbach@mgh.harvard.edu [Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts (United States); Botas, Pablo [Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts (United States); Faculty of Physics, Ruprecht-Karls-Universität Heidelberg, Heidelberg (Germany); Giantsoudi, Drosoula; Gorissen, Bram L.; Paganetti, Harald [Department of Radiation Oncology, Massachusetts General Hospital, Boston, Massachusetts (United States)

    2016-12-01

    Purpose: We describe a treatment plan optimization method for intensity modulated proton therapy (IMPT) that avoids high values of linear energy transfer (LET) in critical structures located within or near the target volume while limiting degradation of the best possible physical dose distribution. Methods and Materials: To allow fast optimization based on dose and LET, a GPU-based Monte Carlo code was extended to provide dose-averaged LET in addition to dose for all pencil beams. After optimizing an initial IMPT plan based on physical dose, a prioritized optimization scheme is used to modify the LET distribution while constraining the physical dose objectives to values close to the initial plan. The LET optimization step is performed based on objective functions evaluated for the product of LET and physical dose (LET×D). To first approximation, LET×D represents a measure of the additional biological dose that is caused by high LET. Results: The method is effective for treatments where serial critical structures with maximum dose constraints are located within or near the target. We report on 5 patients with intracranial tumors (high-grade meningiomas, base-of-skull chordomas, ependymomas) in whom the target volume overlaps with the brainstem and optic structures. In all cases, high LET×D in critical structures could be avoided while minimally compromising physical dose planning objectives. Conclusion: LET-based reoptimization of IMPT plans represents a pragmatic approach to bridge the gap between purely physical dose-based and relative biological effectiveness (RBE)-based planning. The method makes IMPT treatments safer by mitigating a potentially increased risk of side effects resulting from elevated RBE of proton beams near the end of range.

  3. Comparison of depth-dose distributions of proton therapeutic beams calculated by means of logical detectors and ionization chamber modeled in Monte Carlo codes

    Energy Technology Data Exchange (ETDEWEB)

    Pietrzak, Robert [Department of Nuclear Physics and Its Applications, Institute of Physics, University of Silesia, Katowice (Poland); Konefał, Adam, E-mail: adam.konefal@us.edu.pl [Department of Nuclear Physics and Its Applications, Institute of Physics, University of Silesia, Katowice (Poland); Sokół, Maria; Orlef, Andrzej [Department of Medical Physics, Maria Sklodowska-Curie Memorial Cancer Center, Institute of Oncology, Gliwice (Poland)

    2016-08-01

    The success of proton therapy depends strongly on the precision of treatment planning. Dose distribution in biological tissue may be obtained from Monte Carlo simulations using various scientific codes making it possible to perform very accurate calculations. However, there are many factors affecting the accuracy of modeling. One of them is a structure of objects called bins registering a dose. In this work the influence of bin structure on the dose distributions was examined. The MCNPX code calculations of Bragg curve for the 60 MeV proton beam were done in two ways: using simple logical detectors being the volumes determined in water, and using a precise model of ionization chamber used in clinical dosimetry. The results of the simulations were verified experimentally in the water phantom with Marcus ionization chamber. The average local dose difference between the measured relative doses in the water phantom and those calculated by means of the logical detectors was 1.4% at first 25 mm, whereas in the full depth range this difference was 1.6% for the maximum uncertainty in the calculations less than 2.4% and for the maximum measuring error of 1%. In case of the relative doses calculated with the use of the ionization chamber model this average difference was somewhat greater, being 2.3% at depths up to 25 mm and 2.4% in the full range of depths for the maximum uncertainty in the calculations of 3%. In the dose calculations the ionization chamber model does not offer any additional advantages over the logical detectors. The results provided by both models are similar and in good agreement with the measurements, however, the logical detector approach is a more time-effective method. - Highlights: • Influence of the bin structure on the proton dose distributions was examined for the MC simulations. • The considered relative proton dose distributions in water correspond to the clinical application. • MC simulations performed with the logical detectors and the

  4. Monte Carlo simulation and parameterized treatment on the effect of nuclear elastic scattering in high-energy proton radiography

    Science.gov (United States)

    Xu, Hai-Bo; Zheng, Na

    2015-07-01

    A version of Geant4 has been developed to treat high-energy proton radiography. This article presents the results of calculations simulating the effects of nuclear elastic scattering for various test step wedges. Comparisons with experimental data are also presented. The traditional expressions of the transmission should be correct if the angle distribution of the scattering is Gaussian multiple Coulomb scattering. The mean free path (which depends on the collimator angle) and the radiation length are treated as empirical parameters, according to transmission as a function of thickness obtained by simulations. The results can be used in density reconstruction, which depends on the transmission expressions. Supported by NSAF (11176001) and Science and Technology Developing Foundation of China Academy of Engineering Physics (2012A0202006)

  5. The influence of lateral beam profile modifications in scanned proton and carbon ion therapy: a Monte Carlo study

    CERN Document Server

    Parodi, K; Kraemer, M; Sommerer, F; Naumann, J; Mairani, A; Brons, S

    2010-01-01

    Scanned ion beam delivery promises superior flexibility and accuracy for highly conformal tumour therapy in comparison to the usage of passive beam shaping systems. The attainable precision demands correct overlapping of the pencil-like beams which build up the entire dose distribution in the treatment field. In particular, improper dose application due to deviations of the lateral beam profiles from the nominal planning conditions must be prevented via appropriate beam monitoring in the beamline, prior to the entrance in the patient. To assess the necessary tolerance thresholds of the beam monitoring system at the Heidelberg Ion Beam Therapy Center, Germany, this study has investigated several worst-case scenarios for a sensitive treatment plan, namely scanned proton and carbon ion delivery to a small target volume at a shallow depth. Deviations from the nominal lateral beam profiles were simulated, which may occur because of misaligned elements or changes of the beam optic in the beamline. Data have been an...

  6. Optimization of a neutron production target based on the 7Li (p,n)7Be reaction with the Monte Carlo Method

    International Nuclear Information System (INIS)

    Burlon, Alejandro A.; Kreiner, Andres J.; Minsky, Daniel; Valda, Alejandro A.; Somacal, Hector R.

    2003-01-01

    In order to optimize a neutron production target for accelerator-based boron neutron capture therapy (AB-BNCT) a Monte Carlo Neutron and Photon (MCNP) investigation has been performed. Neutron fields from a LiF thick target (with both a D 2 O-graphite and a Al/AlF 3 -graphite moderator/reflector assembly) were evaluated along the centerline in a head phantom. The target neutron beam was simulated from the 7 Li(p,n) 7 Be nuclear reaction for 1.89, 2.0 and 2.3 MeV protons. The results show that it is more advantageous to irradiate the target with near resonance energy protons (2.3 MeV) because of the high neutron yield at this energy. On the other hand, the Al/AlF 3 -graphite exhibits a more efficient performance than D 2 O. (author)

  7. Pentanol-based target material with polarized protons

    International Nuclear Information System (INIS)

    Bunyatova, E.I.

    1992-01-01

    1-pentanol is a promising material for a target with polarized protons owing to its high resistance to radiation damage. To develop the target, the solutions of 1-pentanol or 2-pentanol with complexes of pentavalent chromium ware investigated. The material based EHBA-Cr(V) solution in a glass-like matrix, consisting of 1-pentanol, 3-pentanol and 1,2-propanediol, was proposed as a target material. It was investigated by the electron paramagnetic resonance and differential scanning calorimetry methods. 24 refs.; 3 figs.; 1 tab

  8. Composite proton exchange membrane based on sulfonated organic nanoparticles

    Science.gov (United States)

    Pitia, Emmanuel Sokiri

    exchange was characterized with solid state 13C NMR spectroscopy, FTIR spectroscopy, TGA, elemental analysis, and titration. The results indicate the extent of ion exchange was ~ 70-80%. Due to the mass of QAA, the remaining QAA reduced the IEC of the nanoparticles to < 2.2 meq/g. In fabricating the composite membranes, the nanoparticles and polystyrene were solution cast in a continuous process with and without electric field. The electric field had no effect on the water uptake. Based on the morphology and the proton conductivity, it appears orientation of the nanoparticles did not occur. We hypothesize the lack of orientation was caused by swelling of the particles with the solvent. The solvent inside the particle minimized polarizability, and thus prevented orientation. The composite membranes were limited to low proton conductivity of ~ 10-5 S/cm due to low IEC of the nanoparticles, but good dispersion of the nanoparticles was achieved. Future work should look into eliminating the QAA during synthesis and developing a rigid core for the nanoparticles.

  9. Acceptance and implementation of a system of planning computerized based on Monte Carlo

    International Nuclear Information System (INIS)

    Lopez-Tarjuelo, J.; Garcia-Molla, R.; Suan-Senabre, X. J.; Quiros-Higueras, J. Q.; Santos-Serra, A.; Marco-Blancas, N.; Calzada-Feliu, S.

    2013-01-01

    It has been done the acceptance for use clinical Monaco computerized planning system, based on an on a virtual model of the energy yield of the head of the linear electron Accelerator and that performs the calculation of the dose with an algorithm of x-rays (XVMC) based on Monte Carlo algorithm. (Author)

  10. Dosimetric accuracy of a treatment planning system for actively scanned proton beams and small target volumes: Monte Carlo and experimental validation

    CERN Document Server

    Magro, G; Mairani, A; Mirandola, A; Panizza, D; Russo, S; Ferrari, A; Valvo, F; Fossati, P; Ciocca, M

    2015-01-01

    This study was performed to evaluate the accuracy of a commercial treatment planning system (TPS), in optimising proton pencil beam dose distributions for small targets of different sizes (5–30 mm side) located at increasing depths in water. The TPS analytical algorithm was benchmarked against experimental data and the FLUKA Monte Carlo (MC) code, previously validated for the selected beam-line. We tested the Siemens syngo® TPS plan optimisation module for water cubes fixing the configurable parameters at clinical standards, with homogeneous target coverage to a 2 Gy (RBE) dose prescription as unique goal. Plans were delivered and the dose at each volume centre was measured in water with a calibrated PTW Advanced Markus® chamber. An EBT3® film was also positioned at the phantom entrance window for the acquisition of 2D dose maps. Discrepancies between TPS calculated and MC simulated values were mainly due to the different lateral spread modeling and resulted in being related to the field-to-spot size r...

  11. Monte Carlo simulations and experimental results on neutron production in the spallation target QUINTA irradiated with 660 MeV protons

    International Nuclear Information System (INIS)

    Khushvaktov, J.H.; Yuldashev, B.S.; Adam, J.; Vrzalova, J.; Baldin, A.A.; Furman, W.I.; Gustov, S.A.; Kish, Yu.V.; Solnyshkin, A.A.; Stegailov, V.I.; Tichy, P.; Tsoupko-Sitnikov, V.M.; Tyutyunnikov, S.I.; Zavorka, L.; Svoboda, J.; Zeman, M.; Vespalec, R.; Wagner, V.

    2017-01-01

    The activation experiment was performed using the accelerated beam of the Phasotron accelerator at the Joint Institute for Nuclear Research (JINR). The natural uranium spallation target QUINTA was irradiated with protons of energy 660 MeV. Monte Carlo simulations were performed using the FLUKA and Geant4 codes. The number of leakage neutrons from the sections of the uranium target surrounded by the lead shielding and the number of leakage neutrons from the lead shield were determined. The total number of fissions in the setup QUINTA were determined. Experimental values of reaction rates for the produced nuclei in the 127 I sample were obtained, and several values of the reaction rates were compared with the results of simulations by the FLUKA and Geant4 codes. The experimentally determined fluence of neutrons in the energy range of 10-200 MeV using the (n, xn) reactions in the 127 I(NaI) sample was compared with the results of simulations. Possibility of transmutation of the long-lived radionuclide 129 I in the QUINTA setup was estimated. [ru

  12. Acceptance and implementation of a system of planning computerized based on Monte Carlo; Aceptacion y puesta en marcha de un sistema de planificacion comutarizada basado en Monte Carlo

    Energy Technology Data Exchange (ETDEWEB)

    Lopez-Tarjuelo, J.; Garcia-Molla, R.; Suan-Senabre, X. J.; Quiros-Higueras, J. Q.; Santos-Serra, A.; Marco-Blancas, N.; Calzada-Feliu, S.

    2013-07-01

    It has been done the acceptance for use clinical Monaco computerized planning system, based on an on a virtual model of the energy yield of the head of the linear electron Accelerator and that performs the calculation of the dose with an algorithm of x-rays (XVMC) based on Monte Carlo algorithm. (Author)

  13. Response matrix Monte Carlo based on a general geometry local calculation for electron transport

    International Nuclear Information System (INIS)

    Ballinger, C.T.; Rathkopf, J.A.; Martin, W.R.

    1991-01-01

    A Response Matrix Monte Carlo (RMMC) method has been developed for solving electron transport problems. This method was born of the need to have a reliable, computationally efficient transport method for low energy electrons (below a few hundred keV) in all materials. Today, condensed history methods are used which reduce the computation time by modeling the combined effect of many collisions but fail at low energy because of the assumptions required to characterize the electron scattering. Analog Monte Carlo simulations are prohibitively expensive since electrons undergo coulombic scattering with little state change after a collision. The RMMC method attempts to combine the accuracy of an analog Monte Carlo simulation with the speed of the condensed history methods. Like condensed history, the RMMC method uses probability distributions functions (PDFs) to describe the energy and direction of the electron after several collisions. However, unlike the condensed history method the PDFs are based on an analog Monte Carlo simulation over a small region. Condensed history theories require assumptions about the electron scattering to derive the PDFs for direction and energy. Thus the RMMC method samples from PDFs which more accurately represent the electron random walk. Results show good agreement between the RMMC method and analog Monte Carlo. 13 refs., 8 figs

  14. Determination of the spatial response of neutron based analysers using a Monte Carlo based method

    Science.gov (United States)

    Tickner

    2000-10-01

    One of the principal advantages of using thermal neutron capture (TNC, also called prompt gamma neutron activation analysis or PGNAA) or neutron inelastic scattering (NIS) techniques for measuring elemental composition is the high penetrating power of both the incident neutrons and the resultant gamma-rays, which means that large sample volumes can be interrogated. Gauges based on these techniques are widely used in the mineral industry for on-line determination of the composition of bulk samples. However, attenuation of both neutrons and gamma-rays in the sample and geometric (source/detector distance) effects typically result in certain parts of the sample contributing more to the measured composition than others. In turn, this introduces errors in the determination of the composition of inhomogeneous samples. This paper discusses a combined Monte Carlo/analytical method for estimating the spatial response of a neutron gauge. Neutron propagation is handled using a Monte Carlo technique which allows an arbitrarily complex neutron source and gauge geometry to be specified. Gamma-ray production and detection is calculated analytically which leads to a dramatic increase in the efficiency of the method. As an example, the method is used to study ways of reducing the spatial sensitivity of on-belt composition measurements of cement raw meal.

  15. Proton therapy for tumors of the skull base

    Energy Technology Data Exchange (ETDEWEB)

    Munzenrider, J.E.; Liebsch, N.J. [Dept. of Radiation Oncology, Harvard Univ. Medical School, Boston, MA (United States)

    1999-06-01

    Charged particle beams are ideal for treating skull base and cervical spine tumors: dose can be focused in the target, while achieving significant sparing of the brain, brain stem, cervical cord, and optic nerves and chiasm. For skull base tumors, 10-year local control rates with combined proton-photon therapy are highest for chondrosarcomas, intermediate for male chordomas, and lowest for female chordomas (94%, 65%, and 42%, respectively). For cervical spine tumors, 10-year local control rates are not significantly different for chordomas and chondrosarcomas (54% and 48%, respectively), nor is there any difference in local control between males and females. Observed treatment-related morbidity has been judged acceptable, in view of the major morbidity and mortality which accompany uncontrolled tumor growth. (orig.)

  16. Confronting uncertainty in model-based geostatistics using Markov Chain Monte Carlo simulation

    NARCIS (Netherlands)

    Minasny, B.; Vrugt, J.A.; McBratney, A.B.

    2011-01-01

    This paper demonstrates for the first time the use of Markov Chain Monte Carlo (MCMC) simulation for parameter inference in model-based soil geostatistics. We implemented the recently developed DiffeRential Evolution Adaptive Metropolis (DREAM) algorithm to jointly summarize the posterior

  17. New memory devices based on the proton transfer process

    International Nuclear Information System (INIS)

    Wierzbowska, Małgorzata

    2016-01-01

    Memory devices operating due to the fast proton transfer (PT) process are proposed by the means of first-principles calculations. Writing  information is performed using the electrostatic potential of scanning tunneling microscopy (STM). Reading information is based on the effect of the local magnetization induced at the zigzag graphene nanoribbon (Z-GNR) edge—saturated with oxygen or the hydroxy group—and can be realized with the use of giant magnetoresistance (GMR), a magnetic tunnel junction or spin-transfer torque devices. The energetic barriers for the hop forward and backward processes can be tuned by the distance and potential of the STM tip; this thus enables us to tailor the non-volatile logic states. The proposed system enables very dense packing of the logic cells and could be used in random access and flash memory devices. (paper)

  18. New memory devices based on the proton transfer process

    Science.gov (United States)

    Wierzbowska, Małgorzata

    2016-01-01

    Memory devices operating due to the fast proton transfer (PT) process are proposed by the means of first-principles calculations. Writing information is performed using the electrostatic potential of scanning tunneling microscopy (STM). Reading information is based on the effect of the local magnetization induced at the zigzag graphene nanoribbon (Z-GNR) edge—saturated with oxygen or the hydroxy group—and can be realized with the use of giant magnetoresistance (GMR), a magnetic tunnel junction or spin-transfer torque devices. The energetic barriers for the hop forward and backward processes can be tuned by the distance and potential of the STM tip; this thus enables us to tailor the non-volatile logic states. The proposed system enables very dense packing of the logic cells and could be used in random access and flash memory devices.

  19. Development of Monte Carlo-based pebble bed reactor fuel management code

    International Nuclear Information System (INIS)

    Setiadipura, Topan; Obara, Toru

    2014-01-01

    Highlights: • A new Monte Carlo-based fuel management code for OTTO cycle pebble bed reactor was developed. • The double-heterogeneity was modeled using statistical method in MVP-BURN code. • The code can perform analysis of equilibrium and non-equilibrium phase. • Code-to-code comparisons for Once-Through-Then-Out case were investigated. • Ability of the code to accommodate the void cavity was confirmed. - Abstract: A fuel management code for pebble bed reactors (PBRs) based on the Monte Carlo method has been developed in this study. The code, named Monte Carlo burnup analysis code for PBR (MCPBR), enables a simulation of the Once-Through-Then-Out (OTTO) cycle of a PBR from the running-in phase to the equilibrium condition. In MCPBR, a burnup calculation based on a continuous-energy Monte Carlo code, MVP-BURN, is coupled with an additional utility code to be able to simulate the OTTO cycle of PBR. MCPBR has several advantages in modeling PBRs, namely its Monte Carlo neutron transport modeling, its capability of explicitly modeling the double heterogeneity of the PBR core, and its ability to model different axial fuel speeds in the PBR core. Analysis at the equilibrium condition of the simplified PBR was used as the validation test of MCPBR. The calculation results of the code were compared with the results of diffusion-based fuel management PBR codes, namely the VSOP and PEBBED codes. Using JENDL-4.0 nuclide library, MCPBR gave a 4.15% and 3.32% lower k eff value compared to VSOP and PEBBED, respectively. While using JENDL-3.3, MCPBR gave a 2.22% and 3.11% higher k eff value compared to VSOP and PEBBED, respectively. The ability of MCPBR to analyze neutron transport in the top void of the PBR core and its effects was also confirmed

  20. Accelerated Monte Carlo system reliability analysis through machine-learning-based surrogate models of network connectivity

    International Nuclear Information System (INIS)

    Stern, R.E.; Song, J.; Work, D.B.

    2017-01-01

    The two-terminal reliability problem in system reliability analysis is known to be computationally intractable for large infrastructure graphs. Monte Carlo techniques can estimate the probability of a disconnection between two points in a network by selecting a representative sample of network component failure realizations and determining the source-terminal connectivity of each realization. To reduce the runtime required for the Monte Carlo approximation, this article proposes an approximate framework in which the connectivity check of each sample is estimated using a machine-learning-based classifier. The framework is implemented using both a support vector machine (SVM) and a logistic regression based surrogate model. Numerical experiments are performed on the California gas distribution network using the epicenter and magnitude of the 1989 Loma Prieta earthquake as well as randomly-generated earthquakes. It is shown that the SVM and logistic regression surrogate models are able to predict network connectivity with accuracies of 99% for both methods, and are 1–2 orders of magnitude faster than using a Monte Carlo method with an exact connectivity check. - Highlights: • Surrogate models of network connectivity are developed by machine-learning algorithms. • Developed surrogate models can reduce the runtime required for Monte Carlo simulations. • Support vector machine and logistic regressions are employed to develop surrogate models. • Numerical example of California gas distribution network demonstrate the proposed approach. • The developed models have accuracies 99%, and are 1–2 orders of magnitude faster than MCS.

  1. Configuration and validation of an analytical model predicting secondary neutron radiation in proton therapy using Monte Carlo simulations and experimental measurements.

    Science.gov (United States)

    Farah, J; Bonfrate, A; De Marzi, L; De Oliveira, A; Delacroix, S; Martinetti, F; Trompier, F; Clairand, I

    2015-05-01

    This study focuses on the configuration and validation of an analytical model predicting leakage neutron doses in proton therapy. Using Monte Carlo (MC) calculations, a facility-specific analytical model was built to reproduce out-of-field neutron doses while separately accounting for the contribution of intra-nuclear cascade, evaporation, epithermal and thermal neutrons. This model was first trained to reproduce in-water neutron absorbed doses and in-air neutron ambient dose equivalents, H*(10), calculated using MCNPX. Its capacity in predicting out-of-field doses at any position not involved in the training phase was also checked. The model was next expanded to enable a full 3D mapping of H*(10) inside the treatment room, tested in a clinically relevant configuration and finally consolidated with experimental measurements. Following the literature approach, the work first proved that it is possible to build a facility-specific analytical model that efficiently reproduces in-water neutron doses and in-air H*(10) values with a maximum difference less than 25%. In addition, the analytical model succeeded in predicting out-of-field neutron doses in the lateral and vertical direction. Testing the analytical model in clinical configurations proved the need to separate the contribution of internal and external neutrons. The impact of modulation width on stray neutrons was found to be easily adjustable while beam collimation remains a challenging issue. Finally, the model performance agreed with experimental measurements with satisfactory results considering measurement and simulation uncertainties. Analytical models represent a promising solution that substitutes for time-consuming MC calculations when assessing doses to healthy organs. Copyright © 2015 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  2. Abstract ID: 240 A probabilistic-based nuclear reaction model for Monte Carlo ion transport in particle therapy.

    Science.gov (United States)

    Maria Jose, Gonzalez Torres; Jürgen, Henniger

    2018-01-01

    In order to expand the Monte Carlo transport program AMOS to particle therapy applications, the ion module is being developed in the radiation physics group (ASP) at the TU Dresden. This module simulates the three main interactions of ions in matter for the therapy energy range: elastic scattering, inelastic collisions and nuclear reactions. The simulation of the elastic scattering is based on the Binary Collision Approximation and the inelastic collisions on the Bethe-Bloch theory. The nuclear reactions, which are the focus of the module, are implemented according to a probabilistic-based model developed in the group. The developed model uses probability density functions to sample the occurrence of a nuclear reaction given the initial energy of the projectile particle as well as the energy at which this reaction will take place. The particle is transported until the reaction energy is reached and then the nuclear reaction is simulated. This approach allows a fast evaluation of the nuclear reactions. The theory and application of the proposed model will be addressed in this presentation. The results of the simulation of a proton beam colliding with tissue will also be presented. Copyright © 2017.

  3. A Monte-Carlo study to assess the effect of 1.5 T magnetic fields on the overall robustness of pencil-beam scanning proton radiotherapy plans for prostate cancer

    Science.gov (United States)

    Kurz, Christopher; Landry, Guillaume; Resch, Andreas F.; Dedes, George; Kamp, Florian; Ganswindt, Ute; Belka, Claus; Raaymakers, Bas W.; Parodi, Katia

    2017-11-01

    Combining magnetic-resonance imaging (MRI) and proton therapy (PT) using pencil-beam scanning (PBS) may improve image-guided radiotherapy. We aimed at assessing the impact of a magnetic field on PBS-PT plan quality and robustness. Specifically, the robustness against anatomical changes and positioning errors in an MRI-guided scenario with a 30 cm radius 1.5 T magnetic field was studied for prostate PT. Five prostate cancer patients with three consecutive CT images (CT1-3) were considered. Single-field uniform dose PBS-PT plans were generated on the segmented CT1 with Monte-Carlo-based treatment planning software for inverse optimization. Plans were optimized at 90° gantry angle without B-field (no B), with  ±1.5 T B-field (B and minus B), as well as at 81° gantry angle and  +1.5 T (B G81). Plans were re-calculated on aligned CT2 and CT3 to study the impact of anatomical changes. Dose distributions were compared in terms of changes in DVH parameters, proton range and gamma-index pass-rates. To assess the impact of positioning errors, DVH parameters were compared for  ±5 mm CT1 patient shifts in anterior-posterior (AP) and left-right (LR) direction. Proton beam deflection considerably reduced robustness against inter-fractional changes for the B scenario. Range agreement, gamma-index pass-rates and PTV V95% were significantly lower compared to no B. Improved robustness was obtained for minus B and B G81, the latter showing only minor differences to no B. The magnetic field introduced slight dosimetric changes under LR shifts. The impact of AP shifts was considerably larger, and equivalent for scenarios with and without B-field. Results suggest that robustness equivalent to PT without magnetic field can be achieved by adaptation of the treatment parameters, such as B-field orientation (minus B) with respect to the patient and/or gantry angle (B G81). MRI-guided PT for prostate cancer might thus be implemented without compromising robustness

  4. Research on reactor physics analysis method based on Monte Carlo homogenization

    International Nuclear Information System (INIS)

    Ye Zhimin; Zhang Peng

    2014-01-01

    In order to meet the demand of nuclear energy market in the future, many new concepts of nuclear energy systems has been put forward. The traditional deterministic neutronics analysis method has been challenged in two aspects: one is the ability of generic geometry processing; the other is the multi-spectrum applicability of the multigroup cross section libraries. Due to its strong geometry modeling capability and the application of continuous energy cross section libraries, the Monte Carlo method has been widely used in reactor physics calculations, and more and more researches on Monte Carlo method has been carried out. Neutronics-thermal hydraulics coupling analysis based on Monte Carlo method has been realized. However, it still faces the problems of long computation time and slow convergence which make it not applicable to the reactor core fuel management simulations. Drawn from the deterministic core analysis method, a new two-step core analysis scheme is proposed in this work. Firstly, Monte Carlo simulations are performed for assembly, and the assembly homogenized multi-group cross sections are tallied at the same time. Secondly, the core diffusion calculations can be done with these multigroup cross sections. The new scheme can achieve high efficiency while maintain acceptable precision, so it can be used as an effective tool for the design and analysis of innovative nuclear energy systems. Numeric tests have been done in this work to verify the new scheme. (authors)

  5. A measurement-based generalized source model for Monte Carlo dose simulations of CT scans.

    Science.gov (United States)

    Ming, Xin; Feng, Yuanming; Liu, Ransheng; Yang, Chengwen; Zhou, Li; Zhai, Hezheng; Deng, Jun

    2017-03-07

    The goal of this study is to develop a generalized source model for accurate Monte Carlo dose simulations of CT scans based solely on the measurement data without a priori knowledge of scanner specifications. The proposed generalized source model consists of an extended circular source located at x-ray target level with its energy spectrum, source distribution and fluence distribution derived from a set of measurement data conveniently available in the clinic. Specifically, the central axis percent depth dose (PDD) curves measured in water and the cone output factors measured in air were used to derive the energy spectrum and the source distribution respectively with a Levenberg-Marquardt algorithm. The in-air film measurement of fan-beam dose profiles at fixed gantry was back-projected to generate the fluence distribution of the source model. A benchmarked Monte Carlo user code was used to simulate the dose distributions in water with the developed source model as beam input. The feasibility and accuracy of the proposed source model was tested on a GE LightSpeed and a Philips Brilliance Big Bore multi-detector CT (MDCT) scanners available in our clinic. In general, the Monte Carlo simulations of the PDDs in water and dose profiles along lateral and longitudinal directions agreed with the measurements within 4%/1 mm for both CT scanners. The absolute dose comparison using two CTDI phantoms (16 cm and 32 cm in diameters) indicated a better than 5% agreement between the Monte Carlo-simulated and the ion chamber-measured doses at a variety of locations for the two scanners. Overall, this study demonstrated that a generalized source model can be constructed based only on a set of measurement data and used for accurate Monte Carlo dose simulations of patients' CT scans, which would facilitate patient-specific CT organ dose estimation and cancer risk management in the diagnostic and therapeutic radiology.

  6. Protonation Equilibrium of Linear Homopolyacids

    Directory of Open Access Journals (Sweden)

    Požar J.

    2015-07-01

    Full Text Available The paper presents a short summary of investigations dealing with protonation equilibrium of linear homopolyacids, in particularly those of high charge density. Apart from the review of experimental results which can be found in the literature, a brief description of theoretical models used in processing the dependence of protonation constants on monomer dissociation degree and ionic strength is given (cylindrical model based on Poisson-Boltzmann equation, cylindrical Stern model, the models according to Ising, Högfeldt, Mandel and Katchalsky. The applicability of these models regarding the polyion charge density, electrolyte concentration and counterion type is discussed. The results of Monte Carlo simulations of protonation equilibrium are also briefly mentioned. In addition, frequently encountered errors connected with calibration of of glass electrode and the related unreliability of determined protonation constants are pointed out.

  7. SU-F-J-57: Effectiveness of Daily CT-Based Three-Dimensional Image Guided and Adaptive Proton Therapy

    Energy Technology Data Exchange (ETDEWEB)

    Moriya, S [University of Tsukuba, Tsukuba, Ibaraki (Japan); National Cancer Center, Kashiwa, Chiba (Japan); Tachibana, H; Hotta, K; Baba, H; Kohno, R; Akimoto, T [National Cancer Center, Kashiwa, Chiba (Japan); Nakamura, N [National Cancer Center Hospital East, Kashiwa, Chiba (Japan); Miyakawa, S; Kurosawa, T [Komazawa University, Setagaya, Tokyo (Japan)

    2016-06-15

    Purpose: Daily CT-based three-dimensional image-guided and adaptive (CTIGRT-ART) proton therapy system was designed and developed. We also evaluated the effectiveness of the CTIGRT-ART. Methods: Retrospective analysis was performed in three lung cancer patients: Proton treatment planning was performed using CT image datasets acquired by Toshiba Aquilion ONE. Planning target volume and surrounding organs were contoured by a well-trained radiation oncologist. Dose distribution was optimized using 180-deg. and 270-deg. two fields in passive scattering proton therapy. Well commissioned Simplified Monte Carlo algorithm was used as dose calculation engine. Daily consecutive CT image datasets was acquired by an in-room CT (Toshiba Aquilion LB). In our in-house program, two image registrations for bone and tumor were performed to shift the isocenter using treatment CT image dataset. Subsequently, dose recalculation was performed after the shift of the isocenter. When the dose distribution after the tumor registration exhibits change of dosimetric parameter of CTV D90% compared to the initial plan, an additional process of was performed that the range shifter thickness was optimized. Dose distribution with CTV D90% for the bone registration, the tumor registration only and adaptive plan with the tumor registration was compared to the initial plan. Results: In the bone registration, tumor dose coverage was decreased by 16% on average (Maximum: 56%). The tumor registration shows better coverage than the bone registration, however the coverage was also decreased by 9% (Maximum: 22%) The adaptive plan shows similar dose coverage of the tumor (Average: 2%, Maximum: 7%). Conclusion: There is a high possibility that only image registration for bone and tumor may reduce tumor coverage. Thus, our proposed methodology of image guidance and adaptive planning using the range adaptation after tumor registration would be effective for proton therapy. This research is partially supported

  8. Coal Oxide as a Thermally Robust Carbon-Based Proton Conductor.

    Science.gov (United States)

    Hatakeyama, Kazuto; Ogata, Chikako; Koinuma, Michio; Taniguchi, Takaaki; Hayami, Shinya; Kuroiwa, Keita; Matsumoto, Yasumichi

    2015-10-21

    Inexpensive solid proton conducting materials with high proton conductivity and thermal stability are necessary for practical solid state electrochemical devices. Here we report that coal oxide (CO) is a promising carbon-based proton conductor with remarkable thermal robustness. The CO produced by simple liquid-phase oxidation of coal demonstrates excellent dispersibility in water owing to the surface carboxyl groups. The proton conductivity of CO, 3.9 × 10(-3) S cm(-1) at 90% relative humidity, is as high as that of graphene oxide (GO). Remarkably, CO exhibits much higher thermal stability than GO, with CO retaining the excellent proton conductivity as well as the capacitance performance even after thermal annealing at 200 °C. Our study demonstrates that the chemical modification of the abundant coal provides proton conductors that can be used in practical applications for a wide range of energy devices.

  9. Radial-based tail methods for Monte Carlo simulations of cylindrical interfaces

    Science.gov (United States)

    Goujon, Florent; Bêche, Bruno; Malfreyt, Patrice; Ghoufi, Aziz

    2018-03-01

    In this work, we implement for the first time the radial-based tail methods for Monte Carlo simulations of cylindrical interfaces. The efficiency of this method is then evaluated through the calculation of surface tension and coexisting properties. We show that the inclusion of tail corrections during the course of the Monte Carlo simulation impacts the coexisting and the interfacial properties. We establish that the long range corrections to the surface tension are the same order of magnitude as those obtained from planar interface. We show that the slab-based tail method does not amend the localization of the Gibbs equimolar dividing surface. Additionally, a non-monotonic behavior of surface tension is exhibited as a function of the radius of the equimolar dividing surface.

  10. MBR Monte Carlo Simulation in PYTHIA8

    Science.gov (United States)

    Ciesielski, R.

    We present the MBR (Minimum Bias Rockefeller) Monte Carlo simulation of (anti)proton-proton interactions and its implementation in the PYTHIA8 event generator. We discuss the total, elastic, and total-inelastic cross sections, and three contributions from diffraction dissociation processes that contribute to the latter: single diffraction, double diffraction, and central diffraction or double-Pomeron exchange. The event generation follows a renormalized-Regge-theory model, successfully tested using CDF data. Based on the MBR-enhanced PYTHIA8 simulation, we present cross-section predictions for the LHC and beyond, up to collision energies of 50 TeV.

  11. Development of a space radiation Monte Carlo computer simulation based on the FLUKA and ROOT codes

    CERN Document Server

    Pinsky, L; Ferrari, A; Sala, P; Carminati, F; Brun, R

    2001-01-01

    This NASA funded project is proceeding to develop a Monte Carlo-based computer simulation of the radiation environment in space. With actual funding only initially in place at the end of May 2000, the study is still in the early stage of development. The general tasks have been identified and personnel have been selected. The code to be assembled will be based upon two major existing software packages. The radiation transport simulation will be accomplished by updating the FLUKA Monte Carlo program, and the user interface will employ the ROOT software being developed at CERN. The end-product will be a Monte Carlo-based code which will complement the existing analytic codes such as BRYNTRN/HZETRN presently used by NASA to evaluate the effects of radiation shielding in space. The planned code will possess the ability to evaluate the radiation environment for spacecraft and habitats in Earth orbit, in interplanetary space, on the lunar surface, or on a planetary surface such as Mars. Furthermore, it will be usef...

  12. Molecular modeling of protonic acid doping of emeraldine base polyaniline for chemical sensors

    NARCIS (Netherlands)

    Chen, X.; Yuan, C.A.; Wong, C.K.Y.; Ye, H.; Leung, S.Y.Y.; Zhang, G.

    2012-01-01

    We proposed a molecular modeling methodology to study the protonic acid doping of emeraldine base polyaniline which can used in gas detection. The commercial forcefield COMPASS was used for the polymer and protonic acid molecules. The molecular model, which is capable of representing the polyaniline

  13. A novel PTFE-based proton-conductive membrane

    Science.gov (United States)

    Reichman, S.; Duvdevani, T.; Aharon, A.; Philosoph, M.; Golodnitsky, D.; Peled, E.

    The demand for a solid polymer electrolyte membrane (SPEM) for fuel-cell systems, capable of withstanding temperatures above 130 °C, decreasing the electrode-catalyst loadings and reducing poisoning by carbon monoxide, has prompted this study. A novel, low-cost, highly conductive, nanoporous proton-conducting membrane (NP-PCM) based on a polytetrafluoroethylene (PTFE) backbone has been developed. It comprises non-conductive nano-size ceramic powder, PTFE binder and an aqueous acid. The preparation procedures were studied and the membrane was characterized with the use of: SEM, EDS, pore-size-distribution measurements (PSD), TGA-DTA and electrochemical methods. The ionic conductivity of a membrane doped with 3 M sulfuric acid increases with the ceramic powder content and reaches 0.22 S cm -1 at 50% (v/v) silica. A non-optimized direct-methanol fuel cell (DMFC) with a 250 μm thick membrane has been assembled. It demonstrated 50 and 130 mW cm -2 at 80 and 130 °C, respectively. Future study will be directed to improving the membrane-preparation process, getting thinner membranes and using this membrane in a hydrogen-fed fuel cell.

  14. Present status of transport code development based on Monte Carlo method

    International Nuclear Information System (INIS)

    Nakagawa, Masayuki

    1985-01-01

    The present status of development in Monte Carlo code is briefly reviewed. The main items are the followings; Application fields, Methods used in Monte Carlo code (geometry spectification, nuclear data, estimator and variance reduction technique) and unfinished works, Typical Monte Carlo codes and Merits of continuous energy Monte Carlo code. (author)

  15. Fault Risk Assessment of Underwater Vehicle Steering System Based on Virtual Prototyping and Monte Carlo Simulation

    Directory of Open Access Journals (Sweden)

    He Deyu

    2016-09-01

    Full Text Available Assessing the risks of steering system faults in underwater vehicles is a human-machine-environment (HME systematic safety field that studies faults in the steering system itself, the driver’s human reliability (HR and various environmental conditions. This paper proposed a fault risk assessment method for an underwater vehicle steering system based on virtual prototyping and Monte Carlo simulation. A virtual steering system prototype was established and validated to rectify a lack of historic fault data. Fault injection and simulation were conducted to acquire fault simulation data. A Monte Carlo simulation was adopted that integrated randomness due to the human operator and environment. Randomness and uncertainty of the human, machine and environment were integrated in the method to obtain a probabilistic risk indicator. To verify the proposed method, a case of stuck rudder fault (SRF risk assessment was studied. This method may provide a novel solution for fault risk assessment of a vehicle or other general HME system.

  16. ERSN-OpenMC, a Java-based GUI for OpenMC Monte Carlo code

    Directory of Open Access Journals (Sweden)

    Jaafar EL Bakkali

    2016-07-01

    Full Text Available OpenMC is a new Monte Carlo transport particle simulation code focused on solving two types of neutronic problems mainly the k-eigenvalue criticality fission source problems and external fixed fission source problems. OpenMC does not have any Graphical User Interface and the creation of one is provided by our java-based application named ERSN-OpenMC. The main feature of this application is to provide to the users an easy-to-use and flexible graphical interface to build better and faster simulations, with less effort and great reliability. Additionally, this graphical tool was developed with several features, as the ability to automate the building process of OpenMC code and related libraries as well as the users are given the freedom to customize their installation of this Monte Carlo code. A full description of the ERSN-OpenMC application is presented in this paper.

  17. High accuracy modeling for advanced nuclear reactor core designs using Monte Carlo based coupled calculations

    Science.gov (United States)

    Espel, Federico Puente

    The main objective of this PhD research is to develop a high accuracy modeling tool using a Monte Carlo based coupled system. The presented research comprises the development of models to include the thermal-hydraulic feedback to the Monte Carlo method and speed-up mechanisms to accelerate the Monte Carlo criticality calculation. Presently, deterministic codes based on the diffusion approximation of the Boltzmann transport equation, coupled with channel-based (or sub-channel based) thermal-hydraulic codes, carry out the three-dimensional (3-D) reactor core calculations of the Light Water Reactors (LWRs). These deterministic codes utilize nuclear homogenized data (normally over large spatial zones, consisting of fuel assembly or parts of fuel assembly, and in the best case, over small spatial zones, consisting of pin cell), which is functionalized in terms of thermal-hydraulic feedback parameters (in the form of off-line pre-generated cross-section libraries). High accuracy modeling is required for advanced nuclear reactor core designs that present increased geometry complexity and material heterogeneity. Such high-fidelity methods take advantage of the recent progress in computation technology and coupled neutron transport solutions with thermal-hydraulic feedback models on pin or even on sub-pin level (in terms of spatial scale). The continuous energy Monte Carlo method is well suited for solving such core environments with the detailed representation of the complicated 3-D problem. The major advantages of the Monte Carlo method over the deterministic methods are the continuous energy treatment and the exact 3-D geometry modeling. However, the Monte Carlo method involves vast computational time. The interest in Monte Carlo methods has increased thanks to the improvements of the capabilities of high performance computers. Coupled Monte-Carlo calculations can serve as reference solutions for verifying high-fidelity coupled deterministic neutron transport methods

  18. Monte Carlo SURE-based parameter selection for parallel magnetic resonance imaging reconstruction.

    Science.gov (United States)

    Weller, Daniel S; Ramani, Sathish; Nielsen, Jon-Fredrik; Fessler, Jeffrey A

    2014-05-01

    Regularizing parallel magnetic resonance imaging (MRI) reconstruction significantly improves image quality but requires tuning parameter selection. We propose a Monte Carlo method for automatic parameter selection based on Stein's unbiased risk estimate that minimizes the multichannel k-space mean squared error (MSE). We automatically tune parameters for image reconstruction methods that preserve the undersampled acquired data, which cannot be accomplished using existing techniques. We derive a weighted MSE criterion appropriate for data-preserving regularized parallel imaging reconstruction and the corresponding weighted Stein's unbiased risk estimate. We describe a Monte Carlo approximation of the weighted Stein's unbiased risk estimate that uses two evaluations of the reconstruction method per candidate parameter value. We reconstruct images using the denoising sparse images from GRAPPA using the nullspace method (DESIGN) and L1 iterative self-consistent parallel imaging (L1 -SPIRiT). We validate Monte Carlo Stein's unbiased risk estimate against the weighted MSE. We select the regularization parameter using these methods for various noise levels and undersampling factors and compare the results to those using MSE-optimal parameters. Our method selects nearly MSE-optimal regularization parameters for both DESIGN and L1 -SPIRiT over a range of noise levels and undersampling factors. The proposed method automatically provides nearly MSE-optimal choices of regularization parameters for data-preserving nonlinear parallel MRI reconstruction methods. Copyright © 2013 Wiley Periodicals, Inc.

  19. A dental public health approach based on computational mathematics: Monte Carlo simulation of childhood dental decay.

    Science.gov (United States)

    Tennant, Marc; Kruger, Estie

    2013-02-01

    This study developed a Monte Carlo simulation approach to examining the prevalence and incidence of dental decay using Australian children as a test environment. Monte Carlo simulation has been used for a half a century in particle physics (and elsewhere); put simply, it is the probability for various population-level outcomes seeded randomly to drive the production of individual level data. A total of five runs of the simulation model for all 275,000 12-year-olds in Australia were completed based on 2005-2006 data. Measured on average decayed/missing/filled teeth (DMFT) and DMFT of highest 10% of sample (Sic10) the runs did not differ from each other by more than 2% and the outcome was within 5% of the reported sampled population data. The simulations rested on the population probabilities that are known to be strongly linked to dental decay, namely, socio-economic status and Indigenous heritage. Testing the simulated population found DMFT of all cases where DMFT0 was 2.3 (n = 128,609) and DMFT for Indigenous cases only was 1.9 (n = 13,749). In the simulation population the Sic25 was 3.3 (n = 68,750). Monte Carlo simulations were created in particle physics as a computational mathematical approach to unknown individual-level effects by resting a simulation on known population-level probabilities. In this study a Monte Carlo simulation approach to childhood dental decay was built, tested and validated. © 2013 FDI World Dental Federation.

  20. Proton exchange in acid-base complexes induced by reaction coordinates with heavy atom motions

    Energy Technology Data Exchange (ETDEWEB)

    Alavi, Saman, E-mail: saman.alavi@nrc-cnrc.gc.ca [Steacie Institute for Molecular Sciences, National Research Council of Canada, Ontario, K1A 0R6 (Canada); Taghikhani, Mahdi [Department of Chemistry, Sharif Institute of Technology, Tehran (Iran, Islamic Republic of)

    2012-06-19

    Highlights: Black-Right-Pointing-Pointer Proton exchange in acid-base complexes is studied. Black-Right-Pointing-Pointer The structures, binding energies, and normal mode vibrations are calculated. Black-Right-Pointing-Pointer Transition state structures of proton exchange mechanism are determined. Black-Right-Pointing-Pointer In the complexes studied, the reaction coordinate involves heavy atom rocking. Black-Right-Pointing-Pointer The reaction coordinate is not simply localized in the proton movements. - Abstract: We extend previous work on nitric acid-ammonia and nitric acid-alkylamine complexes to illustrate that proton exchange reaction coordinates involve the rocking motion of the base moiety in many double hydrogen-bonded gas phase strong acid-strong base complexes. The complexes studied involve the biologically and atmospherically relevant glycine, formic, acetic, propionic, and sulfuric acids with ammonia/alkylamine bases. In these complexes, the magnitude of the imaginary frequencies associated with the proton exchange transition states are <400 cm{sup -1}. This contrasts with widely studied proton exchange reactions between symmetric carboxylic acid dimers or asymmetric DNA base pair and their analogs where the reaction coordinate is localized in proton motions and the magnitude of the imaginary frequencies for the transition states are >1100 cm{sup -1}. Calculations on complexes of these acids with water are performed for comparison. Variations of normal vibration modes along the reaction coordinate in the complexes are described.

  1. Proton exchange in acid–base complexes induced by reaction coordinates with heavy atom motions

    International Nuclear Information System (INIS)

    Alavi, Saman; Taghikhani, Mahdi

    2012-01-01

    Highlights: ► Proton exchange in acid–base complexes is studied. ► The structures, binding energies, and normal mode vibrations are calculated. ► Transition state structures of proton exchange mechanism are determined. ► In the complexes studied, the reaction coordinate involves heavy atom rocking. ► The reaction coordinate is not simply localized in the proton movements. - Abstract: We extend previous work on nitric acid–ammonia and nitric acid–alkylamine complexes to illustrate that proton exchange reaction coordinates involve the rocking motion of the base moiety in many double hydrogen-bonded gas phase strong acid–strong base complexes. The complexes studied involve the biologically and atmospherically relevant glycine, formic, acetic, propionic, and sulfuric acids with ammonia/alkylamine bases. In these complexes, the magnitude of the imaginary frequencies associated with the proton exchange transition states are −1 . This contrasts with widely studied proton exchange reactions between symmetric carboxylic acid dimers or asymmetric DNA base pair and their analogs where the reaction coordinate is localized in proton motions and the magnitude of the imaginary frequencies for the transition states are >1100 cm −1 . Calculations on complexes of these acids with water are performed for comparison. Variations of normal vibration modes along the reaction coordinate in the complexes are described.

  2. Proton response of alanine based pellets and films

    Energy Technology Data Exchange (ETDEWEB)

    Onori, S.; De Angelis, C.; Fattibene, P. [Istituto Superiore di Sanita, Rome (Italy). Lab. di Fisica]|[Istituto Nazionale di Fisica Nucleare, Rome (Italy). Sezione Sanita; D`Errico, F. [DCMN, Universita degli Studi di Pisa (Italy); Egger, E. [Paul Scherrer Inst. (PSI), Villigen (Switzerland); Janovsky, I. [Nuclear Research Institute, Rez (Czech Republic)

    1997-01-01

    The performance of two different alanine-ESR systems (1 mm ISS pellets and NRI films) in therapy proton beams was checked against reference dosimetry provided by a Markus parallel plate ionization chamber. Irradiations were performed at the OPTIS facility of the PSI with a 62 MeV proton beam. The energy dependence of the response was derived from the analysis of film stacks irradiated with pristine beams. In fact, 250 {mu}m films allowed for a high resolution sampling of the proton slowing down mechanisms. Moreover pellets were irradiated in unmodulated and modulated beams. Alanine dose distributions were always in good agreement with the PSI Markus chamber results. Simple density scaling allowed for the conversion of the results to the dose distributions in water, consistently with the good tissue equivalence properties of these detectors. (Author).

  3. The future of new calculation concepts in dosimetry based on the Monte Carlo Methods

    International Nuclear Information System (INIS)

    Makovicka, L.; Vasseur, A.; Sauget, M.; Martin, E.; Gschwind, R.; Henriet, J.; Vasseur, A.; Sauget, M.; Martin, E.; Gschwind, R.; Henriet, J.; Salomon, M.

    2009-01-01

    Monte Carlo codes, precise but slow, are very important tools in the vast majority of specialities connected to Radiation Physics, Radiation Protection and Dosimetry. A discussion about some other computing solutions is carried out; solutions not only based on the enhancement of computer power, or on the 'biasing'used for relative acceleration of these codes (in the case of photons), but on more efficient methods (A.N.N. - artificial neural network, C.B.R. - case-based reasoning - or other computer science techniques) already and successfully used for a long time in other scientific or industrial applications and not only Radiation Protection or Medical Dosimetry. (authors)

  4. Energy-based truncation of multi-determinant wavefunctions in quantum Monte Carlo.

    Science.gov (United States)

    Per, Manolo C; Cleland, Deidre M

    2017-04-28

    We present a method for truncating large multi-determinant expansions for use in diffusion Monte Carlo calculations. Current approaches use wavefunction-based criteria to perform the truncation. Our method is more intuitively based on the contribution each determinant makes to the total energy. We show that this approach gives consistent behaviour across systems with varying correlation character, which leads to effective error cancellation in energy differences. This is demonstrated through accurate calculations of the electron affinity of oxygen and the atomisation energy of the carbon dimer. The approach is simple and easy to implement, requiring only quantities already accessible in standard configuration interaction calculations.

  5. Monte Carlo simulation optimisation of zinc sulphide based fast-neutron detector for radiography using a {sup 252}Cf source

    Energy Technology Data Exchange (ETDEWEB)

    Meshkian, Mohsen, E-mail: mohsenm@ethz.ch

    2016-02-01

    Neutron radiography is rapidly extending as one of the methods for non-destructive screening of materials. There are various parameters to be studied for optimising imaging screens and image quality for different fast-neutron radiography systems. Herein, a Geant4 Monte Carlo simulation is employed to evaluate the response of a fast-neutron radiography system using a {sup 252}Cf neutron source. The neutron radiography system is comprised of a moderator as the neutron-to-proton converter with suspended silver-activated zinc sulphide (ZnS(Ag)) as the phosphor material. The neutron-induced protons deposit energy in the phosphor which consequently emits scintillation light. Further, radiographs are obtained by simulating the overall radiography system including source and sample. Two different standard samples are used to evaluate the quality of the radiographs.

  6. Oxide-based protonic conductors: Point defects and transport properties

    DEFF Research Database (Denmark)

    Bonanos, N.

    2001-01-01

    , hydrogen pumps, fuel cells, etc. The extent to which protonic defects form depends mainly on the partial pressure of water vapour, temperature and basicity of the constituent oxides, while their mobility depends, among other factors, on the metal-oxygen bond length and bond energy. The defect equilibria...

  7. Monte Carlo tests of the Rasch model based on scalability coefficients

    DEFF Research Database (Denmark)

    Christensen, Karl Bang; Kreiner, Svend

    2010-01-01

    that summarizes the number of Guttman errors in the data matrix. These coefficients are shown to yield efficient tests of the Rasch model using p-values computed using Markov chain Monte Carlo methods. The power of the tests of unequal item discrimination, and their ability to distinguish between local dependence......For item responses fitting the Rasch model, the assumptions underlying the Mokken model of double monotonicity are met. This makes non-parametric item response theory a natural starting-point for Rasch item analysis. This paper studies scalability coefficients based on Loevinger's H coefficient...

  8. Channel capacity study of underwater wireless optical communications links based on Monte Carlo simulation

    International Nuclear Information System (INIS)

    Li, Jing; Ma, Yong; Zhou, Qunqun; Zhou, Bo; Wang, Hongyuan

    2012-01-01

    Channel capacity of ocean water is limited by propagation distance and optical properties. Previous studies on this problem are based on water-tank experiments with different amounts of Maalox antacid. However, propagation distance is limited by the experimental set-up and the optical properties are different from ocean water. Therefore, the experiment result is not accurate for the physical design of underwater wireless communications links. This letter developed a Monte Carlo model to study channel capacity of underwater optical communications. Moreover, this model can flexibly configure various parameters of transmitter, receiver and channel, and is suitable for physical underwater optical communications links design. (paper)

  9. Monte Carlo simulation based reliability evaluation in a multi-bilateral contracts market

    International Nuclear Information System (INIS)

    Goel, L.; Viswanath, P.A.; Wang, P.

    2004-01-01

    This paper presents a time sequential Monte Carlo simulation technique to evaluate customer load point reliability in multi-bilateral contracts market. The effects of bilateral transactions, reserve agreements, and the priority commitments of generating companies on customer load point reliability have been investigated. A generating company with bilateral contracts is modelled as an equivalent time varying multi-state generation (ETMG). A procedure to determine load point reliability based on ETMG has been developed. The developed procedure is applied to a reliability test system to illustrate the technique. Representing each bilateral contract by an ETMG provides flexibility in determining the reliability at various customer load points. (authors)

  10. Channel capacity study of underwater wireless optical communications links based on Monte Carlo simulation

    Science.gov (United States)

    Li, Jing; Ma, Yong; Zhou, Qunqun; Zhou, Bo; Wang, Hongyuan

    2012-01-01

    Channel capacity of ocean water is limited by propagation distance and optical properties. Previous studies on this problem are based on water-tank experiments with different amounts of Maalox antacid. However, propagation distance is limited by the experimental set-up and the optical properties are different from ocean water. Therefore, the experiment result is not accurate for the physical design of underwater wireless communications links. This letter developed a Monte Carlo model to study channel capacity of underwater optical communications. Moreover, this model can flexibly configure various parameters of transmitter, receiver and channel, and is suitable for physical underwater optical communications links design.

  11. Geant4 based Monte Carlo simulation for verifying the modified sum-peak method.

    Science.gov (United States)

    Aso, Tsukasa; Ogata, Yoshimune; Makino, Ryuta

    2018-04-01

    The modified sum-peak method can practically estimate radioactivity by using solely the peak and the sum peak count rate. In order to efficiently verify the method in various experimental conditions, a Geant4 based Monte Carlo simulation for a high-purity germanium detector system was applied. The energy spectra in the detector were simulated for a 60 Co point source in various source to detector distances. The calculated radioactivity shows good agreement with the number of decays in the simulation. Copyright © 2017 Elsevier Ltd. All rights reserved.

  12. Considerable variation in NNT - A study based on Monte Carlo simulations

    DEFF Research Database (Denmark)

    Wisloff, T.; Aalen, O. O.; Sønbø Kristiansen, Ivar

    2011-01-01

    Objective: The aim of this analysis was to explore the variation in measures of effect, such as the number-needed-to-treat (NNT) and the relative risk (RR). Study Design and Setting: We performed Monte Carlo simulations of therapies using binominal distributions based on different true absolute...... is used to express treatment effectiveness, it has a regular distribution around the expected value for various values of true ARR, n, and p(0). The equivalent distribution of NNT is by definition nonconnected at zero and is also irregular. The probability that the observed treatment effectiveness is zero...

  13. A new anhydrous proton conductor based on polybenzimidazole and tridecyl phosphate

    International Nuclear Information System (INIS)

    Jiang Fengjing; Pu Hongting; Meyer, Wolfgang H.; Guan Yisi; Wan Decheng

    2008-01-01

    Most of the anhydrous proton conducting membranes are based on inorganic or partially inorganic materials, like SrCeO 3 membranes or polybenzimidazole (PBI)/H 3 PO 4 composite membranes. In present work, a new kind of anhydrous proton conducting membrane based on fully organic components of PBI and tridecyl phosphate (TP) was prepared. The interaction between PBI and TP is discussed. The temperature dependence of the proton conductivity of the composite membranes can be modeled by an Arrhenius relation. Thermogravimetric analysis (TGA) illustrates that these composite membranes are chemically stable up to 145 deg. C. The weight loss appearing at 145 deg. C is attributed to the selfcondensation of phosphate, which results in the proton conductivity drop of the membranes occurring at the same temperature. The DC conductivity of the composite membranes can reach ∼10 -4 S/cm for PBI/1.8TP at 140 deg. C and increases with increasing TP content. The proton conductivity of PBI/TP and PBI/H 3 PO 4 composite membranes is compared. The former have higher proton conductivity, however, the proton conductivity of the PBI/H 3 PO 4 membranes increases with temperature more significantly. Compared with PBI/H 3 PO 4 membranes, the migration stability of TP in PBI/TP membranes is improved significantly

  14. PeneloPET, a Monte Carlo PET simulation tool based on PENELOPE: features and validation

    Energy Technology Data Exchange (ETDEWEB)

    Espana, S; Herraiz, J L; Vicente, E; Udias, J M [Grupo de Fisica Nuclear, Departmento de Fisica Atomica, Molecular y Nuclear, Universidad Complutense de Madrid, Madrid (Spain); Vaquero, J J; Desco, M [Unidad de Medicina y CirugIa Experimental, Hospital General Universitario Gregorio Maranon, Madrid (Spain)], E-mail: jose@nuc2.fis.ucm.es

    2009-03-21

    Monte Carlo simulations play an important role in positron emission tomography (PET) imaging, as an essential tool for the research and development of new scanners and for advanced image reconstruction. PeneloPET, a PET-dedicated Monte Carlo tool, is presented and validated in this work. PeneloPET is based on PENELOPE, a Monte Carlo code for the simulation of the transport in matter of electrons, positrons and photons, with energies from a few hundred eV to 1 GeV. PENELOPE is robust, fast and very accurate, but it may be unfriendly to people not acquainted with the FORTRAN programming language. PeneloPET is an easy-to-use application which allows comprehensive simulations of PET systems within PENELOPE. Complex and realistic simulations can be set by modifying a few simple input text files. Different levels of output data are available for analysis, from sinogram and lines-of-response (LORs) histogramming to fully detailed list mode. These data can be further exploited with the preferred programming language, including ROOT. PeneloPET simulates PET systems based on crystal array blocks coupled to photodetectors and allows the user to define radioactive sources, detectors, shielding and other parts of the scanner. The acquisition chain is simulated in high level detail; for instance, the electronic processing can include pile-up rejection mechanisms and time stamping of events, if desired. This paper describes PeneloPET and shows the results of extensive validations and comparisons of simulations against real measurements from commercial acquisition systems. PeneloPET is being extensively employed to improve the image quality of commercial PET systems and for the development of new ones.

  15. Interaction of 14 MeV neutrons with hydrogenated target proton emission calculation

    International Nuclear Information System (INIS)

    Martin, G.; Perez, N.; Desdin.

    1996-01-01

    Using neutron emission data of a 14 MeV neutron generator, a paraffin target, and based on the n + H 1 → n '+ p reaction, have been obtained the characteristics of the proton emission in a proton-neutron mixed field. It was used Monte Carlo simulation and it was obtained the proton output as function of the converter width and the energy spectrum of protons corresponding to different converter thickness. Among 0.07 and 0.2 cm there is a maximum zone for the proton emission. The energy spectrum agrees with obtained on previous papers. Figures showing these results are provided

  16. X-ray imaging plate performance investigation based on a Monte Carlo simulation tool

    Energy Technology Data Exchange (ETDEWEB)

    Yao, M., E-mail: philippe.duvauchelle@insa-lyon.fr [Laboratoire Vibration Acoustique (LVA), INSA de Lyon, 25 Avenue Jean Capelle, 69621 Villeurbanne Cedex (France); Duvauchelle, Ph.; Kaftandjian, V. [Laboratoire Vibration Acoustique (LVA), INSA de Lyon, 25 Avenue Jean Capelle, 69621 Villeurbanne Cedex (France); Peterzol-Parmentier, A. [AREVA NDE-Solutions, 4 Rue Thomas Dumorey, 71100 Chalon-sur-Saône (France); Schumm, A. [EDF R& D SINETICS, 1 Avenue du Général de Gaulle, 92141 Clamart Cedex (France)

    2015-01-01

    Computed radiography (CR) based on imaging plate (IP) technology represents a potential replacement technique for traditional film-based industrial radiography. For investigating the IP performance especially at high energies, a Monte Carlo simulation tool based on PENELOPE has been developed. This tool tracks separately direct and secondary radiations, and monitors the behavior of different particles. The simulation output provides 3D distribution of deposited energy in IP and evaluation of radiation spectrum propagation allowing us to visualize the behavior of different particles and the influence of different elements. A detailed analysis, on the spectral and spatial responses of IP at different energies up to MeV, has been performed. - Highlights: • A Monte Carlo tool for imaging plate (IP) performance investigation is presented. • The tool outputs 3D maps of energy deposition in IP due to different signals. • The tool also provides the transmitted spectra along the radiation propagation. • An industrial imaging case is simulated with the presented tool. • A detailed analysis, on the spectral and spatial responses of IP, is presented.

  17. Proton-conducting polymer electrolytes based on methacrylates

    Czech Academy of Sciences Publication Activity Database

    Reiter, Jakub; Velická, Jana; Míka, M.

    2008-01-01

    Roč. 53, č. 26 (2008), s. 7769-7774 ISSN 0013-4686 R&D Projects: GA ČR GA106/04/1279; GA AV ČR KJB400320701; GA MŠk LC523; GA ČR(CZ) GA104/06/1471 Institutional research plan: CEZ:AV0Z40320502 Keywords : polymer electrolyte * proton conductivity * phosporic acid Subject RIV: CA - Inorganic Chemistry Impact factor: 3.078, year: 2008

  18. Monte Carlo based, patient-specific RapidArc QA using Linac log files.

    Science.gov (United States)

    Teke, Tony; Bergman, Alanah M; Kwa, William; Gill, Bradford; Duzenli, Cheryl; Popescu, I Antoniu

    2010-01-01

    A Monte Carlo (MC) based QA process to validate the dynamic beam delivery accuracy for Varian RapidArc (Varian Medical Systems, Palo Alto, CA) using Linac delivery log files (DynaLog) is presented. Using DynaLog file analysis and MC simulations, the goal of this article is to (a) confirm that adequate sampling is used in the RapidArc optimization algorithm (177 static gantry angles) and (b) to assess the physical machine performance [gantry angle and monitor unit (MU) delivery accuracy]. Ten clinically acceptable RapidArc treatment plans were generated for various tumor sites and delivered to a water-equivalent cylindrical phantom on the treatment unit. Three Monte Carlo simulations were performed to calculate dose to the CT phantom image set: (a) One using a series of static gantry angles defined by 177 control points with treatment planning system (TPS) MLC control files (planning files), (b) one using continuous gantry rotation with TPS generated MLC control files, and (c) one using continuous gantry rotation with actual Linac delivery log files. Monte Carlo simulated dose distributions are compared to both ionization chamber point measurements and with RapidArc TPS calculated doses. The 3D dose distributions were compared using a 3D gamma-factor analysis, employing a 3%/3 mm distance-to-agreement criterion. The dose difference between MC simulations, TPS, and ionization chamber point measurements was less than 2.1%. For all plans, the MC calculated 3D dose distributions agreed well with the TPS calculated doses (gamma-factor values were less than 1 for more than 95% of the points considered). Machine performance QA was supplemented with an extensive DynaLog file analysis. A DynaLog file analysis showed that leaf position errors were less than 1 mm for 94% of the time and there were no leaf errors greater than 2.5 mm. The mean standard deviation in MU and gantry angle were 0.052 MU and 0.355 degrees, respectively, for the ten cases analyzed. The accuracy and

  19. CARMEN: a system Monte Carlo based on linear programming from direct openings; CARMEN: Un sistema de planficiacion Monte Carlo basado en programacion lineal a partir de aberturas directas

    Energy Technology Data Exchange (ETDEWEB)

    Ureba, A.; Pereira-Barbeiro, A. R.; Jimenez-Ortega, E.; Baeza, J. A.; Salguero, F. J.; Leal, A.

    2013-07-01

    The use of Monte Carlo (MC) has shown an improvement in the accuracy of the calculation of the dose compared to other analytics algorithms installed on the systems of business planning, especially in the case of non-standard situations typical of complex techniques such as IMRT and VMAT. Our treatment planning system called CARMEN, is based on the complete simulation, both the beam transport in the head of the accelerator and the patient, and simulation designed for efficient operation in terms of the accuracy of the estimate and the required computation times. (Author)

  20. The Quest for Evidence for Proton Therapy: Model-Based Approach and Precision Medicine

    Energy Technology Data Exchange (ETDEWEB)

    Widder, Joachim, E-mail: j.widder@umcg.nl [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Schaaf, Arjen van der [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands); Lambin, Philippe [Department of Radiation Oncology, School for Oncology and Developmental Biology (GROW), Maastricht University Medical Center, Maastricht (Netherlands); Marijnen, Corrie A.M. [Department of Radiation Oncology, Leiden University Medical Center, Leiden (Netherlands); Pignol, Jean-Philippe [Department of Radiation Oncology, Erasmus Medical Center Cancer Institute, Rotterdam (Netherlands); Rasch, Coen R. [Department of Radiation Oncology, Academic Medical Center, Amsterdam (Netherlands); Slotman, Ben J. [Department of Radiation Oncology, VU Medical Center, Amsterdam (Netherlands); Verheij, Marcel [Department of Radiation Oncology, Netherlands Cancer Institute, Amsterdam (Netherlands); Langendijk, Johannes A. [Department of Radiation Oncology, University of Groningen, University Medical Center Groningen, Groningen (Netherlands)

    2016-05-01

    Purpose: Reducing dose to normal tissues is the advantage of protons versus photons. We aimed to describe a method for translating this reduction into a clinically relevant benefit. Methods and Materials: Dutch scientific and health care governance bodies have recently issued landmark reports regarding generation of relevant evidence for new technologies in health care including proton therapy. An approach based on normal tissue complication probability (NTCP) models has been adopted to select patients who are most likely to experience fewer (serious) adverse events achievable by state-of-the-art proton treatment. Results: By analogy with biologically targeted therapies, the technology needs to be tested in enriched cohorts of patients exhibiting the decisive predictive marker: difference in normal tissue dosimetric signatures between proton and photon treatment plans. Expected clinical benefit is then estimated by virtue of multifactorial NTCP models. In this sense, high-tech radiation therapy falls under precision medicine. As a consequence, randomizing nonenriched populations between photons and protons is predictably inefficient and likely to produce confusing results. Conclusions: Validating NTCP models in appropriately composed cohorts treated with protons should be the primary research agenda leading to urgently needed evidence for proton therapy.

  1. IMRT head and neck treatment planning with a commercially available Monte Carlo based planning system

    International Nuclear Information System (INIS)

    Boudreau, C; Heath, E; Seuntjens, J; Ballivy, O; Parker, W

    2005-01-01

    The PEREGRINE Monte Carlo dose-calculation system (North American Scientific, Cranberry Township, PA) is the first commercially available Monte Carlo dose-calculation code intended specifically for intensity modulated radiotherapy (IMRT) treatment planning and quality assurance. In order to assess the impact of Monte Carlo based dose calculations for IMRT clinical cases, dose distributions for 11 head and neck patients were evaluated using both PEREGRINE and the CORVUS (North American Scientific, Cranberry Township, PA) finite size pencil beam (FSPB) algorithm with equivalent path-length (EPL) inhomogeneity correction. For the target volumes, PEREGRINE calculations predict, on average, a less than 2% difference in the calculated mean and maximum doses to the gross tumour volume (GTV) and clinical target volume (CTV). An average 16% ± 4% and 12% ± 2% reduction in the volume covered by the prescription isodose line was observed for the GTV and CTV, respectively. Overall, no significant differences were noted in the doses to the mandible and spinal cord. For the parotid glands, PEREGRINE predicted a 6% ± 1% increase in the volume of tissue receiving a dose greater than 25 Gy and an increase of 4% ± 1% in the mean dose. Similar results were noted for the brainstem where PEREGRINE predicted a 6% ± 2% increase in the mean dose. The observed differences between the PEREGRINE and CORVUS calculated dose distributions are attributed to secondary electron fluence perturbations, which are not modelled by the EPL correction, issues of organ outlining, particularly in the vicinity of air cavities, and differences in dose reporting (dose to water versus dose to tissue type)

  2. Monte Carlo simulation as a tool to predict blasting fragmentation based on the Kuz Ram model

    Science.gov (United States)

    Morin, Mario A.; Ficarazzo, Francesco

    2006-04-01

    Rock fragmentation is considered the most important aspect of production blasting because of its direct effects on the costs of drilling and blasting and on the economics of the subsequent operations of loading, hauling and crushing. Over the past three decades, significant progress has been made in the development of new technologies for blasting applications. These technologies include increasingly sophisticated computer models for blast design and blast performance prediction. Rock fragmentation depends on many variables such as rock mass properties, site geology, in situ fracturing and blasting parameters and as such has no complete theoretical solution for its prediction. However, empirical models for the estimation of size distribution of rock fragments have been developed. In this study, a blast fragmentation Monte Carlo-based simulator, based on the Kuz-Ram fragmentation model, has been developed to predict the entire fragmentation size distribution, taking into account intact and joints rock properties, the type and properties of explosives and the drilling pattern. Results produced by this simulator were quite favorable when compared with real fragmentation data obtained from a blast quarry. It is anticipated that the use of Monte Carlo simulation will increase our understanding of the effects of rock mass and explosive properties on the rock fragmentation by blasting, as well as increase our confidence in these empirical models. This understanding will translate into improvements in blasting operations, its corresponding costs and the overall economics of open pit mines and rock quarries.

  3. Comment on “A study on tetrahedron-based inhomogeneous Monte-Carlo optical simulation”

    Science.gov (United States)

    Fang, Qianqian

    2011-01-01

    The Monte Carlo (MC) method is a popular approach to modeling photon propagation inside general turbid media, such as human tissue. Progress had been made in the past year with the independent proposals of two mesh-based Monte Carlo methods employing ray-tracing techniques. Both methods have shown improvements in accuracy and efficiency in modeling complex domains. A recent paper by Shen and Wang [Biomed. Opt. Express 2, 44 (2011)] reported preliminary results towards the cross-validation of the two mesh-based MC algorithms and software implementations, showing a 3–6 fold speed difference between the two software packages. In this comment, we share our views on unbiased software comparisons and discuss additional issues such as the use of pre-computed data, interpolation strategies, impact of compiler settings, use of Russian roulette, memory cost and potential pitfalls in measuring algorithm performance. Despite key differences between the two algorithms in handling of non-tetrahedral meshes, we found that they share similar structure and performance for tetrahedral meshes. A significant fraction of the observed speed differences in the mentioned article was the result of inconsistent use of compilers and libraries. PMID:21559136

  4. Comment on "A study on tetrahedron-based inhomogeneous Monte-Carlo optical simulation".

    Science.gov (United States)

    Fang, Qianqian

    2011-04-19

    The Monte Carlo (MC) method is a popular approach to modeling photon propagation inside general turbid media, such as human tissue. Progress had been made in the past year with the independent proposals of two mesh-based Monte Carlo methods employing ray-tracing techniques. Both methods have shown improvements in accuracy and efficiency in modeling complex domains. A recent paper by Shen and Wang [Biomed. Opt. Express 2, 44 (2011)] reported preliminary results towards the cross-validation of the two mesh-based MC algorithms and software implementations, showing a 3-6 fold speed difference between the two software packages. In this comment, we share our views on unbiased software comparisons and discuss additional issues such as the use of pre-computed data, interpolation strategies, impact of compiler settings, use of Russian roulette, memory cost and potential pitfalls in measuring algorithm performance. Despite key differences between the two algorithms in handling of non-tetrahedral meshes, we found that they share similar structure and performance for tetrahedral meshes. A significant fraction of the observed speed differences in the mentioned article was the result of inconsistent use of compilers and libraries.

  5. CAD-based Monte Carlo automatic modeling method based on primitive solid

    International Nuclear Information System (INIS)

    Wang, Dong; Song, Jing; Yu, Shengpeng; Long, Pengcheng; Wang, Yongliang

    2016-01-01

    Highlights: • We develop a method which bi-convert between CAD model and primitive solid. • This method was improved from convert method between CAD model and half space. • This method was test by ITER model and validated the correctness and efficiency. • This method was integrated in SuperMC which could model for SuperMC and Geant4. - Abstract: Monte Carlo method has been widely used in nuclear design and analysis, where geometries are described with primitive solids. However, it is time consuming and error prone to describe a primitive solid geometry, especially for a complicated model. To reuse the abundant existed CAD models and conveniently model with CAD modeling tools, an automatic modeling method for accurate prompt modeling between CAD model and primitive solid is needed. An automatic modeling method for Monte Carlo geometry described by primitive solid was developed which could bi-convert between CAD model and Monte Carlo geometry represented by primitive solids. While converting from CAD model to primitive solid model, the CAD model was decomposed into several convex solid sets, and then corresponding primitive solids were generated and exported. While converting from primitive solid model to the CAD model, the basic primitive solids were created and related operation was done. This method was integrated in the SuperMC and was benchmarked with ITER benchmark model. The correctness and efficiency of this method were demonstrated.

  6. Pattern-oriented Agent-based Monte Carlo simulation of Cellular Redox Environment

    DEFF Research Database (Denmark)

    Tang, Jiaowei; Holcombe, Mike; Boonen, Harrie C.M.

    ] could be very important factors. In our project, an agent-based Monte Carlo modeling [6] is offered to study the dynamic relationship between extracellular and intracellular redox and complex networks of redox reactions. In the model, pivotal redox-related reactions will be included, and the reactants....../CYSS) and mitochondrial redox couples. Evidence suggests that both intracellular and extracellular redox can affect overall cell redox state. How redox is communicated between extracellular and intracellular environments is still a matter of debate. Some researchers conclude based on experimental data....... Because complex networks and dynamics of redox still is not completely understood , results of existing experiments will be used to validate the modeling according to ideas in pattern-oriented agent-based modeling[8]. The simulation of this model is computational intensive, thus an application 'FLAME...

  7. CAD-Based Monte Carlo Neutron Transport KSTAR Analysis for KSTAR

    Science.gov (United States)

    Seo, Geon Ho; Choi, Sung Hoon; Shim, Hyung Jin

    2017-09-01

    The Monte Carlo (MC) neutron transport analysis for a complex nuclear system such as fusion facility may require accurate modeling of its complicated geometry. In order to take advantage of modeling capability of the computer aided design (CAD) system for the MC neutronics analysis, the Seoul National University MC code, McCARD, has been augmented with a CAD-based geometry processing module by imbedding the OpenCASCADE CAD kernel. In the developed module, the CAD geometry data are internally converted to the constructive solid geometry model with help of the CAD kernel. An efficient cell-searching algorithm is devised for the void space treatment. The performance of the CAD-based McCARD calculations are tested for the Korea Superconducting Tokamak Advanced Research device by comparing with results of the conventional MC calculations using a text-based geometry input.

  8. An Approach in Radiation Therapy Treatment Planning: A Fast, GPU-Based Monte Carlo Method.

    Science.gov (United States)

    Karbalaee, Mojtaba; Shahbazi-Gahrouei, Daryoush; Tavakoli, Mohammad B

    2017-01-01

    An accurate and fast radiation dose calculation is essential for successful radiation radiotherapy. The aim of this study was to implement a new graphic processing unit (GPU) based radiation therapy treatment planning for accurate and fast dose calculation in radiotherapy centers. A program was written for parallel running based on GPU. The code validation was performed by EGSnrc/DOSXYZnrc. Moreover, a semi-automatic, rotary, asymmetric phantom was designed and produced using a bone, the lung, and the soft tissue equivalent materials. All measurements were performed using a Mapcheck dosimeter. The accuracy of the code was validated using the experimental data, which was obtained from the anthropomorphic phantom as the gold standard. The findings showed that, compared with those of DOSXYZnrc in the virtual phantom and for most of the voxels (>95%), GPU-based Monte Carlo method in dose calculation may be useful in routine radiation therapy centers as the core and main component of a treatment planning verification system.

  9. GPU-based high performance Monte Carlo simulation in neutron transport

    International Nuclear Information System (INIS)

    Heimlich, Adino; Mol, Antonio C.A.; Pereira, Claudio M.N.A.

    2009-01-01

    Graphics Processing Units (GPU) are high performance co-processors intended, originally, to improve the use and quality of computer graphics applications. Since researchers and practitioners realized the potential of using GPU for general purpose, their application has been extended to other fields out of computer graphics scope. The main objective of this work is to evaluate the impact of using GPU in neutron transport simulation by Monte Carlo method. To accomplish that, GPU- and CPU-based (single and multicore) approaches were developed and applied to a simple, but time-consuming problem. Comparisons demonstrated that the GPU-based approach is about 15 times faster than a parallel 8-core CPU-based approach also developed in this work. (author)

  10. GPU-based high performance Monte Carlo simulation in neutron transport

    Energy Technology Data Exchange (ETDEWEB)

    Heimlich, Adino; Mol, Antonio C.A.; Pereira, Claudio M.N.A. [Instituto de Engenharia Nuclear (IEN/CNEN-RJ), Rio de Janeiro, RJ (Brazil). Lab. de Inteligencia Artificial Aplicada], e-mail: cmnap@ien.gov.br

    2009-07-01

    Graphics Processing Units (GPU) are high performance co-processors intended, originally, to improve the use and quality of computer graphics applications. Since researchers and practitioners realized the potential of using GPU for general purpose, their application has been extended to other fields out of computer graphics scope. The main objective of this work is to evaluate the impact of using GPU in neutron transport simulation by Monte Carlo method. To accomplish that, GPU- and CPU-based (single and multicore) approaches were developed and applied to a simple, but time-consuming problem. Comparisons demonstrated that the GPU-based approach is about 15 times faster than a parallel 8-core CPU-based approach also developed in this work. (author)

  11. An intense neutron generator based on a proton accelerator

    Energy Technology Data Exchange (ETDEWEB)

    Bartholomew, G.A.; Milton, J.C.D.; Vogt, E.W

    1964-07-01

    A study has been made of the demand for a neutron facility with a thermal flux of {>=} 10{sup 16} n cm{sup -2} sec{sup -1} and of possible methods of producing such fluxes with existing or presently developing technology. Experimental projects proposed by neutron users requiring high fluxes call for neutrons of all energies from thermal to 100 MeV with both continuous-wave and pulsed output. Consideration of the heat generated in the source per useful neutron liberated shows that the (p,xn) reaction with 400 1000 MeV bombarding energies and heavy element targets (e.g. bismuth, lead) is capable of greater specific source strength than other possible methods realizable within the time scale. A preliminary parameter optimization carried through for the accelerator currently promising greatest economy (the separated orbit cyclotron or S.O.C.), reveals that a facility delivering a proton beam of about 65 mA at about 1 BeV would satisfy the flux requirement with a neutron cost significantly more favourable than that projected for a high flux reactor. It is suggested that a proton storage ring providing post-acceleration pulsing of the proton beam should be developed for the facility. With this elaboration, and by taking advantage of the intrinsic microscopic pulse structure provided by the radio frequency duty cycle, a very versatile source may be devised capable of producing multiple beams of continuous and pulsed neutrons with a wide range of energies and pulse widths. The source promises to be of great value for high flux irradiations and as a pilot facility for advanced reactor technology. The proposed proton accelerator also constitutes a meson source capable of producing beams of {pi} and {mu} mesons and of neutrinos orders of magnitude more intense than those of any accelerator presently in use. These beams, which can be produced simultaneously with the neutron beams, open vast areas of new research in fundamental nuclear structure, elementary particle physics

  12. Monte-Carlo-based uncertainty propagation with hierarchical models—a case study in dynamic torque

    Science.gov (United States)

    Klaus, Leonard; Eichstädt, Sascha

    2018-04-01

    For a dynamic calibration, a torque transducer is described by a mechanical model, and the corresponding model parameters are to be identified from measurement data. A measuring device for the primary calibration of dynamic torque, and a corresponding model-based calibration approach, have recently been developed at PTB. The complete mechanical model of the calibration set-up is very complex, and involves several calibration steps—making a straightforward implementation of a Monte Carlo uncertainty evaluation tedious. With this in mind, we here propose to separate the complete model into sub-models, with each sub-model being treated with individual experiments and analysis. The uncertainty evaluation for the overall model then has to combine the information from the sub-models in line with Supplement 2 of the Guide to the Expression of Uncertainty in Measurement. In this contribution, we demonstrate how to carry this out using the Monte Carlo method. The uncertainty evaluation involves various input quantities of different origin and the solution of a numerical optimisation problem.

  13. Monte Carlo closure for moment-based transport schemes in general relativistic radiation hydrodynamic simulations

    Science.gov (United States)

    Foucart, Francois

    2018-04-01

    General relativistic radiation hydrodynamic simulations are necessary to accurately model a number of astrophysical systems involving black holes and neutron stars. Photon transport plays a crucial role in radiatively dominated accretion discs, while neutrino transport is critical to core-collapse supernovae and to the modelling of electromagnetic transients and nucleosynthesis in neutron star mergers. However, evolving the full Boltzmann equations of radiative transport is extremely expensive. Here, we describe the implementation in the general relativistic SPEC code of a cheaper radiation hydrodynamic method that theoretically converges to a solution of Boltzmann's equation in the limit of infinite numerical resources. The algorithm is based on a grey two-moment scheme, in which we evolve the energy density and momentum density of the radiation. Two-moment schemes require a closure that fills in missing information about the energy spectrum and higher order moments of the radiation. Instead of the approximate analytical closure currently used in core-collapse and merger simulations, we complement the two-moment scheme with a low-accuracy Monte Carlo evolution. The Monte Carlo results can provide any or all of the missing information in the evolution of the moments, as desired by the user. As a first test of our methods, we study a set of idealized problems demonstrating that our algorithm performs significantly better than existing analytical closures. We also discuss the current limitations of our method, in particular open questions regarding the stability of the fully coupled scheme.

  14. A method based on Monte Carlo simulation for the determination of the G(E) function.

    Science.gov (United States)

    Chen, Wei; Feng, Tiancheng; Liu, Jun; Su, Chuanying; Tian, Yanjie

    2015-02-01

    The G(E) function method is a spectrometric method for the exposure dose estimation; this paper describes a method based on Monte Carlo method to determine the G(E) function of a 4″ × 4″ × 16″ NaI(Tl) detector. Simulated spectrums of various monoenergetic gamma rays in the region of 40 -3200 keV and the corresponding deposited energy in an air ball in the energy region of full-energy peak were obtained using Monte Carlo N-particle Transport Code. Absorbed dose rate in air was obtained according to the deposited energy and divided by counts of corresponding full-energy peak to get the G(E) function value at energy E in spectra. Curve-fitting software 1st0pt was used to determine coefficients of the G(E) function. Experimental results show that the calculated dose rates using the G(E) function determined by the authors' method are accordant well with those values obtained by ionisation chamber, with a maximum deviation of 6.31 %. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  15. CAD-based Monte Carlo Program for Integrated Simulation of Nuclear System SuperMC

    Science.gov (United States)

    Wu, Yican; Song, Jing; Zheng, Huaqing; Sun, Guangyao; Hao, Lijuan; Long, Pengcheng; Hu, Liqin

    2014-06-01

    Monte Carlo (MC) method has distinct advantages to simulate complicated nuclear systems and is envisioned as routine method for nuclear design and analysis in the future. High fidelity simulation with MC method coupled with multi-physical phenomenon simulation has significant impact on safety, economy and sustainability of nuclear systems. However, great challenges to current MC methods and codes prevent its application in real engineering project. SuperMC is a CAD-based Monte Carlo program for integrated simulation of nuclear system developed by FDS Team, China, making use of hybrid MC-deterministic method and advanced computer technologies. The design aim, architecture and main methodology of SuperMC were presented in this paper. SuperMC2.1, the latest version for neutron, photon and coupled neutron and photon transport calculation, has been developed and validated by using a series of benchmarking cases such as the fusion reactor ITER model and the fast reactor BN-600 model. SuperMC is still in its evolution process toward a general and routine tool for nuclear system. Warning, no authors found for 2014snam.conf06023.

  16. Comparison of nonstationary generalized logistic models based on Monte Carlo simulation

    Directory of Open Access Journals (Sweden)

    S. Kim

    2015-06-01

    Full Text Available Recently, the evidences of climate change have been observed in hydrologic data such as rainfall and flow data. The time-dependent characteristics of statistics in hydrologic data are widely defined as nonstationarity. Therefore, various nonstationary GEV and generalized Pareto models have been suggested for frequency analysis of nonstationary annual maximum and POT (peak-over-threshold data, respectively. However, the alternative models are required for nonstatinoary frequency analysis because of analyzing the complex characteristics of nonstationary data based on climate change. This study proposed the nonstationary generalized logistic model including time-dependent parameters. The parameters of proposed model are estimated using the method of maximum likelihood based on the Newton-Raphson method. In addition, the proposed model is compared by Monte Carlo simulation to investigate the characteristics of models and applicability.

  17. Monte Carlo-based dose calculation engine for minibeam radiation therapy.

    Science.gov (United States)

    Martínez-Rovira, I; Sempau, J; Prezado, Y

    2014-02-01

    Minibeam radiation therapy (MBRT) is an innovative radiotherapy approach based on the well-established tissue sparing effect of arrays of quasi-parallel micrometre-sized beams. In order to guide the preclinical trials in progress at the European Synchrotron Radiation Facility (ESRF), a Monte Carlo-based dose calculation engine has been developed and successfully benchmarked with experimental data in anthropomorphic phantoms. Additionally, a realistic example of treatment plan is presented. Despite the micron scale of the voxels used to tally dose distributions in MBRT, the combination of several efficiency optimisation methods allowed to achieve acceptable computation times for clinical settings (approximately 2 h). The calculation engine can be easily adapted with little or no programming effort to other synchrotron sources or for dose calculations in presence of contrast agents. Copyright © 2013 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.

  18. A comparison of two prompt gamma imaging techniques with collimator-based cameras for range verification in proton therapy

    Science.gov (United States)

    Lin, Hsin-Hon; Chang, Hao-Ting; Chao, Tsi-Chian; Chuang, Keh-Shih

    2017-08-01

    In vivo range verification plays an important role in proton therapy to fully utilize the benefits of the Bragg peak (BP) for delivering high radiation dose to tumor, while sparing the normal tissue. For accurately locating the position of BP, camera equipped with collimators (multi-slit and knife-edge collimator) to image prompt gamma (PG) emitted along the proton tracks in the patient have been proposed for range verification. The aim of the work is to compare the performance of multi-slit collimator and knife-edge collimator for non-invasive proton beam range verification. PG imaging was simulated by a validated GATE/GEANT4 Monte Carlo code to model the spot-scanning proton therapy and cylindrical PMMA phantom in detail. For each spot, 108 protons were simulated. To investigate the correlation between the acquired PG profile and the proton range, the falloff regions of PG profiles were fitted with a 3-line-segment curve function as the range estimate. Factors including the energy window setting, proton energy, phantom size, and phantom shift that may influence the accuracy of detecting range were studied. Results indicated that both collimator systems achieve reasonable accuracy and good response to the phantom shift. The accuracy of range predicted by multi-slit collimator system is less affected by the proton energy, while knife-edge collimator system can achieve higher detection efficiency that lead to a smaller deviation in predicting range. We conclude that both collimator systems have potentials for accurately range monitoring in proton therapy. It is noted that neutron contamination has a marked impact on range prediction of the two systems, especially in multi-slit system. Therefore, a neutron reduction technique for improving the accuracy of range verification of proton therapy is needed.

  19. The first proton sponge-based amino acids: synthesis, acid-base properties and some reactivity.

    Science.gov (United States)

    Ozeryanskii, Valery A; Gorbacheva, Anastasia Yu; Pozharskii, Alexander F; Vlasenko, Marina P; Tereznikov, Alexander Yu; Chernov'yants, Margarita S

    2015-08-21

    The first hybrid base constructed from 1,8-bis(dimethylamino)naphthalene (proton sponge or DMAN) and glycine, N-methyl-N-(8-dimethylamino-1-naphthyl)aminoacetic acid, was synthesised in high yield and its hydrobromide was structurally characterised and used to determine the acid-base properties via potentiometric titration. It was found that the basic strength of the DMAN-glycine base (pKa = 11.57, H2O) is on the level of amidine amino acids like arginine and creatine and its structure, zwitterionic vs. neutral, based on the spectroscopic (IR, NMR, mass) and theoretical (DFT) approaches has a strong preference to the zwitterionic form. Unlike glycine, the DMAN-glycine zwitterion is N-chiral and is hydrolytically cleaved with the loss of glycolic acid on heating in DMSO. This reaction together with the mild decarboxylative conversion of proton sponge-based amino acids into 2,3-dihydroperimidinium salts under air-oxygen was monitored with the help of the DMAN-alanine amino acid. The newly devised amino acids are unique as they combine fluorescence, strongly basic and redox-active properties.

  20. Modelling of scintillator based flat-panel detectors with Monte-Carlo simulations

    International Nuclear Information System (INIS)

    Reims, N; Sukowski, F; Uhlmann, N

    2011-01-01

    Scintillator based flat panel detectors are state of the art in the field of industrial X-ray imaging applications. Choosing the proper system and setup parameters for the vast range of different applications can be a time consuming task, especially when developing new detector systems. Since the system behaviour cannot always be foreseen easily, Monte-Carlo (MC) simulations are keys to gain further knowledge of system components and their behaviour for different imaging conditions. In this work we used two Monte-Carlo based models to examine an indirect converting flat panel detector, specifically the Hamamatsu C9312SK. We focused on the signal generation in the scintillation layer and its influence on the spatial resolution of the whole system. The models differ significantly in their level of complexity. The first model gives a global description of the detector based on different parameters characterizing the spatial resolution. With relatively small effort a simulation model can be developed which equates the real detector regarding signal transfer. The second model allows a more detailed insight of the system. It is based on the well established cascade theory, i.e. describing the detector as a cascade of elemental gain and scattering stages, which represent the built in components and their signal transfer behaviour. In comparison to the first model the influence of single components especially the important light spread behaviour in the scintillator can be analysed in a more differentiated way. Although the implementation of the second model is more time consuming both models have in common that a relatively small amount of system manufacturer parameters are needed. The results of both models were in good agreement with the measured parameters of the real system.

  1. GGEMS-Brachy: GPU GEant4-based Monte Carlo simulation for brachytherapy applications

    International Nuclear Information System (INIS)

    Lemaréchal, Yannick; Bert, Julien; Schick, Ulrike; Pradier, Olivier; Garcia, Marie-Paule; Boussion, Nicolas; Visvikis, Dimitris; Falconnet, Claire; Després, Philippe; Valeri, Antoine

    2015-01-01

    In brachytherapy, plans are routinely calculated using the AAPM TG43 formalism which considers the patient as a simple water object. An accurate modeling of the physical processes considering patient heterogeneity using Monte Carlo simulation (MCS) methods is currently too time-consuming and computationally demanding to be routinely used. In this work we implemented and evaluated an accurate and fast MCS on Graphics Processing Units (GPU) for brachytherapy low dose rate (LDR) applications. A previously proposed Geant4 based MCS framework implemented on GPU (GGEMS) was extended to include a hybrid GPU navigator, allowing navigation within voxelized patient specific images and analytically modeled 125 I seeds used in LDR brachytherapy. In addition, dose scoring based on track length estimator including uncertainty calculations was incorporated. The implemented GGEMS-brachy platform was validated using a comparison with Geant4 simulations and reference datasets. Finally, a comparative dosimetry study based on the current clinical standard (TG43) and the proposed platform was performed on twelve prostate cancer patients undergoing LDR brachytherapy. Considering patient 3D CT volumes of 400  × 250  × 65 voxels and an average of 58 implanted seeds, the mean patient dosimetry study run time for a 2% dose uncertainty was 9.35 s (≈500 ms 10 −6 simulated particles) and 2.5 s when using one and four GPUs, respectively. The performance of the proposed GGEMS-brachy platform allows envisaging the use of Monte Carlo simulation based dosimetry studies in brachytherapy compatible with clinical practice. Although the proposed platform was evaluated for prostate cancer, it is equally applicable to other LDR brachytherapy clinical applications. Future extensions will allow its application in high dose rate brachytherapy applications. (paper)

  2. Proton radiography to improve proton therapy treatment

    NARCIS (Netherlands)

    Takatsu, J.; van der Graaf, E. R.; van Goethem, Marc-Jan; van Beuzekom, M.; Klaver, T.; Visser, Jan; Brandenburg, S.; Biegun, A. K.

    The quality of cancer treatment with protons critically depends on an accurate prediction of the proton stopping powers for the tissues traversed by the protons. Today, treatment planning in proton radiotherapy is based on stopping power calculations from densities of X-ray Computed Tomography (CT)

  3. Proton exchange between oxymethyl radical and acids and bases: semiempirical quantum-chemical study

    Directory of Open Access Journals (Sweden)

    Irina Pustolaikina

    2016-12-01

    Full Text Available The reactions with proton participation are widely represented in the analytical, technological and biological chemistry. Quantum-chemical study of the exchange processes in hydrogen bonding complexes will allow us to achieve progress in the understanding of the elementary act mechanism of proton transfer in hydrogen bonding chain as well as the essence of the acid-base interactions. Oxymethyl radical •CH2ОН is small in size and comfortable as a model particle that well transmits protolytic properties of paramagnetic acids having more complex structure. Quantum-chemical modeling of proton exchange reaction oxymethyl radical ∙CH2OH and its diamagnetic analog CH3OH with amines, carboxylic acids and water was carried out using UAM1 method with the help of Gaussian-2009 program. QST2 method was used for the search of transition state, IRC procedure was applied for the calculation of descents along the reaction coordinate. The difference in the structure of transition states of ∙CH2OH/ CH3OH with bases and acids has been shown. It has been confirmed that in the case of bases, consecutive proton exchange mechanism was fixed, and in the case of complexes with carboxylic acids parallel proton exchange mechanism was fixed. The similarity in the reaction behavior of paramagnetic and diamagnetic systems in the proton exchange has been found. It was suggested that the mechanism of proton exchange reaction is determined by the structure of the hydrogen bonding cyclic complex, which is, in turn, depends from the nature of the acid-base interactions partners.

  4. His166 is the Schiff base proton acceptor in attractant phototaxis receptor sensory rhodopsin I.

    Science.gov (United States)

    Sasaki, Jun; Takahashi, Hazuki; Furutani, Yuji; Sineshchekov, Oleg A; Spudich, John L; Kandori, Hideki

    2014-09-23

    Photoactivation of attractant phototaxis receptor sensory rhodopsin I (SRI) in Halobacterium salinarum entails transfer of a proton from the retinylidene chromophore's Schiff base (SB) to an unidentified acceptor residue on the cytoplasmic half-channel, in sharp contrast to other microbial rhodopsins, including the closely related repellent phototaxis receptor SRII and the outward proton pump bacteriorhodopsin, in which the SB proton acceptor is an aspartate residue salt-bridged to the SB in the extracellular (EC) half-channel. His166 on the cytoplasmic side of the SB in SRI has been implicated in the SB proton transfer reaction by mutation studies, and mutants of His166 result in an inverted SB proton release to the EC as well as inversion of the protein's normally attractant phototaxis signal to repellent. Here we found by difference Fourier transform infrared spectroscopy the appearance of Fermi-resonant X-H stretch modes in light-minus-dark difference spectra; their assignment with (15)N labeling and site-directed mutagenesis demonstrates that His166 is the SB proton acceptor during the photochemical reaction cycle of the wild-type SRI-HtrI complex.

  5. Noninvasive spectral imaging of skin chromophores based on multiple regression analysis aided by Monte Carlo simulation

    Science.gov (United States)

    Nishidate, Izumi; Wiswadarma, Aditya; Hase, Yota; Tanaka, Noriyuki; Maeda, Takaaki; Niizeki, Kyuichi; Aizu, Yoshihisa

    2011-08-01

    In order to visualize melanin and blood concentrations and oxygen saturation in human skin tissue, a simple imaging technique based on multispectral diffuse reflectance images acquired at six wavelengths (500, 520, 540, 560, 580 and 600nm) was developed. The technique utilizes multiple regression analysis aided by Monte Carlo simulation for diffuse reflectance spectra. Using the absorbance spectrum as a response variable and the extinction coefficients of melanin, oxygenated hemoglobin, and deoxygenated hemoglobin as predictor variables, multiple regression analysis provides regression coefficients. Concentrations of melanin and total blood are then determined from the regression coefficients using conversion vectors that are deduced numerically in advance, while oxygen saturation is obtained directly from the regression coefficients. Experiments with a tissue-like agar gel phantom validated the method. In vivo experiments with human skin of the human hand during upper limb occlusion and of the inner forearm exposed to UV irradiation demonstrated the ability of the method to evaluate physiological reactions of human skin tissue.

  6. Electric conduction in semiconductors: a pedagogical model based on the Monte Carlo method

    Energy Technology Data Exchange (ETDEWEB)

    Capizzo, M C; Sperandeo-Mineo, R M; Zarcone, M [UoP-PERG, University of Palermo Physics Education Research Group and Dipartimento di Fisica e Tecnologie Relative, Universita di Palermo (Italy)], E-mail: sperandeo@difter.unipa.it

    2008-05-15

    We present a pedagogic approach aimed at modelling electric conduction in semiconductors in order to describe and explain some macroscopic properties, such as the characteristic behaviour of resistance as a function of temperature. A simple model of the band structure is adopted for the generation of electron-hole pairs as well as for the carrier transport in moderate electric fields. The semiconductor behaviour is described by substituting the traditional statistical approach (requiring a deep mathematical background) with microscopic models, based on the Monte Carlo method, in which simple rules applied to microscopic particles and quasi-particles determine the macroscopic properties. We compare measurements of electric properties of matter with 'virtual experiments' built by using some models where the physical concepts can be presented at different formalization levels.

  7. A Monte Carlo-based treatment-planning tool for ion beam therapy

    CERN Document Server

    Böhlen, T T; Dosanjh, M; Ferrari, A; Haberer, T; Parodi, K; Patera, V; Mairan, A

    2013-01-01

    Ion beam therapy, as an emerging radiation therapy modality, requires continuous efforts to develop and improve tools for patient treatment planning (TP) and research applications. Dose and fluence computation algorithms using the Monte Carlo (MC) technique have served for decades as reference tools for accurate dose computations for radiotherapy. In this work, a novel MC-based treatment-planning (MCTP) tool for ion beam therapy using the pencil beam scanning technique is presented. It allows single-field and simultaneous multiple-fields optimization for realistic patient treatment conditions and for dosimetric quality assurance for irradiation conditions at state-of-the-art ion beam therapy facilities. It employs iterative procedures that allow for the optimization of absorbed dose and relative biological effectiveness (RBE)-weighted dose using radiobiological input tables generated by external RBE models. Using a re-implementation of the local effect model (LEM), theMCTP tool is able to perform TP studies u...

  8. Proton therapy physics

    CERN Document Server

    2012-01-01

    Proton Therapy Physics goes beyond current books on proton therapy to provide an in-depth overview of the physics aspects of this radiation therapy modality, eliminating the need to dig through information scattered in the medical physics literature. After tracing the history of proton therapy, the book summarizes the atomic and nuclear physics background necessary for understanding proton interactions with tissue. It describes the physics of proton accelerators, the parameters of clinical proton beams, and the mechanisms to generate a conformal dose distribution in a patient. The text then covers detector systems and measuring techniques for reference dosimetry, outlines basic quality assurance and commissioning guidelines, and gives examples of Monte Carlo simulations in proton therapy. The book moves on to discussions of treatment planning for single- and multiple-field uniform doses, dose calculation concepts and algorithms, and precision and uncertainties for nonmoving and moving targets. It also exami...

  9. Monte Carlo simulation of grating-based neutron phase contrast imaging at CPHS

    International Nuclear Information System (INIS)

    Zhang Ran; Chen Zhiqiang; Huang Zhifeng; Xiao Yongshun; Wang Xuewu; Wie Jie; Loong, C.-K.

    2011-01-01

    Since the launching of the Compact Pulsed Hadron Source (CPHS) project of Tsinghua University in 2009, works have begun on the design and engineering of an imaging/radiography instrument for the neutron source provided by CPHS. The instrument will perform basic tasks such as transmission imaging and computerized tomography. Additionally, we include in the design the utilization of coded-aperture and grating-based phase contrast methodology, as well as the options of prompt gamma-ray analysis and neutron-energy selective imaging. Previously, we had implemented the hardware and data-analysis software for grating-based X-ray phase contrast imaging. Here, we investigate Geant4-based Monte Carlo simulations of neutron refraction phenomena and then model the grating-based neutron phase contrast imaging system according to the classic-optics-based method. The simulated experimental results of the retrieving phase shift gradient information by five-step phase-stepping approach indicate the feasibility of grating-based neutron phase contrast imaging as an option for the cold neutron imaging instrument at the CPHS.

  10. Full modelling of the MOSAIC animal PET system based on the GATE Monte Carlo simulation code

    International Nuclear Information System (INIS)

    Merheb, C; Petegnief, Y; Talbot, J N

    2007-01-01

    within 9%. For a 410-665 keV energy window, the measured sensitivity for a centred point source was 1.53% and mouse and rat scatter fractions were respectively 12.0% and 18.3%. The scattered photons produced outside the rat and mouse phantoms contributed to 24% and 36% of total simulated scattered coincidences. Simulated and measured single and prompt count rates agreed well for activities up to the electronic saturation at 110 MBq for the mouse and rat phantoms. Volumetric spatial resolution was 17.6 μL at the centre of the FOV with differences less than 6% between experimental and simulated spatial resolution values. The comprehensive evaluation of the Monte Carlo modelling of the Mosaic(TM) system demonstrates that the GATE package is adequately versatile and appropriate to accurately describe the response of an Anger logic based animal PET system

  11. Model-based fault detection for proton exchange membrane fuel cell ...

    African Journals Online (AJOL)

    In this paper, an intelligent model-based fault detection (FD) is developed for proton exchange membrane fuel cell (PEMFC) dynamic systems using an independent radial basis function (RBF) networks. The novelty is that this RBF networks is used to model the PEMFC dynamic systems and residuals are generated based ...

  12. A zero-variance based scheme for Monte Carlo criticality simulations

    NARCIS (Netherlands)

    Christoforou, S.

    2010-01-01

    The ability of the Monte Carlo method to solve particle transport problems by simulating the particle behaviour makes it a very useful technique in nuclear reactor physics. However, the statistical nature of Monte Carlo implies that there will always be a variance associated with the estimate

  13. Monte Carlo based dosimetry and treatment planning for neutron capture therapy of brain tumors

    International Nuclear Information System (INIS)

    Zamenhof, R.G.; Clement, S.D.; Harling, O.K.; Brenner, J.F.; Wazer, D.E.; Madoc-Jones, H.; Yanch, J.C.

    1990-01-01

    Monte Carlo based dosimetry and computer-aided treatment planning for neutron capture therapy have been developed to provide the necessary link between physical dosimetric measurements performed on the MITR-II epithermal-neutron beams and the need of the radiation oncologist to synthesize large amounts of dosimetric data into a clinically meaningful treatment plan for each individual patient. Monte Carlo simulation has been employed to characterize the spatial dose distributions within a skull/brain model irradiated by an epithermal-neutron beam designed for neutron capture therapy applications. The geometry and elemental composition employed for the mathematical skull/brain model and the neutron and photon fluence-to-dose conversion formalism are presented. A treatment planning program, NCTPLAN, developed specifically for neutron capture therapy, is described. Examples are presented illustrating both one and two-dimensional dose distributions obtainable within the brain with an experimental epithermal-neutron beam, together with beam quality and treatment plan efficacy criteria which have been formulated for neutron capture therapy. The incorporation of three-dimensional computed tomographic image data into the treatment planning procedure is illustrated. The experimental epithermal-neutron beam has a maximum usable circular diameter of 20 cm, and with 30 ppm of B-10 in tumor and 3 ppm of B-10 in blood, it produces a beam-axis advantage depth of 7.4 cm, a beam-axis advantage ratio of 1.83, a global advantage ratio of 1.70, and an advantage depth RBE-dose rate to tumor of 20.6 RBE-cGy/min (cJ/kg-min). These characteristics make this beam well suited for clinical applications, enabling an RBE-dose of 2,000 RBE-cGy/min (cJ/kg-min) to be delivered to tumor at brain midline in six fractions with a treatment time of approximately 16 minutes per fraction

  14. Study on in situ calibration for neutron flux monitor in the Large Helical Device based on Monte Carlo calculations.

    Science.gov (United States)

    Nakano, Y; Yamazaki, A; Watanabe, K; Uritani, A; Ogawa, K; Isobe, M

    2014-11-01

    Neutron monitoring is important to manage safety of fusion experiment facilities because neutrons are generated in fusion reactions. Monte Carlo simulations play an important role in evaluating the influence of neutron scattering from various structures and correcting differences between deuterium plasma experiments and in situ calibration experiments. We evaluated these influences based on differences between the both experiments at Large Helical Device using Monte Carlo simulation code MCNP5. A difference between the both experiments in absolute detection efficiency of the fission chamber between O-ports is estimated to be the biggest of all monitors. We additionally evaluated correction coefficients for some neutron monitors.

  15. Microalgae dewatering based on forward osmosis employing proton exchange membrane.

    Science.gov (United States)

    Son, Jieun; Sung, Mina; Ryu, Hoyoung; Oh, You-Kwan; Han, Jong-In

    2017-11-01

    In this study, electrically-facilitated forward osmosis (FO) employing proton exchange membrane (PEM) was established for the purpose of microalgae dewatering. An increase in water flux was observed when an external voltage was applied to the FO equipped with the PEM; as expected, the trend became more dramatic with both concentration of draw solution and applied voltage raised. With this FO used for microalgae dewatering, 247% of increase in flux and 86% in final biomass concentration were observed. In addition to the effect on flux improvement, the electrically-facilitated FO exhibited the ability to remove chlorophyll from the dewatered biomass, down to 0.021±0015mg/g cell. All these suggest that the newly suggested electrically-facilitated FO, one particularly employed PEM, can indeed offer a workable way of dewatering of microalgae; it appeared to be so because it can also remove the ever-problematic chlorophyll from extracted lipids in a simultaneous fashion. Copyright © 2017 Elsevier Ltd. All rights reserved.

  16. Analysis by Monte Carlo simulations of the sensitivity to single event upset of SRAM memories under spatial proton or terrestrial neutron environment

    International Nuclear Information System (INIS)

    Lambert, D.

    2006-07-01

    Electronic systems in space and terrestrial environments are subjected to a flow of particles of natural origin, which can induce dysfunctions. These particles can cause Single Event Upsets (SEU) in SRAM memories. Although non-destructive, the SEU can have consequences on the equipment functioning in applications requiring a great reliability (airplane, satellite, launcher, medical, etc). Thus, an evaluation of the sensitivity of the component technology is necessary to predict the reliability of a system. In atmospheric environment, the SEU sensitivity is mainly caused by the secondary ions resulting from the nuclear reactions between the neutrons and the atoms of the component. In space environment, the protons with strong energies induce the same effects as the atmospheric neutrons. In our work, a new code of prediction of the rate of SEU has been developed (MC-DASIE) in order to quantify the sensitivity for a given environment and to explore the mechanisms of failures according to technology. This code makes it possible to study various technologies of memories SRAM (Bulk and SOI) in neutron and proton environment between 1 MeV and 1 GeV. Thus, MC-DASIE was used with experiment data to study the effect of integration on the sensitivity of the memories in terrestrial environment, a comparison between the neutron and proton irradiations and the influence of the modeling of the target component on the calculation of the rate of SEU. (author)

  17. Girsanov's transformation based variance reduced Monte Carlo simulation schemes for reliability estimation in nonlinear stochastic dynamics

    Science.gov (United States)

    Kanjilal, Oindrila; Manohar, C. S.

    2017-07-01

    The study considers the problem of simulation based time variant reliability analysis of nonlinear randomly excited dynamical systems. Attention is focused on importance sampling strategies based on the application of Girsanov's transformation method. Controls which minimize the distance function, as in the first order reliability method (FORM), are shown to minimize a bound on the sampling variance of the estimator for the probability of failure. Two schemes based on the application of calculus of variations for selecting control signals are proposed: the first obtains the control force as the solution of a two-point nonlinear boundary value problem, and, the second explores the application of the Volterra series in characterizing the controls. The relative merits of these schemes, vis-à-vis the method based on ideas from the FORM, are discussed. Illustrative examples, involving archetypal single degree of freedom (dof) nonlinear oscillators, and a multi-degree of freedom nonlinear dynamical system, are presented. The credentials of the proposed procedures are established by comparing the solutions with pertinent results from direct Monte Carlo simulations.

  18. Implementation of GPU accelerated SPECT reconstruction with Monte Carlo-based scatter correction.

    Science.gov (United States)

    Bexelius, Tobias; Sohlberg, Antti

    2018-03-21

    Statistical SPECT reconstruction can be very time-consuming especially when compensations for collimator and detector response, attenuation, and scatter are included in the reconstruction. This work proposes an accelerated SPECT reconstruction algorithm based on graphics processing unit (GPU) processing. Ordered subset expectation maximization (OSEM) algorithm with CT-based attenuation modelling, depth-dependent Gaussian convolution-based collimator-detector response modelling, and Monte Carlo-based scatter compensation was implemented using OpenCL. The OpenCL implementation was compared against the existing multi-threaded OSEM implementation running on a central processing unit (CPU) in terms of scatter-to-primary ratios, standardized uptake values (SUVs), and processing speed using mathematical phantoms and clinical multi-bed bone SPECT/CT studies. The difference in scatter-to-primary ratios, visual appearance, and SUVs between GPU and CPU implementations was minor. On the other hand, at its best, the GPU implementation was noticed to be 24 times faster than the multi-threaded CPU version on a normal 128 × 128 matrix size 3 bed bone SPECT/CT data set when compensations for collimator and detector response, attenuation, and scatter were included. GPU SPECT reconstructions show great promise as an every day clinical reconstruction tool.

  19. Online advertising and marketing claims by providers of proton beam therapy: are they guideline-based?

    Science.gov (United States)

    Corkum, Mark T; Liu, Wei; Palma, David A; Bauman, Glenn S; Dinniwell, Robert E; Warner, Andrew; Mishra, Mark V; Louie, Alexander V

    2018-03-15

    Cancer patients frequently search the Internet for treatment options, and hospital websites are seen as reliable sources of knowledge. Guidelines support the use of proton radiotherapy in specific disease sites or on clinical trials. This study aims to evaluate direct-to-consumer advertising content and claims made by proton therapy centre (PTC) websites worldwide. Operational PTC websites in English were identified through the Particle Therapy Co-Operative Group website. Data abstraction of website content was performed independently by two investigators. Eight international guidelines were consulted to determine guideline-based indications for proton radiotherapy. Univariate and multivariate logistic regression models were used to determine the characteristics of PTC websites that indicated proton radiotherapy offered greater disease control or cure rates. Forty-eight PTCs with 46 English websites were identified. 60·9% of PTC websites claimed proton therapy provided improved disease control or cure. U.S. websites listed more indications than international websites (15·5 ± 5·4 vs. 10·4 ± 5·8, p = 0·004). The most common disease sites advertised were prostate (87·0%), head and neck (87·0%) and pediatrics (82·6%), all of which were indicated in least one international guideline. Several disease sites advertised were not present in any consensus guidelines, including pancreatobiliary (52·2%), breast (50·0%), and esophageal (43·5%) cancers. Multivariate analysis found increasing number of disease sites and claiming their centre was a local or regional leader in proton radiotherapy was associated with indicating proton radiotherapy offers greater disease control or cure. Information from PTC websites often differs from recommendations found in international consensus guidelines. As online marketing information may have significant influence on patient decision-making, alignment of such information with accepted guidelines and consensus

  20. Water oxidation by amorphous cobalt-based oxides: volume activity and proton transfer to electrolyte bases.

    Science.gov (United States)

    Klingan, Katharina; Ringleb, Franziska; Zaharieva, Ivelina; Heidkamp, Jonathan; Chernev, Petko; Gonzalez-Flores, Diego; Risch, Marcel; Fischer, Anna; Dau, Holger

    2014-05-01

    Water oxidation in the neutral pH regime catalyzed by amorphous transition-metal oxides is of high interest in energy science. Crucial determinants of electrocatalytic activity were investigated for a cobalt-based oxide film electrodeposited at various thicknesses on inert electrodes. For water oxidation at low current densities, the turnover frequency (TOF) per cobalt ion of the bulk material stayed fully constant for variation of the thickness of the oxide film by a factor of 100 (from about 15 nm to 1.5 μm). Thickness variation changed neither the nanostructure of the outer film surface nor the atomic structure of the oxide catalyst significantly. These findings imply catalytic activity of the bulk hydrated oxide material. Nonclassical dependence on pH was observed. For buffered electrolytes with pKa values of the buffer base ranging from 4.7 (acetate) to 10.3 (hydrogen carbonate), the catalytic activity reflected the protonation state of the buffer base in the electrolyte solution directly and not the intrinsic catalytic properties of the oxide itself. It is proposed that catalysis of water oxidation occurs within the bulk hydrated oxide film at the margins of cobalt oxide fragments of molecular dimensions. At high current densities, the availability of a proton-accepting base at the catalyst-electrolyte interface controls the rate of water oxidation. The reported findings may be of general relevance for water oxidation catalyzed at moderate pH by amorphous transition-metal oxides. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  1. MRI-Based Computed Tomography Metal Artifact Correction Method for Improving Proton Range Calculation Accuracy

    Energy Technology Data Exchange (ETDEWEB)

    Park, Peter C. [Department of Radiation Physics, University of Texas MD Anderson Cancer Center, Houston, Texas (United States); Schreibmann, Eduard; Roper, Justin; Elder, Eric; Crocker, Ian [Department of Radiation Oncology, Winship Cancer Institute of Emory University, Atlanta, Georgia (United States); Fox, Tim [Varian Medical Systems, Palo Alto, California (United States); Zhu, X. Ronald [Department of Radiation Physics, University of Texas MD Anderson Cancer Center, Houston, Texas (United States); Dong, Lei [Scripps Proton Therapy Center, San Diego, California (United States); Dhabaan, Anees, E-mail: anees.dhabaan@emory.edu [Department of Radiation Oncology, Winship Cancer Institute of Emory University, Atlanta, Georgia (United States)

    2015-03-15

    Purpose: Computed tomography (CT) artifacts can severely degrade dose calculation accuracy in proton therapy. Prompted by the recently increased popularity of magnetic resonance imaging (MRI) in the radiation therapy clinic, we developed an MRI-based CT artifact correction method for improving the accuracy of proton range calculations. Methods and Materials: The proposed method replaces corrupted CT data by mapping CT Hounsfield units (HU number) from a nearby artifact-free slice, using a coregistered MRI. MRI and CT volumetric images were registered with use of 3-dimensional (3D) deformable image registration (DIR). The registration was fine-tuned on a slice-by-slice basis by using 2D DIR. Based on the intensity of paired MRI pixel values and HU from an artifact-free slice, we performed a comprehensive analysis to predict the correct HU for the corrupted region. For a proof-of-concept validation, metal artifacts were simulated on a reference data set. Proton range was calculated using reference, artifactual, and corrected images to quantify the reduction in proton range error. The correction method was applied to 4 unique clinical cases. Results: The correction method resulted in substantial artifact reduction, both quantitatively and qualitatively. On respective simulated brain and head and neck CT images, the mean error was reduced from 495 and 370 HU to 108 and 92 HU after correction. Correspondingly, the absolute mean proton range errors of 2.4 cm and 1.7 cm were reduced to less than 2 mm in both cases. Conclusions: Our MRI-based CT artifact correction method can improve CT image quality and proton range calculation accuracy for patients with severe CT artifacts.

  2. MRI-Based Computed Tomography Metal Artifact Correction Method for Improving Proton Range Calculation Accuracy

    International Nuclear Information System (INIS)

    Park, Peter C.; Schreibmann, Eduard; Roper, Justin; Elder, Eric; Crocker, Ian; Fox, Tim; Zhu, X. Ronald; Dong, Lei; Dhabaan, Anees

    2015-01-01

    Purpose: Computed tomography (CT) artifacts can severely degrade dose calculation accuracy in proton therapy. Prompted by the recently increased popularity of magnetic resonance imaging (MRI) in the radiation therapy clinic, we developed an MRI-based CT artifact correction method for improving the accuracy of proton range calculations. Methods and Materials: The proposed method replaces corrupted CT data by mapping CT Hounsfield units (HU number) from a nearby artifact-free slice, using a coregistered MRI. MRI and CT volumetric images were registered with use of 3-dimensional (3D) deformable image registration (DIR). The registration was fine-tuned on a slice-by-slice basis by using 2D DIR. Based on the intensity of paired MRI pixel values and HU from an artifact-free slice, we performed a comprehensive analysis to predict the correct HU for the corrupted region. For a proof-of-concept validation, metal artifacts were simulated on a reference data set. Proton range was calculated using reference, artifactual, and corrected images to quantify the reduction in proton range error. The correction method was applied to 4 unique clinical cases. Results: The correction method resulted in substantial artifact reduction, both quantitatively and qualitatively. On respective simulated brain and head and neck CT images, the mean error was reduced from 495 and 370 HU to 108 and 92 HU after correction. Correspondingly, the absolute mean proton range errors of 2.4 cm and 1.7 cm were reduced to less than 2 mm in both cases. Conclusions: Our MRI-based CT artifact correction method can improve CT image quality and proton range calculation accuracy for patients with severe CT artifacts

  3. Fast CPU-based Monte Carlo simulation for radiotherapy dose calculation

    Science.gov (United States)

    Ziegenhein, Peter; Pirner, Sven; Kamerling, Cornelis Ph; Oelfke, Uwe

    2015-08-01

    Monte-Carlo (MC) simulations are considered to be the most accurate method for calculating dose distributions in radiotherapy. Its clinical application, however, still is limited by the long runtimes conventional implementations of MC algorithms require to deliver sufficiently accurate results on high resolution imaging data. In order to overcome this obstacle we developed the software-package PhiMC, which is capable of computing precise dose distributions in a sub-minute time-frame by leveraging the potential of modern many- and multi-core CPU-based computers. PhiMC is based on the well verified dose planning method (DPM). We could demonstrate that PhiMC delivers dose distributions which are in excellent agreement to DPM. The multi-core implementation of PhiMC scales well between different computer architectures and achieves a speed-up of up to 37× compared to the original DPM code executed on a modern system. Furthermore, we could show that our CPU-based implementation on a modern workstation is between 1.25× and 1.95× faster than a well-known GPU implementation of the same simulation method on a NVIDIA Tesla C2050. Since CPUs work on several hundreds of GB RAM the typical GPU memory limitation does not apply for our implementation and high resolution clinical plans can be calculated.

  4. SQERTSS: Dynamic rank based throttling of transition probabilities in kinetic Monte Carlo simulations

    International Nuclear Information System (INIS)

    Danielson, Thomas; Sutton, Jonathan E.; Hin, Céline; Virginia Polytechnic Institute and State University; Savara, Aditya

    2017-01-01

    Lattice based Kinetic Monte Carlo (KMC) simulations offer a powerful simulation technique for investigating large reaction networks while retaining spatial configuration information, unlike ordinary differential equations. However, large chemical reaction networks can contain reaction processes with rates spanning multiple orders of magnitude. This can lead to the problem of “KMC stiffness” (similar to stiffness in differential equations), where the computational expense has the potential to be overwhelmed by very short time-steps during KMC simulations, with the simulation spending an inordinate amount of KMC steps / cpu-time simulating fast frivolous processes (FFPs) without progressing the system (reaction network). In order to achieve simulation times that are experimentally relevant or desired for predictions, a dynamic throttling algorithm involving separation of the processes into speed-ranks based on event frequencies has been designed and implemented with the intent of decreasing the probability of FFP events, and increasing the probability of slow process events -- allowing rate limiting events to become more likely to be observed in KMC simulations. This Staggered Quasi-Equilibrium Rank-based Throttling for Steady-state (SQERTSS) algorithm designed for use in achieving and simulating steady-state conditions in KMC simulations. Lastly, as shown in this work, the SQERTSS algorithm also works for transient conditions: the correct configuration space and final state will still be achieved if the required assumptions are not violated, with the caveat that the sizes of the time-steps may be distorted during the transient period.

  5. Monte Carlo vs. Pencil Beam based optimization of stereotactic lung IMRT

    Directory of Open Access Journals (Sweden)

    Weinmann Martin

    2009-12-01

    Full Text Available Abstract Background The purpose of the present study is to compare finite size pencil beam (fsPB and Monte Carlo (MC based optimization of lung intensity-modulated stereotactic radiotherapy (lung IMSRT. Materials and methods A fsPB and a MC algorithm as implemented in a biological IMRT planning system were validated by film measurements in a static lung phantom. Then, they were applied for static lung IMSRT planning based on three different geometrical patient models (one phase static CT, density overwrite one phase static CT, average CT of the same patient. Both 6 and 15 MV beam energies were used. The resulting treatment plans were compared by how well they fulfilled the prescribed optimization constraints both for the dose distributions calculated on the static patient models and for the accumulated dose, recalculated with MC on each of 8 CTs of a 4DCT set. Results In the phantom measurements, the MC dose engine showed discrepancies Conclusions It is feasible to employ the MC dose engine for optimization of lung IMSRT and the plans are superior to fsPB. Use of static patient models introduces a bias in the MC dose distribution compared to the 4D MC recalculated dose, but this bias is predictable and therefore MC based optimization on static patient models is considered safe.

  6. Radioluminescence dosimeters based on doped optical fibers irradiated with accelerated protons

    International Nuclear Information System (INIS)

    Veronese, I.; Cantone, M.C.; Abbas, K.; Holzwarth, U.; Simonelli, F.; Chiodini, N.; Fasoli, M.; Galli, A.; Martini, M.; Moretti, F.; Vedda, A.; Mones, E.

    2008-01-01

    Full text: Radioluminescence (RL) dosimeters, based on SiO 2 optical fibres doped with rare-earth ions, have been recently developed by using the sol-gel and powder in tube techniques. The response of the devices, initially tested with soft X-rays, have been then studied using: i) Standard photon beams of 6 and 15 MV produced by a medical linear accelerator; ii) A 60 Co radiotherapy unit; and iii) Small photon beams of 6 MV energy generated by a stereotactic radiosurgery system. The promising results obtained in these studies induced to extend the use of the radio luminescent doped fibres also for the characterisation of accelerated proton beams, in view of a possible employment of the detectors in proton therapy practices. The present work summarises the results of a feasibility study performed by exposing Ce and Eu doped optical fibres to proton beams of energy up to 40 MeV, produced by the Scanditronix MC-40 Cyclotron at the Joint Research Centre of the European Commission in Ispra (Italy). The sensitivity of the RL response to protons will be presented and compared with the corresponding one obtained using photon beams as source of irradiation. Moreover, the dependence of the signal on the energy and intensity of the proton beam will be show and discussed. (author)

  7. Ultrafast cone-beam CT scatter correction with GPU-based Monte Carlo simulation

    Directory of Open Access Journals (Sweden)

    Yuan Xu

    2014-03-01

    Full Text Available Purpose: Scatter artifacts severely degrade image quality of cone-beam CT (CBCT. We present an ultrafast scatter correction framework by using GPU-based Monte Carlo (MC simulation and prior patient CT image, aiming at automatically finish the whole process including both scatter correction and reconstruction within 30 seconds.Methods: The method consists of six steps: 1 FDK reconstruction using raw projection data; 2 Rigid Registration of planning CT to the FDK results; 3 MC scatter calculation at sparse view angles using the planning CT; 4 Interpolation of the calculated scatter signals to other angles; 5 Removal of scatter from the raw projections; 6 FDK reconstruction using the scatter-corrected projections. In addition to using GPU to accelerate MC photon simulations, we also use a small number of photons and a down-sampled CT image in simulation to further reduce computation time. A novel denoising algorithm is used to eliminate MC noise from the simulated scatter images caused by low photon numbers. The method is validated on one simulated head-and-neck case with 364 projection angles.Results: We have examined variation of the scatter signal among projection angles using Fourier analysis. It is found that scatter images at 31 angles are sufficient to restore those at all angles with < 0.1% error. For the simulated patient case with a resolution of 512 × 512 × 100, we simulated 5 × 106 photons per angle. The total computation time is 20.52 seconds on a Nvidia GTX Titan GPU, and the time at each step is 2.53, 0.64, 14.78, 0.13, 0.19, and 2.25 seconds, respectively. The scatter-induced shading/cupping artifacts are substantially reduced, and the average HU error of a region-of-interest is reduced from 75.9 to 19.0 HU.Conclusion: A practical ultrafast MC-based CBCT scatter correction scheme is developed. It accomplished the whole procedure of scatter correction and reconstruction within 30 seconds.----------------------------Cite this

  8. A research plan based on high intensity proton accelerator Neutron Science Research Center

    Energy Technology Data Exchange (ETDEWEB)

    Mizumoto, Motoharu [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1997-03-01

    A plan called Neutron Science Research Center (NSRC) has been proposed in JAERI. The center is a complex composed of research facilities based on a proton linac with an energy of 1.5GeV and an average current of 10mA. The research facilities will consist of Thermal/Cold Neutron Facility, Neutron Irradiation Facility, Neutron Physics Facility, OMEGA/Nuclear Energy Facility, Spallation RI Beam Facility, Meson/Muon Facility and Medium Energy Experiment Facility, where high intensity proton beam and secondary particle beams such as neutron, pion, muon and unstable radio isotope (RI) beams generated from the proton beam will be utilized for innovative researches in the fields on nuclear engineering and basic sciences. (author)

  9. Radiation sensors based on the generation of mobile protons in organic dielectrics.

    Science.gov (United States)

    Kapetanakis, Eleftherios; Douvas, Antonios M; Argitis, Panagiotis; Normand, Pascal

    2013-06-26

    A sensing scheme based on mobile protons generated by radiation, including ionizing radiation (IonR), in organic gate dielectrics is investigated for the development of metal-insulator-semiconductor (MIS)-type dosimeters. Application of an electric field to the gate dielectric moves the protons and thereby alters the flat band voltage (VFB) of the MIS device. The shift in the VFB is proportional to the IonR-generated protons and, therefore, to the IonR total dose. Triphenylsulfonium nonaflate (TPSNF) photoacid generator (PAG)-containing poly(methyl methacrylate) (PMMA) polymeric films was selected as radiation-sensitive gate dielectrics. The effects of UV (249 nm) and gamma (Co-60) irradiations on the high-frequency capacitance versus the gate voltage (C-VG) curves of the MIS devices were investigated for different total dose values. Systematic improvements in sensitivity can be accomplished by increasing the concentration of the TPSNF molecules embedded in the polymeric matrix.

  10. Is the Photoinduced Isomerization in Retinal Protonated Schiff Bases a Single- or Double-Torsional Process?

    Czech Academy of Sciences Publication Activity Database

    Szymczak, J. J.; Barbatti, M.; Lischka, Hans

    2009-01-01

    Roč. 113, č. 43 (2009), s. 11907-11918 ISSN 1089-5639 R&D Projects: GA MŠk LC512 Institutional research plan: CEZ:AV0Z40550506 Keywords : nonadiabatic photodynamical simulations * retinal protonated Schiff base * cis-trans isomerization processes Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 2.899, year: 2009

  11. Proton affinities of anionic bases: Trends across the periodic table, structural effects, and DFT validation

    NARCIS (Netherlands)

    Swart, M.; Bickelhaupt, F.M.

    2006-01-01

    We have carried out an extensive exploration of the gas-phase basicity of archetypal anionic bases across the periodic system using the generalized gradient approximation of density functional theory (DFT) at BP86/QZ4P//BP86/TZ2P. First, we validate DFT as a reliable tool for computing proton

  12. An UV photochromic memory effect in proton-based WO3 electrochromic devices

    International Nuclear Information System (INIS)

    Zhang Yong; Lee, S.-H.; Mascarenhas, A.; Deb, S. K.

    2008-01-01

    We report an UV photochromic memory effect on a standard proton-based WO 3 electrochromic device. It exhibits two memory states, associated with the colored and bleached states of the device, respectively. Such an effect can be used to enhance device performance (increasing the dynamic range), re-energize commercial electrochromic devices, and develop memory devices

  13. An UV photochromic memory effect in proton-based WO3 electrochromic devices

    Science.gov (United States)

    Zhang, Yong; Lee, S.-H.; Mascarenhas, A.; Deb, S. K.

    2008-11-01

    We report an UV photochromic memory effect on a standard proton-based WO3 electrochromic device. It exhibits two memory states, associated with the colored and bleached states of the device, respectively. Such an effect can be used to enhance device performance (increasing the dynamic range), re-energize commercial electrochromic devices, and develop memory devices.

  14. Molecular modeling of the conductivity changes of the emeraldine base polyaniline due to protonic acid doping

    NARCIS (Netherlands)

    Chen, X.; Yuan, C.A.; Wong, C.K.Y.; Zhang, G.

    2012-01-01

    We propose a molecular modeling strategy, which is capable of predicting the conductivity change of emeraldine base polyaniline polymer due to different degree of protonic acid doping. The method is comprised of two key steps: (1) generating the amorphous unit cells with given number of polymer

  15. Ionomeric membranes based on partially sulfonated poly(styrene) : synthesis, proton conduction and methanol permeation

    NARCIS (Netherlands)

    Picchioni, F.; Tricoli, V.; Carretta, N.

    2000-01-01

    Homogeneuosly sulfonated poly(styrene) (SPS) was prepared with various concentration of sulfonic acid groups in the base polymer. Membranes cast from these materials were investigated in relation to proton conductivity and methanol permeability in the temperature range from 20°C to 60°C. It was

  16. SU-E-T-610: Phosphor-Based Fiber Optic Probes for Proton Beam Characterization

    Energy Technology Data Exchange (ETDEWEB)

    Darafsheh, A; Soldner, A; Liu, H; Kassaee, A; Zhu, T; Finlay, J [Univ Pennsylvania, Philadelphia, PA (United States)

    2015-06-15

    Purpose: To investigate feasibility of using fiber optics probes with rare-earth-based phosphor tips for proton beam radiation dosimetry. We designed and fabricated a fiber probe with submillimeter resolution (<0.5 mm3) based on TbF3 phosphors and evaluated its performance for measurement of proton beam including profiles and range. Methods: The fiber optic probe with TbF3 phosphor tip, embedded in tissue-mimicking phantoms was irradiated with double scattering proton beam with energy of 180 MeV. Luminescence spectroscopy was performed by a CCD-coupled spectrograph to analyze the emission spectra of the fiber tip. In order to measure the spatial beam profile and percentage depth dose, we used singular value decomposition method to spectrally separate the phosphors ionoluminescence signal from the background Cerenkov radiation signal. Results: The spectra of the TbF3 fiber probe showed characteristic ionoluminescence emission peaks at 489, 542, 586, and 620 nm. By using singular value decomposition we found the contribution of the ionoluminescence signal to measure the percentage depth dose in phantoms and compared that with measurements performed with ion chamber. We observed quenching effect at the spread out Bragg peak region, manifested as under-responding of the signal, due to the high LET of the beam. However, the beam profiles were not dramatically affected by the quenching effect. Conclusion: We have evaluated the performance of a fiber optic probe with submillimeter resolution for proton beam dosimetry. We demonstrated feasibility of spectral separation of the Cerenkov radiation from the collected signal. Such fiber probes can be used for measurements of proton beams profile and range. The experimental apparatus and spectroscopy method developed in this work provide a robust platform for characterization of proton-irradiated nanophosphor particles for ultralow fluence photodynamic therapy or molecular imaging applications.

  17. Development of CAD-Based Geometry Processing Module for a Monte Carlo Particle Transport Analysis Code

    International Nuclear Information System (INIS)

    Choi, Sung Hoon; Kwark, Min Su; Shim, Hyung Jin

    2012-01-01

    As The Monte Carlo (MC) particle transport analysis for a complex system such as research reactor, accelerator, and fusion facility may require accurate modeling of the complicated geometry. Its manual modeling by using the text interface of a MC code to define the geometrical objects is tedious, lengthy and error-prone. This problem can be overcome by taking advantage of modeling capability of the computer aided design (CAD) system. There have been two kinds of approaches to develop MC code systems utilizing the CAD data: the external format conversion and the CAD kernel imbedded MC simulation. The first approach includes several interfacing programs such as McCAD, MCAM, GEOMIT etc. which were developed to automatically convert the CAD data into the MCNP geometry input data. This approach makes the most of the existing MC codes without any modifications, but implies latent data inconsistency due to the difference of the geometry modeling system. In the second approach, a MC code utilizes the CAD data for the direct particle tracking or the conversion to an internal data structure of the constructive solid geometry (CSG) and/or boundary representation (B-rep) modeling with help of a CAD kernel. MCNP-BRL and OiNC have demonstrated their capabilities of the CAD-based MC simulations. Recently we have developed a CAD-based geometry processing module for the MC particle simulation by using the OpenCASCADE (OCC) library. In the developed module, CAD data can be used for the particle tracking through primitive CAD surfaces (hereafter the CAD-based tracking) or the internal conversion to the CSG data structure. In this paper, the performances of the text-based model, the CAD-based tracking, and the internal CSG conversion are compared by using an in-house MC code, McSIM, equipped with the developed CAD-based geometry processing module

  18. Photodynamic Cancer Therapy Using Wavelet Based Monte Carlo Computation of Light Absorption with a Genetic Algorithm

    Science.gov (United States)

    R P, Meenaakshi Sundhari

    2018-01-27

    Objective: The method to treating cancer that combines light and light-sensitive drugs to selectively destroy tumour cells without harming healthy tissue is called photodynamic therapy (PDT). It requires accurate data for light dose distribution, generated with scalable algorithms. One of the benchmark approaches involves Monte Carlo (MC) simulations. This gives an accurate assessment of light dose distribution, but is very demanding in computation time, which prevents routine application for treatment planning. Methods: In order to resolve this problem, a design for MC simulation based on the gold standard software in biophotonics was implemented with a large modern wavelet based genetic algorithm search (WGAS). Result: The accuracy of the proposed method was compared to that with the standard optimization method using a realistic skin model. The maximum stop band attenuation of the designed LP, HP, BP and BS filters was assessed using the proposed WGAS algorithm as well as with other methods. Conclusion: In this paper, the proposed methodology employs intermediate wavelets which improve the diversification rate of the charged genetic algorithm search and that leads to significant improvement in design effort efficiency. Creative Commons Attribution License

  19. Iterative reconstruction using a Monte Carlo based system transfer matrix for dedicated breast positron emission tomography

    International Nuclear Information System (INIS)

    Saha, Krishnendu; Straus, Kenneth J.; Glick, Stephen J.; Chen, Yu.

    2014-01-01

    To maximize sensitivity, it is desirable that ring Positron Emission Tomography (PET) systems dedicated for imaging the breast have a small bore. Unfortunately, due to parallax error this causes substantial degradation in spatial resolution for objects near the periphery of the breast. In this work, a framework for computing and incorporating an accurate system matrix into iterative reconstruction is presented in an effort to reduce spatial resolution degradation towards the periphery of the breast. The GATE Monte Carlo Simulation software was utilized to accurately model the system matrix for a breast PET system. A strategy for increasing the count statistics in the system matrix computation and for reducing the system element storage space was used by calculating only a subset of matrix elements and then estimating the rest of the elements by using the geometric symmetry of the cylindrical scanner. To implement this strategy, polar voxel basis functions were used to represent the object, resulting in a block-circulant system matrix. Simulation studies using a breast PET scanner model with ring geometry demonstrated improved contrast at 45% reduced noise level and 1.5 to 3 times resolution performance improvement when compared to MLEM reconstruction using a simple line-integral model. The GATE based system matrix reconstruction technique promises to improve resolution and noise performance and reduce image distortion at FOV periphery compared to line-integral based system matrix reconstruction

  20. Iterative reconstruction using a Monte Carlo based system transfer matrix for dedicated breast positron emission tomography

    Energy Technology Data Exchange (ETDEWEB)

    Saha, Krishnendu [Ohio Medical Physics Consulting, Dublin, Ohio 43017 (United States); Straus, Kenneth J.; Glick, Stephen J. [Department of Radiology, University of Massachusetts Medical School, Worcester, Massachusetts 01655 (United States); Chen, Yu. [Department of Radiation Oncology, Columbia University, New York, New York 10032 (United States)

    2014-08-28

    To maximize sensitivity, it is desirable that ring Positron Emission Tomography (PET) systems dedicated for imaging the breast have a small bore. Unfortunately, due to parallax error this causes substantial degradation in spatial resolution for objects near the periphery of the breast. In this work, a framework for computing and incorporating an accurate system matrix into iterative reconstruction is presented in an effort to reduce spatial resolution degradation towards the periphery of the breast. The GATE Monte Carlo Simulation software was utilized to accurately model the system matrix for a breast PET system. A strategy for increasing the count statistics in the system matrix computation and for reducing the system element storage space was used by calculating only a subset of matrix elements and then estimating the rest of the elements by using the geometric symmetry of the cylindrical scanner. To implement this strategy, polar voxel basis functions were used to represent the object, resulting in a block-circulant system matrix. Simulation studies using a breast PET scanner model with ring geometry demonstrated improved contrast at 45% reduced noise level and 1.5 to 3 times resolution performance improvement when compared to MLEM reconstruction using a simple line-integral model. The GATE based system matrix reconstruction technique promises to improve resolution and noise performance and reduce image distortion at FOV periphery compared to line-integral based system matrix reconstruction.

  1. Monte Carlo based investigation of Berry phase for depth resolved characterization of biomedical scattering samples

    Energy Technology Data Exchange (ETDEWEB)

    Baba, Justin S [ORNL; John, Dwayne O [ORNL; Koju, Vijay [ORNL

    2015-01-01

    The propagation of light in turbid media is an active area of research with relevance to numerous investigational fields, e.g., biomedical diagnostics and therapeutics. The statistical random-walk nature of photon propagation through turbid media is ideal for computational based modeling and simulation. Ready access to super computing resources provide a means for attaining brute force solutions to stochastic light-matter interactions entailing scattering by facilitating timely propagation of sufficient (>10million) photons while tracking characteristic parameters based on the incorporated physics of the problem. One such model that works well for isotropic but fails for anisotropic scatter, which is the case for many biomedical sample scattering problems, is the diffusion approximation. In this report, we address this by utilizing Berry phase (BP) evolution as a means for capturing anisotropic scattering characteristics of samples in the preceding depth where the diffusion approximation fails. We extend the polarization sensitive Monte Carlo method of Ramella-Roman, et al.,1 to include the computationally intensive tracking of photon trajectory in addition to polarization state at every scattering event. To speed-up the computations, which entail the appropriate rotations of reference frames, the code was parallelized using OpenMP. The results presented reveal that BP is strongly correlated to the photon penetration depth, thus potentiating the possibility of polarimetric depth resolved characterization of highly scattering samples, e.g., biological tissues.

  2. Earthquake forecasting based on data assimilation: sequential Monte Carlo methods for renewal point processes

    Directory of Open Access Journals (Sweden)

    M. J. Werner

    2011-02-01

    Full Text Available Data assimilation is routinely employed in meteorology, engineering and computer sciences to optimally combine noisy observations with prior model information for obtaining better estimates of a state, and thus better forecasts, than achieved by ignoring data uncertainties. Earthquake forecasting, too, suffers from measurement errors and partial model information and may thus gain significantly from data assimilation. We present perhaps the first fully implementable data assimilation method for earthquake forecasts generated by a point-process model of seismicity. We test the method on a synthetic and pedagogical example of a renewal process observed in noise, which is relevant for the seismic gap hypothesis, models of characteristic earthquakes and recurrence statistics of large quakes inferred from paleoseismic data records. To address the non-Gaussian statistics of earthquakes, we use sequential Monte Carlo methods, a set of flexible simulation-based methods for recursively estimating arbitrary posterior distributions. We perform extensive numerical simulations to demonstrate the feasibility and benefits of forecasting earthquakes based on data assimilation.

  3. Study on quantification method based on Monte Carlo sampling for multiunit probabilistic safety assessment models

    Energy Technology Data Exchange (ETDEWEB)

    Oh, Kye Min [KHNP Central Research Institute, Daejeon (Korea, Republic of); Han, Sang Hoon; Park, Jin Hee; Lim, Ho Gon; Yang, Joon Yang [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of); Heo, Gyun Young [Kyung Hee University, Yongin (Korea, Republic of)

    2017-06-15

    In Korea, many nuclear power plants operate at a single site based on geographical characteristics, but the population density near the sites is higher than that in other countries. Thus, multiunit accidents are a more important consideration than in other countries and should be addressed appropriately. Currently, there are many issues related to a multiunit probabilistic safety assessment (PSA). One of them is the quantification of a multiunit PSA model. A traditional PSA uses a Boolean manipulation of the fault tree in terms of the minimal cut set. However, such methods have some limitations when rare event approximations cannot be used effectively or a very small truncation limit should be applied to identify accident sequence combinations for a multiunit site. In particular, it is well known that seismic risk in terms of core damage frequency can be overestimated because there are many events that have a high failure probability. In this study, we propose a quantification method based on a Monte Carlo approach for a multiunit PSA model. This method can consider all possible accident sequence combinations in a multiunit site and calculate a more exact value for events that have a high failure probability. An example model for six identical units at a site was also developed and quantified to confirm the applicability of the proposed method.

  4. The development of GPU-based parallel PRNG for Monte Carlo applications in CUDA Fortran

    Directory of Open Access Journals (Sweden)

    Hamed Kargaran

    2016-04-01

    Full Text Available The implementation of Monte Carlo simulation on the CUDA Fortran requires a fast random number generation with good statistical properties on GPU. In this study, a GPU-based parallel pseudo random number generator (GPPRNG have been proposed to use in high performance computing systems. According to the type of GPU memory usage, GPU scheme is divided into two work modes including GLOBAL_MODE and SHARED_MODE. To generate parallel random numbers based on the independent sequence method, the combination of middle-square method and chaotic map along with the Xorshift PRNG have been employed. Implementation of our developed PPRNG on a single GPU showed a speedup of 150x and 470x (with respect to the speed of PRNG on a single CPU core for GLOBAL_MODE and SHARED_MODE, respectively. To evaluate the accuracy of our developed GPPRNG, its performance was compared to that of some other commercially available PPRNGs such as MATLAB, FORTRAN and Miller-Park algorithm through employing the specific standard tests. The results of this comparison showed that the developed GPPRNG in this study can be used as a fast and accurate tool for computational science applications.

  5. The development of GPU-based parallel PRNG for Monte Carlo applications in CUDA Fortran

    Energy Technology Data Exchange (ETDEWEB)

    Kargaran, Hamed, E-mail: h-kargaran@sbu.ac.ir; Minuchehr, Abdolhamid; Zolfaghari, Ahmad [Department of nuclear engineering, Shahid Behesti University, Tehran, 1983969411 (Iran, Islamic Republic of)

    2016-04-15

    The implementation of Monte Carlo simulation on the CUDA Fortran requires a fast random number generation with good statistical properties on GPU. In this study, a GPU-based parallel pseudo random number generator (GPPRNG) have been proposed to use in high performance computing systems. According to the type of GPU memory usage, GPU scheme is divided into two work modes including GLOBAL-MODE and SHARED-MODE. To generate parallel random numbers based on the independent sequence method, the combination of middle-square method and chaotic map along with the Xorshift PRNG have been employed. Implementation of our developed PPRNG on a single GPU showed a speedup of 150x and 470x (with respect to the speed of PRNG on a single CPU core) for GLOBAL-MODE and SHARED-MODE, respectively. To evaluate the accuracy of our developed GPPRNG, its performance was compared to that of some other commercially available PPRNGs such as MATLAB, FORTRAN and Miller-Park algorithm through employing the specific standard tests. The results of this comparison showed that the developed GPPRNG in this study can be used as a fast and accurate tool for computational science applications.

  6. Using gEUD based plan analysis method to evaluate proton vs. photon plans for lung cancer radiation therapy.

    Science.gov (United States)

    Xiao, Zhiyan; Zou, Wei J; Chen, Ting; Yue, Ning J; Jabbour, Salma K; Parikh, Rahul; Zhang, Miao

    2018-03-01

    The goal of this study was to exam the efficacy of current DVH based clinical guidelines draw from photon experience for lung cancer radiation therapy on proton therapy. Comparison proton plans and IMRT plans were generated for 10 lung patients treated in our proton facility. A gEUD based plan evaluation method was developed for plan evaluation. This evaluation method used normal lung gEUD(a) curve in which the model parameter "a" was sampled from the literature reported value. For all patients, the proton plans delivered lower normal lung V 5 Gy with similar V 20 Gy and similar target coverage. Based on current clinical guidelines, proton plans were ranked superior to IMRT plans for all 10 patients. However, the proton and IMRT normal lung gEUD(a) curves crossed for 8 patients within the tested range of "a", which means there was a possibility that proton plan would be worse than IMRT plan for lung sparing. A concept of deficiency index (DI) was introduced to quantify the probability of proton plans doing worse than IMRT plans. By applying threshold on DI, four patients' proton plan was ranked inferior to the IMRT plan. Meanwhile if a threshold to the location of curve crossing was applied, 6 patients' proton plan was ranked inferior to the IMRT plan. The contradictory ranking results between the current clinical guidelines and the gEUD(a) curve analysis demonstrated there is potential pitfalls by applying photon experience directly to the proton world. A comprehensive plan evaluation based on radio-biological models should be carried out to decide if a lung patient would really be benefit from proton therapy. © 2018 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.

  7. Monte Carlo based treatment planning systems for Boron Neutron Capture Therapy in Petten, The Netherlands

    Energy Technology Data Exchange (ETDEWEB)

    Nievaart, V A; Daquino, G G; Moss, R L [JRC European Commission, PO Box 2, 1755ZG Petten (Netherlands)

    2007-06-15

    Boron Neutron Capture Therapy (BNCT) is a bimodal form of radiotherapy for the treatment of tumour lesions. Since the cancer cells in the treatment volume are targeted with {sup 10}B, a higher dose is given to these cancer cells due to the {sup 10}B(n,{alpha}){sup 7}Li reaction, in comparison with the surrounding healthy cells. In Petten (The Netherlands), at the High Flux Reactor, a specially tailored neutron beam has been designed and installed. Over 30 patients have been treated with BNCT in 2 clinical protocols: a phase I study for the treatment of glioblastoma multiforme and a phase II study on the treatment of malignant melanoma. Furthermore, activities concerning the extra-corporal treatment of metastasis in the liver (from colorectal cancer) are in progress. The irradiation beam at the HFR contains both neutrons and gammas that, together with the complex geometries of both patient and beam set-up, demands for very detailed treatment planning calculations. A well designed Treatment Planning System (TPS) should obey the following general scheme: (1) a pre-processing phase (CT and/or MRI scans to create the geometric solid model, cross-section files for neutrons and/or gammas); (2) calculations (3D radiation transport, estimation of neutron and gamma fluences, macroscopic and microscopic dose); (3) post-processing phase (displaying of the results, iso-doses and -fluences). Treatment planning in BNCT is performed making use of Monte Carlo codes incorporated in a framework, which includes also the pre- and post-processing phases. In particular, the glioblastoma multiforme protocol used BNCT{sub r}tpe, while the melanoma metastases protocol uses NCTPlan. In addition, an ad hoc Positron Emission Tomography (PET) based treatment planning system (BDTPS) has been implemented in order to integrate the real macroscopic boron distribution obtained from PET scanning. BDTPS is patented and uses MCNP as the calculation engine. The precision obtained by the Monte Carlo

  8. Muusikamaailm : "Euroopa muusikakuu" Baselis. Leif Ove Andsnes Londonis. Konkursipreemiaid. Monte Pederson lahkunud / Priit Kuusk

    Index Scriptorium Estoniae

    Kuusk, Priit, 1938-

    2001-01-01

    Novembrikuus elab šveitsi linn Basel "Euroopa muusikakuu" tähe all. Noor norra pianist Leif Ove Andsnes kutsuti Londonisse esinema. Konkursipreemiaid erinevatel konkurssidelt. Suri ameerika laulja Monte Pederson

  9. CAD-based Monte Carlo program for integrated simulation of nuclear system SuperMC

    International Nuclear Information System (INIS)

    Wu, Yican; Song, Jing; Zheng, Huaqing; Sun, Guangyao; Hao, Lijuan; Long, Pengcheng; Hu, Liqin

    2015-01-01

    Highlights: • The new developed CAD-based Monte Carlo program named SuperMC for integrated simulation of nuclear system makes use of hybrid MC-deterministic method and advanced computer technologies. SuperMC is designed to perform transport calculation of various types of particles, depletion and activation calculation including isotope burn-up, material activation and shutdown dose, and multi-physics coupling calculation including thermo-hydraulics, fuel performance and structural mechanics. The bi-directional automatic conversion between general CAD models and physical settings and calculation models can be well performed. Results and process of simulation can be visualized with dynamical 3D dataset and geometry model. Continuous-energy cross section, burnup, activation, irradiation damage and material data etc. are used to support the multi-process simulation. Advanced cloud computing framework makes the computation and storage extremely intensive simulation more attractive just as a network service to support design optimization and assessment. The modular design and generic interface promotes its flexible manipulation and coupling of external solvers. • The new developed and incorporated advanced methods in SuperMC was introduced including hybrid MC-deterministic transport method, particle physical interaction treatment method, multi-physics coupling calculation method, geometry automatic modeling and processing method, intelligent data analysis and visualization method, elastic cloud computing technology and parallel calculation method. • The functions of SuperMC2.1 integrating automatic modeling, neutron and photon transport calculation, results and process visualization was introduced. It has been validated by using a series of benchmarking cases such as the fusion reactor ITER model and the fast reactor BN-600 model. - Abstract: Monte Carlo (MC) method has distinct advantages to simulate complicated nuclear systems and is envisioned as a routine

  10. Monte Carlo simulation of hybrid systems: An example

    International Nuclear Information System (INIS)

    Bacha, F.; D'Alencon, H.; Grivelet, J.; Jullien, E.; Jejcic, A.; Maillard, J.; Silva, J.; Zukanovich, R.; Vergnes, J.

    1997-01-01

    Simulation of hybrid systems needs tracking of particles from the GeV (incident proton beam) range down to a fraction of eV (thermic neutrons). We show how a GEANT based Monte-Carlo program can achieve this, with a realistic computer time and accompanying tools. An example of a dedicated original actinide burner is simulated with this chain. 8 refs., 5 figs

  11. Memory and learning behaviors mimicked in nanogranular SiO2-based proton conductor gated oxide-based synaptic transistors.

    Science.gov (United States)

    Wan, Chang Jin; Zhu, Li Qiang; Zhou, Ju Mei; Shi, Yi; Wan, Qing

    2013-11-07

    In neuroscience, signal processing, memory and learning function are established in the brain by modifying ionic fluxes in neurons and synapses. Emulation of memory and learning behaviors of biological systems by nanoscale ionic/electronic devices is highly desirable for building neuromorphic systems or even artificial neural networks. Here, novel artificial synapses based on junctionless oxide-based protonic/electronic hybrid transistors gated by nanogranular phosphorus-doped SiO2-based proton-conducting films are fabricated on glass substrates by a room-temperature process. Short-term memory (STM) and long-term memory (LTM) are mimicked by tuning the pulse gate voltage amplitude. The LTM process in such an artificial synapse is due to the proton-related interfacial electrochemical reaction. Our results are highly desirable for building future neuromorphic systems or even artificial networks via electronic elements.

  12. Application of the measurement-based Monte Carlo method in nasopharyngeal cancer patients for intensity modulated radiation therapy

    International Nuclear Information System (INIS)

    Yeh, C.Y.; Lee, C.C.; Chao, T.C.; Lin, M.H.; Lai, P.A.; Liu, F.H.; Tung, C.J.

    2014-01-01

    This study aims to utilize a measurement-based Monte Carlo (MBMC) method to evaluate the accuracy of dose distributions calculated using the Eclipse radiotherapy treatment planning system (TPS) based on the anisotropic analytical algorithm. Dose distributions were calculated for the nasopharyngeal carcinoma (NPC) patients treated with the intensity modulated radiotherapy (IMRT). Ten NPC IMRT plans were evaluated by comparing their dose distributions with those obtained from the in-house MBMC programs for the same CT images and beam geometry. To reconstruct the fluence distribution of the IMRT field, an efficiency map was obtained by dividing the energy fluence of the intensity modulated field by that of the open field, both acquired from an aS1000 electronic portal imaging device. The integrated image of the non-gated mode was used to acquire the full dose distribution delivered during the IMRT treatment. This efficiency map redistributed the particle weightings of the open field phase-space file for IMRT applications. Dose differences were observed in the tumor and air cavity boundary. The mean difference between MBMC and TPS in terms of the planning target volume coverage was 0.6% (range: 0.0–2.3%). The mean difference for the conformity index was 0.01 (range: 0.0–0.01). In conclusion, the MBMC method serves as an independent IMRT dose verification tool in a clinical setting. - Highlights: ► The patient-based Monte Carlo method serves as a reference standard to verify IMRT doses. ► 3D Dose distributions for NPC patients have been verified by the Monte Carlo method. ► Doses predicted by the Monte Carlo method matched closely with those by the TPS. ► The Monte Carlo method predicted a higher mean dose to the middle ears than the TPS. ► Critical organ doses should be confirmed to avoid overdose to normal organs

  13. The Calculation Of Titanium Buildup Factor Based On Monte Carlo Method

    International Nuclear Information System (INIS)

    Has, Hengky Istianto; Achmad, Balza; Harto, Andang Widi

    2001-01-01

    The objective of radioactive-waste container is to reduce radiation emission to the environment. For that purpose, we need material with ability to shield that radiation and last for 10.000 years. Titanium is one of the materials that can be used to make containers. Unfortunately, its buildup factor, which is an importance factor in setting up radiation shielding, has not been calculated. Therefore, the calculations of titanium buildup factor as a function of other parameters is needed. Buildup factor can be determined either experimentally or by simulation. The purpose of this study is to determine titanium buildup factor using simulation program based on Monte Carlo method. Monte Carlo is a stochastic method, therefore is proper to calculate nuclear radiation which naturally has random characteristic. Simulation program also able to give result while experiments can not be performed, because of their limitations.The result of the simulation is, that by increasing titanium thickness the buildup factor number and dosage increase. In contrary If photon energy is higher, then buildup factor number and dosage are lower. The photon energy used in the simulation was ranged from 0.2 MeV to 2.0 MeV with 0.2 MeV step size, while the thickness was ranged from 0.2 cm to 3.0 cm with step size of 0.2 cm. The highest buildup factor number is β = 1.4540 ± 0.047229 at 0.2 MeV photon energy with titanium thickness of 3.0 cm. The lowest is β = 1.0123 ± 0.000650 at 2.0 MeV photon energy with 0.2 cm thickness of titanium. For the dosage buildup factor, the highest dose is β D = 1.3991 ± 0.013999 at 0.2 MeV of the photon energy with a titanium thickness of 3.0 cm and the lowest is β D = 1.0042 ± 0.000597 at 2.0 MeV with titanium thickness of 0.2 cm. For the photon energy and the thickness of titanium used in simulation, buildup factor and dosage buildup factor as a function of photon energy and titanium thickness can be formulated as follow β = 1.1264 e - 0.0855 E e 0 .0584 T

  14. A GPU-based Monte Carlo dose calculation code for photon transport in a voxel phantom

    Energy Technology Data Exchange (ETDEWEB)

    Bellezzo, M.; Do Nascimento, E.; Yoriyaz, H., E-mail: mbellezzo@gmail.br [Instituto de Pesquisas Energeticas e Nucleares / CNEN, Av. Lineu Prestes 2242, Cidade Universitaria, 05508-000 Sao Paulo (Brazil)

    2014-08-15

    As the most accurate method to estimate absorbed dose in radiotherapy, Monte Carlo method has been widely used in radiotherapy treatment planning. Nevertheless, its efficiency can be improved for clinical routine applications. In this paper, we present the CUBMC code, a GPU-based Mc photon transport algorithm for dose calculation under the Compute Unified Device Architecture platform. The simulation of physical events is based on the algorithm used in Penelope, and the cross section table used is the one generated by the Material routine, als present in Penelope code. Photons are transported in voxel-based geometries with different compositions. To demonstrate the capabilities of the algorithm developed in the present work four 128 x 128 x 128 voxel phantoms have been considered. One of them is composed by a homogeneous water-based media, the second is composed by bone, the third is composed by lung and the fourth is composed by a heterogeneous bone and vacuum geometry. Simulations were done considering a 6 MeV monoenergetic photon point source. There are two distinct approaches that were used for transport simulation. The first of them forces the photon to stop at every voxel frontier, the second one is the Woodcock method, where the photon stop in the frontier will be considered depending on the material changing across the photon travel line. Dose calculations using these methods are compared for validation with Penelope and MCNP5 codes. Speed-up factors are compared using a NVidia GTX 560-Ti GPU card against a 2.27 GHz Intel Xeon CPU processor. (Author)

  15. Reporting and analyzing statistical uncertainties in Monte Carlo-based treatment planning

    International Nuclear Information System (INIS)

    Chetty, Indrin J.; Rosu, Mihaela; Kessler, Marc L.; Fraass, Benedick A.; Haken, Randall K. ten; Kong, Feng-Ming; McShan, Daniel L.

    2006-01-01

    Purpose: To investigate methods of reporting and analyzing statistical uncertainties in doses to targets and normal tissues in Monte Carlo (MC)-based treatment planning. Methods and Materials: Methods for quantifying statistical uncertainties in dose, such as uncertainty specification to specific dose points, or to volume-based regions, were analyzed in MC-based treatment planning for 5 lung cancer patients. The effect of statistical uncertainties on target and normal tissue dose indices was evaluated. The concept of uncertainty volume histograms for targets and organs at risk was examined, along with its utility, in conjunction with dose volume histograms, in assessing the acceptability of the statistical precision in dose distributions. The uncertainty evaluation tools were extended to four-dimensional planning for application on multiple instances of the patient geometry. All calculations were performed using the Dose Planning Method MC code. Results: For targets, generalized equivalent uniform doses and mean target doses converged at 150 million simulated histories, corresponding to relative uncertainties of less than 2% in the mean target doses. For the normal lung tissue (a volume-effect organ), mean lung dose and normal tissue complication probability converged at 150 million histories despite the large range in the relative organ uncertainty volume histograms. For 'serial' normal tissues such as the spinal cord, large fluctuations exist in point dose relative uncertainties. Conclusions: The tools presented here provide useful means for evaluating statistical precision in MC-based dose distributions. Tradeoffs between uncertainties in doses to targets, volume-effect organs, and 'serial' normal tissues must be considered carefully in determining acceptable levels of statistical precision in MC-computed dose distributions

  16. A GPU-based Monte Carlo dose calculation code for photon transport in a voxel phantom

    International Nuclear Information System (INIS)

    Bellezzo, M.; Do Nascimento, E.; Yoriyaz, H.

    2014-08-01

    As the most accurate method to estimate absorbed dose in radiotherapy, Monte Carlo method has been widely used in radiotherapy treatment planning. Nevertheless, its efficiency can be improved for clinical routine applications. In this paper, we present the CUBMC code, a GPU-based Mc photon transport algorithm for dose calculation under the Compute Unified Device Architecture platform. The simulation of physical events is based on the algorithm used in Penelope, and the cross section table used is the one generated by the Material routine, als present in Penelope code. Photons are transported in voxel-based geometries with different compositions. To demonstrate the capabilities of the algorithm developed in the present work four 128 x 128 x 128 voxel phantoms have been considered. One of them is composed by a homogeneous water-based media, the second is composed by bone, the third is composed by lung and the fourth is composed by a heterogeneous bone and vacuum geometry. Simulations were done considering a 6 MeV monoenergetic photon point source. There are two distinct approaches that were used for transport simulation. The first of them forces the photon to stop at every voxel frontier, the second one is the Woodcock method, where the photon stop in the frontier will be considered depending on the material changing across the photon travel line. Dose calculations using these methods are compared for validation with Penelope and MCNP5 codes. Speed-up factors are compared using a NVidia GTX 560-Ti GPU card against a 2.27 GHz Intel Xeon CPU processor. (Author)

  17. Abstract ID: 197 Monte Carlo simulations of X-ray grating interferometry based imaging systems.

    Science.gov (United States)

    Tessarini, Stefan; Fix, Michael K; Volken, Werner; Frei, Daniel; Stampanoni, Marco F M

    2018-01-01

    Over the last couple of years the implementation of Monte Carlo (MC) methods of grating based imaging techniques is of increasing interest. Several different approaches were taken to include coherent effects into MC in order to simulate the radiation transport of the image forming procedure. These include full MC using FLUKA [1], which however are only considering monochromatic sources. Alternatively, ray-tracing based MC [2] allow fast simulations with the limitation to provide only qualitative results, i.e. this technique is not suitable for dose calculation in the imaged object. Finally, hybrid models [3] were used allowing quantitative results in reasonable computation time, however only two-dimensional implementations are available. Thus, this work aims to develop a full MC framework for X-ray grating interferometry imaging systems using polychromatic sources suitable for large-scale samples. For this purpose the EGSnrc C++ MC code system is extended to take Snell's law, the optical path length and Huygens principle into account. Thereby the EGSnrc library was modified, e.g. the complex index of refraction has to be assigned to each region depending on the material. The framework is setup to be user-friendly and robust with respect to future updates of the EGSnrc package. These implementations have to be tested using dedicated academic situations. Next steps include the validation by comparisons of measurements for different setups with the corresponding MC simulations. Furthermore, the newly developed implementation will be compared with other simulation approaches. This framework will then serve as bases for dose calculation on CT data and has further potential to investigate the image formation process in grating based imaging systems. Copyright © 2017.

  18. Efficiency of respiratory-gated delivery of synchrotron-based pulsed proton irradiation

    International Nuclear Information System (INIS)

    Tsunashima, Yoshikazu; Vedam, Sastry; Dong, Lei; Bues, Martin; Balter, Peter; Smith, Alfred; Mohan, Radhe; Umezawa, Masumi; Sakae, Takeji

    2008-01-01

    Significant differences exist in respiratory-gated proton beam delivery with a synchrotron-based accelerator system when compared to photon therapy with a conventional linear accelerator. Delivery of protons with a synchrotron accelerator is governed by a magnet excitation cycle pattern. Optimal synchronization of the magnet excitation cycle pattern with the respiratory motion pattern is critical to the efficiency of respiratory-gated proton delivery. There has been little systematic analysis to optimize the accelerator's operational parameters to improve gated treatment efficiency. The goal of this study was to estimate the overall efficiency of respiratory-gated synchrotron-based proton irradiation through realistic simulation. Using 62 respiratory motion traces from 38 patients, we simulated respiratory gating for duty cycles of 30%, 20% and 10% around peak exhalation for various fixed and variable magnet excitation patterns. In each case, the time required to deliver 100 monitor units in both non-gated and gated irradiation scenarios was determined. Based on results from this study, the minimum time required to deliver 100 MU was 1.1 min for non-gated irradiation. For respiratory-gated delivery at a 30% duty cycle around peak exhalation, corresponding average delivery times were typically three times longer with a fixed magnet excitation cycle pattern. However, when a variable excitation cycle was allowed in synchrony with the patient's respiratory cycle, the treatment time only doubled. Thus, respiratory-gated delivery of synchrotron-based pulsed proton irradiation is feasible and more efficient when a variable magnet excitation cycle pattern is used

  19. Absorbed dose calculations using mesh-based human phantoms and Monte Carlo methods

    International Nuclear Information System (INIS)

    Kramer, Richard

    2010-01-01

    Full text. Health risks attributable to ionizing radiation are considered to be a function of the absorbed dose to radiosensitive organs and tissues of the human body. However, as human tissue cannot express itself in terms of absorbed dose, exposure models have to be used to determine the distribution of absorbed dose throughout the human body. An exposure model, be it physical or virtual, consists of a representation of the human body, called phantom, plus a method for transporting ionizing radiation through the phantom and measuring or calculating the absorbed dose to organ and tissues of interest. Female Adult meSH (FASH) and the Male Adult meSH (MASH) virtual phantoms have been developed at the University of Pernambuco in Recife/Brazil based on polygon mesh surfaces using open source software tools. Representing standing adults, FASH and MASH have organ and tissue masses, body height and mass adjusted to the anatomical data published by the International Commission on Radiological Protection for the reference male and female adult. For the purposes of absorbed dose calculations the phantoms have been coupled to the EGSnrc Monte Carlo code, which transports photons, electrons and positrons through arbitrary media. This presentation reports on the development of the FASH and the MASH phantoms and will show dosimetric applications for X-ray diagnosis and for prostate brachytherapy. (author)

  20. Absorbed Dose Calculations Using Mesh-based Human Phantoms And Monte Carlo Methods

    International Nuclear Information System (INIS)

    Kramer, Richard

    2011-01-01

    Health risks attributable to the exposure to ionizing radiation are considered to be a function of the absorbed or equivalent dose to radiosensitive organs and tissues. However, as human tissue cannot express itself in terms of equivalent dose, exposure models have to be used to determine the distribution of equivalent dose throughout the human body. An exposure model, be it physical or computational, consists of a representation of the human body, called phantom, plus a method for transporting ionizing radiation through the phantom and measuring or calculating the equivalent dose to organ and tissues of interest. The FASH2 (Female Adult meSH) and the MASH2 (Male Adult meSH) computational phantoms have been developed at the University of Pernambuco in Recife/Brazil based on polygon mesh surfaces using open source software tools and anatomical atlases. Representing standing adults, FASH2 and MASH2 have organ and tissue masses, body height and body mass adjusted to the anatomical data published by the International Commission on Radiological Protection for the reference male and female adult. For the purposes of absorbed dose calculations the phantoms have been coupled to the EGSnrc Monte Carlo code, which can transport photons, electrons and positrons through arbitrary media. This paper reviews the development of the FASH2 and the MASH2 phantoms and presents dosimetric applications for X-ray diagnosis and for prostate brachytherapy.

  1. Absorbed Dose Calculations Using Mesh-based Human Phantoms And Monte Carlo Methods

    Science.gov (United States)

    Kramer, Richard

    2011-08-01

    Health risks attributable to the exposure to ionizing radiation are considered to be a function of the absorbed or equivalent dose to radiosensitive organs and tissues. However, as human tissue cannot express itself in terms of equivalent dose, exposure models have to be used to determine the distribution of equivalent dose throughout the human body. An exposure model, be it physical or computational, consists of a representation of the human body, called phantom, plus a method for transporting ionizing radiation through the phantom and measuring or calculating the equivalent dose to organ and tissues of interest. The FASH2 (Female Adult meSH) and the MASH2 (Male Adult meSH) computational phantoms have been developed at the University of Pernambuco in Recife/Brazil based on polygon mesh surfaces using open source software tools and anatomical atlases. Representing standing adults, FASH2 and MASH2 have organ and tissue masses, body height and body mass adjusted to the anatomical data published by the International Commission on Radiological Protection for the reference male and female adult. For the purposes of absorbed dose calculations the phantoms have been coupled to the EGSnrc Monte Carlo code, which can transport photons, electrons and positrons through arbitrary media. This paper reviews the development of the FASH2 and the MASH2 phantoms and presents dosimetric applications for X-ray diagnosis and for prostate brachytherapy.

  2. Absorbed dose calculations using mesh-based human phantoms and Monte Carlo methods

    Energy Technology Data Exchange (ETDEWEB)

    Kramer, Richard [Universidade Federal de Pernambuco (UFPE), Recife, PE (Brazil)

    2010-07-01

    Full text. Health risks attributable to ionizing radiation are considered to be a function of the absorbed dose to radiosensitive organs and tissues of the human body. However, as human tissue cannot express itself in terms of absorbed dose, exposure models have to be used to determine the distribution of absorbed dose throughout the human body. An exposure model, be it physical or virtual, consists of a representation of the human body, called phantom, plus a method for transporting ionizing radiation through the phantom and measuring or calculating the absorbed dose to organ and tissues of interest. Female Adult meSH (FASH) and the Male Adult meSH (MASH) virtual phantoms have been developed at the University of Pernambuco in Recife/Brazil based on polygon mesh surfaces using open source software tools. Representing standing adults, FASH and MASH have organ and tissue masses, body height and mass adjusted to the anatomical data published by the International Commission on Radiological Protection for the reference male and female adult. For the purposes of absorbed dose calculations the phantoms have been coupled to the EGSnrc Monte Carlo code, which transports photons, electrons and positrons through arbitrary media. This presentation reports on the development of the FASH and the MASH phantoms and will show dosimetric applications for X-ray diagnosis and for prostate brachytherapy. (author)

  3. Optimization of Control Points Number at Coordinate Measurements based on the Monte-Carlo Method

    Science.gov (United States)

    Korolev, A. A.; Kochetkov, A. V.; Zakharov, O. V.

    2018-01-01

    Improving the quality of products causes an increase in the requirements for the accuracy of the dimensions and shape of the surfaces of the workpieces. This, in turn, raises the requirements for accuracy and productivity of measuring of the workpieces. The use of coordinate measuring machines is currently the most effective measuring tool for solving similar problems. The article proposes a method for optimizing the number of control points using Monte Carlo simulation. Based on the measurement of a small sample from batches of workpieces, statistical modeling is performed, which allows one to obtain interval estimates of the measurement error. This approach is demonstrated by examples of applications for flatness, cylindricity and sphericity. Four options of uniform and uneven arrangement of control points are considered and their comparison is given. It is revealed that when the number of control points decreases, the arithmetic mean decreases, the standard deviation of the measurement error increases and the probability of the measurement α-error increases. In general, it has been established that it is possible to repeatedly reduce the number of control points while maintaining the required measurement accuracy.

  4. Monte Carlo efficiency calibration of a neutron generator-based total-body irradiator

    International Nuclear Information System (INIS)

    Shypailo, R.J.; Ellis, K.J.

    2009-01-01

    Many body composition measurement systems are calibrated against a single-sized reference phantom. Prompt-gamma neutron activation (PGNA) provides the only direct measure of total body nitrogen (TBN), an index of the body's lean tissue mass. In PGNA systems, body size influences neutron flux attenuation, induced gamma signal distribution, and counting efficiency. Thus, calibration based on a single-sized phantom could result in inaccurate TBN values. We used Monte Carlo simulations (MCNP-5; Los Alamos National Laboratory) in order to map a system's response to the range of body weights (65-160 kg) and body fat distributions (25-60%) in obese humans. Calibration curves were constructed to derive body-size correction factors relative to a standard reference phantom, providing customized adjustments to account for differences in body habitus of obese adults. The use of MCNP-generated calibration curves should allow for a better estimate of the true changes in lean tissue mass that many occur during intervention programs focused only on weight loss. (author)

  5. dsmcFoam+: An OpenFOAM based direct simulation Monte Carlo solver

    Science.gov (United States)

    White, C.; Borg, M. K.; Scanlon, T. J.; Longshaw, S. M.; John, B.; Emerson, D. R.; Reese, J. M.

    2018-03-01

    dsmcFoam+ is a direct simulation Monte Carlo (DSMC) solver for rarefied gas dynamics, implemented within the OpenFOAM software framework, and parallelised with MPI. It is open-source and released under the GNU General Public License in a publicly available software repository that includes detailed documentation and tutorial DSMC gas flow cases. This release of the code includes many features not found in standard dsmcFoam, such as molecular vibrational and electronic energy modes, chemical reactions, and subsonic pressure boundary conditions. Since dsmcFoam+ is designed entirely within OpenFOAM's C++ object-oriented framework, it benefits from a number of key features: the code emphasises extensibility and flexibility so it is aimed first and foremost as a research tool for DSMC, allowing new models and test cases to be developed and tested rapidly. All DSMC cases are as straightforward as setting up any standard OpenFOAM case, as dsmcFoam+ relies upon the standard OpenFOAM dictionary based directory structure. This ensures that useful pre- and post-processing capabilities provided by OpenFOAM remain available even though the fully Lagrangian nature of a DSMC simulation is not typical of most OpenFOAM applications. We show that dsmcFoam+ compares well to other well-known DSMC codes and to analytical solutions in terms of benchmark results.

  6. A comprehensive revisit of the ρ meson with improved Monte-Carlo based QCD sum rules

    Science.gov (United States)

    Wang, Qi-Nan; Zhang, Zhu-Feng; Steele, T. G.; Jin, Hong-Ying; Huang, Zhuo-Ran

    2017-07-01

    We improve the Monte-Carlo based QCD sum rules by introducing the rigorous Hölder-inequality-determined sum rule window and a Breit-Wigner type parametrization for the phenomenological spectral function. In this improved sum rule analysis methodology, the sum rule analysis window can be determined without any assumptions on OPE convergence or the QCD continuum. Therefore, an unbiased prediction can be obtained for the phenomenological parameters (the hadronic mass and width etc.). We test the new approach in the ρ meson channel with re-examination and inclusion of α s corrections to dimension-4 condensates in the OPE. We obtain results highly consistent with experimental values. We also discuss the possible extension of this method to some other channels. Supported by NSFC (11175153, 11205093, 11347020), Open Foundation of the Most Important Subjects of Zhejiang Province, and K. C. Wong Magna Fund in Ningbo University, TGS is Supported by the Natural Sciences and Engineering Research Council of Canada (NSERC), Z. F. Zhang and Z. R. Huang are Grateful to the University of Saskatchewan for its Warm Hospitality

  7. Voxel-based Monte Carlo simulation of X-ray imaging and spectroscopy experiments

    International Nuclear Information System (INIS)

    Bottigli, U.; Brunetti, A.; Golosio, B.; Oliva, P.; Stumbo, S.; Vincze, L.; Randaccio, P.; Bleuet, P.; Simionovici, A.; Somogyi, A.

    2004-01-01

    A Monte Carlo code for the simulation of X-ray imaging and spectroscopy experiments in heterogeneous samples is presented. The energy spectrum, polarization and profile of the incident beam can be defined so that X-ray tube systems as well as synchrotron sources can be simulated. The sample is modeled as a 3D regular grid. The chemical composition and density is given at each point of the grid. Photoelectric absorption, fluorescent emission, elastic and inelastic scattering are included in the simulation. The core of the simulation is a fast routine for the calculation of the path lengths of the photon trajectory intersections with the grid voxels. The voxel representation is particularly useful for samples that cannot be well described by a small set of polyhedra. This is the case of most naturally occurring samples. In such cases, voxel-based simulations are much less expensive in terms of computational cost than simulations on a polygonal representation. The efficient scheme used for calculating the path lengths in the voxels and the use of variance reduction techniques make the code suitable for the detailed simulation of complex experiments on generic samples in a relatively short time. Examples of applications to X-ray imaging and spectroscopy experiments are discussed

  8. IVF cycle cost estimation using Activity Based Costing and Monte Carlo simulation.

    Science.gov (United States)

    Cassettari, Lucia; Mosca, Marco; Mosca, Roberto; Rolando, Fabio; Costa, Mauro; Pisaturo, Valerio

    2016-03-01

    The Authors present a new methodological approach in stochastic regime to determine the actual costs of an healthcare process. The paper specifically shows the application of the methodology for the determination of the cost of an Assisted reproductive technology (ART) treatment in Italy. The reason of this research comes from the fact that deterministic regime is inadequate to implement an accurate estimate of the cost of this particular treatment. In fact the durations of the different activities involved are unfixed and described by means of frequency distributions. Hence the need to determine in addition to the mean value of the cost, the interval within which it is intended to vary with a known confidence level. Consequently the cost obtained for each type of cycle investigated (in vitro fertilization and embryo transfer with or without intracytoplasmic sperm injection), shows tolerance intervals around the mean value sufficiently restricted as to make the data obtained statistically robust and therefore usable also as reference for any benchmark with other Countries. It should be noted that under a methodological point of view the approach was rigorous. In fact it was used both the technique of Activity Based Costing for determining the cost of individual activities of the process both the Monte Carlo simulation, with control of experimental error, for the construction of the tolerance intervals on the final result.

  9. Learning Algorithm of Boltzmann Machine Based on Spatial Monte Carlo Integration Method

    Directory of Open Access Journals (Sweden)

    Muneki Yasuda

    2018-04-01

    Full Text Available The machine learning techniques for Markov random fields are fundamental in various fields involving pattern recognition, image processing, sparse modeling, and earth science, and a Boltzmann machine is one of the most important models in Markov random fields. However, the inference and learning problems in the Boltzmann machine are NP-hard. The investigation of an effective learning algorithm for the Boltzmann machine is one of the most important challenges in the field of statistical machine learning. In this paper, we study Boltzmann machine learning based on the (first-order spatial Monte Carlo integration method, referred to as the 1-SMCI learning method, which was proposed in the author’s previous paper. In the first part of this paper, we compare the method with the maximum pseudo-likelihood estimation (MPLE method using a theoretical and a numerical approaches, and show the 1-SMCI learning method is more effective than the MPLE. In the latter part, we compare the 1-SMCI learning method with other effective methods, ratio matching and minimum probability flow, using a numerical experiment, and show the 1-SMCI learning method outperforms them.

  10. Limits on the Efficiency of Event-Based Algorithms for Monte Carlo Neutron Transport

    Energy Technology Data Exchange (ETDEWEB)

    Romano, Paul K.; Siegel, Andrew R.

    2017-04-16

    The traditional form of parallelism in Monte Carlo particle transport simulations, wherein each individual particle history is considered a unit of work, does not lend itself well to data-level parallelism. Event-based algorithms, which were originally used for simulations on vector processors, may offer a path toward better utilizing data-level parallelism in modern computer architectures. In this study, a simple model is developed for estimating the efficiency of the event-based particle transport algorithm under two sets of assumptions. Data collected from simulations of four reactor problems using OpenMC was then used in conjunction with the models to calculate the speedup due to vectorization as a function of two parameters: the size of the particle bank and the vector width. When each event type is assumed to have constant execution time, the achievable speedup is directly related to the particle bank size. We observed that the bank size generally needs to be at least 20 times greater than vector size in order to achieve vector efficiency greater than 90%. When the execution times for events are allowed to vary, however, the vector speedup is also limited by differences in execution time for events being carried out in a single event-iteration. For some problems, this implies that vector effciencies over 50% may not be attainable. While there are many factors impacting performance of an event-based algorithm that are not captured by our model, it nevertheless provides insights into factors that may be limiting in a real implementation.

  11. Independent Monte-Carlo dose calculation for MLC based CyberKnife radiotherapy

    Science.gov (United States)

    Mackeprang, P.-H.; Vuong, D.; Volken, W.; Henzen, D.; Schmidhalter, D.; Malthaner, M.; Mueller, S.; Frei, D.; Stampanoni, M. F. M.; Dal Pra, A.; Aebersold, D. M.; Fix, M. K.; Manser, P.

    2018-01-01

    This work aims to develop, implement and validate a Monte Carlo (MC)-based independent dose calculation (IDC) framework to perform patient-specific quality assurance (QA) for multi-leaf collimator (MLC)-based CyberKnife® (Accuray Inc., Sunnyvale, CA) treatment plans. The IDC framework uses an XML-format treatment plan as exported from the treatment planning system (TPS) and DICOM format patient CT data, an MC beam model using phase spaces, CyberKnife MLC beam modifier transport using the EGS++ class library, a beam sampling and coordinate transformation engine and dose scoring using DOSXYZnrc. The framework is validated against dose profiles and depth dose curves of single beams with varying field sizes in a water tank in units of cGy/Monitor Unit and against a 2D dose distribution of a full prostate treatment plan measured with Gafchromic EBT3 (Ashland Advanced Materials, Bridgewater, NJ) film in a homogeneous water-equivalent slab phantom. The film measurement is compared to IDC results by gamma analysis using 2% (global)/2 mm criteria. Further, the dose distribution of the clinical treatment plan in the patient CT is compared to TPS calculation by gamma analysis using the same criteria. Dose profiles from IDC calculation in a homogeneous water phantom agree within 2.3% of the global max dose or 1 mm distance to agreement to measurements for all except the smallest field size. Comparing the film measurement to calculated dose, 99.9% of all voxels pass gamma analysis, comparing dose calculated by the IDC framework to TPS calculated dose for the clinical prostate plan shows 99.0% passing rate. IDC calculated dose is found to be up to 5.6% lower than dose calculated by the TPS in this case near metal fiducial markers. An MC-based modular IDC framework was successfully developed, implemented and validated against measurements and is now available to perform patient-specific QA by IDC.

  12. Modeling parameterized geometry in GPU-based Monte Carlo particle transport simulation for radiotherapy.

    Science.gov (United States)

    Chi, Yujie; Tian, Zhen; Jia, Xun

    2016-08-07

    Monte Carlo (MC) particle transport simulation on a graphics-processing unit (GPU) platform has been extensively studied recently due to the efficiency advantage achieved via massive parallelization. Almost all of the existing GPU-based MC packages were developed for voxelized geometry. This limited application scope of these packages. The purpose of this paper is to develop a module to model parametric geometry and integrate it in GPU-based MC simulations. In our module, each continuous region was defined by its bounding surfaces that were parameterized by quadratic functions. Particle navigation functions in this geometry were developed. The module was incorporated to two previously developed GPU-based MC packages and was tested in two example problems: (1) low energy photon transport simulation in a brachytherapy case with a shielded cylinder applicator and (2) MeV coupled photon/electron transport simulation in a phantom containing several inserts of different shapes. In both cases, the calculated dose distributions agreed well with those calculated in the corresponding voxelized geometry. The averaged dose differences were 1.03% and 0.29%, respectively. We also used the developed package to perform simulations of a Varian VS 2000 brachytherapy source and generated a phase-space file. The computation time under the parameterized geometry depended on the memory location storing the geometry data. When the data was stored in GPU's shared memory, the highest computational speed was achieved. Incorporation of parameterized geometry yielded a computation time that was ~3 times of that in the corresponding voxelized geometry. We also developed a strategy to use an auxiliary index array to reduce frequency of geometry calculations and hence improve efficiency. With this strategy, the computational time ranged in 1.75-2.03 times of the voxelized geometry for coupled photon/electron transport depending on the voxel dimension of the auxiliary index array, and in 0

  13. Monte Carlo study of particle production in diffractive proton-proton collisions at √(s) = 13 TeV with the very forward detector combined with central information

    Energy Technology Data Exchange (ETDEWEB)

    Zhou, Qi-Dong [Nagoya University, Institute for Space-Earth Environmental Research, Nagoya (Japan); Itow, Yoshitaka; Sako, Takashi [Nagoya University, Institute for Space-Earth Environmental Research, Nagoya (Japan); Nagoya University, Kobayashi-Maskawa Institute, Nagoya (Japan); Menjo, Hiroaki [Nagoya University, Graduate School of Science, Nagoya (Japan)

    2017-04-15

    Very forward (VF) detectors in hadron colliders, having unique sensitivity to diffractive processes, can be a powerful tool for studying diffractive dissociation by combining them with central detectors. Several Monte Carlo simulation samples in p-p collisions at √(s) = 13 TeV were analyzed, and different nondiffractive and diffractive contributions were clarified through differential cross sections of forward neutral particles. Diffraction selection criteria in the VF-triggered-event samples were determined by using the central track information. The corresponding selection applicable in real experiments has ∼ 100% purity and 30-70% efficiency. Consequently, the central information enables classification of the forward productions into diffraction and nondiffraction categories; in particular, most of the surviving events from the selection belong to low-mass diffraction events at log{sub 10}(ξ{sub x}) < -5.5. Therefore, the combined method can uniquely access the low-mass diffraction regime experimentally. (orig.)

  14. An exercise in model validation: Comparing univariate statistics and Monte Carlo-based multivariate statistics

    International Nuclear Information System (INIS)

    Weathers, J.B.; Luck, R.; Weathers, J.W.

    2009-01-01

    The complexity of mathematical models used by practicing engineers is increasing due to the growing availability of sophisticated mathematical modeling tools and ever-improving computational power. For this reason, the need to define a well-structured process for validating these models against experimental results has become a pressing issue in the engineering community. This validation process is partially characterized by the uncertainties associated with the modeling effort as well as the experimental results. The net impact of the uncertainties on the validation effort is assessed through the 'noise level of the validation procedure', which can be defined as an estimate of the 95% confidence uncertainty bounds for the comparison error between actual experimental results and model-based predictions of the same quantities of interest. Although general descriptions associated with the construction of the noise level using multivariate statistics exists in the literature, a detailed procedure outlining how to account for the systematic and random uncertainties is not available. In this paper, the methodology used to derive the covariance matrix associated with the multivariate normal pdf based on random and systematic uncertainties is examined, and a procedure used to estimate this covariance matrix using Monte Carlo analysis is presented. The covariance matrices are then used to construct approximate 95% confidence constant probability contours associated with comparison error results for a practical example. In addition, the example is used to show the drawbacks of using a first-order sensitivity analysis when nonlinear local sensitivity coefficients exist. Finally, the example is used to show the connection between the noise level of the validation exercise calculated using multivariate and univariate statistics.

  15. CT-Based Brachytherapy Treatment Planning using Monte Carlo Simulation Aided by an Interface Software

    Directory of Open Access Journals (Sweden)

    Vahid Moslemi

    2011-03-01

    Full Text Available Introduction: In brachytherapy, radioactive sources are placed close to the tumor, therefore, small changes in their positions can cause large changes in the dose distribution. This emphasizes the need for computerized treatment planning. The usual method for treatment planning of cervix brachytherapy uses conventional radiographs in the Manchester system. Nowadays, because of their advantages in locating the source positions and the surrounding tissues, CT and MRI images are replacing conventional radiographs. In this study, we used CT images in Monte Carlo based dose calculation for brachytherapy treatment planning, using an interface software to create the geometry file required in the MCNP code. The aim of using the interface software is to facilitate and speed up the geometry set-up for simulations based on the patient’s anatomy. This paper examines the feasibility of this method in cervix brachytherapy and assesses its accuracy and speed. Material and Methods: For dosimetric measurements regarding the treatment plan, a pelvic phantom was made from polyethylene in which the treatment applicators could be placed. For simulations using CT images, the phantom was scanned at 120 kVp. Using an interface software written in MATLAB, the CT images were converted into MCNP input file and the simulation was then performed. Results: Using the interface software, preparation time for the simulations of the applicator and surrounding structures was approximately 3 minutes; the corresponding time needed in the conventional MCNP geometry entry being approximately 1 hour. The discrepancy in the simulated and measured doses to point A was 1.7% of the prescribed dose.  The corresponding dose differences between the two methods in rectum and bladder were 3.0% and 3.7% of the prescribed dose, respectively. Comparing the results of simulation using the interface software with those of simulation using the standard MCNP geometry entry showed a less than 1

  16. Nanostructure-based proton exchange membrane for fuel cell applications at high temperature.

    Science.gov (United States)

    Li, Junsheng; Wang, Zhengbang; Li, Junrui; Pan, Mu; Tang, Haolin

    2014-02-01

    As a clean and highly efficient energy source, the proton exchange membrane fuel cell (PEMFC) has been considered an ideal alternative to traditional fossil energy sources. Great efforts have been devoted to realizing the commercialization of the PEMFC in the past decade. To eliminate some technical problems that are associated with the low-temperature operation (such as catalyst poisoning and poor water management), PEMFCs are usually operated at elevated temperatures (e.g., > 100 degrees C). However, traditional proton exchange membrane (PEM) shows poor performance at elevated temperature. To achieve a high-performance PEM for high temperature fuel cell applications, novel PEMs, which are based on nanostructures, have been developed recently. In this review, we discuss and summarize the methods for fabricating the nanostructure-based PEMs for PEMFC operated at elevated temperatures and the high temperature performance of these PEMs. We also give an outlook on the rational design and development of the nanostructure-based PEMs.

  17. First 3D measurements of proton beams in a deformable silicone-based dosimeter

    Science.gov (United States)

    Høye, E. M.; Sadel, M.; Kaplan, L.; Skyt, P. S.; Muren, L. P.; Petersen, J. B. B.; Swakoń, J.; Mierzwińska, G.; Rydygier, M.; Malinowski, L.; Balling, P.

    2017-05-01

    3D dosimetry provides high-resolution dose information of radiation therapy (RT), and is explored to enable and secure high-quality delivery of advanced RT modalities, including proton therapy. We present the first 3D measurements of spot-scanning proton plans in a silicone-based, radiochromic dosimeter with deformation properties. The dose information was read-out by optical CT-scanning. We found that the dosimeter signal was quenched close to the Bragg peak, and that this had a large impact on a measured spread-out Bragg peak. The dose response was linear both in the entrance region and in the Bragg peak, however, the dose response significantly reduced in the Bragg peak. Quenching was attributed to a linear-energy-transfer dependent dose response. Linear energy transfer distributions for each proton treatment plan will provide a means for calibrating the optical measurement to linear energy transfer, as well as dose. This might enable use of the silicone-dosimeter in quality assurance of proton beams.

  18. Proton therapy

    Science.gov (United States)

    Proton beam therapy; Cancer - proton therapy; Radiation therapy - proton therapy; Prostate cancer - proton therapy ... that use x-rays to destroy cancer cells, proton therapy uses a beam of special particles called ...

  19. A scintillator-based online detector for the angularly resolved measurement of laser-accelerated proton spectra.

    Science.gov (United States)

    Metzkes, J; Karsch, L; Kraft, S D; Pawelke, J; Richter, C; Schürer, M; Sobiella, M; Stiller, N; Zeil, K; Schramm, U

    2012-12-01

    In recent years, a new generation of high repetition rate (~10 Hz), high power (~100 TW) laser systems has stimulated intense research on laser-driven sources for fast protons. Considering experimental instrumentation, this development requires online diagnostics for protons to be added to the established offline detection tools such as solid state track detectors or radiochromic films. In this article, we present the design and characterization of a scintillator-based online detector that gives access to the angularly resolved proton distribution along one spatial dimension and resolves 10 different proton energy ranges. Conceived as an online detector for key parameters in laser-proton acceleration, such as the maximum proton energy and the angular distribution, the detector features a spatial resolution of ~1.3 mm and a spectral resolution better than 1.5 MeV for a maximum proton energy above 12 MeV in the current design. Regarding its areas of application, we consider the detector a useful complement to radiochromic films and Thomson parabola spectrometers, capable to give immediate feedback on the experimental performance. The detector was characterized at an electrostatic Van de Graaff tandetron accelerator and tested in a laser-proton acceleration experiment, proving its suitability as a diagnostic device for laser-accelerated protons.

  20. Volume Measurement Algorithm for Food Product with Irregular Shape using Computer Vision based on Monte Carlo Method

    Directory of Open Access Journals (Sweden)

    Joko Siswantoro

    2014-11-01

    Full Text Available Volume is one of important issues in the production and processing of food product. Traditionally, volume measurement can be performed using water displacement method based on Archimedes’ principle. Water displacement method is inaccurate and considered as destructive method. Computer vision offers an accurate and nondestructive method in measuring volume of food product. This paper proposes algorithm for volume measurement of irregular shape food product using computer vision based on Monte Carlo method. Five images of object were acquired from five different views and then processed to obtain the silhouettes of object. From the silhouettes of object, Monte Carlo method was performed to approximate the volume of object. The simulation result shows that the algorithm produced high accuracy and precision for volume measurement.

  1. Monte-Carlo Simulation for PDC-Based Optical CDMA System

    Directory of Open Access Journals (Sweden)

    FAHIM AZIZ UMRANI

    2010-10-01

    Full Text Available This paper presents the Monte-Carlo simulation of Optical CDMA (Code Division Multiple Access systems, and analyse its performance in terms of the BER (Bit Error Rate. The spreading sequence chosen for CDMA is Perfect Difference Codes. Furthermore, this paper derives the expressions of noise variances from first principles to calibrate the noise for both bipolar (electrical domain and unipolar (optical domain signalling required for Monte-Carlo simulation. The simulated results conform to the theory and show that the receiver gain mismatch and splitter loss at the transceiver degrades the system performance.

  2. A new Monte Carlo-based treatment plan optimization approach for intensity modulated radiation therapy.

    Science.gov (United States)

    Li, Yongbao; Tian, Zhen; Shi, Feng; Song, Ting; Wu, Zhaoxia; Liu, Yaqiang; Jiang, Steve; Jia, Xun

    2015-04-07

    Intensity-modulated radiation treatment (IMRT) plan optimization needs beamlet dose distributions. Pencil-beam or superposition/convolution type algorithms are typically used because of their high computational speed. However, inaccurate beamlet dose distributions may mislead the optimization process and hinder the resulting plan quality. To solve this problem, the Monte Carlo (MC) simulation method has been used to compute all beamlet doses prior to the optimization step. The conventional approach samples the same number of particles from each beamlet. Yet this is not the optimal use of MC in this problem. In fact, there are beamlets that have very small intensities after solving the plan optimization problem. For those beamlets, it may be possible to use fewer particles in dose calculations to increase efficiency. Based on this idea, we have developed a new MC-based IMRT plan optimization framework that iteratively performs MC dose calculation and plan optimization. At each dose calculation step, the particle numbers for beamlets were adjusted based on the beamlet intensities obtained through solving the plan optimization problem in the last iteration step. We modified a GPU-based MC dose engine to allow simultaneous computations of a large number of beamlet doses. To test the accuracy of our modified dose engine, we compared the dose from a broad beam and the summed beamlet doses in this beam in an inhomogeneous phantom. Agreement within 1% for the maximum difference and 0.55% for the average difference was observed. We then validated the proposed MC-based optimization schemes in one lung IMRT case. It was found that the conventional scheme required 10(6) particles from each beamlet to achieve an optimization result that was 3% difference in fluence map and 1% difference in dose from the ground truth. In contrast, the proposed scheme achieved the same level of accuracy with on average 1.2 × 10(5) particles per beamlet. Correspondingly, the computation

  3. Optimization of FIBMOS Through 2D Silvaco ATLAS and 2D Monte Carlo Particle-based Device Simulations

    OpenAIRE

    Kang, J.; He, X.; Vasileska, D.; Schroder, D. K.

    2001-01-01

    Focused Ion Beam MOSFETs (FIBMOS) demonstrate large enhancements in core device performance areas such as output resistance, hot electron reliability and voltage stability upon channel length or drain voltage variation. In this work, we describe an optimization technique for FIBMOS threshold voltage characterization using the 2D Silvaco ATLAS simulator. Both ATLAS and 2D Monte Carlo particle-based simulations were used to show that FIBMOS devices exhibit enhanced current drive ...

  4. Test Population Selection from Weibull-Based, Monte Carlo Simulations of Fatigue Life

    Science.gov (United States)

    Vlcek, Brian L.; Zaretsky, Erwin V.; Hendricks, Robert C.

    2012-01-01

    Fatigue life is probabilistic and not deterministic. Experimentally establishing the fatigue life of materials, components, and systems is both time consuming and costly. As a result, conclusions regarding fatigue life are often inferred from a statistically insufficient number of physical tests. A proposed methodology for comparing life results as a function of variability due to Weibull parameters, variability between successive trials, and variability due to size of the experimental population is presented. Using Monte Carlo simulation of randomly selected lives from a large Weibull distribution, the variation in the L10 fatigue life of aluminum alloy AL6061 rotating rod fatigue tests was determined as a function of population size. These results were compared to the L10 fatigue lives of small (10 each) populations from AL2024, AL7075 and AL6061. For aluminum alloy AL6061, a simple algebraic relationship was established for the upper and lower L10 fatigue life limits as a function of the number of specimens failed. For most engineering applications where less than 30 percent variability can be tolerated in the maximum and minimum values, at least 30 to 35 test samples are necessary. The variability of test results based on small sample sizes can be greater than actual differences, if any, that exists between materials and can result in erroneous conclusions. The fatigue life of AL2024 is statistically longer than AL6061 and AL7075. However, there is no statistical difference between the fatigue lives of AL6061 and AL7075 even though AL7075 had a fatigue life 30 percent greater than AL6061.

  5. The use of tetrahedral mesh geometries in Monte Carlo simulation of applicator based brachytherapy dose distributions

    International Nuclear Information System (INIS)

    Fonseca, Gabriel Paiva; Yoriyaz, Hélio; Landry, Guillaume; White, Shane; Reniers, Brigitte; Verhaegen, Frank; D’Amours, Michel; Beaulieu, Luc

    2014-01-01

    Accounting for brachytherapy applicator attenuation is part of the recommendations from the recent report of AAPM Task Group 186. To do so, model based dose calculation algorithms require accurate modelling of the applicator geometry. This can be non-trivial in the case of irregularly shaped applicators such as the Fletcher Williamson gynaecological applicator or balloon applicators with possibly irregular shapes employed in accelerated partial breast irradiation (APBI) performed using electronic brachytherapy sources (EBS). While many of these applicators can be modelled using constructive solid geometry (CSG), the latter may be difficult and time-consuming. Alternatively, these complex geometries can be modelled using tessellated geometries such as tetrahedral meshes (mesh geometries (MG)). Recent versions of Monte Carlo (MC) codes Geant4 and MCNP6 allow for the use of MG. The goal of this work was to model a series of applicators relevant to brachytherapy using MG. Applicators designed for 192 Ir sources and 50 kV EBS were studied; a shielded vaginal applicator, a shielded Fletcher Williamson applicator and an APBI balloon applicator. All applicators were modelled in Geant4 and MCNP6 using MG and CSG for dose calculations. CSG derived dose distributions were considered as reference and used to validate MG models by comparing dose distribution ratios. In general agreement within 1% for the dose calculations was observed for all applicators between MG and CSG and between codes when considering volumes inside the 25% isodose surface. When compared to CSG, MG required longer computation times by a factor of at least 2 for MC simulations using the same code. MCNP6 calculation times were more than ten times shorter than Geant4 in some cases. In conclusion we presented methods allowing for high fidelity modelling with results equivalent to CSG. To the best of our knowledge MG offers the most accurate representation of an irregular APBI balloon applicator. (paper)

  6. Monte Carlo based protocol for cell survival and tumour control probability in BNCT.

    Science.gov (United States)

    Ye, S J

    1999-02-01

    A mathematical model to calculate the theoretical cell survival probability (nominally, the cell survival fraction) is developed to evaluate preclinical treatment conditions for boron neutron capture therapy (BNCT). A treatment condition is characterized by the neutron beam spectra, single or bilateral exposure, and the choice of boron carrier drug (boronophenylalanine (BPA) or boron sulfhydryl hydride (BSH)). The cell survival probability defined from Poisson statistics is expressed with the cell-killing yield, the 10B(n,alpha)7Li reaction density, and the tolerable neutron fluence. The radiation transport calculation from the neutron source to tumours is carried out using Monte Carlo methods: (i) reactor-based BNCT facility modelling to yield the neutron beam library at an irradiation port; (ii) dosimetry to limit the neutron fluence below a tolerance dose (10.5 Gy-Eq); (iii) calculation of the 10B(n,alpha)7Li reaction density in tumours. A shallow surface tumour could be effectively treated by single exposure producing an average cell survival probability of 10(-3)-10(-5) for probable ranges of the cell-killing yield for the two drugs, while a deep tumour will require bilateral exposure to achieve comparable cell kills at depth. With very pure epithermal beams eliminating thermal, low epithermal and fast neutrons, the cell survival can be decreased by factors of 2-10 compared with the unmodified neutron spectrum. A dominant effect of cell-killing yield on tumour cell survival demonstrates the importance of choice of boron carrier drug. However, these calculations do not indicate an unambiguous preference for one drug, due to the large overlap of tumour cell survival in the probable ranges of the cell-killing yield for the two drugs. The cell survival value averaged over a bulky tumour volume is used to predict the overall BNCT therapeutic efficacy, using a simple model of tumour control probability (TCP).

  7. Monte Carlo N-particle simulation of neutron-based sterilisation of anthrax contamination.

    Science.gov (United States)

    Liu, B; Xu, J; Liu, T; Ouyang, X

    2012-10-01

    To simulate the neutron-based sterilisation of anthrax contamination by Monte Carlo N-particle (MCNP) 4C code. Neutrons are elementary particles that have no charge. They are 20 times more effective than electrons or γ-rays in killing anthrax spores on surfaces and inside closed containers. Neutrons emitted from a (252)Cf neutron source are in the 100 keV to 2 MeV energy range. A 2.5 MeV D-D neutron generator can create neutrons at up to 10(13) n s(-1) with current technology. All these enable an effective and low-cost method of killing anthrax spores. There is no effect on neutron energy deposition on the anthrax sample when using a reflector that is thicker than its saturation thickness. Among all three reflecting materials tested in the MCNP simulation, paraffin is the best because it has the thinnest saturation thickness and is easy to machine. The MCNP radiation dose and fluence simulation calculation also showed that the MCNP-simulated neutron fluence that is needed to kill the anthrax spores agrees with previous analytical estimations very well. The MCNP simulation indicates that a 10 min neutron irradiation from a 0.5 g (252)Cf neutron source or a 1 min neutron irradiation from a 2.5 MeV D-D neutron generator may kill all anthrax spores in a sample. This is a promising result because a 2.5 MeV D-D neutron generator output >10(13) n s(-1) should be attainable in the near future. This indicates that we could use a D-D neutron generator to sterilise anthrax contamination within several seconds.

  8. Photoresponse of the protonated Schiff-base retinal chromophore in the gas phase

    DEFF Research Database (Denmark)

    Toker, Jonathan; Rahbek, Dennis Bo; Kiefer, H V

    2013-01-01

    The fragmentation, initiated by photoexcitation as well as collisionally-induced excitation, of several retinal chromophores was studied in the gas phase. The chromophore in the protonated Schiff-base form (RPSB), essential for mammalian vision, shows a remarkably selective photoresponse. The sel......The fragmentation, initiated by photoexcitation as well as collisionally-induced excitation, of several retinal chromophores was studied in the gas phase. The chromophore in the protonated Schiff-base form (RPSB), essential for mammalian vision, shows a remarkably selective photoresponse...... modifications of the chromophore. We propose that isomerizations play an important role in the photoresponse of gas-phase retinal chromophores and guide internal conversion through conical intersections. The role of protein interactions is then to control the specificity of the photoisomerization in the primary...

  9. Recent Development of Pd-Based Electrocatalysts for Proton Exchange Membrane Fuel Cells

    Directory of Open Access Journals (Sweden)

    Hui Meng

    2015-07-01

    Full Text Available This review selectively summarizes the latest developments in the Pd-based cataysts for low temperature proton exchange membrane fuel cells, especially in the application of formic acid oxidation, alcohol oxidation and oxygen reduction reaction. The advantages and shortcomings of the Pd-based catalysts for electrocatalysis are analyzed. The influence of the structure and morphology of the Pd materials on the performance of the Pd-based catalysts were described. Finally, the perspectives of future trends on Pd-based catalysts for different applications were considered.

  10. Development of three-dimensional program based on Monte Carlo and discrete ordinates bidirectional coupling method

    International Nuclear Information System (INIS)

    Han Jingru; Chen Yixue; Yuan Longjun

    2013-01-01

    The Monte Carlo (MC) and discrete ordinates (SN) are the commonly used methods in the design of radiation shielding. Monte Carlo method is able to treat the geometry exactly, but time-consuming in dealing with the deep penetration problem. The discrete ordinate method has great computational efficiency, but it is quite costly in computer memory and it suffers from ray effect. Single discrete ordinates method or single Monte Carlo method has limitation in shielding calculation for large complex nuclear facilities. In order to solve the problem, the Monte Carlo and discrete ordinates bidirectional coupling method is developed. The bidirectional coupling method is implemented in the interface program to transfer the particle probability distribution of MC and angular flux of discrete ordinates. The coupling method combines the advantages of MC and SN. The test problems of cartesian and cylindrical coordinate have been calculated by the coupling methods. The calculation results are performed with comparison to MCNP and TORT and satisfactory agreements are obtained. The correctness of the program is proved. (authors)

  11. SU-E-T-656: Quantitative Analysis of Proton Boron Fusion Therapy (PBFT) in Various Conditions

    Energy Technology Data Exchange (ETDEWEB)

    Yoon, D; Jung, J; Shin, H; Kim, M; Suh, T [The Catholic University, Seoul (Korea, Republic of); Jang, H [Seoul St. Mary Hospital, Seoul (Korea, Republic of)

    2015-06-15

    Purpose: Three alpha particles are concomitant of proton boron interaction, which can be used in radiotherapy applications. We performed simulation studies to determine the effectiveness of proton boron fusion therapy (PBFT) under various conditions. Methods: Boron uptake regions (BURs) of various widths and densities were implemented in Monte Carlo n-particle extended (MCNPX) simulation code. The effect of proton beam energy was considered for different BURs. Four simulation scenarios were designed to verify the effectiveness of integrated boost that was observed in the proton boron reaction. In these simulations, the effect of proton beam energy was determined for different physical conditions, such as size, location, and boron concentration. Results: Proton dose amplification was confirmed for all proton beam energies considered (< 96.62%). Based on the simulation results for different physical conditions, the threshold for the range in which proton dose amplification occurred was estimated as 0.3 cm. Effective proton boron reaction requires the boron concentration to be equal to or greater than 14.4 mg/g. Conclusion: We established the effects of the PBFT with various conditions by using Monte Carlo simulation. The results of our research can be used for providing a PBFT dose database.

  12. Development of a consistent Monte Carlo-deterministic transport methodology based on the method of characteristics and MCNP5

    International Nuclear Information System (INIS)

    Karriem, Z.; Ivanov, K.; Zamonsky, O.

    2011-01-01

    This paper presents work that has been performed to develop an integrated Monte Carlo- Deterministic transport methodology in which the two methods make use of exactly the same general geometry and multigroup nuclear data. The envisioned application of this methodology is in reactor lattice physics methods development and shielding calculations. The methodology will be based on the Method of Long Characteristics (MOC) and the Monte Carlo N-Particle Transport code MCNP5. Important initial developments pertaining to ray tracing and the development of an MOC flux solver for the proposed methodology are described. Results showing the viability of the methodology are presented for two 2-D general geometry transport problems. The essential developments presented is the use of MCNP as geometry construction and ray tracing tool for the MOC, verification of the ray tracing indexing scheme that was developed to represent the MCNP geometry in the MOC and the verification of the prototype 2-D MOC flux solver. (author)

  13. Three-Dimensional Simulation of DRIE Process Based on the Narrow Band Level Set and Monte Carlo Method

    Directory of Open Access Journals (Sweden)

    Jia-Cheng Yu

    2018-02-01

    Full Text Available A three-dimensional topography simulation of deep reactive ion etching (DRIE is developed based on the narrow band level set method for surface evolution and Monte Carlo method for flux distribution. The advanced level set method is implemented to simulate the time-related movements of etched surface. In the meanwhile, accelerated by ray tracing algorithm, the Monte Carlo method incorporates all dominant physical and chemical mechanisms such as ion-enhanced etching, ballistic transport, ion scattering, and sidewall passivation. The modified models of charged particles and neutral particles are epitomized to determine the contributions of etching rate. The effects such as scalloping effect and lag effect are investigated in simulations and experiments. Besides, the quantitative analyses are conducted to measure the simulation error. Finally, this simulator will be served as an accurate prediction tool for some MEMS fabrications.

  14. Atlas-based estimation of lung and lobar anatomy in proton MRI.

    Science.gov (United States)

    Tustison, Nicholas J; Qing, Kun; Wang, Chengbo; Altes, Talissa A; Mugler, John P

    2016-07-01

    To propose an accurate methodological framework for automatically segmenting pulmonary proton MRI based on an optimal consensus of a spatially normalized library of annotated lung atlases. A library of 62 manually annotated lung atlases comprising 48 mixed healthy, chronic obstructive pulmonary disease, and asthmatic subjects of a large age range with multiple ventilation levels is used to produce an optimal segmentation in proton MRI, based on a consensus of the spatially normalized library. An extension of this methodology is used to provide best-guess estimates of lobar subdivisions in proton MRI from annotated computed tomography data. A leave-one-out evaluation strategy was used for evaluation. Jaccard overlap measures for the left and right lungs were used for performance comparisons relative to the current state-of-the-art (0.966 ± 0.018 and 0.970 ± 0.016, respectively). Best-guess estimates for the lobes exhibited comparable performance levels (left upper: 0.882 ± 0.059, left lower: 0.868 ± 0.06, right upper: 0.852 ± 0.067, right middle: 0.657 ± 0.130, right lower: 0.873 ± 0.063). An annotated atlas library approach can be used to provide good lung and lobe estimation in proton MRI. The proposed framework is useful for subsequent anatomically based analysis of structural and/or functional pulmonary image data. Magn Reson Med 76:315-320, 2016. © 2015 Wiley Periodicals, Inc. © 2015 Wiley Periodicals, Inc.

  15. Pseudo-diode based on protonic/electronic hybrid oxide transistor

    Science.gov (United States)

    Fu, Yang Ming; Liu, Yang Hui; Zhu, Li Qiang; Xiao, Hui; Song, An Ran

    2018-01-01

    Current rectification behavior has been proved to be essential in modern electronics. Here, a pseudo-diode is proposed based on protonic/electronic hybrid indium-gallium-zinc oxide electric-double-layer (EDL) transistor. The oxide EDL transistors are fabricated by using phosphorous silicate glass (PSG) based proton conducting electrolyte as gate dielectric. A diode operation mode is established on the transistor, originating from field configurable proton fluxes within the PSG electrolyte. Current rectification ratios have been modulated to values ranged between ˜4 and ˜50 000 with gate electrode biased at voltages ranged between -0.7 V and 0.1 V. Interestingly, the proposed pseudo-diode also exhibits field reconfigurable threshold voltages. When the gate is biased at -0.5 V and 0.3 V, threshold voltages are set to ˜-1.3 V and -0.55 V, respectively. The proposed pseudo-diode may find potential applications in brain-inspired platforms and low-power portable systems.

  16. A Nonvolatile MOSFET Memory Device Based on Mobile Protons in SiO(2) Thin Films

    Energy Technology Data Exchange (ETDEWEB)

    Vanheusden, K.; Warren, W.L.; Devine, R.A.B.; Fleetwood, D.M.; Draper, B.L.; Schwank, J.R.

    1999-03-02

    It is shown how mobile H{sup +} ions can be generated thermally inside the oxide layer of Si/SiO{sub 2}/Si structures. The technique involves only standard silicon processing steps: the nonvolatile field effect transistor (NVFET) is based on a standard MOSFET with thermally grown SiO{sub 2} capped with a poly-silicon layer. The capped thermal oxide receives an anneal at {approximately}1100 C that enables the incorporation of the mobile protons into the gate oxide. The introduction of the protons is achieved by a subsequent 500-800 C anneal in a hydrogen-containing ambient, such as forming gas (N{sub 2}:H{sub 2} 95:5). The mobile protons are stable and entrapped inside the oxide layer, and unlike alkali ions, their space-charge distribution can be controlled and rapidly rearranged at room temperature by an applied electric field. Using this principle, a standard MOS transistor can be converted into a nonvolatile memory transistor that can be switched between normally on and normally off. Switching speed, retention, endurance, and radiation tolerance data are presented showing that this non-volatile memory technology can be competitive with existing Si-based non-volatile memory technologies such as the floating gate technologies (e.g. Flash memory).

  17. An easily sintered, chemically stable, barium zirconate-based proton conductor for high-performance proton-conducting solid oxide fuel cells

    KAUST Repository

    Sun, Wenping

    2014-07-25

    Yttrium and indium co-doped barium zirconate is investigated to develop a chemically stable and sintering active proton conductor for solid oxide fuel cells (SOFCs). BaZr0.8Y0.2-xInxO3- δ possesses a pure cubic perovskite structure. The sintering activity of BaZr0.8Y0.2-xInxO3- δ increases significantly with In concentration. BaZr0.8Y0.15In0.05O3- δ (BZYI5) exhibits the highest total electrical conductivity among the sintered oxides. BZYI5 also retains high chemical stability against CO2, vapor, and reduction of H2. The good sintering activity, high conductivity, and chemical stability of BZYI5 facilitate the fabrication of durable SOFCs based on a highly conductive BZYI5 electrolyte film by cost-effective ceramic processes. Fully dense BZYI5 electrolyte film is successfully prepared on the anode substrate by a facile drop-coating technique followed by co-firing at 1400 °C for 5 h in air. The BZYI5 film exhibits one of the highest conductivity among the BaZrO3-based electrolyte films with various sintering aids. BZYI5-based single cells output very encouraging and by far the highest peak power density for BaZrO3-based proton-conducting SOFCs, reaching as high as 379 mW cm-2 at 700 °C. The results demonstrate that Y and In co-doping is an effective strategy for exploring sintering active and chemically stable BaZrO3-based proton conductors for high performance proton-conducting SOFCs. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  18. Development of a hybrid multi-scale phantom for Monte-Carlo based internal dosimetry

    International Nuclear Information System (INIS)

    Marcatili, S.; Villoing, D.; Bardies, M.

    2015-01-01

    Full text of publication follows. Aim: in recent years several phantoms were developed for radiopharmaceutical dosimetry in clinical and preclinical settings. Voxel-based models (Zubal, Max/Fax, ICRP110) were developed to reach a level of realism that could not be achieved by mathematical models. In turn, 'hybrid' models (XCAT, MOBY/ROBY, Mash/Fash) allow a further degree of versatility by offering the possibility to finely tune each model according to various parameters. However, even 'hybrid' models require the generation of a voxel version for Monte-Carlo modeling of radiation transport. Since absorbed dose simulation time is strictly related to geometry spatial sampling, a compromise should be made between phantom realism and simulation speed. This trade-off leads on one side in an overestimation of the size of small radiosensitive structures such as the skin or hollow organs' walls, and on the other hand to unnecessarily detailed voxellization of large, homogeneous structures. The Aim of this work is to develop a hybrid multi-resolution phantom model for Geant4 and Gate, to better characterize energy deposition in small structures while preserving reasonable computation times. Materials and Methods: we have developed a pipeline for the conversion of preexisting phantoms into a multi-scale Geant4 model. Meshes of each organ are created from raw binary images of a phantom and then voxellized to the smallest spatial sampling required by the user. The user can then decide to re-sample the internal part of each organ, while leaving a layer of smallest voxels at the edge of the organ. In this way, the realistic shape of the organ is maintained while reducing the voxel number in the inner part. For hollow organs, the wall is always modeled using the smallest voxel sampling. This approach allows choosing different voxel resolutions for each organ according to a specific application. Results: preliminary results show that it is possible to

  19. Monte Carol-Based Dosimetry of Beta-Emitters for Intravascular Brachytherapy

    Energy Technology Data Exchange (ETDEWEB)

    Choi, C.K.

    2002-06-25

    Monte Carlo simulations for radiation dosimetry and the experimental verifications of the simulations have been developed for the treatment geometry of intravascular brachytherapy, a form of radionuclide therapy for occluded coronary disease (restenosis). Monte Carlo code, MCNP4C, has been used to calculate the radiation dose from the encapsulated array of B-emitting seeds (Sr/Y-source train). Solid water phantoms have been fabricated to measure the dose on the radiochromic films that were exposed to the beta source train for both linear and curved coronary vessel geometries. While the dose difference for the 5-degree curved vessel at the prescription point of f+2.0 mm is within the 10% guideline set by the AAPM, however, the difference increased dramatically to 16.85% for the 10-degree case which requires additional adjustment for the acceptable dosimetry planning. The experimental dose measurements agree well with the simulation results

  20. Development of a Monte-Carlo based method for calculating the effect of stationary fluctuations

    DEFF Research Database (Denmark)

    Pettersen, E. E.; Demazire, C.; Jareteg, K.

    2015-01-01

    equivalent problems nevertheless requires the possibility to modify the macroscopic cross-sections, and we use the work of Kuijper, van der Marck and Hogenbirk to define group-wise macroscopic cross-sections in MCNP [1]. The method is illustrated in this paper at a frequency of 1 Hz, for which only the real......This paper deals with the development of a novel method for performing Monte Carlo calculations of the effect, on the neutron flux, of stationary fluctuations in macroscopic cross-sections. The basic principle relies on the formulation of two equivalent problems in the frequency domain: one...... stationary dynamic calculations, the presented method does not require any modification of the Monte Carlo code....

  1. Monte Carlo based simulation of LIAC intraoperative radiotherapy accelerator along with beam shaper applicator

    Directory of Open Access Journals (Sweden)

    N Heidarloo

    2017-08-01

    Full Text Available Intraoperative electron radiotherapy is one of the radiotherapy methods that delivers a high single fraction of radiation dose to the patient in one session during the surgery. Beam shaper applicator is one of the applicators that is recently employed with this radiotherapy method. This applicator has a considerable application in treatment of large tumors. In this study, the dosimetric characteristics of the electron beam produced by LIAC intraoperative radiotherapy accelerator in conjunction with this applicator have been evaluated through Monte Carlo simulation by MCNP code. The results showed that the electron beam produced by the beam shaper applicator would have the desirable dosimetric characteristics, so that the mentioned applicator can be considered for clinical purposes. Furthermore, the good agreement between the results of simulation and practical dosimetry, confirms the applicability of Monte Carlo method in determining the dosimetric parameters of electron beam  intraoperative radiotherapy

  2. Simulation based sequential Monte Carlo methods for discretely observed Markov processes

    OpenAIRE

    Neal, Peter

    2014-01-01

    Parameter estimation for discretely observed Markov processes is a challenging problem. However, simulation of Markov processes is straightforward using the Gillespie algorithm. We exploit this ease of simulation to develop an effective sequential Monte Carlo (SMC) algorithm for obtaining samples from the posterior distribution of the parameters. In particular, we introduce two key innovations, coupled simulations, which allow us to study multiple parameter values on the basis of a single sim...

  3. Phase constitution of Ni-based quaternary alloys studied by Monte Carlo simulation

    Czech Academy of Sciences Publication Activity Database

    Buršík, Jiří

    2002-01-01

    Roč. 147, 1-2 (2002), s. 162-165 ISSN 0010-4655. [Computational Modeling and Simulation of Complex Systems, Europhysics Conference on Computational Physics (CCP 2001). Aachen, 05.09.2001-08.09.2001] R&D Projects: GA ČR GA202/01/0383 Institutional research plan: CEZ:AV0Z2041904 Keywords : Monte Carlo simulation * ordering * pairwise interaction Subject RIV: BM - Solid Matter Physics ; Magnetism Impact factor: 1.204, year: 2002

  4. Development of a practical fuel management system for PSBR based on advanced three-dimensional Monte Carlo coupled depletion methodology

    Science.gov (United States)

    Tippayakul, Chanatip

    The main objective of this research is to develop a practical fuel management system for the Pennsylvania State University Breazeale research reactor (PSBR) based on several advanced Monte Carlo coupled depletion methodologies. Primarily, this research involved two major activities: model and method developments and analyses and validations of the developed models and methods. The starting point of this research was the utilization of the earlier developed fuel management tool, TRIGSIM, to create the Monte Carlo model of core loading 51 (end of the core loading). It was found when comparing the normalized power results of the Monte Carlo model to those of the current fuel management system (using HELIOS/ADMARC-H) that they agreed reasonably well (within 2%--3% differences on average). Moreover, the reactivity of some fuel elements was calculated by the Monte Carlo model and it was compared with measured data. It was also found that the fuel element reactivity results of the Monte Carlo model were in good agreement with the measured data. However, the subsequent task of analyzing the conversion from the core loading 51 to the core loading 52 using TRIGSIM showed quite significant difference of each control rod worth between the Monte Carlo model and the current methodology model. The differences were mainly caused by inconsistent absorber atomic number densities between the two models. Hence, the model of the first operating core (core loading 2) was revised in light of new information about the absorber atomic densities to validate the Monte Carlo model with the measured data. With the revised Monte Carlo model, the results agreed better to the measured data. Although TRIGSIM showed good modeling and capabilities, the accuracy of TRIGSIM could be further improved by adopting more advanced algorithms. Therefore, TRIGSIM was planned to be upgraded. The first task of upgrading TRIGSIM involved the improvement of the temperature modeling capability. The new TRIGSIM was

  5. Design and evaluation of a Monte Carlo based model of an orthovoltage treatment system

    International Nuclear Information System (INIS)

    Penchev, Petar; Maeder, Ulf; Fiebich, Martin; Zink, Klemens; University Hospital Marburg

    2015-01-01

    The aim of this study was to develop a flexible framework of an orthovoltage treatment system capable of calculating and visualizing dose distributions in different phantoms and CT datasets. The framework provides a complete set of various filters, applicators and X-ray energies and therefore can be adapted to varying studies or be used for educational purposes. A dedicated user friendly graphical interface was developed allowing for easy setup of the simulation parameters and visualization of the results. For the Monte Carlo simulations the EGSnrc Monte Carlo code package was used. Building the geometry was accomplished with the help of the EGSnrc C++ class library. The deposited dose was calculated according to the KERMA approximation using the track-length estimator. The validation against measurements showed a good agreement within 4-5% deviation, down to depths of 20% of the depth dose maximum. Furthermore, to show its capabilities, the validated model was used to calculate the dose distribution on two CT datasets. Typical Monte Carlo calculation time for these simulations was about 10 minutes achieving an average statistical uncertainty of 2% on a standard PC. However, this calculation time depends strongly on the used CT dataset, tube potential, filter material/thickness and applicator size.

  6. Cloud-based Monte Carlo modelling of BSSRDF for the rendering of human skin appearance (Conference Presentation)

    Science.gov (United States)

    Doronin, Alexander; Rushmeier, Holly E.; Meglinski, Igor; Bykov, Alexander V.

    2016-03-01

    We present a new Monte Carlo based approach for the modelling of Bidirectional Scattering-Surface Reflectance Distribution Function (BSSRDF) for accurate rendering of human skin appearance. The variations of both skin tissues structure and the major chromophores are taken into account correspondingly to the different ethnic and age groups. The computational solution utilizes HTML5, accelerated by the graphics processing units (GPUs), and therefore is convenient for the practical use at the most of modern computer-based devices and operating systems. The results of imitation of human skin reflectance spectra, corresponding skin colours and examples of 3D faces rendering are presented and compared with the results of phantom studies.

  7. Sampling-based nuclear data uncertainty quantification for continuous energy Monte-Carlo codes

    International Nuclear Information System (INIS)

    Zhu, T.

    2015-01-01

    Research on the uncertainty of nuclear data is motivated by practical necessity. Nuclear data uncertainties can propagate through nuclear system simulations into operation and safety related parameters. The tolerance for uncertainties in nuclear reactor design and operation can affect the economic efficiency of nuclear power, and essentially its sustainability. The goal of the present PhD research is to establish a methodology of nuclear data uncertainty quantification (NDUQ) for MCNPX, the continuous-energy Monte-Carlo (M-C) code. The high fidelity (continuous-energy treatment and flexible geometry modelling) of MCNPX makes it the choice of routine criticality safety calculations at PSI/LRS, but also raises challenges for NDUQ by conventional sensitivity/uncertainty (S/U) methods. For example, only recently in 2011, the capability of calculating continuous energy κ eff sensitivity to nuclear data was demonstrated in certain M-C codes by using the method of iterated fission probability. The methodology developed during this PhD research is fundamentally different from the conventional S/U approach: nuclear data are treated as random variables and sampled in accordance to presumed probability distributions. When sampled nuclear data are used in repeated model calculations, the output variance is attributed to the collective uncertainties of nuclear data. The NUSS (Nuclear data Uncertainty Stochastic Sampling) tool is based on this sampling approach and implemented to work with MCNPX’s ACE format of nuclear data, which also gives NUSS compatibility with MCNP and SERPENT M-C codes. In contrast, multigroup uncertainties are used for the sampling of ACE-formatted pointwise-energy nuclear data in a groupwise manner due to the more limited quantity and quality of nuclear data uncertainties. Conveniently, the usage of multigroup nuclear data uncertainties allows consistent comparison between NUSS and other methods (both S/U and sampling-based) that employ the same

  8. Novel proton exchange membrane based on crosslinked poly(vinyl alcohol) for direct methanol fuel cells

    Science.gov (United States)

    Liu, Chien-Pan; Dai, Chi-An; Chao, Chi-Yang; Chang, Shoou-Jinn

    2014-03-01

    In this study, we report the synthesis and the characterization of poly (vinyl alcohol) based proton conducting membranes. In particular, we describe a novel physically and chemically PVA/HFA (poly (vinyl alcohol)/hexafluoroglutaric acid) blending membranes with BASANa (Benzenesulfonic acid sodium salt) and GA (Glutaraldehyde) as binary reaction agents. The key PEM parameters such as ion exchange capacity (IEC), water uptake, proton conductivity, and methanol permeability were controlled by adjusting the chemical composition of the membranes. The IEC value of the membrane is found to be an important parameter in affecting water uptake, conductivity as well as the permeability of the resulting membrane. Plots of the water uptake, conductivity, and methanol permeability vs. IEC of the membranes show a distinct change in the slope of their curves at roughly the same IEC value which suggests a transition of structural changes in the network. The proton conductivities and the methanol permeability of all the membranes are in the range of 10-3-10-2 S cm-1 and 10-8-10-7 cm2 s-1, respectively, depending on its binary crosslinking density, and it shows great selectivity compared with those of Nafion®-117. The membranes display good mechanical properties which suggest a good lifetime usage of the membranes applied in DMFCs.

  9. SPL-based Proton Driver for a nu-Factory at CERN

    CERN Document Server

    Benedetto, E; Garoby, R; Meddahi, M

    2010-01-01

    The conceptual design and feasibility studies for a nu-Factory Proton Driver based on the CERN Superconducting Proton Linac (SPL) have been com- pleted. In the proposed scenario, the 4 MW proton beam (H- beam) is acceler- ated with the upgraded High Power (HP)-SPL to 5 GeV, stored in an accumu- lator ring and Þnally transported to a compressor ring, where bunch rotation takes place, in order to achieve the speciÞc time structure. We here summa- rize the choices in terms of lattice, magnet technology and RF manipulations in the two rings. The possible critical issues, such as heating of the foil for the charge-exchange injection, space-charge problems in the compressor and beam stability in the accumulator ring, have been addressed and are shown not to be show-stoppers. The analysis focuses on the baseline scenario, consider- ing 6 bunches in the accumulator, and preliminary studies are discussed for the option of 3 or a single bunch per burst.

  10. Two solar proton fluence models based on ground level enhancement observations

    Science.gov (United States)

    Raukunen, Osku; Vainio, Rami; Tylka, Allan J.; Dietrich, William F.; Jiggens, Piers; Heynderickx, Daniel; Dierckxsens, Mark; Crosby, Norma; Ganse, Urs; Siipola, Robert

    2018-01-01

    Solar energetic particles (SEPs) constitute an important component of the radiation environment in interplanetary space. Accurate modeling of SEP events is crucial for the mitigation of radiation hazards in spacecraft design. In this study we present two new statistical models of high energy solar proton fluences based on ground level enhancement (GLE) observations during solar cycles 19-24. As the basis of our modeling, we utilize a four parameter double power law function (known as the Band function) fits to integral GLE fluence spectra in rigidity. In the first model, the integral and differential fluences for protons with energies between 10 MeV and 1 GeV are calculated using the fits, and the distributions of the fluences at certain energies are modeled with an exponentially cut-off power law function. In the second model, we use a more advanced methodology: by investigating the distributions and relationships of the spectral fit parameters we find that they can be modeled as two independent and two dependent variables. Therefore, instead of modeling the fluences separately at different energies, we can model the shape of the fluence spectrum. We present examples of modeling results and show that the two methodologies agree well except for a short mission duration (1 year) at low confidence level. We also show that there is a reasonable agreement between our models and three well-known solar proton models (JPL, ESP and SEPEM), despite the differences in both the modeling methodologies and the data used to construct the models.

  11. An analytic linear accelerator source model for GPU-based Monte Carlo dose calculations

    Science.gov (United States)

    Tian, Zhen; Li, Yongbao; Folkerts, Michael; Shi, Feng; Jiang, Steve B.; Jia, Xun

    2015-10-01

    Recently, there has been a lot of research interest in developing fast Monte Carlo (MC) dose calculation methods on graphics processing unit (GPU) platforms. A good linear accelerator (linac) source model is critical for both accuracy and efficiency considerations. In principle, an analytical source model should be more preferred for GPU-based MC dose engines than a phase-space file-based model, in that data loading and CPU-GPU data transfer can be avoided. In this paper, we presented an analytical field-independent source model specifically developed for GPU-based MC dose calculations, associated with a GPU-friendly sampling scheme. A key concept called phase-space-ring (PSR) was proposed. Each PSR contained a group of particles that were of the same type, close in energy and reside in a narrow ring on the phase-space plane located just above the upper jaws. The model parameterized the probability densities of particle location, direction and energy for each primary photon PSR, scattered photon PSR and electron PSR. Models of one 2D Gaussian distribution or multiple Gaussian components were employed to represent the particle direction distributions of these PSRs. A method was developed to analyze a reference phase-space file and derive corresponding model parameters. To efficiently use our model in MC dose calculations on GPU, we proposed a GPU-friendly sampling strategy, which ensured that the particles sampled and transported simultaneously are of the same type and close in energy to alleviate GPU thread divergences. To test the accuracy of our model, dose distributions of a set of open fields in a water phantom were calculated using our source model and compared to those calculated using the reference phase-space files. For the high dose gradient regions, the average distance-to-agreement (DTA) was within 1 mm and the maximum DTA within 2 mm. For relatively low dose gradient regions, the root-mean-square (RMS) dose difference was within 1.1% and the maximum

  12. An analytic linear accelerator source model for GPU-based Monte Carlo dose calculations.

    Science.gov (United States)

    Tian, Zhen; Li, Yongbao; Folkerts, Michael; Shi, Feng; Jiang, Steve B; Jia, Xun

    2015-10-21

    Recently, there has been a lot of research interest in developing fast Monte Carlo (MC) dose calculation methods on graphics processing unit (GPU) platforms. A good linear accelerator (linac) source model is critical for both accuracy and efficiency considerations. In principle, an analytical source model should be more preferred for GPU-based MC dose engines than a phase-space file-based model, in that data loading and CPU-GPU data transfer can be avoided. In this paper, we presented an analytical field-independent source model specifically developed for GPU-based MC dose calculations, associated with a GPU-friendly sampling scheme. A key concept called phase-space-ring (PSR) was proposed. Each PSR contained a group of particles that were of the same type, close in energy and reside in a narrow ring on the phase-space plane located just above the upper jaws. The model parameterized the probability densities of particle location, direction and energy for each primary photon PSR, scattered photon PSR and electron PSR. Models of one 2D Gaussian distribution or multiple Gaussian components were employed to represent the particle direction distributions of these PSRs. A method was developed to analyze a reference phase-space file and derive corresponding model parameters. To efficiently use our model in MC dose calculations on GPU, we proposed a GPU-friendly sampling strategy, which ensured that the particles sampled and transported simultaneously are of the same type and close in energy to alleviate GPU thread divergences. To test the accuracy of our model, dose distributions of a set of open fields in a water phantom were calculated using our source model and compared to those calculated using the reference phase-space files. For the high dose gradient regions, the average distance-to-agreement (DTA) was within 1 mm and the maximum DTA within 2 mm. For relatively low dose gradient regions, the root-mean-square (RMS) dose difference was within 1.1% and the maximum

  13. Monte Carlo Based Calibration and Uncertainty Analysis of a Coupled Plant Growth and Hydrological Model

    Science.gov (United States)

    Houska, Tobias; Multsch, Sebastian; Kraft, Philipp; Frede, Hans-Georg; Breuer, Lutz

    2014-05-01

    Computer simulations are widely used to support decision making and planning in the agriculture sector. On the one hand, many plant growth models use simplified hydrological processes and structures, e.g. by the use of a small number of soil layers or by the application of simple water flow approaches. On the other hand, in many hydrological models plant growth processes are poorly represented. Hence, fully coupled models with a high degree of process representation would allow a more detailed analysis of the dynamic behaviour of the soil-plant interface. We used the Python programming language to couple two of such high process oriented independent models and to calibrate both models simultaneously. The Catchment Modelling Framework (CMF) simulated soil hydrology based on the Richards equation and the Van-Genuchten-Mualem retention curve. CMF was coupled with the Plant growth Modelling Framework (PMF), which predicts plant growth on the basis of radiation use efficiency, degree days, water shortage and dynamic root biomass allocation. The Monte Carlo based Generalised Likelihood Uncertainty Estimation (GLUE) method was applied to parameterize the coupled model and to investigate the related uncertainty of model predictions to it. Overall, 19 model parameters (4 for CMF and 15 for PMF) were analysed through 2 x 106 model runs randomly drawn from an equally distributed parameter space. Three objective functions were used to evaluate the model performance, i.e. coefficient of determination (R2), bias and model efficiency according to Nash Sutcliffe (NSE). The model was applied to three sites with different management in Muencheberg (Germany) for the simulation of winter wheat (Triticum aestivum L.) in a cross-validation experiment. Field observations for model evaluation included soil water content and the dry matters of roots, storages, stems and leaves. Best parameter sets resulted in NSE of 0.57 for the simulation of soil moisture across all three sites. The shape

  14. Monte Carlo-based treatment planning system calculation engine for microbeam radiation therapy

    Energy Technology Data Exchange (ETDEWEB)

    Martinez-Rovira, I.; Sempau, J.; Prezado, Y. [Institut de Tecniques Energetiques, Universitat Politecnica de Catalunya, Diagonal 647, Barcelona E-08028 (Spain) and ID17 Biomedical Beamline, European Synchrotron Radiation Facility (ESRF), 6 rue Jules Horowitz B.P. 220, F-38043 Grenoble Cedex (France); Institut de Tecniques Energetiques, Universitat Politecnica de Catalunya, Diagonal 647, Barcelona E-08028 (Spain); Laboratoire Imagerie et modelisation en neurobiologie et cancerologie, UMR8165, Centre National de la Recherche Scientifique (CNRS), Universites Paris 7 et Paris 11, Bat 440., 15 rue Georges Clemenceau, F-91406 Orsay Cedex (France)

    2012-05-15

    Purpose: Microbeam radiation therapy (MRT) is a synchrotron radiotherapy technique that explores the limits of the dose-volume effect. Preclinical studies have shown that MRT irradiations (arrays of 25-75-{mu}m-wide microbeams spaced by 200-400 {mu}m) are able to eradicate highly aggressive animal tumor models while healthy tissue is preserved. These promising results have provided the basis for the forthcoming clinical trials at the ID17 Biomedical Beamline of the European Synchrotron Radiation Facility (ESRF). The first step includes irradiation of pets (cats and dogs) as a milestone before treatment of human patients. Within this context, accurate dose calculations are required. The distinct features of both beam generation and irradiation geometry in MRT with respect to conventional techniques require the development of a specific MRT treatment planning system (TPS). In particular, a Monte Carlo (MC)-based calculation engine for the MRT TPS has been developed in this work. Experimental verification in heterogeneous phantoms and optimization of the computation time have also been performed. Methods: The penelope/penEasy MC code was used to compute dose distributions from a realistic beam source model. Experimental verification was carried out by means of radiochromic films placed within heterogeneous slab-phantoms. Once validation was completed, dose computations in a virtual model of a patient, reconstructed from computed tomography (CT) images, were performed. To this end, decoupling of the CT image voxel grid (a few cubic millimeter volume) to the dose bin grid, which has micrometer dimensions in the transversal direction of the microbeams, was performed. Optimization of the simulation parameters, the use of variance-reduction (VR) techniques, and other methods, such as the parallelization of the simulations, were applied in order to speed up the dose computation. Results: Good agreement between MC simulations and experimental results was achieved, even at

  15. Basic features of proton-proton interactions at ultra-relativistic energies and RFT-based quark-gluon string model

    Directory of Open Access Journals (Sweden)

    Zabrodin E.

    2017-01-01

    Full Text Available Proton-proton collisions at energies from √s = 200 GeV up to √s = 14 TeV are studied within the microscopic quark-gluon string model. The model is based on Gribov’s Reggeon Field Theory accomplished by string phenomenology. Comparison with experimental data shows that QGSM describes well particle yields, rapidity - and transverse momentum spectra, rise of mean 〈 pT 〉 and forward-backward multiplicity correlations. The latter arise in QGSM because of the addition of various processes with different mean multiplicities. The model also indicates fulfillment of extended longitudinal scaling and violation of Koba-Nielsen-Olesen scaling at LHC. The origin of both features is traced to short-range particle correlations in the strings. Predictions are made for √s = 14 TeV.

  16. Basic features of proton-proton interactions at ultra-relativistic energies and RFT-based quark-gluon string model

    Science.gov (United States)

    Zabrodin, E.; Bravina, L.; Bleibel, J.

    2017-12-01

    Proton-proton collisions at energies from √s = 200 GeV up to √s = 14 TeV are studied within the microscopic quark-gluon string model. The model is based on Gribov's Reggeon Field Theory accomplished by string phenomenology. Comparison with experimental data shows that QGSM describes well particle yields, rapidity - and transverse momentum spectra, rise of mean 〈 pT 〉 and forward-backward multiplicity correlations. The latter arise in QGSM because of the addition of various processes with different mean multiplicities. The model also indicates fulfillment of extended longitudinal scaling and violation of Koba-Nielsen-Olesen scaling at LHC. The origin of both features is traced to short-range particle correlations in the strings. Predictions are made for √s = 14 TeV.

  17. Demonstration of a zero-variance based scheme for variance reduction to a mini-core Monte Carlo calculation

    International Nuclear Information System (INIS)

    Christoforou, Stavros; Hoogenboom, J. Eduard

    2011-01-01

    A zero-variance based scheme is implemented and tested in the MCNP5 Monte Carlo code. The scheme is applied to a mini-core reactor using the adjoint function obtained from a deterministic calculation for biasing the transport kernels. It is demonstrated that the variance of the k eff estimate is halved compared to a standard criticality calculation. In addition, the biasing does not affect source distribution convergence of the system. However, since the code lacked optimisations for speed, we were not able to demonstrate an appropriate increase in the efficiency of the calculation, because of the higher CPU time cost. (author)

  18. Development of a Monte-Carlo based method for calculating the effect of stationary fluctuations

    DEFF Research Database (Denmark)

    Pettersen, E. E.; Demazire, C.; Jareteg, K.

    2015-01-01

    that corresponds to the real part of the neutron balance, and one that corresponds to the imaginary part. The two equivalent problems are in nature similar to two subcritical systems driven by external neutron sources, and can thus be treated as such in a Monte Carlo framework. The definition of these two...... equivalent problems nevertheless requires the possibility to modify the macroscopic cross-sections, and we use the work of Kuijper, van der Marck and Hogenbirk to define group-wise macroscopic cross-sections in MCNP [1]. The method is illustrated in this paper at a frequency of 1 Hz, for which only the real...

  19. Inorganic proton conducting electrolyte coupled oxide-based dendritic transistors for synaptic electronics.

    Science.gov (United States)

    Wan, Chang Jin; Zhu, Li Qiang; Zhou, Ju Mei; Shi, Yi; Wan, Qing

    2014-05-07

    Ionic/electronic hybrid devices with synaptic functions are considered to be the essential building blocks for neuromorphic systems and brain-inspired computing. Here, artificial synapses based on indium-zinc-oxide (IZO) transistors gated by nanogranular SiO2 proton-conducting electrolyte films are fabricated on glass substrates. Spike-timing dependent plasticity and paired-pulse facilitation are successfully mimicked in an individual bottom-gate transistor. Most importantly, dynamic logic and dendritic integration established by spatiotemporally correlated spikes are also mimicked in dendritic transistors with two in-plane gates as the presynaptic input terminals.

  20. Search for proton decay and supernova neutrino bursts with a lunar base neutron detector

    International Nuclear Information System (INIS)

    Cline, D.B.

    1989-06-01

    We describe the current status of the search for proton decay on earth, emphasizing the decay mode P → K + ν - and discuss the possibility of detecting this mode with a single detector on a lunar base station. The same detector could be used to search for neutrino bursts from distant supernova using the neutral current signature ν μ,τ +N → n+ν x by detecting the produced neutrons. The key advantage of the lunar experiment is the low neutrino flux and possible low radioactive background. (author). 5 refs, 4 tabs, 3 figs

  1. Durability Issues of High Temperature Proton Exchange Membrane Fuel Cells Based on Acid Doped Polybenzimidazole Membranes

    DEFF Research Database (Denmark)

    To achieve high temperature operation of proton exchange membrane fuel cells (PEMFC), preferably under ambient pressure, phosphoric acid doped polybenzimidazole (PBI) membrane represents an effective approach, which in recent years has motivated extensive research activities with great progress....... As a critical concern, issues of long term durability of PBI based fuel cells are addressed in this talk, including oxidative degradation of the polymer, mechanical failures of the membrane, acid leaching out, corrosion of carbon support and sintering of catalysts particles. Excellent polymer durability has...

  2. Intramolecular hydrogen bond in molecular and proton-transfer forms of Schiff bases

    Energy Technology Data Exchange (ETDEWEB)

    Filarowski, A.; Koll, A.; Karpfen, A.; Wolschann, P

    2004-02-16

    The force field and structural parameters modifications upon the formation of intramolecular hydrogen bond and proton transfer reaction in N-methyl-2-hydroxybenzylidene amine (HBZA) are determined on the basis of ab initio and DFT calculations. Reliability of the calculations is verified by comparing of the theoretical vibrational spectra with those experimentally determined in the gas phase. A model of resonance interactions is applied and the quantitative contribution of ortho-quinoid structure in the particular conformers is estimated. A comparison is also made to the systems without {pi}-electron coupling (Mannich bases)

  3. Evaluation of the interindividual human variation in bioactivation of methyleugenol using physiologically based kinetic modeling and Monte Carlo simulations

    Energy Technology Data Exchange (ETDEWEB)

    Al-Subeihi, Ala' A.A., E-mail: subeihi@yahoo.com [Division of Toxicology, Wageningen University, Tuinlaan 5, 6703 HE Wageningen (Netherlands); BEN-HAYYAN-Aqaba International Laboratories, Aqaba Special Economic Zone Authority (ASEZA), P. O. Box 2565, Aqaba 77110 (Jordan); Alhusainy, Wasma; Kiwamoto, Reiko; Spenkelink, Bert [Division of Toxicology, Wageningen University, Tuinlaan 5, 6703 HE Wageningen (Netherlands); Bladeren, Peter J. van [Division of Toxicology, Wageningen University, Tuinlaan 5, 6703 HE Wageningen (Netherlands); Nestec S.A., Avenue Nestlé 55, 1800 Vevey (Switzerland); Rietjens, Ivonne M.C.M.; Punt, Ans [Division of Toxicology, Wageningen University, Tuinlaan 5, 6703 HE Wageningen (Netherlands)

    2015-03-01

    The present study aims at predicting the level of formation of the ultimate carcinogenic metabolite of methyleugenol, 1′-sulfooxymethyleugenol, in the human population by taking variability in key bioactivation and detoxification reactions into account using Monte Carlo simulations. Depending on the metabolic route, variation was simulated based on kinetic constants obtained from incubations with a range of individual human liver fractions or by combining kinetic constants obtained for specific isoenzymes with literature reported human variation in the activity of these enzymes. The results of the study indicate that formation of 1′-sulfooxymethyleugenol is predominantly affected by variation in i) P450 1A2-catalyzed bioactivation of methyleugenol to 1′-hydroxymethyleugenol, ii) P450 2B6-catalyzed epoxidation of methyleugenol, iii) the apparent kinetic constants for oxidation of 1′-hydroxymethyleugenol, and iv) the apparent kinetic constants for sulfation of 1′-hydroxymethyleugenol. Based on the Monte Carlo simulations a so-called chemical-specific adjustment factor (CSAF) for intraspecies variation could be derived by dividing different percentiles by the 50th percentile of the predicted population distribution for 1′-sulfooxymethyleugenol formation. The obtained CSAF value at the 90th percentile was 3.2, indicating that the default uncertainty factor of 3.16 for human variability in kinetics may adequately cover the variation within 90% of the population. Covering 99% of the population requires a larger uncertainty factor of 6.4. In conclusion, the results showed that adequate predictions on interindividual human variation can be made with Monte Carlo-based PBK modeling. For methyleugenol this variation was observed to be in line with the default variation generally assumed in risk assessment. - Highlights: • Interindividual human differences in methyleugenol bioactivation were simulated. • This was done using in vitro incubations, PBK modeling

  4. A Monte Carlo and continuum study of mechanical properties of nanoparticle based films

    Energy Technology Data Exchange (ETDEWEB)

    Ogunsola, Oluwatosin; Ehrman, Sheryl [University of Maryland, Department of Chemical and Biomolecular Engineering, Chemical and Nuclear Engineering Building (United States)], E-mail: sehrman@eng.umd.edu

    2008-01-15

    A combination Monte Carlo and equivalent-continuum simulation approach was used to investigate the structure-mechanical property relationships of titania nanoparticle deposits. Films of titania composed of nanoparticle aggregates were simulated using a Monte Carlo approach with diffusion-limited aggregation. Each aggregate in the simulation is fractal-like and random in structure. In the film structure, it is assumed that bond strength is a function of distance with two limiting values for the bond strengths: one representing the strong chemical bond between the particles at closest proximity in the aggregate and the other representing the weak van der Waals bond between particles from different aggregates. The Young's modulus of the film is estimated using an equivalent-continuum modeling approach, and the influences of particle diameter (5-100 nm) and aggregate size (3-400 particles per aggregate) on predicted Young's modulus are investigated. The Young's modulus is observed to increase with a decrease in primary particle size and is independent of the size of the aggregates deposited. Decreasing porosity resulted in an increase in Young's modulus as expected from results reported previously in the literature.

  5. A Monte Carlo and continuum study of mechanical properties of nanoparticle based films

    International Nuclear Information System (INIS)

    Ogunsola, Oluwatosin; Ehrman, Sheryl

    2008-01-01

    A combination Monte Carlo and equivalent-continuum simulation approach was used to investigate the structure-mechanical property relationships of titania nanoparticle deposits. Films of titania composed of nanoparticle aggregates were simulated using a Monte Carlo approach with diffusion-limited aggregation. Each aggregate in the simulation is fractal-like and random in structure. In the film structure, it is assumed that bond strength is a function of distance with two limiting values for the bond strengths: one representing the strong chemical bond between the particles at closest proximity in the aggregate and the other representing the weak van der Waals bond between particles from different aggregates. The Young's modulus of the film is estimated using an equivalent-continuum modeling approach, and the influences of particle diameter (5-100 nm) and aggregate size (3-400 particles per aggregate) on predicted Young's modulus are investigated. The Young's modulus is observed to increase with a decrease in primary particle size and is independent of the size of the aggregates deposited. Decreasing porosity resulted in an increase in Young's modulus as expected from results reported previously in the literature

  6. GPU Acceleration of Mean Free Path Based Kernel Density Estimators for Monte Carlo Neutronics Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Burke, TImothy P. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Kiedrowski, Brian C. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Martin, William R. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Brown, Forrest B. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2015-11-19

    Kernel Density Estimators (KDEs) are a non-parametric density estimation technique that has recently been applied to Monte Carlo radiation transport simulations. Kernel density estimators are an alternative to histogram tallies for obtaining global solutions in Monte Carlo tallies. With KDEs, a single event, either a collision or particle track, can contribute to the score at multiple tally points with the uncertainty at those points being independent of the desired resolution of the solution. Thus, KDEs show potential for obtaining estimates of a global solution with reduced variance when compared to a histogram. Previously, KDEs have been applied to neutronics for one-group reactor physics problems and fixed source shielding applications. However, little work was done to obtain reaction rates using KDEs. This paper introduces a new form of the MFP KDE that is capable of handling general geometries. Furthermore, extending the MFP KDE to 2-D problems in continuous energy introduces inaccuracies to the solution. An ad-hoc solution to these inaccuracies is introduced that produces errors smaller than 4% at material interfaces.

  7. Multivariate Monte-Carlo Simulation and Economic Valuation of Complex Financial Contracts: An Excel Based Implementation.

    Directory of Open Access Journals (Sweden)

    Timothy J Kyng

    2014-03-01

    Full Text Available The economic valuation of complex financial contracts is often done using Monte-Carlo simulation. We show how to implement this approach using Excel. We discuss Monte-Carlo evaluation for standard single asset European options and then demonstrate how the basic ideas may be extended to evaluate options with exotic multi-asset multi-period features. Single asset option evaluation becomes a special case. We use a typical Executive Stock Option to motivate the discussion, which we analyse using novel theory developed in our previous works. We demonstrate the simulation of the multivariate normal distribution and the multivariate Log-Normal distribution using the Cholesky Square Root of a covariance matrix for replicating the correlation structure in the multi-asset, multi period simulation required for estimating the economic value of the contract. We do this in the standard Black Scholes framework with constant parameters. Excel implementation provides many pedagogical merits due to its relative transparency and simplicity for students. This approach also has relevance to industry due to the widespread use of Excel by practitioners and for graduates who may desire to work in the finance industry. This allows students to be able to price complex financial contracts for which an analytic approach is intractable.

  8. Investigation of SIBM driven recrystallization in alpha Zirconium based on EBSD data and Monte Carlo modeling

    Science.gov (United States)

    Jedrychowski, M.; Bacroix, B.; Salman, O. U.; Tarasiuk, J.; Wronski, S.

    2015-08-01

    The work focuses on the influence of moderate plastic deformation on subsequent partial recrystallization of hexagonal zirconium (Zr702). In the considered case, strain induced boundary migration (SIBM) is assumed to be the dominating recrystallization mechanism. This hypothesis is analyzed and tested in detail using experimental EBSD-OIM data and Monte Carlo computer simulations. An EBSD investigation is performed on zirconium samples, which were channel-die compressed in two perpendicular directions: normal direction (ND) and transverse direction (TD) of the initial material sheet. The maximal applied strain was below 17%. Then, samples were briefly annealed in order to achieve a partly recrystallized state. Obtained EBSD data were analyzed in terms of texture evolution associated with a microstructural characterization, including: kernel average misorientation (KAM), grain orientation spread (GOS), twinning, grain size distributions, description of grain boundary regions. In parallel, Monte Carlo Potts model combined with experimental microstructures was employed in order to verify two main recrystallization scenarios: SIBM driven growth from deformed sub-grains and classical growth of recrystallization nuclei. It is concluded that simulation results provided by the SIBM model are in a good agreement with experimental data in terms of texture as well as microstructural evolution.

  9. Accounting for inhomogeneous broadening in nano-optics by electromagnetic modeling based on Monte Carlo methods

    Science.gov (United States)

    Gudjonson, Herman; Kats, Mikhail A.; Liu, Kun; Nie, Zhihong; Kumacheva, Eugenia; Capasso, Federico

    2014-01-01

    Many experimental systems consist of large ensembles of uncoupled or weakly interacting elements operating as a single whole; this is particularly the case for applications in nano-optics and plasmonics, including colloidal solutions, plasmonic or dielectric nanoparticles on a substrate, antenna arrays, and others. In such experiments, measurements of the optical spectra of ensembles will differ from measurements of the independent elements as a result of small variations from element to element (also known as polydispersity) even if these elements are designed to be identical. In particular, sharp spectral features arising from narrow-band resonances will tend to appear broader and can even be washed out completely. Here, we explore this effect of inhomogeneous broadening as it occurs in colloidal nanopolymers comprising self-assembled nanorod chains in solution. Using a technique combining finite-difference time-domain simulations and Monte Carlo sampling, we predict the inhomogeneously broadened optical spectra of these colloidal nanopolymers and observe significant qualitative differences compared with the unbroadened spectra. The approach combining an electromagnetic simulation technique with Monte Carlo sampling is widely applicable for quantifying the effects of inhomogeneous broadening in a variety of physical systems, including those with many degrees of freedom that are otherwise computationally intractable. PMID:24469797

  10. Proton-beam window design for a transmutation facility operating with a liquid lead target

    Energy Technology Data Exchange (ETDEWEB)

    Jansen, C.; Lypsch, F.; Lizana, P. [Institute for Safety Research and Reactor Technology, Juelich (Germany)] [and others

    1995-10-01

    The proton beam target of an accelerator-driven transmutation facility can be designed as a vertical liquid lead column. To prevent lead vapor from entering the accelerator vacuum, a proton-beam window has to separate the area above the lead surface from the accelerator tube. Two radiation-cooled design alternatives have been investigated which should withstand a proton beam of 1.6 GeV and 25 mA. Temperature calculations based on energy deposition calculations with the Monte Carlo code HETC, stability analysis and spallation-induced damage calculations have been performed showing the applicability of both designs.

  11. Adaptive Kernel-density Independence Sampling based Monte Carlo Sampling (A-KISMCS) for inverse hydrological modelling

    Science.gov (United States)

    Pande, S.; Shafiei, M.

    2016-12-01

    Markov chain Monte Carlo (MCMC) methods have been applied in many hydrologic studies to explore posterior parameter distributions within a Bayesian framework. Accurate estimation of posterior parameter distributions is key to reliably estimate marginal likelihood functions and hence to reliably estimate measures of Bayesian complexity. This paper introduces an alternative to well-known random walk based MCMC samplers. An Adaptive Kernel Density Independence Sampling based Monte Carlo Sampling (A-KISMCS) is proposed. A-KISMCS uses an independence sampler with Metropolis-Hastings (M-H) updates which ensures that candidate observations are drawn independently of the current state of a chain. This ensures efficient exploration of the target distribution. The bandwidth of the kernel density estimator is also adapted online in order to increase its accuracy and ensure fast convergence to a target distribution. The performance of A-KISMCS is tested on one several case studies, including synthetic and real world case studies of hydrological modelling and compared with Differential Evolution Adaptive Metropolis (DREAM-zs), which is fundamentally based on random walk sampling with differential evolution. Results show that while DREAM-zs converges to slightly sharper posterior densities, A-KISMCS is slightly more efficient in tracking the mode of the posteriors.

  12. A Treatment Planning Comparison of Combined Photon-Proton Beams Versus Proton Beams-Only for the Treatment of Skull Base Tumors

    International Nuclear Information System (INIS)

    Feuvret, Loic; Noel, Georges; Weber, Damien C.; Pommier, Pascal; Ferrand, Regis; De Marzi, Ludovic; Dhermain, Frederic; Alapetite, Claire; Mammar, Hamid; Boisserie, Gilbert; Habrand, Jean-Louis; Mazeron, Jean-Jacques

    2007-01-01

    Purpose: To compare treatment planning between combined photon-proton planning (CP) and proton planning (PP) for skull base tumors, so as to assess the potential limitations of CP for these tumors. Methods and Materials: Plans for 10 patients were computed for both CP and PP. Prescribed dose was 67 cobalt Gray equivalent (CGE) for PP; 45 Gy (photons) and 22 CGE (protons) for CP. Dose-volume histograms (DVHs) were calculated for gross target volume (GTV), clinical target volume (CTV), normal tissues (NT), and organs at risk (OARs) for each plan. Results were analyzed using DVH parameters, inhomogeneity coefficient (IC), and conformity index (CI). Results: Mean doses delivered to the GTVs and CTVs with CP (65.0 and 61.7 CGE) and PP (65.3 and 62.2 Gy CGE) were not significantly different (p > 0.1 and p = 0.72). However, the dose inhomogeneity was drastically increased with CP, with a mean significant incremental IC value of 10.5% and CP of 6.8%, for both the GTV (p = 0.01) and CTV (p = 0.04), respectively. The CI 80% values for the GTV and CTV were significantly higher with PP compared with CP. Compared with CP, the use of protons only led to a significant reduction of NT and OAR irradiation, in the intermediate-to-low dose (≤80% isodose line) range. Conclusions: These results suggest that the use of CP results in levels of target dose conformation similar to those with PP. Use of PP significantly reduced the tumor dose inhomogeneity and the delivered intermediate-to-low dose to NT and OARs, leading us to conclude that this treatment is mainly appropriate for tumors in children

  13. Commissioning of a compact laser-based proton beam line for high intensity bunches around 10Â MeV

    Science.gov (United States)

    Busold, S.; Schumacher, D.; Deppert, O.; Brabetz, C.; Kroll, F.; Blažević, A.; Bagnoud, V.; Roth, M.

    2014-03-01

    We report on the first results of experiments with a new laser-based proton beam line at the GSI accelerator facility in Darmstadt. It delivers high current bunches at proton energies around 9.6 MeV, containing more than 109 particles in less than 10 ns and with tunable energy spread down to 2.7% (ΔE/E0 at FWHM). A target normal sheath acceleration stage serves as a proton source and a pulsed solenoid provides for beam collimation and energy selection. Finally a synchronous radio frequency (rf) field is applied via a rf cavity for energy compression at a synchronous phase of -90 deg. The proton bunch is characterized at the end of the very compact beam line, only 3 m behind the laser matter interaction point, which defines the particle source.

  14. Commissioning of a compact laser-based proton beam line for high intensity bunches around 10 MeV

    Directory of Open Access Journals (Sweden)

    S. Busold

    2014-03-01

    Full Text Available We report on the first results of experiments with a new laser-based proton beam line at the GSI accelerator facility in Darmstadt. It delivers high current bunches at proton energies around 9.6 MeV, containing more than 10^{9} particles in less than 10 ns and with tunable energy spread down to 2.7% (ΔE/E_{0} at FWHM. A target normal sheath acceleration stage serves as a proton source and a pulsed solenoid provides for beam collimation and energy selection. Finally a synchronous radio frequency (rf field is applied via a rf cavity for energy compression at a synchronous phase of -90  deg. The proton bunch is characterized at the end of the very compact beam line, only 3 m behind the laser matter interaction point, which defines the particle source.

  15. Optimization of mass of plastic scintillator film for flow-cell based tritium monitoring: a Monte Carlo study

    International Nuclear Information System (INIS)

    Roy, Arup Singha; Palani Selvam, T.; Raman, Anand; Raja, V.; Chaudhury, Probal

    2014-01-01

    Over the years, various types of tritium-in-air monitors have been designed and developed based on different principles. Ionization chamber, proportional counter and scintillation detector systems are few among them. A plastic scintillator based, flow-cell type online tritium-in-air monitoring system was developed for online monitoring of tritium in air. The value of the scintillator mass inside the cell-volume, which maximizes the response of the detector system, should be obtained to get maximum efficiency. The present study is aimed to optimize the amount of mass of the plastic scintillator film for the flow-cell based tritium monitoring instrument so that maximum efficiency is achieved. The Monte Carlo based EGSnrc code system has been used for this purpose

  16. Effect of a proton conducting filler on the physico-chemical properties of SPEEK-based membranes

    Energy Technology Data Exchange (ETDEWEB)

    Mecheri, B.; Chen, F.; Traversa, E. [Department of Chemical Science and Technology, University of Rome ' ' Tor Vergata' ' , Via della Ricerca Scientifica 1, 00133 Roma (Italy); D' Epifanio, A. [Department of Chemical Science and Technology, University of Rome ' ' Tor Vergata' ' , Via della Ricerca Scientifica 1, 00133 Roma (Italy); Hunter College of the City University of New York, New York, NY 10065 (United States); Pisani, L. [CRS4 Parco Scientifico e Tecnologico, POLARIS, 09010 Pula(CA) (Italy); Weise, F.C.; Greenbaum, S. [Hunter College of the City University of New York, New York, NY 10065 (United States); Licoccia, S.

    2009-08-15

    Composite membranes based on sulphonated polyetherether ketone (SPEEK) having a 60% degree of sulphonation (DS=0.6) and containing 23 and 50 wt.-% hydrated tin oxide (SnO{sub 2}.nH{sub 2}O) were prepared and characterised. The lower water uptake (WU) and the higher conductivity values recorded for the composite membranes with respect to pure SPEEK reference suggested the involvement of SnO{sub 2}.nH{sub 2}O in the proton conduction mechanism. Pulsed-field-gradient spin-echo (PFGSE) NMR was employed to obtain a direct measurement of water self-diffusion coefficient in the membranes. Differences were observed between the unfilled SPEEK and the composites, including departures from the normal correlation between water diffusivity and proton conductivity in the case of composites. To better understand the SnO{sub 2}.nH{sub 2}O effect on the proton transport properties of the SPEEK-based membrane, we employed an analytical model that predicts the membrane conductivity as a function of its hydration level and porous structure. The comparison of the model results with the experimental proton conductivity values demonstrated that the tin oxide phase provides additional paths between the water clusters for proton transport, resulting in reduced tortuosity and enhanced proton conductivity. Moreover, the composite showed reduced methanol crossover with respect to the unfilled membrane. (Abstract Copyright [2009], Wiley Periodicals, Inc.)

  17. Two proton-conductive hybrids based on 2-(3-pyridyl)benzimidazole molecules and Keggin-type heteropolyacids

    Energy Technology Data Exchange (ETDEWEB)

    Wei, Mei-Lin, E-mail: weimeilinhd@163.com; Wang, Yu-Xia; Wang, Xin-Jun

    2014-01-15

    Two proton-conductive organic/inorganic complexes were constructed by Keggin-type heteropolyacids and 2-(3-pyridyl)benzimidazole molecules. Single-crystal X-ray diffraction analyses revealed that two complexes crystallized in the monoclinic space group P2{sub 1}/c, exhibited different unit cell parameters, and presented different hydrogen-bonded networks constructed by 2-(3-pyridyl)benzimidazole molecules, [PMo{sub 12}O{sub 40}]{sup 3−} anions and solvent molecules. The results of thermogravimetric analyses suggest that two supramolecular complexes have different thermal stability based on the different hydrogen-bonded networks. Two complexes at 100 °C under 35–98% relative humidity showed a good proton conductivity of about 10{sup −3} S cm{sup −1}. The proton conductivities of two complexes under 98% relative humidity both increase on a logarithmic scale with temperature range from 25 to 100 °C. At 100 °C, both complexes showed poor proton conductivities of 10{sup −8}–10{sup −9} S cm{sup −1} under acetonitrile or methanol vapor. - Graphical abstract: Two molecular hybrids constructed by Keggin-type heteropolyacids and 2-(3-pyridyl)benzimidazole molecules showed good proton conductivities of 10{sup −3} S cm{sup −1} at 100 °C under 35–98% relative humidity. Display Omitted - Highlights: • 2-(3-Pyridyl)benzimidazole could form hydrogen bonds via the N–H groups. • Heteropolyacids have suitable characteristics to be used excellent proton conductors. • Two proton-conductive hybrids based on Keggin HPAs and 3-PyBim were constructed. • The structures were determined by using single-crystal X-ray diffraction data. • They showed good proton conductivities of 10{sup −3} S cm{sup −1} at 100 °C under 35–98% RH.

  18. Analysis of accelerator based neutron spectra for BNCT using proton recoil spectroscopy

    International Nuclear Information System (INIS)

    Wielopolski, L.; Ludewig, H.; Powell, J.R.; Raparia, D.; Alessi, J.G.; Lowenstein, D.I.

    1998-01-01

    Boron Neutron Capture Therapy (BNCT) is a promising binary treatment modality for high-grade primary brain tumors (glioblastoma multiforme, GM) and other cancers. BNCT employs a boron-10 containing compound that preferentially accumulates in the cancer cells in the brain. Upon neutron capture by 10 B energetic alpha particles and triton released at the absorption site kill the cancer cell. In order to gain penetration depth in the brain Fairchild proposed, for this purpose, the use of energetic epithermal neutrons at about 10 keV. Phase I/II clinical trials of BNCT for GM are underway at the Brookhaven Medical Research Reactor (BMRR) and at the MIT Reactor, using these nuclear reactors as the source for epithermal neutrons. In light of the limitations of new reactor installations, e.g. cost, safety and licensing, and limited capability for modulating the reactor based neutron beam energy spectra alternative neutron sources are being contemplated for wider implementation of this modality in a hospital environment. For example, accelerator based neutron sources offer the possibility of tailoring the neutron beams, in terms of improved depth-dose distributions, to the individual and offer, with relative ease, the capability of modifying the neutron beam energy and port size. In previous work new concepts for compact accelerator/target configuration were published. In this work, using the Van de Graaff accelerator the authors have explored different materials for filtering and reflecting neutron beams produced by irradiating a thick Li target with 1.8 to 2.5 MeV proton beams. However, since the yield and the maximum neutron energy emerging from the Li-7(p,n)Be-7 reaction increase with increase in the proton beam energy, there is a need for optimization of the proton energy versus filter and shielding requirements to obtain the desired epithermal neutron beam. The MCNP-4A computer code was used for the initial design studies that were verified with benchmark experiments

  19. A Monte Carlo simulation based two-stage adaptive resonance theory mapping approach for offshore oil spill vulnerability index classification.

    Science.gov (United States)

    Li, Pu; Chen, Bing; Li, Zelin; Zheng, Xiao; Wu, Hongjing; Jing, Liang; Lee, Kenneth

    2014-09-15

    In this paper, a Monte Carlo simulation based two-stage adaptive resonance theory mapping (MC-TSAM) model was developed to classify a given site into distinguished zones representing different levels of offshore Oil Spill Vulnerability Index (OSVI). It consisted of an adaptive resonance theory (ART) module, an ART Mapping module, and a centroid determination module. Monte Carlo simulation was integrated with the TSAM approach to address uncertainties that widely exist in site conditions. The applicability of the proposed model was validated by classifying a large coastal area, which was surrounded by potential oil spill sources, based on 12 features. Statistical analysis of the results indicated that the classification process was affected by multiple features instead of one single feature. The classification results also provided the least or desired number of zones which can sufficiently represent the levels of offshore OSVI in an area under uncertainty and complexity, saving time and budget in spill monitoring and response. Copyright © 2014 Elsevier Ltd. All rights reserved.

  20. Two solar proton fluence models based on ground level enhancement observations

    Directory of Open Access Journals (Sweden)

    Raukunen Osku

    2018-01-01

    Full Text Available Solar energetic particles (SEPs constitute an important component of the radiation environment in interplanetary space. Accurate modeling of SEP events is crucial for the mitigation of radiation hazards in spacecraft design. In this study we present two new statistical models of high energy solar proton fluences based on ground level enhancement (GLE observations during solar cycles 19–24. As the basis of our modeling, we utilize a four parameter double power law function (known as the Band function fits to integral GLE fluence spectra in rigidity. In the first model, the integral and differential fluences for protons with energies between 10 MeV and 1 GeV are calculated using the fits, and the distributions of the fluences at certain energies are modeled with an exponentially cut-off power law function. In the second model, we use a more advanced methodology: by investigating the distributions and relationships of the spectral fit parameters we find that they can be modeled as two independent and two dependent variables. Therefore, instead of modeling the fluences separately at different energies, we can model the shape of the fluence spectrum. We present examples of modeling results and show that the two methodologies agree well except for a short mission duration (1 year at low confidence level. We also show that there is a reasonable agreement between our models and three well-known solar proton models (JPL, ESP and SEPEM, despite the differences in both the modeling methodologies and the data used to construct the models.

  1. Experimental and Monte Carlo-based determination of the beam quality specifier for TomoTherapyHD treatment units.

    Science.gov (United States)

    Howitz, S; Schwedas, M; Wiezorek, T; Zink, K

    2017-10-12

    Reference dosimetry by means of clinical linear accelerators in high-energy photon fields requires the determination of the beam quality specifier TPR 20,10 , which characterizes the relative particle flux density of the photon beam. The measurement of TPR 20,10 has to be performed in homogenous photon beams of size 10×10cm 2 with a focus-detector distance of 100cm. These requirements cannot be fulfilled by TomoTherapy treatment units from Accuray. The TomoTherapy unit provides a flattening-filter-free photon fan beam with a maximum field width of 40cm and field lengths of 1.0cm, 2.5cm and 5.0cm at a focus-isocenter distance of 85cm. For the determination of the beam quality specifier from measurements under nonstandard reference conditions Sauer and Palmans proposed experiment-based fit functions. Moreover, Sauer recommends considering the impact of the flattening-filter-free beam on the measured data. To verify these fit functions, in the present study a Monte Carlo based model of the treatment head of a TomoTherapyHD unit was designed and commissioned with existing beam data of our clinical TomoTherapy machine. Depth dose curves and dose profiles were in agreement within 1.5% between experimental and Monte Carlo-based data. Based on the fit functions from Sauer and Palmans the beam quality specifier TPR 20,10 was determined from field sizes 5×5cm 2 , 10×5cm 2 , 20×5cm 2 and 40×5cm 2 based on dosimetric measurements and Monte Carlo simulations. The mean value from all experimental values of TPR 20,10 resulted in TPR 20,10 ¯=0.635±0.4%. The impact of the non-homogenous field due to the flattening-filter-free beam was negligible for field sizes below 20×5cm 2 . The beam quality specifier calculated by Monte Carlo simulations was TPR 20,10 =0.628 and TPR 20,10 =0.631 for two different calculation methods. The stopping power ratio water-to-air s w,a Δ directly depends on the beam quality specifier. The value determined from all experimental TPR 20,10 data

  2. PC-based process distribution to solve iterative Monte Carlo simulations in physical dosimetry

    International Nuclear Information System (INIS)

    Leal, A.; Sanchez-Doblado, F.; Perucha, M.; Rincon, M.; Carrasco, E.; Bernal, C.

    2001-01-01

    A distribution model to simulate physical dosimetry measurements with Monte Carlo (MC) techniques has been developed. This approach is indicated to solve the simulations where there are continuous changes of measurement conditions (and hence of the input parameters) such as a TPR curve or the estimation of the resolution limit of an optimal densitometer in the case of small field profiles. As a comparison, a high resolution scan for narrow beams with no iterative process is presented. The model has been installed on a network PCs without any resident software. The only requirement for these PCs has been a small and temporal Linux partition in the hard disks and to be connecting by the net with our server PC. (orig.)

  3. Monte Carlo based geometrical model for efficiency calculation of an n-type HPGe detector

    Energy Technology Data Exchange (ETDEWEB)

    Padilla Cabal, Fatima, E-mail: fpadilla@instec.c [Instituto Superior de Tecnologias y Ciencias Aplicadas, ' Quinta de los Molinos' Ave. Salvador Allende, esq. Luaces, Plaza de la Revolucion, Ciudad de la Habana, CP 10400 (Cuba); Lopez-Pino, Neivy; Luis Bernal-Castillo, Jose; Martinez-Palenzuela, Yisel; Aguilar-Mena, Jimmy; D' Alessandro, Katia; Arbelo, Yuniesky; Corrales, Yasser; Diaz, Oscar [Instituto Superior de Tecnologias y Ciencias Aplicadas, ' Quinta de los Molinos' Ave. Salvador Allende, esq. Luaces, Plaza de la Revolucion, Ciudad de la Habana, CP 10400 (Cuba)

    2010-12-15

    A procedure to optimize the geometrical model of an n-type detector is described. Sixteen lines from seven point sources ({sup 241}Am, {sup 133}Ba, {sup 22}Na, {sup 60}Co, {sup 57}Co, {sup 137}Cs and {sup 152}Eu) placed at three different source-to-detector distances (10, 20 and 30 cm) were used to calibrate a low-background gamma spectrometer between 26 and 1408 keV. Direct Monte Carlo techniques using the MCNPX 2.6 and GEANT 4 9.2 codes, and a semi-empirical procedure were performed to obtain theoretical efficiency curves. Since discrepancies were found between experimental and calculated data using the manufacturer parameters of the detector, a detail study of the crystal dimensions and the geometrical configuration is carried out. The relative deviation with experimental data decreases from a mean value of 18-4%, after the parameters were optimized.

  4. Hospital-based proton linear accelerator for particle therapy and radioisotope production

    Science.gov (United States)

    Lennox, Arlene J.

    1991-05-01

    Taking advantage of recent advances in linear accelerator technology, it is possible for a hospital to use a 70 MeV proton linac for fast neutron therapy, boron neutron capture therapy, proton therapy for ocular melanomas, and production of radiopharmaceuticals. The linac can also inject protons into a synchrotron for proton therapy of deep-seated tumors. With 180 μA average current, a single linac can support all these applications. This paper presents a conceptual design for a medical proton linac, switchyard, treatment rooms, and isotope production rooms. Special requirements for each application are outlined and a layout for sharing beam among the applications is suggested.

  5. Poly(dA-dT).poly(dA-dT) two-pathway proton exchange mechanism. Effect of general and specific base catalysis on deuteration rates

    International Nuclear Information System (INIS)

    Hartmann, B.; Leng, M.; Ramstein, J.

    1986-01-01

    The deuteration rates of the poly(dA-dT).poly(dA-dT) amino and imino protons have been measured with stopped-flow spectrophotometry as a function of general and specific base catalyst concentration. Two proton exchange classes are found with time constants differing by a factor of 10 (4 and 0.4 s-1). The slower class represents the exchange of the adenine amino protons whereas the proton of the faster class has been assigned to the thymine imino proton. The exchange rates of these two classes of protons are independent of general and specific base catalyst concentration. This very characteristic behavior demonstrates that in our experimental conditions the exchange rates of the imino and amino protons in poly(dA-dT).poly(dA-dT) are limited by two different conformational fluctuations. We present a three-state exchange mechanism accounting for our experimental results

  6. Accelerating parameter identification of proton exchange membrane fuel cell model with ranking-based differential evolution

    International Nuclear Information System (INIS)

    Gong, Wenyin; Cai, Zhihua

    2013-01-01

    Parameter identification of PEM (proton exchange membrane) fuel cell model is a very active area of research. Generally, it can be treated as a numerical optimization problem with complex nonlinear and multi-variable features. DE (differential evolution), which has been successfully used in various fields, is a simple yet efficient evolutionary algorithm for global numerical optimization. In this paper, with the objective of accelerating the process of parameter identification of PEM fuel cell models and reducing the necessary computational efforts, we firstly present a generic and simple ranking-based mutation operator for the DE algorithm. Then, the ranking-based mutation operator is incorporated into five highly-competitive DE variants to solve the PEM fuel cell model parameter identification problems. The main contributions of this work are the proposed ranking-based DE variants and their application to the parameter identification problems of PEM fuel cell models. Experiments have been conducted by using both the simulated voltage–current data and the data obtained from the literature to validate the performance of our approach. The results indicate that the ranking-based DE methods provide better results with respect to the solution quality, the convergence rate, and the success rate compared with their corresponding original DE methods. In addition, the voltage–current characteristics obtained by our approach are in good agreement with the original voltage–current curves in all cases. - Highlights: • A simple and generic ranking-based mutation operator is presented in this paper. • Several DE (differential evolution) variants are used to solve the parameter identification of PEMFC (proton exchange membrane fuel cells) model. • Results show that our method accelerates the process of parameter identification. • The V–I characteristics are in very good agreement with experimental data

  7. Computational Model of D-Region Ion Production Caused by Energetic Electron Precipitations Based on General Monte Carlo Transport Calculations

    Science.gov (United States)

    Kouznetsov, A.; Cully, C. M.

    2017-12-01

    During enhanced magnetic activities, large ejections of energetic electrons from radiation belts are deposited in the upper polar atmosphere where they play important roles in its physical and chemical processes, including VLF signals subionospheric propagation. Electron deposition can affect D-Region ionization, which are estimated based on ionization rates derived from energy depositions. We present a model of D-region ion production caused by an arbitrary (in energy and pitch angle) distribution of fast (10 keV - 1 MeV) electrons. The model relies on a set of pre-calculated results obtained using a general Monte Carlo approach with the latest version of the MCNP6 (Monte Carlo N-Particle) code for the explicit electron tracking in magnetic fields. By expressing those results using the ionization yield functions, the pre-calculated results are extended to cover arbitrary magnetic field inclinations and atmospheric density profiles, allowing ionization rate altitude profile computations in the range of 20 and 200 km at any geographic point of interest and date/time by adopting results from an external atmospheric density model (e.g. NRLMSISE-00). The pre-calculated MCNP6 results are stored in a CDF (Common Data Format) file, and IDL routines library is written to provide an end-user interface to the model.

  8. Generic FPGA-Based Platform for Distributed IO in Proton Therapy Patient Safety Interlock System

    Science.gov (United States)

    Eichin, Michael; Carmona, Pablo Fernandez; Johansen, Ernst; Grossmann, Martin; Mayor, Alexandre; Erhardt, Daniel; Gomperts, Alexander; Regele, Harald; Bula, Christian; Sidler, Christof

    2017-06-01

    At the Paul Scherrer Institute (PSI) in Switzerland, cancer patients are treated with protons. Proton therapy at PSI has a long history and started in the 1980s. More than 30 years later, a new gantry has recently been installed in the existing facility. This new machine has been delivered by an industry partner. A big challenge is the integration of the vendor's safety system into the existing PSI environment. Different interface standards and the complexity of the system made it necessary to find a technical solution connecting an industry system to the existing PSI infrastructure. A novel very flexible distributed IO system based on field-programmable gate array (FPGA) technology was developed, supporting many different IO interface standards and high-speed communication links connecting the device to a PSI standard versa module eurocard-bus input output controller. This paper summarizes the features of the hardware technology, the FPGA framework with its high-speed communication link protocol, and presents our first measurement results.

  9. Fabrication of high-transmission microporous membranes by proton beam writing-based molding technique

    Science.gov (United States)

    Wang, Liping; Meyer, Clemens; Guibert, Edouard; Homsy, Alexandra; Whitlow, Harry J.

    2017-08-01

    Porous membranes are widely used as filters in a broad range of micro and nanofluidic applications, e.g. organelle sorters, permeable cell growth substrates, and plasma filtration. Conventional silicon fabrication approaches are not suitable for microporous membranes due to the low mechanical stability of thin film substrates. Other techniques like ion track etching are limited to the production of randomly distributed and randomly orientated pores with non-uniform pore sizes. In this project, we developed a procedure for fabricating high-transmission microporous membranes by proton beam writing (PBW) with a combination of spin-casting and soft lithography. In this approach, focused 2 MeV protons were used to lithographically write patterns consisting of hexagonal arrays of high-density pillars of few μm size in a SU-8 layer coated on a silicon wafer. After development, the pillars were conformably coated with a thin film of poly-para-xylylene (Parylene)-C release agent and spin-coated with polydimethylsiloxane (PDMS). To facilitate demolding, a special technique based on the use of a laser-cut sealing tape ring was developed. This method facilitated the successful delamination of 20-μm thick PDMS membrane with high-density micropores from the mold without rupture or damage.

  10. Optimization of dual-energy CT acquisitions for proton therapy using projection-based decomposition.

    Science.gov (United States)

    Vilches-Freixas, Gloria; Létang, Jean Michel; Ducros, Nicolas; Rit, Simon

    2017-09-01

    Dual-energy computed tomography (DECT) has been presented as a valid alternative to single-energy CT to reduce the uncertainty of the conversion of patient CT numbers to proton stopping power ratio (SPR) of tissues relative to water. The aim of this work was to optimize DECT acquisition protocols from simulations of X-ray images for the treatment planning of proton therapy using a projection-based dual-energy decomposition algorithm. We have investigated the effect of various voltages and tin filtration combinations on the SPR map accuracy and precision, and the influence of the dose allocation between the low-energy (LE) and the high-energy (HE) acquisitions. For all spectra combinations, virtual CT projections of the Gammex phantom were simulated with a realistic energy-integrating detector response model. Two situations were simulated: an ideal case without noise (infinite dose) and a realistic situation with Poisson noise corresponding to a 20 mGy total central dose. To determine the optimal dose balance, the proportion of LE-dose with respect to the total dose was varied from 10% to 90% while keeping the central dose constant, for four dual-energy spectra. SPR images were derived using a two-step projection-based decomposition approach. The ranges of 70 MeV, 90 MeV, and 100 MeV proton beams onto the adult female (AF) reference computational phantom of the ICRP were analytically determined from the reconstructed SPR maps. The energy separation between the incident spectra had a strong impact on the SPR precision. Maximizing the incident energy gap reduced image noise. However, the energy gap was not a good metric to evaluate the accuracy of the SPR. In terms of SPR accuracy, a large variability of the optimal spectra was observed when studying each phantom material separately. The SPR accuracy was almost flat in the 30-70% LE-dose range, while the precision showed a minimum slightly shifted in favor of lower LE-dose. Photon noise in the SPR images (20 mGy dose

  11. Monte Carlo simulation methods in moment-based scale-bridging algorithms for thermal radiative-transfer problems

    International Nuclear Information System (INIS)

    Densmore, J.D.; Park, H.; Wollaber, A.B.; Rauenzahn, R.M.; Knoll, D.A.

    2015-01-01

    We present a moment-based acceleration algorithm applied to Monte Carlo simulation of thermal radiative-transfer problems. Our acceleration algorithm employs a continuum system of moments to accelerate convergence of stiff absorption–emission physics. The combination of energy-conserving tallies and the use of an asymptotic approximation in optically thick regions remedy the difficulties of local energy conservation and mitigation of statistical noise in such regions. We demonstrate the efficiency and accuracy of the developed method. We also compare directly to the standard linearization-based method of Fleck and Cummings [1]. A factor of 40 reduction in total computational time is achieved with the new algorithm for an equivalent (or more accurate) solution as compared with the Fleck–Cummings algorithm

  12. Proton therapy of uveal melanomas. Intercomparison of MRI-based and conventional treatment planning

    International Nuclear Information System (INIS)

    Marnitz, S.; Hinkelbein, W.; Cordini, D.; Heufelder, J.; Simiantonakis, I.; Kluge, H.; Bendl, R.; Lemke, A.J.; Bechrakis, N.E.; Foerster, M.H.

    2006-01-01

    Background and purpose: proton therapy for uveal melanoma provides high-conformal dose application to the target volume and, thus, an optimal saving of the organs at risk nearby. Treatment planning is done with the model-based treatment-planning system eyeplan. Tumor reconstruction is based only on a fundus composite, which often leads to an overestimation of the clinical target volume (CTV). The purpose was to exploit MRI on trial in a proton therapy-planning system by using the novel image-based treatment-planning system octopus. Patients and methods: ten patients with uveal melanomas received both a high-resolution planning CT and MRI of the eye. MR examinations were made with an eye coil. Eyeplan requires eye geometry data for modeling, and tantalum marker clips for submillimeter positioning and additional information from ultrasound and 3-D imaging. By contrast, octopus provides the full integration of 3-D imaging (e.g., CT, MRI). CTVs were delineated in each slice. For all patients, CTVs (eyeplan vs. octopus) were compared intraindividually. Results: octopus planning led to a mean reduction of the target volume by a factor of 1.7 (T1-weighted [T1w]) and 2.2 (T2w) without compromising safety. The corresponding field size could be scaled down on average by a factor of 1.2 (T1w) and 1.4 (T2w), respectively. Conclusion: compared with the conventional eyeplan, MRI-based treatment planning of ocular tumors with octopus could be a powerful tool for reducing the CTV and, consequently, the treatment volume and the field size. This might be translated into a better patient compliance during treatment and a decreased late toxicity. (orig.)

  13. Validation of proton stopping power ratio estimation based on dual energy CT using fresh tissue samples

    Science.gov (United States)

    Taasti, Vicki T.; Michalak, Gregory J.; Hansen, David C.; Deisher, Amanda J.; Kruse, Jon J.; Krauss, Bernhard; Muren, Ludvig P.; Petersen, Jørgen B. B.; McCollough, Cynthia H.

    2018-01-01

    Dual energy CT (DECT) has been shown, in theoretical and phantom studies, to improve the stopping power ratio (SPR) determination used for proton treatment planning compared to the use of single energy CT (SECT). However, it has not been shown that this also extends to organic tissues. The purpose of this study was therefore to investigate the accuracy of SPR estimation for fresh pork and beef tissue samples used as surrogates of human tissues. The reference SPRs for fourteen tissue samples, which included fat, muscle and femur bone, were measured using proton pencil beams. The tissue samples were subsequently CT scanned using four different scanners with different dual energy acquisition modes, giving in total six DECT-based SPR estimations for each sample. The SPR was estimated using a proprietary algorithm (syngo.via DE Rho/Z Maps, Siemens Healthcare, Forchheim, Germany) for extracting the electron density and the effective atomic number. SECT images were also acquired and SECT-based SPR estimations were performed using a clinical Hounsfield look-up table. The mean and standard deviation of the SPR over large volume-of-interests were calculated. For the six different DECT acquisition methods, the root-mean-square errors (RMSEs) for the SPR estimates over all tissue samples were between 0.9% and 1.5%. For the SECT-based SPR estimation the RMSE was 2.8%. For one DECT acquisition method, a positive bias was seen in the SPR estimates, having a mean error of 1.3%. The largest errors were found in the very dense cortical bone from a beef femur. This study confirms the advantages of DECT-based SPR estimation although good results were also obtained using SECT for most tissues.

  14. Proton and carbon ion radiotherapy for primary brain tumors and tumors of the skull base

    International Nuclear Information System (INIS)

    Combs, Stephanie E.; Kessel, Kerstin; Habermehl, Daniel; Debus, Jurgen; Haberer, Thomas; Jaekel, Oliver

    2013-01-01

    To analyze clinical concepts, toxicity and treatment outcome in patients with brain and skull base tumors treated with photons and particle therapy. Material and methods: In total 260 patients with brain tumors and tumors of the skull base were treated at the Heidelberg Ion Therapy Center (HIT). Patients enrolled in and randomized within prospective clinical trials as well as bony or soft tissue tumors are not included in this analysis. Treatment was delivered as protons, carbon ions, or combinations of photons and a carbon ion boost. All patients are included in a tight follow-up program. The median follow-up time is 12 months (range 2-39 months). Results: Main histologies included meningioma (n = 107) for skull base lesions, pituitary adenomas (n = 14), low-grade gliomas (n = 51) as well as high-grade gliomas (n = 55) for brain tumors. In all patients treatment could be completed without any unexpected severe toxicities. No side effects > CTC Grade III were observed. To date, no severe late toxicities were observed, however, for endpoints such as secondary malignancies or neuro cognitive side effects follow-up time still remains too short. Local recurrences were mainly seen in the group of high-grade gliomas or atypical meningiomas; for benign skull base meningiomas, to date, no recurrences were observed during follow-up. Conclusion: The specific benefit of particle therapy will potentially reduce the risk of secondary malignancies as well as improve neuro cognitive outcome and quality of life (QOL); thus, longer follow-up will be necessary to confirm these endpoints. Indication-specific trials on meningiomas and gliomas are underway to elucidate the role of protons and carbon ions in these indications

  15. Proton and carbon ion radiotherapy for primary brain tumors and tumors of the skull base

    Energy Technology Data Exchange (ETDEWEB)

    Combs, Stephanie E.; Kessel, Kerstin; Habermehl, Daniel; Debus, Jurgen [Univ. Hospital of Heidelberg, Dept. of Radiation Oncology, Heidelberg (Germany)], e-mail: Stephanie.Combs@med.uni-heidelberg.de; Haberer, Thomas [Heidelberger Ionenstrahl Therapiezentrum (HIT), Heidelberg (Germany); Jaekel, Oliver [Univ. Hospital of Heidelberg, Dept. of Radiation Oncology, Heidelberg (Germany); Heidelberger Ionenstrahl Therapiezentrum (HIT), Heidelberg (Germany)

    2013-10-15

    To analyze clinical concepts, toxicity and treatment outcome in patients with brain and skull base tumors treated with photons and particle therapy. Material and methods: In total 260 patients with brain tumors and tumors of the skull base were treated at the Heidelberg Ion Therapy Center (HIT). Patients enrolled in and randomized within prospective clinical trials as well as bony or soft tissue tumors are not included in this analysis. Treatment was delivered as protons, carbon ions, or combinations of photons and a carbon ion boost. All patients are included in a tight follow-up program. The median follow-up time is 12 months (range 2-39 months). Results: Main histologies included meningioma (n = 107) for skull base lesions, pituitary adenomas (n = 14), low-grade gliomas (n = 51) as well as high-grade gliomas (n = 55) for brain tumors. In all patients treatment could be completed without any unexpected severe toxicities. No side effects > CTC Grade III were observed. To date, no severe late toxicities were observed, however, for endpoints such as secondary malignancies or neuro cognitive side effects follow-up time still remains too short. Local recurrences were mainly seen in the group of high-grade gliomas or atypical meningiomas; for benign skull base meningiomas, to date, no recurrences were observed during follow-up. Conclusion: The specific benefit of particle therapy will potentially reduce the risk of secondary malignancies as well as improve neuro cognitive outcome and quality of life (QOL); thus, longer follow-up will be necessary to confirm these endpoints. Indication-specific trials on meningiomas and gliomas are underway to elucidate the role of protons and carbon ions in these indications.

  16. Proton therapy of uveal melanomas. Intercomparison of MRI-based and conventional treatment planning

    Energy Technology Data Exchange (ETDEWEB)

    Marnitz, S.; Hinkelbein, W. [Dept. of Radiooncology, Charite Univ. Medicine, Berlin (Germany); Cordini, D.; Heufelder, J.; Simiantonakis, I.; Kluge, H. [Eye Tumor Therapy, Hahn-Meitner Inst., Berlin (Germany); Bendl, R. [Dept. of Medical Physics, German Cancer Research Center (DKFZ), Heidelberg (Germany); Lemke, A.J. [Dept. of Diagnostic Radiology, Charite Univ. Medicine, Berlin (Germany); Bechrakis, N.E.; Foerster, M.H. [Dept. of Ophthalmology, Charite Univ. Medicine, Berlin (Germany)

    2006-07-15

    Background and purpose: proton therapy for uveal melanoma provides high-conformal dose application to the target volume and, thus, an optimal saving of the organs at risk nearby. Treatment planning is done with the model-based treatment-planning system eyeplan. Tumor reconstruction is based only on a fundus composite, which often leads to an overestimation of the clinical target volume (CTV). The purpose was to exploit MRI on trial in a proton therapy-planning system by using the novel image-based treatment-planning system octopus. Patients and methods: ten patients with uveal melanomas received both a high-resolution planning CT and MRI of the eye. MR examinations were made with an eye coil. Eyeplan requires eye geometry data for modeling, and tantalum marker clips for submillimeter positioning and additional information from ultrasound and 3-D imaging. By contrast, octopus provides the full integration of 3-D imaging (e.g., CT, MRI). CTVs were delineated in each slice. For all patients, CTVs (eyeplan vs. octopus) were compared intraindividually. Results: octopus planning led to a mean reduction of the target volume by a factor of 1.7 (T1-weighted [T1w]) and 2.2 (T2w) without compromising safety. The corresponding field size could be scaled down on average by a factor of 1.2 (T1w) and 1.4 (T2w), respectively. Conclusion: compared with the conventional eyeplan, MRI-based treatment planning of ocular tumors with octopus could be a powerful tool for reducing the CTV and, consequently, the treatment volume and the field size. This might be translated into a better patient compliance during treatment and a decreased late toxicity. (orig.)

  17. Low energy proton and electron response of an n-octane liquid scintillator

    International Nuclear Information System (INIS)

    Wood, M.H.; Blankenship, S.M.; Lang, T.P.; Avignone, F.T. III; Morgan, G.L.; Raman, S.

    1981-01-01

    The proton and electron responses of an n-octane based liquid scintillator were measured. Knock-on protons by neutrons from the Oak Ridge Electron Linear Accelerator (ORELA) were used to determine the proton response. Neutron energies from 82 to 1195 keV were selected by time-of-flight with a resolution of better than 1%. The electron response was measured using γ-rays and fluorescence X-rays with energies from 13.4 to 1275 keV. Data interpretation required Monte Carlo simulation. The proton response deviates from linearity below 470 keV while the electron data are consistent with linearity down to 40 keV. The proton to electron response ratio measured in this work is 0.20 +- 0.02 at 1 MeV and 0.13 +- 0.01 at 100 keV. (orig.)

  18. Monte Carlo simulation based study of a proposed multileaf collimator for a telecobalt machine

    International Nuclear Information System (INIS)

    Sahani, G.; Dash Sharma, P. K.; Hussain, S. A.; Dutt Sharma, Sunil; Sharma, D. N.

    2013-01-01

    Purpose: The objective of the present work was to propose a design of a secondary multileaf collimator (MLC) for a telecobalt machine and optimize its design features through Monte Carlo simulation. Methods: The proposed MLC design consists of 72 leaves (36 leaf pairs) with additional jaws perpendicular to leaf motion having the capability of shaping a maximum square field size of 35 × 35 cm 2 . The projected widths at isocenter of each of the central 34 leaf pairs and 2 peripheral leaf pairs are 10 and 5 mm, respectively. The ends of the leaves and the x-jaws were optimized to obtain acceptable values of dosimetric and leakage parameters. Monte Carlo N-Particle code was used for generating beam profiles and depth dose curves and estimating the leakage radiation through the MLC. A water phantom of dimension 50 × 50 × 40 cm 3 with an array of voxels (4 × 0.3 × 0.6 cm 3 = 0.72 cm 3 ) was used for the study of dosimetric and leakage characteristics of the MLC. Output files generated for beam profiles were exported to the PTW radiation field analyzer software through locally developed software for analysis of beam profiles in order to evaluate radiation field width, beam flatness, symmetry, and beam penumbra. Results: The optimized version of the MLC can define radiation fields of up to 35 × 35 cm 2 within the prescribed tolerance values of 2 mm. The flatness and symmetry were found to be well within the acceptable tolerance value of 3%. The penumbra for a 10 × 10 cm 2 field size is 10.7 mm which is less than the generally acceptable value of 12 mm for a telecobalt machine. The maximum and average radiation leakage through the MLC were found to be 0.74% and 0.41% which are well below the International Electrotechnical Commission recommended tolerance values of 2% and 0.75%, respectively. The maximum leakage through the leaf ends in closed condition was observed to be 8.6% which is less than the values reported for other MLCs designed for medical linear

  19. Monte Carlo simulations of the radiation environment for the CMS Experiment

    CERN Document Server

    AUTHOR|(CDS)2068566; Bayshev, I.; Bergstrom, I.; Cooijmans, T.; Dabrowski, A.; Glöggler, L.; Guthoff, M.; Kurochkin, I.; Vincke, H.; Tajeda, S.

    2016-01-01

    Monte Carlo radiation transport codes are used by the CMS Beam Radiation Instrumentation and Luminosity (BRIL) project to estimate the radiation levels due to proton-proton collisions and machine induced background. Results are used by the CMS collaboration for various applications: comparison with detector hit rates, pile-up studies, predictions of radiation damage based on various models (Dose, NIEL, DPA), shielding design, estimations of residual dose environment. Simulation parameters, and the maintenance of the input files are summarised, and key results are presented. Furthermore, an overview of additional programs developed by the BRIL project to meet the specific needs of CMS community is given.

  20. The Virtual Monte Carlo

    CERN Document Server

    Hrivnacova, I; Berejnov, V V; Brun, R; Carminati, F; Fassò, A; Futo, E; Gheata, A; Caballero, I G; Morsch, Andreas

    2003-01-01

    The concept of Virtual Monte Carlo (VMC) has been developed by the ALICE Software Project to allow different Monte Carlo simulation programs to run without changing the user code, such as the geometry definition, the detector response simulation or input and output formats. Recently, the VMC classes have been integrated into the ROOT framework, and the other relevant packages have been separated from the AliRoot framework and can be used individually by any other HEP project. The general concept of the VMC and its set of base classes provided in ROOT will be presented. Existing implementations for Geant3, Geant4 and FLUKA and simple examples of usage will be described.

  1. Sustainable Queuing-Network Design for Airport Security Based on the Monte Carlo Method

    Directory of Open Access Journals (Sweden)

    Xiangqian Xu

    2018-01-01

    Full Text Available The design of airport queuing networks is a significant research field currently for researchers. Many factors must to be considered in order to achieve the optimized strategies, including the passenger flow volume, boarding time, and boarding order of passengers. Optimizing these factors lead to the sustainable development of the queuing network, which currently faces a few difficulties. In particular, the high variance in checkpoint lines can be extremely costly to passengers as they arrive unduly early or possibly miss their scheduled flights. In this article, the Monte Carlo method is used to design the queuing network so as to achieve sustainable development. Thereafter, a network diagram is used to determine the critical working point, and design a structurally and functionally sustainable network. Finally, a case study for a sustainable queuing-network design in the airport is conducted to verify the efficiency of the proposed model. Specifically, three sustainable queuing-network design solutions are proposed, all of which not only maintain the same standards of security, but also increase checkpoint throughput and reduce passenger waiting time variance.

  2. Monte Carlo based unit commitment procedures for the deregulated market environment

    International Nuclear Information System (INIS)

    Granelli, G.P.; Marannino, P.; Montagna, M.; Zanellini, F.

    2006-01-01

    The unit commitment problem, originally conceived in the framework of short term operation of vertically integrated utilities, needs a thorough re-examination in the light of the ongoing transition towards the open electricity market environment. In this work the problem is re-formulated to adapt unit commitment to the viewpoint of a generation company (GENCO) which is no longer bound to satisfy its load, but is willing to maximize its profits. Moreover, with reference to the present day situation in many countries, the presence of a GENCO (the former monopolist) which is in the position of exerting the market power, requires a careful analysis to be carried out considering the different perspectives of a price taker and of the price maker GENCO. Unit commitment is thus shown to lead to a couple of distinct, yet slightly different problems. The unavoidable uncertainties in load profile and price behaviour over the time period of interest are also taken into account by means of a Monte Carlo simulation. Both the forecasted loads and prices are handled as random variables with a normal multivariate distribution. The correlation between the random input variables corresponding to successive hours of the day was considered by carrying out a statistical analysis of actual load and price data. The whole procedure was tested making use of reasonable approximations of the actual data of the thermal generation units available to come actual GENCOs operating in Italy. (author)

  3. Monte Carlo based water/medium stopping-power ratios for various ICRP and ICRU tissues

    International Nuclear Information System (INIS)

    Fernandez-Varea, Jose M; Carrasco, Pablo; Panettieri, Vanessa; Brualla, Lorenzo

    2007-01-01

    Water/medium stopping-power ratios, s w,m , have been calculated for several ICRP and ICRU tissues, namely adipose tissue, brain, cortical bone, liver, lung (deflated and inflated) and spongiosa. The considered clinical beams were 6 and 18 MV x-rays and the field size was 10 x 10 cm 2 . Fluence distributions were scored at a depth of 10 cm using the Monte Carlo code PENELOPE. The collision stopping powers for the studied tissues were evaluated employing the formalism of ICRU Report 37 (1984 Stopping Powers for Electrons and Positrons (Bethesda, MD: ICRU)). The Bragg-Gray values of s w,m calculated with these ingredients range from about 0.98 (adipose tissue) to nearly 1.14 (cortical bone), displaying a rather small variation with beam quality. Excellent agreement, to within 0.1%, is found with stopping-power ratios reported by Siebers et al (2000a Phys. Med. Biol. 45 983-95) for cortical bone, inflated lung and spongiosa. In the case of cortical bone, s w,m changes approximately 2% when either ICRP or ICRU compositions are adopted, whereas the stopping-power ratios of lung, brain and adipose tissue are less sensitive to the selected composition. The mass density of lung also influences the calculated values of s w,m , reducing them by around 1% (6 MV) and 2% (18 MV) when going from deflated to inflated lung

  4. Radiation dose performance in the triple-source CT based on a Monte Carlo method

    Science.gov (United States)

    Yang, Zhenyu; Zhao, Jun

    2012-10-01

    Multiple-source structure is promising in the development of computed tomography, for it could effectively eliminate motion artifacts in the cardiac scanning and other time-critical implementations with high temporal resolution. However, concerns about the dose performance shade this technique, as few reports on the evaluation of dose performance of multiple-source CT have been proposed for judgment. Our experiments focus on the dose performance of one specific multiple-source CT geometry, the triple-source CT scanner, whose theories and implementations have already been well-established and testified by our previous work. We have modeled the triple-source CT geometry with the help of EGSnrc Monte Carlo radiation transport code system, and simulated the CT examinations of a digital chest phantom with our modified version of the software, using x-ray spectrum according to the data of physical tube. Single-source CT geometry is also estimated and tested for evaluation and comparison. Absorbed dose of each organ is calculated according to its real physics characteristics. Results show that the absorbed radiation dose of organs with the triple-source CT is almost equal to that with the single-source CT system. As the advantage of temporal resolution, the triple-source CT would be a better choice in the x-ray cardiac examination.

  5. Accuracy assessment of a new Monte Carlo based burnup computer code

    International Nuclear Information System (INIS)

    El Bakkari, B.; ElBardouni, T.; Nacir, B.; ElYounoussi, C.; Boulaich, Y.; Meroun, O.; Zoubair, M.; Chakir, E.

    2012-01-01

    Highlights: ► A new burnup code called BUCAL1 was developed. ► BUCAL1 uses the MCNP tallies directly in the calculation of the isotopic inventories. ► Validation of BUCAL1 was done by code to code comparison using VVER-1000 LEU Benchmark Assembly. ► Differences from BM value were found to be ± 600 pcm for k ∞ and ±6% for the isotopic compositions. ► The effect on reactivity due to the burnup of Gd isotopes is well reproduced by BUCAL1. - Abstract: This study aims to test for the suitability and accuracy of a new home-made Monte Carlo burnup code, called BUCAL1, by investigating and predicting the neutronic behavior of a “VVER-1000 LEU Assembly Computational Benchmark”, at lattice level. BUCAL1 uses MCNP tally information directly in the computation; this approach allows performing straightforward and accurate calculation without having to use the calculated group fluxes to perform transmutation analysis in a separate code. ENDF/B-VII evaluated nuclear data library was used in these calculations. Processing of the data library is performed using recent updates of NJOY99 system. Code to code comparisons with the reported Nuclear OECD/NEA results are presented and analyzed.

  6. Calculation of Credit Valuation Adjustment Based on Least Square Monte Carlo Methods

    Directory of Open Access Journals (Sweden)

    Qian Liu

    2015-01-01

    Full Text Available Counterparty credit risk has become one of the highest-profile risks facing participants in the financial markets. Despite this, relatively little is known about how counterparty credit risk is actually priced mathematically. We examine this issue using interest rate swaps. This largely traded financial product allows us to well identify the risk profiles of both institutions and their counterparties. Concretely, Hull-White model for rate and mean-reverting model for default intensity have proven to be in correspondence with the reality and to be well suited for financial institutions. Besides, we find that least square Monte Carlo method is quite efficient in the calculation of credit valuation adjustment (CVA, for short as it avoids the redundant step to generate inner scenarios. As a result, it accelerates the convergence speed of the CVA estimators. In the second part, we propose a new method to calculate bilateral CVA to avoid double counting in the existing bibliographies, where several copula functions are adopted to describe the dependence of two first to default times.

  7. A molecular dynamics study of intramolecular proton transfer reaction of malonaldehyde in solutions based upon mixed quantum-classical approximation. I. Proton transfer reaction in water.

    Science.gov (United States)

    Yamada, Atsushi; Kojima, Hidekazu; Okazaki, Susumu

    2014-08-28

    In order to investigate proton transfer reaction in solution, mixed quantum-classical molecular dynamics calculations have been carried out based on our previously proposed quantum equation of motion for the reacting system [A. Yamada and S. Okazaki, J. Chem. Phys. 128, 044507 (2008)]. Surface hopping method was applied to describe forces acting on the solvent classical degrees of freedom. In a series of our studies, quantum and solvent effects on the reaction dynamics in solutions have been analysed in detail. Here, we report our mixed quantum-classical molecular dynamics calculations for intramolecular proton transfer of malonaldehyde in water. Thermally activated proton transfer process, i.e., vibrational excitation in the reactant state followed by transition to the product state and vibrational relaxation in the product state, as well as tunneling reaction can be described by solving the equation of motion. Zero point energy is, of course, included, too. The quantum simulation in water has been compared with the fully classical one and the wave packet calculation in vacuum. The calculated quantum reaction rate in water was 0.70 ps(-1), which is about 2.5 times faster than that in vacuum, 0.27 ps(-1). This indicates that the solvent water accelerates the reaction. Further, the quantum calculation resulted in the reaction rate about 2 times faster than the fully classical calculation, which indicates that quantum effect enhances the reaction rate, too. Contribution from three reaction mechanisms, i.e., tunneling, thermal activation, and barrier vanishing reactions, is 33:46:21 in the mixed quantum-classical calculations. This clearly shows that the tunneling effect is important in the reaction.

  8. Study of an intense proton beam profiler based on laser absorption; Etude d'un profileur optique de faisceaux intenses de protons par absorption laser

    Energy Technology Data Exchange (ETDEWEB)

    Pottin, B

    2001-10-01

    Among the challenges of high current proton accelerators, the development of new beam diagnostics is of major importance. The main difficulty for these instruments, is the beam power which deteriorates any instruments used to catch it. The chosen detectors are therefore 'non-interceptive systems. After an introduction concerning characteristics of the used accelerator (chapter I), parameters defining a beam of particles are presented (chapter II). Among these ones, the profile is an important beam characteristic for its transport. After the description of the different types of beam profilers, their problematic application to intense beams is discussed. New physical phenomena have to be used for profilers. Thus, we have prospected optical luminescence phenomena. The light produced during the interaction of protons with the residual gas and/or locally injected is a source of informations on beam characteristics. In chapters III and IV, there is an experimental and theoretical analysis of the luminescence. Chapter V is a direct application of spectroscopic measurements to estimate the output of protons with a non-interceptive technique. With the spectral analysis, the idea of a profiler based on laser absorption is developed. This presentation is both theoretical and experimental (chapters 6 and 7). The laser absorption needs the use of metastable states we define in the chapter 6. The evolution of the metastable states, with time and space, has been rigorously studied to discuss the concept of an optical profiler. Chapter VII presents all the necessary instrumentation for the use of a laser and the first measurements with the beam. At the thesis end, the first recorded profile is presented. An experimental critic is presented with a description of the different sources of errors and the proposed cures. (author)

  9. Influence of Geant4 parameters on proton dose distribution

    Directory of Open Access Journals (Sweden)

    Asad Merouani

    2015-09-01

    Full Text Available Purpose: The proton therapy presents a great precision during the radiation dose delivery. It is useful when the tumor is located in a sensitive area like brain or eyes. The Monte Carlo (MC simulations are usually used in treatment planning system (TPS to estimate the radiation dose. In this paper we are interested in estimating the proton dose statistical uncertainty generated by the MC simulations. Methods: Geant4 was used in the simulation of the eye’s treatment room for 62 MeV protons therapy, installed in the Istituto Nazionale Fisica Nucleare Laboratori Nazionali del Sud (LNS-INFN facility in Catania. This code is a Monte Carlo based on software dedicated to simulate the passage of particles through the matter. In this work, we are interested in optimizing the Geant4 parameters on energy deposit distribution by proton to achieve the spatial resolution of dose distribution required for cancer therapy. We propose various simulations and compare the corresponding dose distribution inside water to evaluate the statistical uncertainties. Results: The simulated Bragg peak, based on facility model is in agreement with the experimental data, The calculations show that the mean statistical uncertainty is less than 1% for a simulation set with 5 × 104 events, 10-3 mm production threshold and a 10-2 mm step limit. Conclusion: The set of Geant4 cut and step limit values can be chosen in combination with the number of events to reach precision recommended from International Commission on Radiation Units and measurements (ICRU in Monte Carlo codes for proton therapy treatment.

  10. Catalyst Degradation in High Temperature Proton Exchange Membrane Fuel Cells Based on Acid Doped Polybenzimidazole Membranes

    DEFF Research Database (Denmark)

    Cleemann, Lars Nilausen; Buazar, F.; Li, Qingfeng

    2013-01-01

    confirmed by the post TEM and XRD analysis. A strong dependence of the fuel cell performance degradation on the catalyst supports was observed. Graphitization of the carbon blacks improved the stability and catalyst durability though at the expense of a significant decrease in the specific surface area......Degradation of carbon supported platinum catalysts is a major failure mode for the long term durability of high temperature proton exchange membrane fuel cells based on phosphoric acid doped polybenzimidazole membranes. With Vulcan carbon black as a reference, thermally treated carbon black...... and multi‐walled carbon nanotubes were used as supports for electrode catalysts and evaluated in accelerated durability tests under potential cycling at 150 °C. Measurements of open circuit voltage, area specific resistance and hydrogen permeation through the membrane were carried out, indicating little...

  11. Fatigue Analysis of Proton Exchange Membrane Fuel Cell Stacks Based on Structural Stress Distribution

    Science.gov (United States)

    Wu, C. W.; Liu, B.; Wei, M. Y.; Liu, L. F.

    2017-05-01

    Proton exchange membrane fuel cell (PEMFC) stack usually undergoes various vibrations during packing, transportation and serving time, in particular for those used in the automobiles and portable equipment. Based on the Miner fatigue damage theory, the fatigue lives of the fuel cell components are first assessed. Then the component fatigue life contours of the stack are obtained under four working conditions, i.e. the three single-axial (in X-, Y- and Z-axis separately) and multi-axial random vibrations. Accordingly, the component damage under various vibrations is evaluated. The stress distribution on the gasket and PEM will greatly affect their fatigue lives. Finally, we compare the fatigue lives of 4-bolt- and 6-bolt-clamping stacks under the same total clamping force, and find that increasing the bolt number could improve the bolt fatigue lives.

  12. Geant4 simulations of proton beam transport through a carbon or beryllium degrader and following a beam line

    NARCIS (Netherlands)

    van Goethem, M. J.; van der Meer, R.; Reist, H. W.; Schippers, J. M.

    2009-01-01

    Monte Carlo simulations based on the Geant4 simulation toolkit were performed for the carbon wedge degrader used in the beam line at the Center of Proton Therapy of the Paul Scherrer Institute (PSI). The simulations are part of the beam line studies for the development and understanding of the

  13. Two schemes for quantitative photoacoustic tomography based on Monte Carlo simulation

    International Nuclear Information System (INIS)

    Liu, Yubin; Yuan, Zhen; Jiang, Huabei

    2016-01-01

    Purpose: The aim of this study was to develop novel methods for photoacoustically determining the optical absorption coefficient of biological tissues using Monte Carlo (MC) simulation. Methods: In this study, the authors propose two quantitative photoacoustic tomography (PAT) methods for mapping the optical absorption coefficient. The reconstruction methods combine conventional PAT with MC simulation in a novel way to determine the optical absorption coefficient of biological tissues or organs. Specifically, the authors’ two schemes were theoretically and experimentally examined using simulations, tissue-mimicking phantoms, ex vivo, and in vivo tests. In particular, the authors explored these methods using several objects with different absorption contrasts embedded in turbid media and by using high-absorption media when the diffusion approximation was not effective at describing the photon transport. Results: The simulations and experimental tests showed that the reconstructions were quantitatively accurate in terms of the locations, sizes, and optical properties of the targets. The positions of the recovered targets were accessed by the property profiles, where the authors discovered that the off center error was less than 0.1 mm for the circular target. Meanwhile, the sizes and quantitative optical properties of the targets were quantified by estimating the full width half maximum of the optical absorption property. Interestingly, for the reconstructed sizes, the authors discovered that the errors ranged from 0 for relatively small-size targets to 26% for relatively large-size targets whereas for the recovered optical properties, the errors ranged from 0% to 12.5% for different cases. Conclusions: The authors found that their methods can quantitatively reconstruct absorbing objects of different sizes and optical contrasts even when the diffusion approximation is unable to accurately describe the photon propagation in biological tissues. In particular, their

  14. GPU-BASED MONTE CARLO DUST RADIATIVE TRANSFER SCHEME APPLIED TO ACTIVE GALACTIC NUCLEI

    International Nuclear Information System (INIS)

    Heymann, Frank; Siebenmorgen, Ralf

    2012-01-01

    A three-dimensional parallel Monte Carlo (MC) dust radiative transfer code is presented. To overcome the huge computing-time requirements of MC treatments, the computational power of vectorized hardware is used, utilizing either multi-core computer power or graphics processing units. The approach is a self-consistent way to solve the radiative transfer equation in arbitrary dust configurations. The code calculates the equilibrium temperatures of two populations of large grains and stochastic heated polycyclic aromatic hydrocarbons. Anisotropic scattering is treated applying the Heney-Greenstein phase function. The spectral energy distribution (SED) of the object is derived at low spatial resolution by a photon counting procedure and at high spatial resolution by a vectorized ray tracer. The latter allows computation of high signal-to-noise images of the objects at any frequencies and arbitrary viewing angles. We test the robustness of our approach against other radiative transfer codes. The SED and dust temperatures of one- and two-dimensional benchmarks are reproduced at high precision. The parallelization capability of various MC algorithms is analyzed and included in our treatment. We utilize the Lucy algorithm for the optical thin case where the Poisson noise is high, the iteration-free Bjorkman and Wood method to reduce the calculation time, and the Fleck and Canfield diffusion approximation for extreme optical thick cells. The code is applied to model the appearance of active galactic nuclei (AGNs) at optical and infrared wavelengths. The AGN torus is clumpy and includes fluffy composite grains of various sizes made up of silicates and carbon. The dependence of the SED on the number of clumps in the torus and the viewing angle is studied. The appearance of the 10 μm silicate features in absorption or emission is discussed. The SED of the radio-loud quasar 3C 249.1 is fit by the AGN model and a cirrus component to account for the far-infrared emission.

  15. Introducing ab initio based neural networks for transition-rate prediction in kinetic Monte Carlo simulations

    Science.gov (United States)

    Messina, Luca; Castin, Nicolas; Domain, Christophe; Olsson, Pär

    2017-02-01

    The quality of kinetic Monte Carlo (KMC) simulations of microstructure evolution in alloys relies on the parametrization of point-defect migration rates, which are complex functions of the local chemical composition and can be calculated accurately with ab initio methods. However, constructing reliable models that ensure the best possible transfer of physical information from ab initio to KMC is a challenging task. This work presents an innovative approach, where the transition rates are predicted by artificial neural networks trained on a database of 2000 migration barriers, obtained with density functional theory (DFT) in place of interatomic potentials. The method is tested on copper precipitation in thermally aged iron alloys, by means of a hybrid atomistic-object KMC model. For the object part of the model, the stability and mobility properties of copper-vacancy clusters are analyzed by means of independent atomistic KMC simulations, driven by the same neural networks. The cluster diffusion coefficients and mean free paths are found to increase with size, confirming the dominant role of coarsening of medium- and large-sized clusters in the precipitation kinetics. The evolution under thermal aging is in better agreement with experiments with respect to a previous interatomic-potential model, especially concerning the experiment time scales. However, the model underestimates the solubility of copper in iron due to the excessively high solution energy predicted by the chosen DFT method. Nevertheless, this work proves the capability of neural networks to transfer complex ab initio physical properties to higher-scale models, and facilitates the extension to systems with increasing chemical complexity, setting the ground for reliable microstructure evolution simulations in a wide range of alloys and applications.

  16. Peak Skin and Eye Lens Radiation Dose From Brain Perfusion CT Based on Monte Carlo Simulation

    Science.gov (United States)

    Zhang, Di; Cagnon, Chris H.; Pablo Villablanca, J.; McCollough, Cynthia H.; Cody, Dianna D.; Stevens, Donna M.; Zankl, Maria; Demarco, John J.; Turner, Adam C.; Khatonabadi, Maryam; McNitt-Gray, Michael F.

    2014-01-01

    OBJECTIVE. The purpose of our study was to accurately estimate the radiation dose to skin and the eye lens from clinical CT brain perfusion studies, investigate how well scanner output (expressed as volume CT dose index [CTDIvol]) matches these estimated doses, and investigate the efficacy of eye lens dose reduction techniques. MATERIALS AND METHODS. Peak skin dose and eye lens dose were estimated using Monte Carlo simulation methods on a voxelized patient model and 64-MDCT scanners from four major manufacturers. A range of clinical protocols was evaluated. CTDIvol for each scanner was obtained from the scanner console. Dose reduction to the eye lens was evaluated for various gantry tilt angles as well as scan locations. RESULTS. Peak skin dose and eye lens dose ranged from 81 mGy to 348 mGy, depending on the scanner and protocol used. Peak skin dose and eye lens dose were observed to be 66–79% and 59–63%, respectively, of the CTDIvol values reported by the scanners. The eye lens dose was significantly reduced when the eye lenses were not directly irradiated. CONCLUSION. CTDIvol should not be interpreted as patient dose; this study has shown it to overestimate dose to the skin or eye lens. These results may be used to provide more accurate estimates of actual dose to ensure that protocols are operated safely below thresholds. Tilting the gantry or moving the scanning region further away from the eyes are effective for reducing lens dose in clinical practice. These actions should be considered when they are consistent with the clinical task and patient anatomy. PMID:22268186

  17. Peak skin and eye lens radiation dose from brain perfusion CT based on Monte Carlo simulation.

    Science.gov (United States)

    Zhang, Di; Cagnon, Chris H; Villablanca, J Pablo; McCollough, Cynthia H; Cody, Dianna D; Stevens, Donna M; Zankl, Maria; Demarco, John J; Turner, Adam C; Khatonabadi, Maryam; McNitt-Gray, Michael F

    2012-02-01

    The purpose of our study was to accurately estimate the radiation dose to skin and the eye lens from clinical CT brain perfusion studies, investigate how well scanner output (expressed as volume CT dose index [CTDI(vol)]) matches these estimated doses, and investigate the efficacy of eye lens dose reduction techniques. Peak skin dose and eye lens dose were estimated using Monte Carlo simulation methods on a voxelized patient model and 64-MDCT scanners from four major manufacturers. A range of clinical protocols was evaluated. CTDI(vol) for each scanner was obtained from the scanner console. Dose reduction to the eye lens was evaluated for various gantry tilt angles as well as scan locations. Peak skin dose and eye lens dose ranged from 81 mGy to 348 mGy, depending on the scanner and protocol used. Peak skin dose and eye lens dose were observed to be 66-79% and 59-63%, respectively, of the CTDI(vol) values reported by the scanners. The eye lens dose was significantly reduced when the eye lenses were not directly irradiated. CTDI(vol) should not be interpreted as patient dose; this study has shown it to overestimate dose to the skin or eye lens. These results may be used to provide more accurate estimates of actual dose to ensure that protocols are operated safely below thresholds. Tilting the gantry or moving the scanning region further away from the eyes are effective for reducing lens dose in clinical practice. These actions should be considered when they are consistent with the clinical task and patient anatomy.

  18. Proton-air and proton-proton cross sections

    Directory of Open Access Journals (Sweden)

    Ulrich Ralf

    2013-06-01

    Full Text Available Different attempts to measure hadronic cross sections with cosmic ray data are reviewed. The major results are compared to each other and the differences in the corresponding analyses are discussed. Besides some important differences, it is crucial to see that all analyses are based on the same fundamental relation of longitudinal air shower development to the observed fluctuation of experimental observables. Furthermore, the relation of the measured proton-air to the more fundamental proton-proton cross section is discussed. The current global picture combines hadronic proton-proton cross section data from accelerator and cosmic ray measurements and indicates a good consistency with predictions of models up to the highest energies.

  19. Dosimetry for ocular proton beam therapy at the Harvard Cyclotron Laboratory based on the ICRU Report 59

    International Nuclear Information System (INIS)

    Newhauser, W.D.; Burns, J.; Smith, A.R.

    2002-01-01

    The Massachusetts General Hospital, the Harvard Cyclotron Laboratory (HCL), and the Massachusetts Eye and Ear Infirmary have treated almost 3000 patients with ocular disease using high-energy external-beam proton radiation therapy since 1975. The absorbed dose standard for ocular proton therapy beams at HCL was based on a fluence measurement with a Faraday cup (FC). A majority of proton therapy centers worldwide, however, use an absorbed dose standard that is based on an ionization chamber (IC) technique. The ion chamber calibration is deduced from a measurement in a reference 60 Co photon field together with a calculated correction factor that takes into account differences in a chamber's response in 60 Co and proton fields. In this work, we implemented an ionization chamber-based absolute dosimetry system for the HCL ocular beamline based on the recommendations given in Report 59 by the International Commission on Radiation Units and Measurements. Comparative measurements revealed that the FC system yields an absorbed dose to water value that is 1.1% higher than was obtained with the IC system. That difference is small compared with the experimental uncertainties and is clinically insignificant. In June of 1998, we adopted the IC-based method as our standard practice for the ocular beam

  20. Design Choices of the MedAustron Nozzles and Proton Gantry based on Modeling of Particle Scattering

    CERN Document Server

    Palm, M; Benedikt, M

    2011-01-01

    MedAustron, the Austrian hadron therapy center is currently under construction. Irradiations will be performed using active scanning with a proton or carbon ion pencil beam which is subject to scattering in vacuum windows, beam monitors and air gap. For applications where sharp lateral beam penumbras are required in order to spare critical organs from unwanted dose, scattering should be minimal. A semi-empirical scattering model has been established to evaluate beam size growth at the patient due to upstream scattering. Major design choices for proton gantry and nozzle based on the scattering calculations are presented.

  1. Proton-carbon elastic scattering in the intermediate energy range based on the. alpha. -particle model

    Energy Technology Data Exchange (ETDEWEB)

    Li Qingrun (CCAST (World Lab.), Beijing (China) Academia Sinica, Beijing, BJ (China). Inst. of High Energy Physics); Zhou Jinli (Guangxi Normal Univ., Guilin (China). Dept. of Physics)

    1991-05-01

    The {alpha}-particle model of {sup 12}C is examined by means of proton-{sup 12}C elastic scattering in the intermediate energy range. The results show that the model gives a satisfactory account of the experimental data. The parametrized proton-{sup 4}He amplitudes in the intermediate energy region are presented. (author).

  2. Proton-detected MAS NMR experiments based on dipolar transfers for backbone assignment of highly deuterated proteins

    Science.gov (United States)

    Chevelkov, Veniamin; Habenstein, Birgit; Loquet, Antoine; Giller, Karin; Becker, Stefan; Lange, Adam

    2014-05-01

    Proton-detected solid-state NMR was applied to a highly deuterated insoluble, non-crystalline biological assembly, the Salmonella typhimurium type iii secretion system (T3SS) needle. Spectra of very high resolution and sensitivity were obtained at a low protonation level of 10-20% at exchangeable amide positions. We developed efficient experimental protocols for resonance assignment tailored for this system and the employed experimental conditions. Using exclusively dipolar-based interspin magnetization transfers, we recorded two sets of 3D spectra allowing for an almost complete backbone resonance assignment of the needle subunit PrgI. The additional information provided by the well-resolved proton dimension revealed the presence of two sets of resonances in the N-terminal helix of PrgI, while in previous studies employing 13C detection only a single set of resonances was observed.

  3. Proton Radiation Therapy in the Hospital Environment: Conception, Development, and Operation of the Initial Hospital-Based Facility

    Science.gov (United States)

    Slater, James M.; Slater, Jerry D.; Wroe, Andrew J.

    The world's first hospital-based proton treatment center opened at Loma Linda University Medical Center in 1990, following two decades of development. Patients' needs were the driving force behind its conception, development, and execution; the primary needs were delivery of effective conformal doses of ionizing radiation and avoidance of normal tissue to the maximum extent possible. The facility includes a proton synchrotron and delivery system developed in collaboration with physicists and engineers at Fermi National Accelerator Laboratory and from other high-energy-physics laboratories worldwide. The system, operated and maintained by Loma Linda personnel, was designed to be safe, reliable, flexible in utilization, efficient in use, and upgradeable to meet demands of changing patient needs and advances in technology. Since the facility opened, nearly 14,000 adults and children have been treated for a wide range of cancers and other diseases. Ongoing research is expanding the applications of proton therapy, while reducing costs.

  4. Monte Carlo-based diode design for correction-less small field dosimetry.

    Science.gov (United States)

    Charles, P H; Crowe, S B; Kairn, T; Knight, R T; Hill, B; Kenny, J; Langton, C M; Trapp, J V

    2013-07-07

    Due to their small collecting volume, diodes are commonly used in small field dosimetry. However, the relative sensitivity of a diode increases with decreasing small field size. Conversely, small air gaps have been shown to cause a significant decrease in the sensitivity of a detector as the field size is decreased. Therefore, this study uses Monte Carlo simulations to look at introducing air upstream to diodes such that they measure with a constant sensitivity across all field sizes in small field dosimetry. Varying thicknesses of air were introduced onto the upstream end of two commercial diodes (PTW 60016 photon diode and PTW 60017 electron diode), as well as a theoretical unenclosed silicon chip using field sizes as small as 5 mm × 5 mm. The metric D(w,Q)/D(Det,Q) used in this study represents the ratio of the dose to a point of water to the dose to the diode active volume, for a particular field size and location. The optimal thickness of air required to provide a constant sensitivity across all small field sizes was found by plotting D(w,Q)/D(Det,Q) as a function of introduced air gap size for various field sizes, and finding the intersection point of these plots. That is, the point at which D(w,Q)/D(Det,Q) was constant for all field sizes was found. The optimal thickness of air was calculated to be 3.3, 1.15 and 0.10 mm for the photon diode, electron diode and unenclosed silicon chip, respectively. The variation in these results was due to the different design of each detector. When calculated with the new diode design incorporating the upstream air gap, k(f(clin),f(msr))(Q(clin),Q(msr)) was equal to unity to within statistical uncertainty (0.5%) for all three diodes. Cross-axis profile measurements were also improved with the new detector design. The upstream air gap could be implanted on the commercial diodes via a cap consisting of the air cavity surrounded by water equivalent material. The results for the unclosed silicon chip show that an ideal small

  5. Monte Carlo-based development of a shield and total background estimation for the COBRA experiment

    International Nuclear Information System (INIS)

    Heidrich, Nadine

    2014-11-01

    The COBRA experiment aims for the measurement of the neutrinoless double beta decay and thus for the determination the effective Majorana mass of the neutrino. To be competitive with other next-generation experiments the background rate has to be in the order of 10 -3 counts/kg/keV/yr, which is a challenging criterion. This thesis deals with the development of a shield design and the calculation of the expected total background rate for the large scale COBRA experiment containing 13824 6 cm 3 CdZnTe detectors. For the development of a shield single-layer and multi-layer shields were investigated and a shield design was optimized concerning high-energy muon-induced neutrons. As the best design the combination of 10 cm boron doped polyethylene as outermost layer, 20 cm lead and 10 cm copper as innermost layer were determined. It showed the best performance regarding neutron attenuation as well as (n, γ) self-shielding effects leading to a negligible background rate of less than 2.10 -6 counts/kg/keV/yr. Additionally. the shield with a thickness of 40 cm is compact and costeffective. In the next step the expected total background rate was computed taking into account individual setup parts and various background sources including natural and man-made radioactivity, cosmic ray-induced background and thermal neutrons. Furthermore, a comparison of measured data from the COBRA demonstrator setup with Monte Carlo data was used to calculate reliable contamination levels of the single setup parts. The calculation was performed conservatively to prevent an underestimation. In addition, the contribution to the total background rate regarding the individual detector parts and background sources was investigated. The main portion arise from the Delrin support structure, the Glyptal lacquer followed by the circuit board of the high voltage supply. Most background events originate from particles with a quantity of 99 % in total. Regarding surface events a contribution of 26

  6. A method to generate equivalent energy spectra and filtration models based on measurement for multidetector CT Monte Carlo dosimetry simulations

    International Nuclear Information System (INIS)

    Turner, Adam C.; Zhang Di; Kim, Hyun J.; DeMarco, John J.; Cagnon, Chris H.; Angel, Erin; Cody, Dianna D.; Stevens, Donna M.; Primak, Andrew N.; McCollough, Cynthia H.; McNitt-Gray, Michael F.

    2009-01-01

    The purpose of this study was to present a method for generating x-ray source models for performing Monte Carlo (MC) radiation dosimetry simulations of multidetector row CT (MDCT) scanners. These so-called ''equivalent'' source models consist of an energy spectrum and filtration description that are generated based wholly on the measured values and can be used in place of proprietary manufacturer's data for scanner-specific MDCT MC simulations. Required measurements include the half value layers (HVL 1 and HVL 2 ) and the bowtie profile (exposure values across the fan beam) for the MDCT scanner of interest. Using these measured values, a method was described (a) to numerically construct a spectrum with the calculated HVLs approximately equal to those measured (equivalent spectrum) and then (b) to determine a filtration scheme (equivalent filter) that attenuates the equivalent spectrum in a similar fashion as the actual filtration attenuates the actual x-ray beam, as measured by the bowtie profile measurements. Using this method, two types of equivalent source models were generated: One using a spectrum based on both HVL 1 and HVL 2 measurements and its corresponding filtration scheme and the second consisting of a spectrum based only on the measured HVL 1 and its corresponding filtration scheme. Finally, a third type of source model was built based on the spectrum and filtration data provided by the scanner's manufacturer. MC simulations using each of these three source model types were evaluated by comparing the accuracy of multiple CT dose index (CTDI) simulations to measured CTDI values for 64-slice scanners from the four major MDCT manufacturers. Comprehensive evaluations were carried out for each scanner using each kVp and bowtie filter combination available. CTDI experiments were performed for both head (16 cm in diameter) and body (32 cm in diameter) CTDI phantoms using both central and peripheral measurement positions. Both equivalent source model types

  7. A method to generate equivalent energy spectra and filtration models based on measurement for multidetector CT Monte Carlo dosimetry simulations

    Science.gov (United States)

    Turner, Adam C.; Zhang, Di; Kim, Hyun J.; DeMarco, John J.; Cagnon, Chris H.; Angel, Erin; Cody, Dianna D.; Stevens, Donna M.; Primak, Andrew N.; McCollough, Cynthia H.; McNitt-Gray, Michael F.

    2009-01-01

    The purpose of this study was to present a method for generating x-ray source models for performing Monte Carlo (MC) radiation dosimetry simulations of multidetector row CT (MDCT) scanners. These so-called “equivalent” source models consist of an energy spectrum and filtration description that are generated based wholly on the measured values and can be used in place of proprietary manufacturer’s data for scanner-specific MDCT MC simulations. Required measurements include the half value layers (HVL1 and HVL2) and the bowtie profile (exposure values across the fan beam) for the MDCT scanner of interest. Using these measured values, a method was described (a) to numerically construct a spectrum with the calculated HVLs approximately equal to those measured (equivalent spectrum) and then (b) to determine a filtration scheme (equivalent filter) that attenuates the equivalent spectrum in a similar fashion as the actual filtration attenuates the actual x-ray beam, as measured by the bowtie profile measurements. Using this method, two types of equivalent source models were generated: One using a spectrum based on both HVL1 and HVL2 measurements and its corresponding filtration scheme and the second consisting of a spectrum based only on the measured HVL1 and its corresponding filtration scheme. Finally, a third type of source model was built based on the spectrum and filtration data provided by the scanner’s manufacturer. MC simulations using each of these three source model types were evaluated by comparing the accuracy of multiple CT dose index (CTDI) simulations to measured CTDI values for 64-slice scanners from the four major MDCT manufacturers. Comprehensive evaluations were carried out for each scanner using each kVp and bowtie filter combination available. CTDI experiments were performed for both head (16 cm in diameter) and body (32 cm in diameter) CTDI phantoms using both central and peripheral measurement positions. Both equivalent source model types

  8. A method to generate equivalent energy spectra and filtration models based on measurement for multidetector CT Monte Carlo dosimetry simulations.

    Science.gov (United States)

    Turner, Adam C; Zhang, Di; Kim, Hyun J; DeMarco, John J; Cagnon, Chris H; Angel, Erin; Cody, Dianna D; Stevens, Donna M; Primak, Andrew N; McCollough, Cynthia H; McNitt-Gray, Michael F

    2009-06-01

    The purpose of this study was to present a method for generating x-ray source models for performing Monte Carlo (MC) radiation dosimetry simulations of multidetector row CT (MDCT) scanners. These so-called "equivalent" source models consist of an energy spectrum and filtration description that are generated based wholly on the measured values and can be used in place of proprietary manufacturer's data for scanner-specific MDCT MC simulations. Required measurements include the half value layers (HVL1 and HVL2) and the bowtie profile (exposure values across the fan beam) for the MDCT scanner of interest. Using these measured values, a method was described (a) to numerically construct a spectrum with the calculated HVLs approximately equal to those measured (equivalent spectrum) and then (b) to determine a filtration scheme (equivalent filter) that attenuates the equivalent spectrum in a similar fashion as the actual filtration attenuates the actual x-ray beam, as measured by the bowtie profile measurements. Using this method, two types of equivalent source models were generated: One using a spectrum based on both HVL1 and HVL2 measurements and its corresponding filtration scheme and the second consisting of a spectrum based only on the measured HVL1 and its corresponding filtration scheme. Finally, a third type of source model was built based on the spectrum and filtration data provided by the scanner's manufacturer. MC simulations using each of these three source model types were evaluated by comparing the accuracy of multiple CT dose index (CTDI) simulations to measured CTDI values for 64-slice scanners from the four major MDCT manufacturers. Comprehensive evaluations were carried out for each scanner using each kVp and bowtie filter combination available. CTDI experiments were performed for both head (16 cm in diameter) and body (32 cm in diameter) CTDI phantoms using both central and peripheral measurement positions. Both equivalent source model types result in

  9. Monte Carlo simulation of explosive detection system based on a Deuterium-Deuterium (D-D) neutron generator.

    Science.gov (United States)

    Bergaoui, K; Reguigui, N; Gary, C K; Brown, C; Cremer, J T; Vainionpaa, J H; Piestrup, M A

    2014-12-01

    An explosive detection system based on a Deuterium-Deuterium (D-D) neutron generator has been simulated using the Monte Carlo N-Particle Transport Code (MCNP5). Nuclear-based explosive detection methods can detect explosives by identifying their elemental components, especially nitrogen. Thermal neutron capture reactions have been used for detecting prompt gamma emission (10.82MeV) following radiative neutron capture by (14)N nuclei. The explosive detection system was built based on a fully high-voltage-shielded, axial D-D neutron generator with a radio frequency (RF) driven ion source and nominal yield of about 10(10) fast neutrons per second (E=2.5MeV). Polyethylene and paraffin were used as moderators with borated polyethylene and lead as neutron and gamma ray shielding, respectively. The shape and the thickness of the moderators and shields are optimized to produce the highest thermal neutron flux at the position of the explosive and the minimum total dose at the outer surfaces of the explosive detection system walls. In addition, simulation of the response functions of NaI, BGO, and LaBr3-based γ-ray detectors to different explosives is described. Copyright © 2014 Elsevier Ltd. All rights reserved.

  10. Girsanov's transformation based variance reduced Monte Carlo simulation schemes for reliability estimation in nonlinear stochastic dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Kanjilal, Oindrila, E-mail: oindrila@civil.iisc.ernet.in; Manohar, C.S., E-mail: manohar@civil.iisc.ernet.in

    2017-07-15

    The study considers the problem of simulation based time variant reliability analysis of nonlinear randomly excited dynamical systems. Attention is focused on importance sampling strategies based on the application of Girsanov's transformation method. Controls which minimize the distance function, as in the first order reliability method (FORM), are shown to minimize a bound on the sampling variance of the estimator for the probability of failure. Two schemes based on the application of calculus of variations for selecting control signals are proposed: the first obtains the control force as the solution of a two-point nonlinear boundary value problem, and, the second explores the application of the Volterra series in characterizing the controls. The relative merits of these schemes, vis-à-vis the method based on ideas from the FORM, are discussed. Illustrative examples, involving archetypal single degree of freedom (dof) nonlinear oscillators, and a multi-degree of freedom nonlinear dynamical system, are presented. The credentials of the proposed procedures are established by comparing the solutions with pertinent results from direct Monte Carlo simulations. - Highlights: • The distance minimizing control forces minimize a bound on the sampling variance. • Establishing Girsanov controls via solution of a two-point boundary value problem. • Girsanov controls via Volterra's series representation for the transfer functions.

  11. Accurate classification of brain gliomas by discriminate dictionary learning based on projective dictionary pair learning of proton magnetic resonance spectra.

    Science.gov (United States)

    Adebileje, Sikiru Afolabi; Ghasemi, Keyvan; Aiyelabegan, Hammed Tanimowo; Saligheh Rad, Hamidreza

    2017-04-01

    Proton magnetic resonance spectroscopy is a powerful noninvasive technique that complements the structural images of cMRI, which aids biomedical and clinical researches, by identifying and visualizing the compositions of various metabolites within the tissues of interest. However, accurate classification of proton magnetic resonance spectroscopy is still a challenging issue in clinics due to low signal-to-noise ratio, overlapping peaks of metabolites, and the presence of background macromolecules. This paper evaluates the performance of a discriminate dictionary learning classifiers based on projective dictionary pair learning method for brain gliomas proton magnetic resonance spectroscopy spectra classification task, and the result were compared with the sub-dictionary learning methods. The proton magnetic resonance spectroscopy data contain a total of 150 spectra (74 healthy, 23 grade II, 23 grade III, and 30 grade IV) from two databases. The datasets from both databases were first coupled together, followed by column normalization. The Kennard-Stone algorithm was used to split the datasets into its training and test sets. Performance comparison based on the overall accuracy, sensitivity, specificity, and precision was conducted. Based on the overall accuracy of our classification scheme, the dictionary pair learning method was found to outperform the sub-dictionary learning methods 97.78% compared with 68.89%, respectively. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.

  12. Local structure and hydrogen bond characteristics of imidazole molecules for proton conduction in acid and base proton-conducting composite materials.

    Science.gov (United States)

    Hori, Yuta; Chikai, Takuma; Ida, Tomonori; Mizuno, Motohiro

    2018-04-18

    Composite materials of acidic polymers and basic molecules have high proton-conductivity. Understanding the proton conduction mechanism of the composite materials, which depends on hydrogen bond characteristics, is an important task for developing materials with high proton-conductivity. This work is focused on poly(vinylphosphonic acid)-imidazole and alginic acid-imidazole as examples of composite materials of acidic polymers and basic molecules and examines the local structure and hydrogen bond characteristics of imidazole (Im) molecules in composite materials using density functional theory. The results show that Im molecules interact strongly with polymeric acids in these composite materials and that the interaction energy increases with the increase in the number of Im molecules. The rotational motion of Im molecules occurs in the segment where only Im molecules without excess protons are hydrogen-bonded to each other. The calculation results for the various segments, which depend on the hydrogen bonding environment, show that the proton conduction process in composite materials consists of the following steps: proton transfer in the segment where Im molecules interact with polymeric acids, proton transfer in the segment where Im molecules are affected by excess protons, and Grotthuss diffusion with reorientation of Im molecules in the segment where only Im molecules without excess protons are bonded to each other.

  13. Proton Linear Energy Transfer measurement using Emulsion Cloud Chamber

    Science.gov (United States)

    Shin, Jae-ik; Park, Seyjoon; Kim, Haksoo; Kim, Meyoung; Jeong, Chiyoung; Cho, Sungkoo; Lim, Young Kyung; Shin, Dongho; Lee, Se Byeong; Morishima, Kunihiro; Naganawa, Naotaka; Sato, Osamu; Kwak, Jungwon; Kim, Sung Hyun; Cho, Jung Sook; Ahn, Jung Keun; Kim, Ji Hyun; Yoon, Chun Sil; Incerti, Sebastien

    2015-04-01

    This study proposes to determine the correlation between the Volume Pulse Height (VPH) measured by nuclear emulsion and Linear Energy Transfer (LET) calculated by Monte Carlo simulation based on Geant4. The nuclear emulsion was irradiated at the National Cancer Center (NCC) with a therapeutic proton beam and was installed at 5.2 m distance from the beam nozzle structure with various thicknesses of water-equivalent material (PMMA) blocks to position with specific positions along the Bragg curve. After the beam exposure and development of the emulsion films, the films were scanned by S-UTS developed in Nagoya University. The proton tracks in the scanned films were reconstructed using the 'NETSCAN' method. Through this procedure, the VPH can be derived from each reconstructed proton track at each position along the Bragg curve. The VPH value indicates the magnitude of energy loss in proton track. By comparison with the simulation results obtained using Geant4, we found the correlation between the LET calculated by Monte Carlo simulation and the VPH measured by the nuclear emulsion.

  14. Proton Linear Energy Transfer measurement using Emulsion Cloud Chamber

    International Nuclear Information System (INIS)

    Shin, Jae-ik; Park, Seyjoon; Kim, Haksoo; Kim, Meyoung; Jeong, Chiyoung; Cho, Sungkoo; Lim, Young Kyung; Shin, Dongho; Lee, Se Byeong; Morishima, Kunihiro; Naganawa, Naotaka; Sato, Osamu; Kwak, Jungwon; Kim, Sung Hyun; Cho, Jung Sook; Ahn, Jung Keun; Kim, Ji Hyun; Yoon, Chun Sil; Incerti, Sebastien

    2015-01-01

    This study proposes to determine the correlation between the Volume Pulse Height (VPH) measured by nuclear emulsion and Linear Energy Transfer (LET) calculated by Monte Carlo simulation based on Geant4. The nuclear emulsion was irradiated at the National Cancer Center (NCC) with a therapeutic proton beam and was installed at 5.2 m distance from the beam nozzle structure with various thicknesses of water-equivalent material (PMMA) blocks to position with specific positions along the Bragg curve. After the beam exposure and development of the emulsion films, the films were scanned by S-UTS developed in Nagoya University. The proton tracks in the scanned films were reconstructed using the ‘NETSCAN’ method. Through this procedure, the VPH can be derived from each reconstructed proton track at each position along the Bragg curve. The VPH value indicates the magnitude of energy loss in proton track. By comparison with the simulation results obtained using Geant4, we found the correlation between the LET calculated by Monte Carlo simulation and the VPH measured by the nuclear emulsion

  15. Proton Linear Energy Transfer measurement using Emulsion Cloud Chamber

    Energy Technology Data Exchange (ETDEWEB)

    Shin, Jae-ik [Proton Therapy Center, National Cancer Center (Korea, Republic of); Division of Heavy Ion Clinical Research, Korea Institute of Radiological & Medical Sciences (KIRAMS), Seoul (Korea, Republic of); Park, Seyjoon [Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University, School of Medicine, Seoul (Korea, Republic of); Kim, Haksoo; Kim, Meyoung [Proton Therapy Center, National Cancer Center (Korea, Republic of); Jeong, Chiyoung [Department of Radiation Oncology, Asan Medical Center, University of Ulsan College of Medicine, Seoul (Korea, Republic of); Cho, Sungkoo [Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University, School of Medicine, Seoul (Korea, Republic of); Lim, Young Kyung; Shin, Dongho [Proton Therapy Center, National Cancer Center (Korea, Republic of); Lee, Se Byeong, E-mail: sblee@ncc.re.kr [Proton Therapy Center, National Cancer Center (Korea, Republic of); Morishima, Kunihiro; Naganawa, Naotaka; Sato, Osamu [Department of Physics, Nagoya University, Nagoya (Japan); Kwak, Jungwon [Department of Radiation Oncology, Asan Medical Center, University of Ulsan College of Medicine, Seoul (Korea, Republic of); Kim, Sung Hyun [Center for Underground Physics, Institute for Basic Science (IBS), Daejeon (Korea, Republic of); Cho, Jung Sook [Department of refinement education, Dongseo University, Busan (Korea, Republic of); Ahn, Jung Keun [Department of Physics, Korea University, Seoul (Korea, Republic of); Kim, Ji Hyun; Yoon, Chun Sil [Gyeongsang National University, Jinju (Korea, Republic of); Incerti, Sebastien [CNRS, IN2P3, CENBG, UMR 5797, F-33170 Gradignan (France); Université Bordeaux 1, CENBG, UMR 5797, F-33170 Gradignan (France)

    2015-04-15

    This study proposes to determine the correlation between the Volume Pulse Height (VPH) measured by nuclear emulsion and Linear Energy Transfer (LET) calculated by Monte Carlo simulation based on Geant4. The nuclear emulsion was irradiated at the National Cancer Center (NCC) with a therapeutic proton beam and was installed at 5.2 m distance from the beam nozzle structure with various thicknesses of water-equivalent material (PMMA) blocks to position with specific positions along the Bragg curve. After the beam exposure and development of the emulsion films, the films were scanned by S-UTS developed in Nagoya University. The proton tracks in the scanned films were reconstructed using the ‘NETSCAN’ method. Through this procedure, the VPH can be derived from each reconstructed proton track at each position along the Bragg curve. The VPH value indicates the magnitude of energy loss in proton track. By comparison with the simulation results obtained using Geant4, we found the correlation between the LET calculated by Monte Carlo simulation and the VPH measured by the nuclear emulsion.

  16. Development of modern approach to absorbed dose assessment in radionuclide therapy, based on Monte Carlo method simulation of patient scintigraphy

    Science.gov (United States)

    Lysak, Y. V.; Klimanov, V. A.; Narkevich, B. Ya

    2017-01-01

    One of the most difficult problems of modern radionuclide therapy (RNT) is control of the absorbed dose in pathological volume. This research presents new approach based on estimation of radiopharmaceutical (RP) accumulated activity value in tumor volume, based on planar scintigraphic images of the patient and calculated radiation transport using Monte Carlo method, including absorption and scattering in biological tissues of the patient, and elements of gamma camera itself. In our research, to obtain the data, we performed modeling scintigraphy of the vial with administered to the patient activity of RP in gamma camera, the vial was placed at the certain distance from the collimator, and the similar study was performed in identical geometry, with the same values of activity of radiopharmaceuticals in the pathological target in the body of the patient. For correct calculation results, adapted Fisher-Snyder human phantom was simulated in MCNP program. In the context of our technique, calculations were performed for different sizes of pathological targets and various tumors deeps inside patient’s body, using radiopharmaceuticals based on a mixed β-γ-radiating (131I, 177Lu), and clear β- emitting (89Sr, 90Y) therapeutic radionuclides. Presented method can be used for adequate implementing in clinical practice estimation of absorbed doses in the regions of interest on the basis of planar scintigraphy of the patient with sufficient accuracy.

  17. Trigger Algorithm Design for a SUSY Lepton Trigger based on Forward Proton Tagging

    Energy Technology Data Exchange (ETDEWEB)

    Gronberg, J; Hollar, J

    2010-03-29

    At the Large Hadron Collider (LHC) pair production of SUSY leptons in gamma-gamma interactions will often include intact off-energy protons. Including detectors in the beampipe to measure these protons can give additional information to separate these events from background. We report on expected event rates and background rejection for a slepton trigger design in the Compact Muon Solenoid (CMS) experiment incorporating forward proton information. We conclude that a trigger that can observe an interesting number of events is feasible with the appropriate detector hardware.

  18. Seismic wavefield imaging in the Tokyo metropolitan area, Japan, based on the replica exchange Monte Carlo method

    Science.gov (United States)

    Kano, Masayuki; Nagao, Hiromichi; Nagata, Kenji; Ito, Shin-ichi; Sakai, Shin'ichi; Nakagawa, Shigeki; Hori, Muneo; Hirata, Naoshi

    2017-04-01

    Earthquakes sometimes cause serious disasters not only directly by ground motion itself but also secondarily by infrastructure damage, particularly in densely populated urban areas. To reduce these secondary disasters, it is important to rapidly evaluate seismic hazards by analyzing the seismic responses of individual structures due to the input ground motions. Such input motions are estimated utilizing an array of seismometers that are distributed more sparsely than the structures. We propose a methodology that integrates physics-based and data-driven approaches in order to obtain the seismic wavefield to be input into seismic response analysis. This study adopts the replica exchange Monte Carlo (REMC) method, which is one of the Markov chain Monte Carlo (MCMC) methods, for the estimation of the seismic wavefield together with one-dimensional local subsurface structure and source information. Numerical tests show that the REMC method is able to search the parameters related to the source and the local subsurface structure in broader parameter space than the Metropolis method, which is an ordinary MCMC method. The REMC method well reproduces the seismic wavefield consistent with the true one. In contrast, the ordinary kriging, which is a classical data-driven interpolation method for spatial data, is hardly able to reproduce the true wavefield even at low frequencies. This indicates that it is essential to take both physics-based and data-driven approaches into consideration for seismic wavefield imaging. Then the REMC method is applied to the actual waveforms observed by a dense seismic array MeSO-net (Metropolitan Seismic Observation network), in which 296 accelerometers are continuously in operation with several kilometer intervals in the Tokyo metropolitan area, Japan. The estimated wavefield within a frequency band of 0.10-0.20 Hz is absolutely consistent with the observed waveforms. Further investigation suggests that the seismic wavefield is successfully

  19. Efficient Monte Carlo sampling of inverse problems using a neural network-based forward—applied to GPR crosshole traveltime inversion

    Science.gov (United States)

    Hansen, T. M.; Cordua, K. S.

    2017-12-01

    Probabilistically formulated inverse problems can be solved using Monte Carlo-based sampling methods. In principle, both advanced prior information, based on for example, complex geostatistical models and non-linear forward models can be considered using such methods. However, Monte Carlo methods may be associated with huge computational costs that, in practice, limit their application. This is not least due to the computational requirements related to solving the forward problem, where the physical forward response of some earth model has to be evaluated. Here, it is suggested to replace a numerical complex evaluation of the forward problem, with a trained neural network that can be evaluated very fast. This will introduce a modeling error that is quantified probabilistically such that it can be accounted for during inversion. This allows a very fast and efficient Monte Carlo sampling of the solution to an inverse problem. We demonstrate the methodology for first arrival traveltime inversion of crosshole ground penetrating radar data. An accurate forward model, based on 2-D full-waveform modeling followed by automatic traveltime picking, is replaced by a fast neural network. This provides a sampling algorithm three orders of magnitude faster than using the accurate and computationally expensive forward model, and also considerably faster and more accurate (i.e. with better resolution), than commonly used approximate forward models. The methodology has the potential to dramatically change the complexity of non-linear and non-Gaussian inverse problems that have to be solved using Monte Carlo sampling techniques.

  20. Nonvolatile field effect transistors based on protons and Si/SiO2Si structures

    International Nuclear Information System (INIS)

    Warren, W.L.; Vanheusden, K.; Fleetwood, D.M.; Schwank, J.R.; Winokur, P.S.; Knoll, M.G.; Devine, R.A.B.

    1997-01-01

    Recently, the authors have demonstrated that annealing Si/SiO 2 /Si structures in a hydrogen containing ambient introduces mobile H + ions into the buried SiO 2 layer. Changes in the H + spatial distribution within the SiO 2 layer were electrically monitored by current-voltage (I-V) measurements. The ability to directly probe reversible protonic motion in Si/SiO 2 /Si structures makes this an exemplar system to explore the physics and chemistry of hydrogen in the technologically relevant Si/SiO 2 structure. In this work, they illustrate that this effect can be used as the basis for a programmable nonvolatile field effect transistor (NVFET) memory that may compete with other Si-based memory devices. The power of this novel device is its simplicity; it is based upon standard Si/SiO 2 /Si technology and forming gas annealing, a common treatment used in integrated circuit processing. They also briefly discuss the effects of radiation on its retention properties

  1. A CT-based software tool for evaluating compensator quality in passively scattered proton therapy

    Science.gov (United States)

    Li, Heng; Zhang, Lifei; Dong, Lei; Sahoo, Narayan; Gillin, Michael T.; Zhu, X. Ronald

    2010-11-01

    We have developed a quantitative computed tomography (CT)-based quality assurance (QA) tool for evaluating the accuracy of manufactured compensators used in passively scattered proton therapy. The thickness of a manufactured compensator was measured from its CT images and compared with the planned thickness defined by the treatment planning system. The difference between the measured and planned thicknesses was calculated with use of the Euclidean distance transformation and the kd-tree search method. Compensator accuracy was evaluated by examining several parameters including mean distance, maximum distance, global thickness error and central axis shifts. Two rectangular phantoms were used to validate the performance of the QA tool. Nine patients and 20 compensators were included in this study. We found that mean distances, global thickness errors and central axis shifts were all within 1 mm for all compensators studied, with maximum distances ranging from 1.1 to 3.8 mm. Although all compensators passed manual verification at selected points, about 5% of the pixels still had maximum distances of >2 mm, most of which correlated with large depth gradients. The correlation between the mean depth gradient of the compensator and the percentage of pixels with mean distance based compensator QA tool can be used to quantitatively evaluate manufactured compensators.

  2. Design of a 10 MeV normal conducting CW proton linac based on equidistant multi-gap CH cavities

    Science.gov (United States)

    Li, Zhi-Hui

    2015-09-01

    Continuous wave (CW) high current proton linacs have wide applications as the front end of high power proton machines. The low energy part of such a linac is the most difficult and there is currently no widely accepted solution. Based on the analysis of the focusing properties of the CW low energy proton linac, a 10 MeV low energy normal conducting proton linac based on equidistant seven-gap Cross-bar H-type (CH) cavities is proposed. The linac is composed of ten 7-gap CH cavities and the transverse focusing is maintained by quadrupole doublets located between the cavities. The total length of the linac is less than 6 meters and the average acceleration gradient is about 1.2 MeV/m. The electromagnetic properties of the cavities are investigated by Microwave Studio. At the nominal acceleration gradient the maximum surface electric field in the cavities is less than 1.3 times the Kilpatrick limit, and the Ohmic loss of each cavity is less than 35 kW. Multi-particle beam dynamics simulations are performed with Tracewin code, and the results show that the beam dynamics of the linac are quite stable, the linac has the capability to accelerate up to 30 mA beam with acceptable dynamics behavior. Supported by National Natural Science Foundation of China (11375122, 91126003)

  3. The influence of patient positioning uncertainties in proton radiotherapy on proton range and dose distributions.

    Science.gov (United States)

    Liebl, Jakob; Paganetti, Harald; Zhu, Mingyao; Winey, Brian A

    2014-09-01

    Proton radiotherapy allows radiation treatment delivery with high dose gradients. The nature of such dose distributions increases the influence of patient positioning uncertainties on their fidelity when compared to photon radiotherapy. The present work quantitatively analyzes the influence of setup uncertainties on proton range and dose distributions. Thirty-eight clinical passive scattering treatment fields for small lesions in the head were studied. Dose distributions for shifted and rotated patient positions were Monte Carlo-simulated. Proton range uncertainties at the 50%- and 90%-dose falloff position were calculated considering 18 arbitrary combinations of maximal patient position shifts and rotations for two patient positioning methods. Normal tissue complication probabilities (NTCPs), equivalent uniform doses (EUDs), and tumor control probabilities (TCPs) were studied for organs at risk (OARs) and target volumes of eight patients. The authors identified a median 1σ proton range uncertainty at the 50%-dose falloff of 2.8 mm for anatomy-based patient positioning and 1.6 mm for fiducial-based patient positioning as well as 7.2 and 5.8 mm for the 90%-dose falloff position, respectively. These range uncertainties were correlated to heterogeneity indices (HIs) calculated for each treatment field (38%positioning) and 1.2% (fiducial-based positioning) of the studied OARs and patient shifts. For target volumes TCP decreases by more than 10% (absolute) occurred in less than 2.2% of the considered treatment scenarios for anatomy-based patient positioning and were nonexistent for fiducial-based patient positioning. EUD changes for target volumes were up to 35% (anatomy-based positioning) and 16% (fiducial-based positioning). The influence of patient positioning uncertainties on proton range in therapy of small lesions in the human brain as well as target and OAR dosimetry were studied. Observed range uncertainties were correlated with HIs. The clinical practice of

  4. Advanced Monte Carlo procedure for the IFMIF d-Li neutron source term based on evaluated cross section data

    International Nuclear Information System (INIS)

    Simakov, S.P.; Fischer, U.; Moellendorff, U. von; Schmuck, I.; Konobeev, A.Yu.; Korovin, Yu.A.; Pereslavtsev, P.

    2002-01-01

    A newly developed computational procedure is presented for the generation of d-Li source neutrons in Monte Carlo transport calculations based on the use of evaluated double-differential d+ 6,7 Li cross section data. A new code M c DeLicious was developed as an extension to MCNP4C to enable neutronics design calculations for the d-Li based IFMIF neutron source making use of the evaluated deuteron data files. The M c DeLicious code was checked against available experimental data and calculation results of M c DeLi and MCNPX, both of which use built-in analytical models for the Li(d, xn) reaction. It is shown that M c DeLicious along with newly evaluated d+ 6,7 Li data is superior in predicting the characteristics of the d-Li neutron source. As this approach makes use of tabulated Li(d, xn) cross sections, the accuracy of the IFMIF d-Li neutron source term can be steadily improved with more advanced and validated data

  5. Advanced Monte Carlo procedure for the IFMIF d-Li neutron source term based on evaluated cross section data

    CERN Document Server

    Simakov, S P; Moellendorff, U V; Schmuck, I; Konobeev, A Y; Korovin, Y A; Pereslavtsev, P

    2002-01-01

    A newly developed computational procedure is presented for the generation of d-Li source neutrons in Monte Carlo transport calculations based on the use of evaluated double-differential d+ sup 6 sup , sup 7 Li cross section data. A new code M sup c DeLicious was developed as an extension to MCNP4C to enable neutronics design calculations for the d-Li based IFMIF neutron source making use of the evaluated deuteron data files. The M sup c DeLicious code was checked against available experimental data and calculation results of M sup c DeLi and MCNPX, both of which use built-in analytical models for the Li(d, xn) reaction. It is shown that M sup c DeLicious along with newly evaluated d+ sup 6 sup , sup 7 Li data is superior in predicting the characteristics of the d-Li neutron source. As this approach makes use of tabulated Li(d, xn) cross sections, the accuracy of the IFMIF d-Li neutron source term can be steadily improved with more advanced and validated data.

  6. Warranty optimisation based on the prediction of costs to the manufacturer using neural network model and Monte Carlo simulation

    Science.gov (United States)

    Stamenkovic, Dragan D.; Popovic, Vladimir M.

    2015-02-01

    Warranty is a powerful marketing tool, but it always involves additional costs to the manufacturer. In order to reduce these costs and make use of warranty's marketing potential, the manufacturer needs to master the techniques for warranty cost prediction according to the reliability characteristics of the product. In this paper a combination free replacement and pro rata warranty policy is analysed as warranty model for one type of light bulbs. Since operating conditions have a great impact on product reliability, they need to be considered in such analysis. A neural network model is used to predict light bulb reliability characteristics based on the data from the tests of light bulbs in various operating conditions. Compared with a linear regression model used in the literature for similar tasks, the neural network model proved to be a more accurate method for such prediction. Reliability parameters obtained in this way are later used in Monte Carlo simulation for the prediction of times to failure needed for warranty cost calculation. The results of the analysis make possible for the manufacturer to choose the optimal warranty policy based on expected product operating conditions. In such a way, the manufacturer can lower the costs and increase the profit.

  7. Whole body counter calibration using Monte Carlo modeling with an array of phantom sizes based on national anthropometric reference data

    Science.gov (United States)

    Shypailo, R. J.; Ellis, K. J.

    2011-05-01

    During construction of the whole body counter (WBC) at the Children's Nutrition Research Center (CNRC), efficiency calibration was needed to translate acquired counts of 40K to actual grams of potassium for measurement of total body potassium (TBK) in a diverse subject population. The MCNP Monte Carlo n-particle simulation program was used to describe the WBC (54 detectors plus shielding), test individual detector counting response, and create a series of virtual anthropomorphic phantoms based on national reference anthropometric data. Each phantom included an outer layer of adipose tissue and an inner core of lean tissue. Phantoms were designed for both genders representing ages 3.5 to 18.5 years with body sizes from the 5th to the 95th percentile based on body weight. In addition, a spherical surface source surrounding the WBC was modeled in order to measure the effects of subject mass on room background interference. Individual detector measurements showed good agreement with the MCNP model. The background source model came close to agreement with empirical measurements, but showed a trend deviating from unity with increasing subject size. Results from the MCNP simulation of the CNRC WBC agreed well with empirical measurements using BOMAB phantoms. Individual detector efficiency corrections were used to improve the accuracy of the model. Nonlinear multiple regression efficiency calibration equations were derived for each gender. Room background correction is critical in improving the accuracy of the WBC calibration.

  8. Non-noble metal based electro-catalyst compositions for proton exchange membrane based water electrolysis and methods of making

    Energy Technology Data Exchange (ETDEWEB)

    Kumta, Prashant N.; Kadakia, Karan Sandeep; Datta, Moni Kanchan; Velikokhatnyi, Oleg

    2017-02-07

    The invention provides electro-catalyst compositions for an anode electrode of a proton exchange membrane-based water electrolysis system. The compositions include a noble metal component selected from the group consisting of iridium oxide, ruthenium oxide, rhenium oxide and mixtures thereof, and a non-noble metal component selected from the group consisting of tantalum oxide, tin oxide, niobium oxide, titanium oxide, tungsten oxide, molybdenum oxide, yttrium oxide, scandium oxide, cooper oxide, zirconium oxide, nickel oxide and mixtures thereof. Further, the non-noble metal component can include a dopant. The dopant can be at least one element selected from Groups III, V, VI and VII of the Periodic Table. The compositions can be prepared using a surfactant approach or a sol gel approach. Further, the compositions are prepared using noble metal and non-noble metal precursors. Furthermore, a thin film containing the compositions can be deposited onto a substrate to form the anode electrode.

  9. TestDose: A nuclear medicine software based on Monte Carlo modeling for generating gamma camera acquisitions and dosimetry.

    Science.gov (United States)

    Garcia, Marie-Paule; Villoing, Daphnée; McKay, Erin; Ferrer, Ludovic; Cremonesi, Marta; Botta, Francesca; Ferrari, Mahila; Bardiès, Manuel

    2015-12-01

    The TestDose platform was developed to generate scintigraphic imaging protocols and associated dosimetry by Monte Carlo modeling. TestDose is part of a broader project (www.dositest.com) whose aim is to identify the biases induced by different clinical dosimetry protocols. The TestDose software allows handling the whole pipeline from virtual patient generation to resulting planar and SPECT images and dosimetry calculations. The originality of their approach relies on the implementation of functional segmentation for the anthropomorphic model representing a virtual patient. Two anthropomorphic models are currently available: 4D XCAT and ICRP 110. A pharmacokinetic model describes the biodistribution of a given radiopharmaceutical in each defined compartment at various time-points. The Monte Carlo simulation toolkit gate offers the possibility to accurately simulate scintigraphic images and absorbed doses in volumes of interest. The TestDose platform relies on gate to reproduce precisely any imaging protocol and to provide reference dosimetry. For image generation, TestDose stores user's imaging requirements and generates automatically command files used as input for gate. Each compartment is simulated only once and the resulting output is weighted using pharmacokinetic data. Resulting compartment projections are aggregated to obtain the final image. For dosimetry computation, emission data are stored in the platform database and relevant gate input files are generated for the virtual patient model and associated pharmacokinetics. Two samples of software runs are given to demonstrate the potential of TestDose. A clinical imaging protocol for the Octreoscan™ therapeutical treatment was implemented using the 4D XCAT model. Whole-body "step and shoot" acquisitions at different times postinjection and one SPECT acquisition were generated within reasonable computation times. Based on the same Octreoscan™ kinetics, a dosimetry computation performed on the ICRP 110

  10. TestDose: A nuclear medicine software based on Monte Carlo modeling for generating gamma camera acquisitions and dosimetry

    International Nuclear Information System (INIS)

    Garcia, Marie-Paule; Villoing, Daphnée; McKay, Erin; Ferrer, Ludovic; Cremonesi, Marta; Botta, Francesca; Ferrari, Mahila; Bardiès, Manuel

    2015-01-01

    Purpose: The TestDose platform was developed to generate scintigraphic imaging protocols and associated dosimetry by Monte Carlo modeling. TestDose is part of a broader project (www.dositest.com) whose aim is to identify the biases induced by different clinical dosimetry protocols. Methods: The TestDose software allows handling the whole pipeline from virtual patient generation to resulting planar and SPECT images and dosimetry calculations. The originality of their approach relies on the implementation of functional segmentation for the anthropomorphic model representing a virtual patient. Two anthropomorphic models are currently available: 4D XCAT and ICRP 110. A pharmacokinetic model describes the biodistribution of a given radiopharmaceutical in each defined compartment at various time-points. The Monte Carlo simulation toolkit GATE offers the possibility to accurately simulate scintigraphic images and absorbed doses in volumes of interest. The TestDose platform relies on GATE to reproduce precisely any imaging protocol and to provide reference dosimetry. For image generation, TestDose stores user’s imaging requirements and generates automatically command files used as input for GATE. Each compartment is simulated only once and the resulting output is weighted using pharmacokinetic data. Resulting compartment projections are aggregated to obtain the final image. For dosimetry computation, emission data are stored in the platform database and relevant GATE input files are generated for the virtual patient model and associated pharmacokinetics. Results: Two samples of software runs are given to demonstrate the potential of TestDose. A clinical imaging protocol for the Octreoscan™ therapeutical treatment was implemented using the 4D XCAT model. Whole-body “step and shoot” acquisitions at different times postinjection and one SPECT acquisition were generated within reasonable computation times. Based on the same Octreoscan™ kinetics, a dosimetry

  11. TestDose: A nuclear medicine software based on Monte Carlo modeling for generating gamma camera acquisitions and dosimetry

    Energy Technology Data Exchange (ETDEWEB)

    Garcia, Marie-Paule, E-mail: marie-paule.garcia@univ-brest.fr; Villoing, Daphnée [UMR 1037 INSERM/UPS, CRCT, 133 Route de Narbonne, 31062 Toulouse (France); McKay, Erin [St George Hospital, Gray Street, Kogarah, New South Wales 2217 (Australia); Ferrer, Ludovic [ICO René Gauducheau, Boulevard Jacques Monod, St Herblain 44805 (France); Cremonesi, Marta; Botta, Francesca; Ferrari, Mahila [European Institute of Oncology, Via Ripamonti 435, Milano 20141 (Italy); Bardiès, Manuel [UMR 1037 INSERM/UPS, CRCT, 133 Route de Narbonne, Toulouse 31062 (France)

    2015-12-15

    Purpose: The TestDose platform was developed to generate scintigraphic imaging protocols and associated dosimetry by Monte Carlo modeling. TestDose is part of a broader project (www.dositest.com) whose aim is to identify the biases induced by different clinical dosimetry protocols. Methods: The TestDose software allows handling the whole pipeline from virtual patient generation to resulting planar and SPECT images and dosimetry calculations. The originality of their approach relies on the implementation of functional segmentation for the anthropomorphic model representing a virtual patient. Two anthropomorphic models are currently available: 4D XCAT and ICRP 110. A pharmacokinetic model describes the biodistribution of a given radiopharmaceutical in each defined compartment at various time-points. The Monte Carlo simulation toolkit GATE offers the possibility to accurately simulate scintigraphic images and absorbed doses in volumes of interest. The TestDose platform relies on GATE to reproduce precisely any imaging protocol and to provide reference dosimetry. For image generation, TestDose stores user’s imaging requirements and generates automatically command files used as input for GATE. Each compartment is simulated only once and the resulting output is weighted using pharmacokinetic data. Resulting compartment projections are aggregated to obtain the final image. For dosimetry computation, emission data are stored in the platform database and relevant GATE input files are generated for the virtual patient model and associated pharmacokinetics. Results: Two samples of software runs are given to demonstrate the potential of TestDose. A clinical imaging protocol for the Octreoscan™ therapeutical treatment was implemented using the 4D XCAT model. Whole-body “step and shoot” acquisitions at different times postinjection and one SPECT acquisition were generated within reasonable computation times. Based on the same Octreoscan™ kinetics, a dosimetry

  12. Dose optimization based on linear programming implemented in a system for treatment planning in Monte Carlo

    International Nuclear Information System (INIS)

    Ureba, A.; Palma, B. A.; Leal, A.

    2011-01-01

    Develop a more efficient method of optimization in relation to time, based on linear programming designed to implement a multi objective penalty function which also permits a simultaneous solution integrated boost situations considering two white volumes simultaneously.

  13. Proton absorbed dose distribution in human eye simulated by SRNA-2KG code

    International Nuclear Information System (INIS)

    Ilic, R. D.; Pavlovic, R.

    2004-01-01

    The model of Monte Carlo SRNA code is described together with some numerical experiments to show feasibility of this code to be used in proton therapy, especially for tree dimensional proton absorption dose calculation in human eye. (author) [sr

  14. A Monte Carlo method for the simulation of coagulation and nucleation based on weighted particles and the concepts of stochastic resolution and merging

    Energy Technology Data Exchange (ETDEWEB)

    Kotalczyk, G., E-mail: Gregor.Kotalczyk@uni-due.de; Kruis, F.E.

    2017-07-01

    Monte Carlo simulations based on weighted simulation particles can solve a variety of population balance problems and allow thus to formulate a solution-framework for many chemical engineering processes. This study presents a novel concept for the calculation of coagulation rates of weighted Monte Carlo particles by introducing a family of transformations to non-weighted Monte Carlo particles. The tuning of the accuracy (named ‘stochastic resolution’ in this paper) of those transformations allows the construction of a constant-number coagulation scheme. Furthermore, a parallel algorithm for the inclusion of newly formed Monte Carlo particles due to nucleation is presented in the scope of a constant-number scheme: the low-weight merging. This technique is found to create significantly less statistical simulation noise than the conventional technique (named ‘random removal’ in this paper). Both concepts are combined into a single GPU-based simulation method which is validated by comparison with the discrete-sectional simulation technique. Two test models describing a constant-rate nucleation coupled to a simultaneous coagulation in 1) the free-molecular regime or 2) the continuum regime are simulated for this purpose.

  15. A Monte Carlo method for the simulation of coagulation and nucleation based on weighted particles and the concepts of stochastic resolution and merging

    Science.gov (United States)

    Kotalczyk, G.; Kruis, F. E.

    2017-07-01

    Monte Carlo simulations based on weighted simulation particles can solve a variety of population balance problems and allow thus to formulate a solution-framework for many chemical engineering processes. This study presents a novel concept for the calculation of coagulation rates of weighted Monte Carlo particles by introducing a family of transformations to non-weighted Monte Carlo particles. The tuning of the accuracy (named 'stochastic resolution' in this paper) of those transformations allows the construction of a constant-number coagulation scheme. Furthermore, a parallel algorithm for the inclusion of newly formed Monte Carlo particles due to nucleation is presented in the scope of a constant-number scheme: the low-weight merging. This technique is found to create significantly less statistical simulation noise than the conventional technique (named 'random removal' in this paper). Both concepts are combined into a single GPU-based simulation method which is validated by comparison with the discrete-sectional simulation technique. Two test models describing a constant-rate nucleation coupled to a simultaneous coagulation in 1) the free-molecular regime or 2) the continuum regime are simulated for this purpose.

  16. Prototype system for proton beam range measurement based on gamma electron vertex imaging

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Han Rim [Neutron Utilization Technology Division, Korea Atomic Energy Research Institute, 111, Daedeok-daero 989beon-gil, Yuseong-gu, Daejeon 34057 (Korea, Republic of); Kim, Sung Hun; Park, Jong Hoon [Department of Nuclear Engineering, Hanyang University, Seongdong-gu, Seoul 04763 (Korea, Republic of); Jung, Won Gyun [Heavy-ion Clinical Research Division, Korean Institute of Radiological & Medical Sciences, Seoul 01812 (Korea, Republic of); Lim, Hansang [Department of Electronics Convergence Engineering, Kwangwoon University, Seoul 01897 (Korea, Republic of); Kim, Chan Hyeong, E-mail: chkim@hanyang.ac.kr [Department of Nuclear Engineering, Hanyang University, Seongdong-gu, Seoul 04763 (Korea, Republic of)

    2017-06-11

    In proton therapy, for both therapeutic effectiveness and patient safety, it is very important to accurately measure the proton dose distribution, especially the range of the proton beam. For this purpose, recently we proposed a new imaging method named gamma electron vertex imaging (GEVI), in which the prompt gammas emitting from the nuclear reactions of the proton beam in the patient are converted to electrons, and then the converted electrons are tracked to determine the vertices of the prompt gammas, thereby producing a 2D image of the vertices. In the present study, we developed a prototype GEVI system, including dedicated signal processing and data acquisition systems, which consists of a beryllium plate (= electron converter) to convert the prompt gammas to electrons, two double-sided silicon strip detectors (= hodoscopes) to determine the trajectories of those converted electrons, and a plastic scintillation detector (= calorimeter) to measure their kinetic energies. The system uses triple coincidence logic and multiple energy windows to select only the events from prompt gammas. The detectors of the prototype GEVI system were evaluated for electronic noise level, energy resolution, and time resolution. Finally, the imaging capability of the GEVI system was tested by imaging a {sup 90}Sr beta source, a {sup 60}Co gamma source, and a 45-MeV proton beam in a PMMA phantom. The overall results of the present study generally show that the prototype GEVI system can image the vertices of the prompt gammas produced by the proton nuclear interactions.

  17. Microscopic Gold Particle-Based Fiducial Markers for Proton Therapy of Prostate Cancer

    International Nuclear Information System (INIS)

    Lim, Young Kyung; Kwak, Jungwon; Kim, Dong Wook; Shin, Dongho; Yoon, Myonggeun; Park, Soah; Kim, Jin Sung; Ahn, Sung Hwan; Shin, Jungwook; Lee, Se Byeong; Park, Sung Yong; Pyo, Hong Ryeol; Kim, Dae Yong M.D.; Cho, Kwan Ho

    2009-01-01

    Purpose: We examined the feasibility of using fiducial markers composed of microscopic gold particles and human-compatible polymers as a means to overcome current problems with conventional macroscopic gold fiducial markers, such as dose reduction and artifact generation, in proton therapy for prostate cancer. Methods and Materials: We examined two types of gold particle fiducial marker interactions: that with diagnostic X-rays and with a therapeutic proton beam. That is, we qualitatively and quantitatively compared the radiographic visibility of conventional gold and gold particle fiducial markers and the CT artifacts and dose reduction associated with their use. Results: The gold particle fiducials could be easily distinguished from high-density structures, such as the pelvic bone, in diagnostic X-rays but were nearly transparent to a proton beam. The proton dose distribution was distorted <5% by the gold particle fiducials with a 4.9% normalized gold density; this was the case even in the worst configuration (i.e., parallel alignment with a single-direction proton beam). In addition, CT artifacts were dramatically reduced for the gold particle mixture. Conclusion: Mixtures of microscopic gold particles and human-compatible polymers have excellent potential as fiducial markers for proton therapy for prostate cancer. These include good radiographic visibility, low distortion of the depth-dose distribution, and few CT artifacts.

  18. Proton and electron deep dose profiles for retinoblastoma based on GEANT 4 code

    Energy Technology Data Exchange (ETDEWEB)

    Braga, Flavia V., E-mail: flaviafisica@gmail.co [Universidade Federal de Minas Gerais (UFMG), Belo Horizonte, MG (Brazil). Programa de Pos-graduacao em Ciencias e Tecnicas Nucleares; Centro de Desenvolvimento da Tecnologia Nuclear (CDTN/CNEN-MG), Belo Horizonte, MG (Brazil); Campos, Tarcisio P.R. de [Universidade Federal de Minas Gerais (UFMG), Belo Horizonte, MG (Brazil). Programa de Pos-graduacao em Ciencias e Tecnicas Nucleares; Ribeiro, Kilder L., E-mail: kilderlr@gmail.co [Universidade Estadual de Feira de Santana (UEFS), BA (Brazil). Dept. de Fisica

    2009-07-01

    Herein, the dosimetry responses to a retinoblastoma proton and electron radiation therapy were investigated. The computational tool applied to this simulation was the Geant4 code, version 4.9.1. The code allows simulating the charge particle interaction with eyeball tissue. In the present simulation, a box of 4 cm side water filled had represented the human eye. The simulation was performed considering mono energetic beams of protons and electrons with spectra of 57 to 70 MeV for protons and 2 to 8 MeV for electrons. The simulation was guide by the advanced hadron therapy example distributed with the Geant4 code. The phantom was divided in voxels with 0.2 mm side. The energy deposited in each voxel was evaluated taken the direct beam at one face. The simulation results show the delivery energy and therefore the dose deposited in each voxel. The deep dose profiles to proton and electron were plotted. The well known Bragg peak was reproduced for protons. The maximum delivered dose defined the position at the proton stopped. However, to electrons, the absorbed energies were delivered along its path producing a more continuous distribution following the water depth, but also being stopped in the end of its path. (author)

  19. Prototype system for proton beam range measurement based on gamma electron vertex imaging

    Science.gov (United States)

    Lee, Han Rim; Kim, Sung Hun; Park, Jong Hoon; Jung, Won Gyun; Lim, Hansang; Kim, Chan Hyeong

    2017-06-01

    In proton therapy, for both therapeutic effectiveness and patient safety, it is very important to accurately measure the proton dose distribution, especially the range of the proton beam. For this purpose, recently we proposed a new imaging method named gamma electron vertex imaging (GEVI), in which the prompt gammas emitting from the nuclear reactions of the proton beam in the patient are converted to electrons, and then the converted electrons are tracked to determine the vertices of the prompt gammas, thereby producing a 2D image of the vertices. In the present study, we developed a prototype GEVI system, including dedicated signal processing and data acquisition systems, which consists of a beryllium plate (= electron converter) to convert the prompt gammas to electrons, two double-sided silicon strip detectors (= hodoscopes) to determine the trajectories of those converted electrons, and a plastic scintillation detector (= calorimeter) to measure their kinetic energies. The system uses triple coincidence logic and multiple energy windows to select only the events from prompt gammas. The detectors of the prototype GEVI system were evaluated for electronic noise level, energy resolution, and time resolution. Finally, the imaging capability of the GEVI system was tested by imaging a 90Sr beta source, a 60Co gamma source, and a 45-MeV proton beam in a PMMA phantom. The overall results of the present study generally show that the prototype GEVI system can image the vertices of the prompt gammas produced by the proton nuclear interactions.

  20. Proton and electron deep dose profiles for retinoblastoma based on GEANT 4 code

    International Nuclear Information System (INIS)

    Braga, Flavia V.; Campos, Tarcisio P.R. de; Ribeiro, Kilder L.

    2009-01-01

    Herein, the dosimetry responses to a retinoblastoma proton and electron radiation therapy were investigated. The computational tool applied to this simulation was the Geant4 code, version 4.9.1. The code allows simulating the charge particle interaction with eyeball tissue. In the present simulation, a box of 4 cm side water filled had represented the human eye. The simulation was performed considering mono energetic beams of protons and electrons with spectra of 57 to 70 MeV for protons and 2 to 8 MeV for electrons. The simulation was guide by the advanced hadron therapy example distributed with the Geant4 code. The phantom was divided in voxels with 0.2 mm side. The energy deposited in each voxel was evaluated taken the direct beam at one face. The simulation results show the delivery energy and therefore the dose deposited in each voxel. The deep dose profiles to proton and electron were plotted. The well known Bragg peak was reproduced for protons. The maximum delivered dose defined the position at the proton stopped. However, to electrons, the absorbed energies were delivered along its path producing a more continuous distribution following the water depth, but also being stopped in the end of its path. (author)

  1. Heavy quarks in proton

    CERN Document Server

    AUTHOR|(SzGeCERN)655637

    The measurement of prompt photon associated with a b jet in proton-proton interactions can provide us insight into the inner structure of proton. This is because precision of determination of parton distribution functions of b quark and gluon can be increased by such a measurement. The measurement of cross-section of prompt photon associated with a b jet (process $pp\\longrightarrow \\gamma + b + X$) at $\\sqrt{s}$= 8 TeV with the ATLAS detector is presented. Full 8 TeV dataset collected by ATLAS during the year 2012 was used in this analysis. Corresponding integrated luminosity is 20.3 $fb^{-1}$. Fiducial differential cross-section as a function of photon transverse momentum at particle level was extracted from data and compared with the prediction of leading order event generator Pythia 8. Cross-section extracted from data is normalised independently on the Monte Carlo prediction. Values of data distribution lie above Monte Carlo values. The difference can be explained by presence of higher order effects not ...

  2. 15 MeV proton irradiation effects on Bi-based high temperature superconductors

    Energy Technology Data Exchange (ETDEWEB)

    Alinejad, N.; Sohrabi, D. [Nuclear Science and Technology Research Institute, Tehran (Iran, Islamic Republic of). Plasma and Nuclear Fusion Research School; Bolori, F. [Karaj Agricultural, Medical, and Industrial Research School, Karaj (Iran, Islamic Republic of)

    2015-11-15

    Nowadays, superconducting magnetic coils are used in some tokamaks such as EAST, KSTAR, JT-60, and T-15 to generate strong magnetic fields and also in ITER magnetic fields of about 13 tesla will be produced with the help of superconductors. The tokamak superconductors are exposed to the variety of radiations (neutron, ions beam, and gamma) from plasma nuclear reactions which will affect some of the superconductor properties. Therefore, study of the irradiation effects on the superconductor structure and properties are very crucial from technological and scientific point of view. One of the superconductor irradiation effects to be investigated under different conditions of energy and dosage is the potential resistance of the material used in tokamak reactor magnetic coils against activation by radiation. In this work, pellets of high T{sub c} Bi-based superconductors have been prepared and after measurement of parameters, a sample of pellet has been irradiated with 15 MeV protons using Karaj cyclotron facility. The sample's parameters have been measured again after irradiation treatment. X-ray diffraction patterns and SEM images of the sample before and after irradiation treatment have been studied.

  3. A LabVIEWTM-based scanning and control system for proton beam micromachining

    International Nuclear Information System (INIS)

    Bettiol, Andrew A.; Kan, J.A. van; Sum, T.C.; Watt, F.

    2001-01-01

    LabVIEW TM is steadily gaining in popularity as the programming language of choice for scientific data acquisition and control. This is due to the vast array of measurement instruments and data acquisition cards supported by the LabVIEW TM environment, and the relative ease with which advanced software can be programmed. Furthermore, virtual instruments that are designed for a given system can be easily ported to other LabVIEW TM platforms and hardware. This paper describes the new LabVIEW TM based scanning and control system developed specifically for proton beam micromachining (PBM) applications. The new system is capable of scanning figures at 16-bit resolution with improved sub-microsecond scan rates. Support for electrostatic beam blanking and external dose normalization using a TTL signal have been implemented. The new software incorporates a semi-automated dose calibration system, and a number of novel dose normalization methods. Limitations of the current beam scanning hardware are discussed in light of new results obtained from micromachining experiments performed in SU-8 photoresist

  4. Tungsten oxide proton conducting films for low-voltage transparent oxide-based thin-film transistors

    International Nuclear Information System (INIS)

    Zhang, Hongliang; Wan, Qing; Wan, Changjin; Wu, Guodong; Zhu, Liqiang

    2013-01-01

    Tungsten oxide (WO x ) electrolyte films deposited by reactive magnetron sputtering showed a high room temperature proton conductivity of 1.38 × 10 −4 S/cm with a relative humidity of 60%. Low-voltage transparent W-doped indium-zinc-oxide thin-film transistors gated by WO x -based electrolytes were self-assembled on glass substrates by one mask diffraction method. Enhancement mode operation with a large current on/off ratio of 4.7 × 10 6 , a low subthreshold swing of 108 mV/decade, and a high field-effect mobility 42.6 cm 2 /V s was realized. Our results demonstrated that WO x -based proton conducting films were promising gate dielectric candidates for portable low-voltage oxide-based devices.

  5. Development of a micro-XRF system for biological samples based on proton-induced quasimonochromatic X-rays

    Energy Technology Data Exchange (ETDEWEB)

    Ploykrachang, K., E-mail: ploykrachang.k.aa@m.titech.ac.jp [Research Laboratory for Nuclear Reactors, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8550 (Japan); Hasegawa, J. [Department of Energy Sciences, Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology, 4259 Nagatsuta-cho, Midori-ku, Yokohama 226-8502 (Japan); Kondo, K.; Fukuda, H.; Oguri, Y. [Research Laboratory for Nuclear Reactors, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8550 (Japan)

    2014-07-15

    We have developed a micro-XRF system based on a proton-induced quasimonochromatic X-ray (QMXR) microbeam for in vivo measurement of biological samples. A 2.5-MeV proton beam impinged normally on a Cu foil target that was slightly thicker than the proton range. The emitted QMXR behind the Cu target was focused with a polycapillary X-ray half lens. For application to analysis of wet or aquatic samples, we prepared a QMXR beam with an incident angle of 45° with respect to the horizontal plane by using a dipole magnet in order to bend the primary proton beam downward by 45°. The focal spot size of the QMXR microbeam on a horizontal sample surface was evaluated to be 250 × 350 μm by a wire scanning method. A microscope camera with a long working distance was installed perpendicular to the sample surface to identify the analyzed position on the sample. The fluorescent radiation from the sample was collected by a Si-PIN photodiode X-ray detector. Using the setup above, we were able to successfully measure the accumulation and distribution of Co in the leaves of a free-floating aquatic plant on a dilute Co solution surface.

  6. Multi-period mean–variance portfolio optimization based on Monte-Carlo simulation

    NARCIS (Netherlands)

    F. Cong (Fei); C.W. Oosterlee (Cornelis)

    2016-01-01

    htmlabstractWe propose a simulation-based approach for solving the constrained dynamic mean– variance portfolio managemen tproblem. For this dynamic optimization problem, we first consider a sub-optimal strategy, called the multi-stage strategy, which can be utilized in a forward fashion. Then,

  7. Report: GPU Based Massive Parallel Kawasaki Kinetics In Monte Carlo Modelling of Lipid Microdomains

    OpenAIRE

    Lis, M.; Pintal, L.

    2013-01-01

    This paper introduces novel method of simulation of lipid biomembranes based on Metropolis Hastings algorithm and Graphic Processing Unit computational power. Method gives up to 55 times computational boost in comparison to classical computations. Extensive study of algorithm correctness is provided. Analysis of simulation results and results obtained with classical simulation methodologies are presented.

  8. Monte carlo efficiency calibration of a neutron generator-based total-body irradiator

    Science.gov (United States)

    The increasing prevalence of obesity world-wide has focused attention on the need for accurate body composition assessments, especially of large subjects. However, many body composition measurement systems are calibrated against a single-sized phantom, often based on the standard Reference Man mode...

  9. Kernel Density Independence Sampling based Monte Carlo Scheme (KISMCS) for inverse hydrological modeling

    NARCIS (Netherlands)

    Shafiei, M.; Gharari, S.; Pande, S.; Bhulai, S.

    2014-01-01

    Posterior sampling methods are increasingly being used to describe parameter and model predictive uncertainty in hydrologic modelling. This paper proposes an alternative to random walk chains (such as DREAM-zs). We propose a sampler based on independence chains with an embedded feature of

  10. Empirical performance of RSSI-based Monte Carlo localisation for active RFID patient tracking systems

    NARCIS (Netherlands)

    Cully, William P.L.; Cotton, Simon L.; Scanlon, W.G.

    2012-01-01

    The range of potential applications for indoor and campus based personnel localisation has led researchers to create a wide spectrum of different algorithmic approaches and systems. However, the majority of the proposed systems overlook the unique radio environment presented by the human body

  11. Wormhole Hamiltonian Monte Carlo

    OpenAIRE

    Lan, S; Streets, J; Shahbaba, B

    2014-01-01

    Copyright © 2014, Association for the Advancement of Artificial Intelligence. In machine learning and statistics, probabilistic inference involving multimodal distributions is quite difficult. This is especially true in high dimensional problems, where most existing algorithms cannot easily move from one mode to another. To address this issue, we propose a novel Bayesian inference approach based on Markov Chain Monte Carlo. Our method can effectively sample from multimodal distributions, espe...

  12. Size dependent thermal hysteresis in spin crossover nanoparticles reflected within a Monte Carlo based Ising-like model

    International Nuclear Information System (INIS)

    Atitoaie, Alexandru; Tanasa, Radu; Enachescu, Cristian

    2012-01-01

    Spin crossover compounds are photo-magnetic bistable molecular magnets with two states in thermodynamic competition: the diamagnetic low-spin state and paramagnetic high-spin state. The thermal transition between the two states is often accompanied by a wide hysteresis, premise for possible application of these materials as recording media. In this paper we study the influence of the system's size on the thermal hysteresis loops using Monte Carlo simulations based on an Arrhenius dynamics applied for an Ising like model with long- and short-range interactions. We show that using appropriate boundary conditions it is possible to reproduce both the drop of hysteresis width with decreasing particle size, the hysteresis shift towards lower temperatures and the incomplete transition, as in the available experimental data. The case of larger systems composed by several sublattices is equally treated reproducing the shrinkage of the hysteresis loop's width experimentally observed. - Highlights: ► A study concerning size effects in spin crossover nanoparticles hysteresis is presented. ► An Ising like model with short- and long-range interactions and Arrhenius dynamics is employed. ► In open boundary system the hysteresis width decreases with particle size. ► With appropriate environment, hysteresis loop is shifted towards lower temperature and transition is incomplete.

  13. Sequential Monte Carlo filter for state estimation of LiFePO4 batteries based on an online updated model

    Science.gov (United States)

    Li, Jiahao; Klee Barillas, Joaquin; Guenther, Clemens; Danzer, Michael A.

    2014-02-01

    Battery state monitoring is one of the key techniques in battery management systems e.g. in electric vehicles. An accurate estimation can help to improve the system performance and to prolong the battery remaining useful life. Main challenges for the state estimation for LiFePO4 batteries are the flat characteristic of open-circuit-voltage over battery state of charge (SOC) and the existence of hysteresis phenomena. Classical estimation approaches like Kalman filtering show limitations to handle nonlinear and non-Gaussian error distribution problems. In addition, uncertainties in the battery model parameters must be taken into account to describe the battery degradation. In this paper, a novel model-based method combining a Sequential Monte Carlo filter with adaptive control to determine the cell SOC and its electric impedance is presented. The applicability of this dual estimator is verified using measurement data acquired from a commercial LiFePO4 cell. Due to a better handling of the hysteresis problem, results show the benefits of the proposed method against the estimation with an Extended Kalman filter.

  14. Monte Carlo simulation of neutron irradiation facility developed for accelerator based in vivo neutron activation measurements in human hand bones

    International Nuclear Information System (INIS)

    Aslam; Prestwich, W.V.; McNeill, F.E.; Waker, A.J.

    2006-01-01

    The neutron irradiation facility developed at the McMaster University 3 MV Van de Graaff accelerator was employed to assess in vivo elemental content of aluminum and manganese in human hands. These measurements were carried out to monitor the long-term exposure of these potentially toxic trace elements through hand bone levels. The dose equivalent delivered to a patient during irradiation procedure is the limiting factor for IVNAA measurements. This article describes a method to estimate the average radiation dose equivalent delivered to the patient's hand during irradiation. The computational method described in this work augments the dose measurements carried out earlier [Arnold et al., 2002. Med. Phys. 29(11), 2718-2724]. This method employs the Monte Carlo simulation of hand irradiation facility using MCNP4B. Based on the estimated dose equivalents received by the patient hand, the proposed irradiation procedure for the IVNAA measurement of manganese in human hands [Arnold et al., 2002. Med. Phys. 29(11), 2718-2724] with normal (1 ppm) and elevated manganese content can be carried out with a reasonably low dose of 31 mSv to the hand. Sixty-three percent of the total dose equivalent is delivered by non-useful fast group (>10 keV); the filtration of this neutron group from the beam will further decrease the dose equivalent to the patient's hand

  15. SU-E-J-69: Iterative Deconvolution of the Initial Photon Fluence for EPID Dosimetry: A Monte Carlo Based Study

    Energy Technology Data Exchange (ETDEWEB)

    Czarnecki, D; Voigts-Rhetz, P von; Shishechian, D Uchimura [Technische Hochschule Mittelhessen - University of Applied Sciences, Giessen (Germany); Zink, K [Technische Hochschule Mittelhessen - University of Applied Sciences, Giessen (Germany); Germany and Department of Radiotherapy and Radiooncology, University Medical Center Giessen-Marburg, Marburg (Germany)

    2015-06-15

    Purpose: Developing a fast and accurate calculation model to reconstruct the applied photon fluence from an external photon radiation therapy treatment based on an image recorded by an electronic portal image device (EPID). Methods: To reconstruct the initial photon fluence the 2D EPID image was corrected for scatter from the patient/phantom and EPID to generate the transmitted primary photon fluence. This was done by an iterative deconvolution using precalculated point spread functions (PSF). The transmitted primary photon fluence was then backprojected through the patient/phantom geometry considering linear attenuation to receive the initial photon fluence applied for the treatment.The calculation model was verified using Monte Carlo simulations performed with the EGSnrc code system. EPID images were produced by calculating the dose deposition in the EPID from a 6 MV photon beam irradiating a water phantom with air and bone inhomogeneities and the ICRP anthropomorphic voxel phantom. Results: The initial photon fluence was reconstructed using a single PSF and position dependent PSFs which depend on the radiological thickness of the irradiated object. Appling position dependent point spread functions the mean uncertainty of the reconstructed initial photon fluence could be reduced from 1.13 % to 0.13 %. Conclusion: This study presents a calculation model for fluence reconstruction from EPID images. The{sup Result} show a clear advantage when position dependent PSF are used for the iterative reconstruction. The basic work of a reconstruction method was established and further evaluations must be made in an experimental study.

  16. Photon detection efficiency of laboratory-based x-ray phase contrast imaging techniques for mammography: a Monte Carlo study

    Science.gov (United States)

    Saghamanesh, S.; Aghamiri, S. M.; Kamali-Asl, A.; Yashiro, W.

    2017-09-01

    An important challenge in real-world biomedical applications of x-ray phase contrast imaging (XPCI) techniques is the efficient use of the photon flux generated by an incoherent and polychromatic x-ray source. This efficiency can directly influence dose and exposure time and ideally should not affect the superior contrast and sensitivity of XPCI. In this paper, we present a quantitative evaluation of the photon detection efficiency of two laboratory-based XPCI methods, grating interferometry (GI) and coded-aperture (CA). We adopt a Monte Carlo approach to simulate existing prototypes of those systems, tailored for mammography applications. Our simulations were validated by means of a simple experiment performed on a CA XPCI system. Our results show that the fraction of detected photons in the standard energy range of mammography are about 1.4% and 10% for the GI and CA techniques, respectively. The simulations indicate that the design of the optical components plays an important role in the higher efficiency of CA compared to the GI method. It is shown that the use of lower absorbing materials as the substrates for GI gratings can improve its flux efficiency by up to four times. Along similar lines, we also show that an optimized and compact configuration of GI could lead to a 3.5 times higher fraction of detected counts compared to a standard and non-optimised GI implementation.

  17. Analysis of the neutrons dispersion in a semi-infinite medium based in transport theory and the Monte Carlo method

    International Nuclear Information System (INIS)

    Arreola V, G.; Vazquez R, R.; Guzman A, J. R.

    2012-10-01

    In this work a comparative analysis of the results for the neutrons dispersion in a not multiplicative semi-infinite medium is presented. One of the frontiers of this medium is located in the origin of coordinates, where a neutrons source in beam form, i.e., μο=1 is also. The neutrons dispersion is studied on the statistical method of Monte Carlo and through the unidimensional transport theory and for an energy group. The application of transport theory gives a semi-analytic solution for this problem while the statistical solution for the flow was obtained applying the MCNPX code. The dispersion in light water and heavy water was studied. A first remarkable result is that both methods locate the maximum of the neutrons distribution to less than two mean free trajectories of transport for heavy water, while for the light water is less than ten mean free trajectories of transport; the differences between both methods is major for the light water case. A second remarkable result is that the tendency of both distributions is similar in small mean free trajectories, while in big mean free trajectories the transport theory spreads to an asymptote value and the solution in base statistical method spreads to zero. The existence of a neutron current of low energy and toward the source is demonstrated, in contrary sense to the neutron current of high energy coming from the own source. (Author)

  18. Calculation of effective dose for nuclear medicine applications using an image-base body model and the Monte Carlo Method

    International Nuclear Information System (INIS)

    Yorulmaz, N.; Bozkurt, A.

    2009-01-01

    In nuclear medicine applications, the aim is to obtain diagnostic information about the organs and tissues of the patient with the help of some radiopharmaceuticals administered to him/her. Because some organs of the patient other than those under investigation will also be exposed to the radiation, it is important for radiation risk assessment to know how much radiation is received by the vital or radio-sensitive organs or tissues. In this study, an image-based body model created from the realistic images of a human is used together with the Monte Carlo code MCNP to compute the radiation doses absorbed by organs and tissues for some nuclear medicine procedures at gamma energies of 0.01, 0.015, 0.02, 0.03, 0.05, 0.1, 0.2, 0.5 and 1 MeV. Later, these values are used in conjunction with radiation weighting factors and organ weighting factors to estimate the effective dose for each diagnostic application.

  19. Scalability tests of R-GMA based Grid job monitoring system for CMS Monte Carlo data production

    CERN Document Server

    Bonacorsi, D; Field, L; Fisher, S; Grandi, C; Hobson, P R; Kyberd, P; MacEvoy, B; Nebrensky, J J; Tallini, H; Traylen, S

    2004-01-01

    High Energy Physics experiments such as CMS (Compact Muon Solenoid) at the Large Hadron Collider have unprecedented, large-scale data processing computing requirements, with data accumulating at around 1 Gbyte/s. The Grid distributed computing paradigm has been chosen as the solution to provide the requisite computing power. The demanding nature of CMS software and computing requirements, such as the production of large quantities of Monte Carlo simulated data, makes them an ideal test case for the Grid and a major driver for the development of Grid technologies. One important challenge when using the Grid for large-scale data analysis is the ability to monitor the large numbers of jobs that are being executed simultaneously at multiple remote sites. R-GMA is a monitoring and information management service for distributed resources based on the Grid Monitoring Architecture of the Global Grid Forum. In this paper we report on the first measurements of R-GMA as part of a monitoring architecture to be used for b...

  20. Atomic structure of Mg-based metallic glass investigated with neutron diffraction, reverse Monte Carlo modeling and electron microscopy.

    Science.gov (United States)

    Babilas, Rafał; Łukowiec, Dariusz; Temleitner, Laszlo

    2017-01-01

    The structure of a multicomponent metallic glass, Mg 65 Cu 20 Y 10 Ni 5 , was investigated by the combined methods of neutron diffraction (ND), reverse Monte Carlo modeling (RMC) and high-resolution transmission electron microscopy (HRTEM). The RMC method, based on the results of ND measurements, was used to develop a realistic structure model of a quaternary alloy in a glassy state. The calculated model consists of a random packing structure of atoms in which some ordered regions can be indicated. The amorphous structure was also described by peak values of partial pair correlation functions and coordination numbers, which illustrated some types of cluster packing. The N = 9 clusters correspond to the tri-capped trigonal prisms, which are one of Bernal's canonical clusters, and atomic clusters with N = 6 and N = 12 are suitable for octahedral and icosahedral atomic configurations. The nanocrystalline character of the alloy after annealing was also studied by HRTEM. The selected HRTEM images of the nanocrystalline regions were also processed by inverse Fourier transform analysis. The high-angle annular dark-field (HAADF) technique was used to determine phase separation in the studied glass after heat treatment. The HAADF mode allows for the observation of randomly distributed, dark contrast regions of about 4-6 nm. The interplanar spacing identified for the orthorhombic Mg 2 Cu crystalline phase is similar to the value of the first coordination shell radius from the short-range order.

  1. Commissioning and Validation of the First Monte Carlo Based Dose Calculation Algorithm Commercial Treatment Planning System in Mexico

    International Nuclear Information System (INIS)

    Larraga-Gutierrez, J. M.; Garcia-Garduno, O. A.; Hernandez-Bojorquez, M.; Galvan de la Cruz, O. O.; Ballesteros-Zebadua, P.

    2010-01-01

    This work presents the beam data commissioning and dose calculation validation of the first Monte Carlo (MC) based treatment planning system (TPS) installed in Mexico. According to the manufacturer specifications, the beam data commissioning needed for this model includes: several in-air and water profiles, depth dose curves, head-scatter factors and output factors (6x6, 12x12, 18x18, 24x24, 42x42, 60x60, 80x80 and 100x100 mm 2 ). Radiographic and radiochromic films, diode and ionization chambers were used for data acquisition. MC dose calculations in a water phantom were used to validate the MC simulations using comparisons with measured data. Gamma index criteria 2%/2 mm were used to evaluate the accuracy of MC calculations. MC calculated data show an excellent agreement for field sizes from 18x18 to 100x100 mm 2 . Gamma analysis shows that in average, 95% and 100% of the data passes the gamma index criteria for these fields, respectively. For smaller fields (12x12 and 6x6 mm 2 ) only 92% of the data meet the criteria. Total scatter factors show a good agreement ( 2 ) that show a error of 4.7%. MC dose calculations are accurate and precise for clinical treatment planning up to a field size of 18x18 mm 2 . Special care must be taken for smaller fields.

  2. A high-quality multilayer structure characterization method based on X-ray fluorescence and Monte Carlo simulation

    Energy Technology Data Exchange (ETDEWEB)

    Brunetti, Antonio; Golosio, Bruno [Universita degli Studi di Sassari, Dipartimento di Scienze Politiche, Scienze della Comunicazione e Ingegneria dell' Informazione, Sassari (Italy); Melis, Maria Grazia [Universita degli Studi di Sassari, Dipartimento di Storia, Scienze dell' Uomo e della Formazione, Sassari (Italy); Mura, Stefania [Universita degli Studi di Sassari, Dipartimento di Agraria e Nucleo di Ricerca sulla Desertificazione, Sassari (Italy)

    2014-11-08

    X-ray fluorescence (XRF) is a well known nondestructive technique. It is also applied to multilayer characterization, due to its possibility of estimating both composition and thickness of the layers. Several kinds of cultural heritage samples can be considered as a complex multilayer, such as paintings or decorated objects or some types of metallic samples. Furthermore, they often have rough surfaces and this makes a precise determination of the structure and composition harder. The standard quantitative XRF approach does not take into account this aspect. In this paper, we propose a novel approach based on a combined use of X-ray measurements performed with a polychromatic beam and Monte Carlo simulations. All the information contained in an X-ray spectrum is used. This approach allows obtaining a very good estimation of the sample contents both in terms of chemical elements and material thickness, and in this sense, represents an improvement of the possibility of XRF measurements. Some examples will be examined and discussed. (orig.)

  3. The first private-hospital based proton therapy center in Korea; status of the Proton Therapy Center at Samsung Medical Center.

    Science.gov (United States)

    Chung, Kwangzoo; Han, Youngyih; Kim, Jinsung; Ahn, Sung Hwan; Ju, Sang Gyu; Jung, Sang Hoon; Chung, Yoonsun; Cho, Sungkoo; Jo, Kwanghyun; Shin, Eun Hyuk; Hong, Chae-Seon; Shin, Jung Suk; Park, Seyjoon; Kim, Dae-Hyun; Kim, Hye Young; Lee, Boram; Shibagaki, Gantaro; Nonaka, Hideki; Sasai, Kenzo; Koyabu, Yukio; Choi, Changhoon; Huh, Seung Jae; Ahn, Yong Chan; Pyo, Hong Ryull; Lim, Do Hoon; Park, Hee Chul; Park, Won; Oh, Dong Ryul; Noh, Jae Myung; Yu, Jeong Il; Song, Sanghyuk; Lee, Ji Eun; Lee, Bomi; Choi, Doo Ho

    2015-12-01

    The purpose of this report is to describe the proton therapy system at Samsung Medical Center (SMC-PTS) including the proton beam generator, irradiation system, patient positioning system, patient position verification system, respiratory gating system, and operating and safety control system, and review the current status of the SMC-PTS. The SMC-PTS has a cyclotron (230 MeV) and two treatment rooms: one treatment room is equipped with a multi-purpose nozzle and the other treatment room is equipped with a dedicated pencil beam scanning nozzle. The proton beam generator including the cyclotron and the energy selection system can lower the energy of protons down to 70 MeV from the maximum 230 MeV. The multi-purpose nozzle can deliver both wobbling proton beam and active scanning proton beam, and a multi-leaf collimator has been installed in the downstream of the nozzle. The dedicated scanning nozzle can deliver active scanning proton beam with a helium gas filled pipe minimizing unnecessary interactions with the air in the beam path. The equipment was provided by Sumitomo Heavy Industries Ltd., RayStation from RaySearch Laboratories AB is the selected treatment planning system, and data management will be handled by the MOSAIQ system from Elekta AB. The SMC-PTS located in Seoul, Korea, is scheduled to begin treating cancer patients in 2015.

  4. The first private-hospital based proton therapy center in Korea; Status of the proton therapy center at Samsung Medical Center

    Energy Technology Data Exchange (ETDEWEB)

    Chung, Kwang Zoo; Han, Young Yih; Kim, Jin Sung [Dept. of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul (Korea, Republic of); and others

    2015-12-15

    The purpose of this report is to describe the proton therapy system at Samsung Medical Center (SMC-PTS) including the proton beam generator, irradiation system, patient positioning system, patient position verification system, respiratory gating system, and operating and safety control system, and review the current status of the SMC-PTS. The SMC-PTS has a cyclotron (230 MeV) and two treatment rooms: one treatment room is equipped with a multi-purpose nozzle and the other treatment room is equipped with a dedicated pencil beam scanning nozzle. The proton beam generator including the cyclotron and the energy selection system can lower the energy of protons down to 70 MeV from the maximum 230 MeV. The multi-purpose nozzle can deliver both wobbling proton beam and active scanning proton beam, and a multi-leaf collimator has been installed in the downstream of the nozzle. The dedicated scanning nozzle can deliver active scanning proton beam with a helium gas filled pipe minimizing unnecessary interactions with the air in the beam path. The equipment was provided by Sumitomo Heavy Industries Ltd., RayStation from RaySearch Laboratories AB is the selected treatment planning system, and data management will be handled by the MOSAIQ system from Elekta AB. The SMC-PTS located in Seoul, Korea, is scheduled to begin treating cancer patients in 2015.

  5. The first private-hospital based proton therapy center in Korea; Status of the proton therapy center at Samsung Medical Center

    International Nuclear Information System (INIS)

    Chung, Kwang Zoo; Han, Young Yih; Kim, Jin Sung

    2015-01-01

    The purpose of this report is to describe the proton therapy system at Samsung Medical Center (SMC-PTS) including the proton beam generator, irradiation system, patient positioning system, patient position verification system, respiratory gating system, and operating and safety control system, and review the current status of the SMC-PTS. The SMC-PTS has a cyclotron (230 MeV) and two treatment rooms: one treatment room is equipped with a multi-purpose nozzle and the other treatment room is equipped with a dedicated pencil beam scanning nozzle. The proton beam generator including the cyclotron and the energy selection system can lower the energy of protons down to 70 MeV from the maximum 230 MeV. The multi-purpose nozzle can deliver both wobbling proton beam and active scanning proton beam, and a multi-leaf collimator has been installed in the downstream of the nozzle. The dedicated scanning nozzle can deliver active scanning proton beam with a helium gas filled pipe minimizing unnecessary interactions with the air in the beam path. The equipment was provided by Sumitomo Heavy Industries Ltd., RayStation from RaySearch Laboratories AB is the selected treatment planning system, and data management will be handled by the MOSAIQ system from Elekta AB. The SMC-PTS located in Seoul, Korea, is scheduled to begin treating cancer patients in 2015

  6. Core physics design calculation of mini-type fast reactor based on Monte Carlo method

    International Nuclear Information System (INIS)

    He Keyu; Han Weishi

    2007-01-01

    An accurate physics calculation model has been set up for the mini-type sodium-cooled fast reactor (MFR) based on MCNP-4C code, then a detailed calculation of its critical physics characteristics, neutron flux distribution, power distribution and reactivity control has been carried out. The results indicate that the basic physics characteristics of MFR can satisfy the requirement and objectives of the core design. The power density and neutron flux distribution are symmetrical and reasonable. The control system is able to make a reliable reactivity balance efficiently and meets the request for long-playing operation. (authors)

  7. Integrated Building Energy Design of a Danish Office Building Based on Monte Carlo Simulation Method

    DEFF Research Database (Denmark)

    Sørensen, Mathias Juul; Myhre, Sindre Hammer; Hansen, Kasper Kingo

    2017-01-01

    The focus on reducing buildings energy consumption is gradually increasing, and the optimization of a building’s performance and maximizing its potential leads to great challenges between architects and engineers. In this study, we collaborate with a group of architects on a design project of a new...... office building located in Aarhus, Denmark. Building geometry, floor plans and employee schedules were obtained from the architects which is the basis for this study. This study aims to simplify the iterative design process that is based on the traditional trial and error method in the late design phases...

  8. Estimation of the dose deposited by electron beams in radiotherapy in voxelised phantoms using the Monte Carlo simulation platform GATE based on GEANT4 in a grid environment

    International Nuclear Information System (INIS)

    Perrot, Y.

    2011-01-01

    Radiation therapy treatment planning requires accurate determination of absorbed dose in the patient. Monte Carlo simulation is the most accurate method for solving the transport problem of particles in matter. This thesis is the first study dealing with the validation of the Monte Carlo simulation platform GATE (GEANT4 Application for Tomographic Emission), based on GEANT4 (Geometry And Tracking) libraries, for the computation of absorbed dose deposited by electron beams. This thesis aims at demonstrating that GATE/GEANT4 calculations are able to reach treatment planning requirements in situations where analytical algorithms are not satisfactory. The goal is to prove that GATE/GEANT4 is useful for treatment planning using electrons and competes with well validated Monte Carlo codes. This is demonstrated by the simulations with GATE/GEANT4 of realistic electron beams and electron sources used for external radiation therapy or targeted radiation therapy. The computed absorbed dose distributions are in agreement with experimental measurements and/or calculations from other Monte Carlo codes. Furthermore, guidelines are proposed to fix the physics parameters of the GATE/GEANT4 simulations in order to ensure the accuracy of absorbed dose calculations according to radiation therapy requirements. (author)

  9. Measurement of Dijet Production in Diffractive Deep-Inelastic Scattering with a Leading Proton at HERA

    CERN Document Server

    Aaron, F.D.

    2012-04-18

    The cross section of diffractive deep-inelastic scattering ep \\rightarrow eXp is measured, where the system X contains at least two jets and the leading final state proton is detected in the H1 Forward Proton Spectrometer. The measurement is performed for fractional proton longitudinal momentum loss xIP < 0.1 and covers the range 0.1 < |t| < 0.7 GeV2 in squared four-momentum transfer at the proton vertex and 4 < Q2 < 110 GeV2 in photon virtuality. The differential cross sections extrapolated to |t| < 1 GeV2 are in agreement with next-toleading order QCD predictions based on diffractive parton distribution functions extracted from measurements of inclusive and dijet cross sections in diffractive deep-inelastic scattering. The data are also compared with leading order Monte Carlo models.

  10. Measurement of dijet production in diffractive deep-inelastic scattering with a leading proton at HERA

    Energy Technology Data Exchange (ETDEWEB)

    Aaron, F.D.; Alexa, C.; Rotaru, M.; Stoicea, G. [National Inst. for Physics and Nuclear Engineering, Bucharest (Romania); Andreev, V.; Belousov, A.; Eliseev, A.; Fomenko, A.; Gogitidze, N.; Lebedev, A.; Malinovski, E.; Rusakov, S.; Shtarkov, L.N.; Soloviev, Y.; Vazdik, Y. [Lebedev Physical Inst., Moscow (Russian Federation); Backovic, S.; Dubak, A.; Lastovicka-Medin, G.; Picuric, I.; Raicevic, N. [Univ. of Montenegro, Faculty of Science, Podgorica (ME); Baghdasaryan, A.; Baghdasaryan, S.; Zohrabyan, H. [Yerevan Physics Inst., Yerevan (Armenia); Barrelet, E. [CNRS/IN2P3, LPNHE, Univ. Pierre et Marie Curie Paris 6, Univ. Denis Diderot Paris 7, Paris (France); Bartel, W.; Belov, P.; Brandt, G.; Brinkmann, M.; Britzger, D.; Campbell, A.J.; Eckerlin, G.; Elsen, E.; Felst, R.; Fischer, D.J.; Fleischer, M.; Gayler, J.; Ghazaryan, S.; Glazov, A.; Gouzevitch, M.; Grebenyuk, A.; Grell, B.R.; Habib, S.; Haidt, D.; Helebrant, C.; Kleinwort, C.; Kogler, R.; Kraemer, M.; Levonian, S.; Lipka, K.; List, B.; List, J.; Meyer, A.B.; Meyer, J.; Niebuhr, C.; Nowak, K.; Olsson, J.E.; Pahl, P.; Panagoulias, I.; Papadopoulou, T.; Petrukhin, A.; Piec, S.; Pitzl, D.; Schmitt, S.; Sefkow, F.; Shushkevich, S.; South, D.; Steder, M.; Wuensch, E. [DESY, Hamburg (Germany); Begzsuren, K.; Ravdandorj, T.; Tseepeldorj, B. [Inst. of Physics and Technology of the Mongolian Academy of Sciences, Ulaanbaatar (Mongolia); Bizot, J.C.; Brisson, V.; Delcourt, B.; Jacquet, M.; Pascaud, C.; Tran, T.H.; Zhang, Z.; Zomer, F. [CNRS/IN2P3, LAL, Univ. Paris-Sud, Orsay (France); Boudry, V.; Moreau, F.; Specka, A. [CNRS/IN2P3, LLR, Ecole Polytechnique, Palaiseau (France); Bozovic-Jelisavcic, I.; Mudrinic, M.; Pandurovic, M.; Smiljanic, I. [Univ. of Belgrade, Vinca Inst. of Nuclear Sciences, Belgrade (RS); Bracinik, J.; Kenyon, I.R.; Newman, P.R.; Thompson, P.D. [Univ. of Birmingham (United Kingdom); Bruncko, D.; Cerny, V.; Ferencei, J. [Slovak Academy of Sciences, Kosice (Slovakia)] [and others

    2012-04-15

    The cross section of diffractive deep-inelastic scattering ep{yields}eXp is measured, where the system X contains at least two jets and the leading final state proton is detected in the H1 Forward Proton Spectrometer. The measurement is performed for fractional proton longitudinal momentum loss x{sub P}<0.1 and covers the range 0.1< vertical stroke t vertical stroke <0.7 GeV{sup 2} in squared four-momentum transfer at the proton vertex and 4based on diffractive parton distribution functions extracted from measurements of inclusive and dijet cross sections in diffractive deep-inelastic scattering. The data are also compared with leading order Monte Carlo models. (orig.)

  11. An automated Monte-Carlo based method for the calculation of cascade summing factors

    Science.gov (United States)

    Jackson, M. J.; Britton, R.; Davies, A. V.; McLarty, J. L.; Goodwin, M.

    2016-10-01

    A versatile method has been developed to calculate cascade summing factors for use in quantitative gamma-spectrometry analysis procedures. The proposed method is based solely on Evaluated Nuclear Structure Data File (ENSDF) nuclear data, an X-ray energy library, and accurate efficiency characterisations for single detector counting geometries. The algorithm, which accounts for γ-γ, γ-X, γ-511 and γ-e- coincidences, can be applied to any design of gamma spectrometer and can be expanded to incorporate any number of nuclides. Efficiency characterisations can be derived from measured or mathematically modelled functions, and can accommodate both point and volumetric source types. The calculated results are shown to be consistent with an industry standard gamma-spectrometry software package. Additional benefits including calculation of cascade summing factors for all gamma and X-ray emissions, not just the major emission lines, are also highlighted.

  12. Free standing diamond-like carbon thin films by PLD for laser based electrons/protons acceleration

    Energy Technology Data Exchange (ETDEWEB)

    Thema, F.T.; Beukes, P.; Ngom, B.D. [UNESCO Africa Chair in Nanosciences-Nanotechnology, College of Graduate Studies, University of South Africa, Muckleneuk Ridge, PO Box 392, Pretoria (South Africa); Nanosciences African Network (NANOAFNET), iThemba LABS-National Research Foundation, 1 Old Faure Road, Somerset West, 7129, PO Box722, Western Cape Province (South Africa); Manikandan, E., E-mail: mani@tlabs.ac.za [UNESCO Africa Chair in Nanosciences-Nanotechnology, College of Graduate Studies, University of South Africa, Muckleneuk Ridge, PO Box 392, Pretoria (South Africa); Nanosciences African Network (NANOAFNET), iThemba LABS-National Research Foundation, 1 Old Faure Road, Somerset West, 7129, PO Box722, Western Cape Province (South Africa); Central Research Laboratory, Sree Balaji Medical College & Hospital (SBMCH), Chrompet, Bharath University, Chennai, 600044 (India); Maaza, M., E-mail: maaza@tlabs.ac.za [UNESCO Africa Chair in Nanosciences-Nanotechnology, College of Graduate Studies, University of South Africa, Muckleneuk Ridge, PO Box 392, Pretoria (South Africa); Nanosciences African Network (NANOAFNET), iThemba LABS-National Research Foundation, 1 Old Faure Road, Somerset West, 7129, PO Box722, Western Cape Province (South Africa)

    2015-11-05

    This study we reports for the first time on the synthesis and optical characteristics of free standing diamond-like carbon (DLC) deposited by pulsed laser deposition (PLD) onto graphene buffer layers for ultrahigh intensity laser based electron/proton acceleration applications. The fingerprint techniques of micro-Raman, UV–VIS–NIR and the IR spectroscopic investigations indicate that the suitability of such free standing DLC thin-films within the laser window and long wave infrared (LWIR) spectral range and hence their appropriateness for the targeted applications. - Highlights: • We report for the first time synthesis of free standing diamond-like carbon. • Pulsed laser deposition onto graphene buffer layers. • Fingerprint techniques of micro-Raman, UV–VIS–NIR and the IR spectroscopic investigations. • Ultrahigh intensity laser based electron/proton acceleration applications. • This material's suitable for the laser window and long wave infrared (LWIR) spectral range.

  13. Molecular Dynamics of Hexamethylbenzene at Low Temperatures: Evidence of Unconventional Magnetism Based on Rotational Motion of Protons.

    Science.gov (United States)

    Yen, Fei; Zhao, Zhenzheng; Hu, Sixia; Chen, Lang

    2017-10-23

    The types of magnetism known to date are all mainly based on contributions from electron motion. We show how rotational motion of protons (H + ) within the methyl groups in hexamethylbenzene (C 6 (CH 3 ) 6 ) also contribute significantly to the magnetic susceptibility. Starting from below 118 K, as the rotational motion of the methyl groups set in, an associated magnetic moment positive in nature due to charge of the protons renders the susceptibility to become anomalously dependent on temperature. Starting from 20 K, the susceptibility diverges with decreasing temperature indicative of spin-spin interactions between methyl groups aligned in a previously unclassified type of anti-ferromagnetic configuration. Complementary dielectric constant measurements also show the existence of magneto-dielectric coupling. Our findings allow for the study of strongly correlated systems that are based on a species that possesses much slower dynamics. © 2017 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  14. A dual resolution measurement based Monte Carlo simulation technique for detailed dose analysis of small volume organs in the skull base region

    International Nuclear Information System (INIS)

    Yeh, Chi-Yuan; Tung, Chuan-Jung; Chao, Tsi-Chain; Lin, Mu-Han; Lee, Chung-Chi

    2014-01-01

    The purpose of this study was to examine dose distribution of a skull base tumor and surrounding critical structures in response to high dose intensity-modulated radiosurgery (IMRS) with Monte Carlo (MC) simulation using a dual resolution sandwich phantom. The measurement-based Monte Carlo (MBMC) method (Lin et al., 2009) was adopted for the study. The major components of the MBMC technique involve (1) the BEAMnrc code for beam transport through the treatment head of a Varian 21EX linear accelerator, (2) the DOSXYZnrc code for patient dose simulation and (3) an EPID-measured efficiency map which describes non-uniform fluence distribution of the IMRS treatment beam. For the simulated case, five isocentric 6 MV photon beams were designed to deliver a total dose of 1200 cGy in two fractions to the skull base tumor. A sandwich phantom for the MBMC simulation was created based on the patient's CT scan of a skull base tumor [gross tumor volume (GTV)=8.4 cm 3 ] near the right 8th cranial nerve. The phantom, consisted of a 1.2-cm thick skull base region, had a voxel resolution of 0.05×0.05×0.1 cm 3 and was sandwiched in between 0.05×0.05×0.3 cm 3 slices of a head phantom. A coarser 0.2×0.2×0.3 cm 3 single resolution (SR) phantom was also created for comparison with the sandwich phantom. A particle history of 3×10 8 for each beam was used for simulations of both the SR and the sandwich phantoms to achieve a statistical uncertainty of <2%. Our study showed that the planning target volume (PTV) receiving at least 95% of the prescribed dose (VPTV95) was 96.9%, 96.7% and 99.9% for the TPS, SR, and sandwich phantom, respectively. The maximum and mean doses to large organs such as the PTV, brain stem, and parotid gland for the TPS, SR and sandwich MC simulations did not show any significant difference; however, significant dose differences were observed for very small structures like the right 8th cranial nerve, right cochlea, right malleus and right semicircular

  15. The underlying event in proton-proton collisions

    Energy Technology Data Exchange (ETDEWEB)

    Bechtel, F.

    2009-05-15

    In this thesis, studies of the underlying event in proton-proton collisions at a center-of-mass energy of {radical}(s) = 10 TeV are presented. Crucial ingredient to underlying event models are multiple parton-parton scatters in single proton-proton collisions. The feasibility of measuring the underlying event was investigated with the Compact Muon Solenoid (CMS) detector at the Large Hadron Collider (LHC) using charged particles and charged-particle jets. Systematic uncertainties of the underlying event measurement due to detector misalignment and imperfect track reconstruction are found to be negligible after {integral}Ldt=1 pb{sup -1} of data are available. Different model predictions are compared with each other using fully simulated Monte Carlo samples. It is found, that distinct models differ strongly enough to tell them apart with early data. (orig.)

  16. Simulation of Ni-63 based nuclear micro battery using Monte Carlo modeling

    International Nuclear Information System (INIS)

    Kim, Tae Ho; Kim, Ji Hyun

    2013-01-01

    The radioisotope batteries have an energy density of 100-10000 times greater than chemical batteries. Also, Li ion battery has the fundamental problems such as short life time and requires recharge system. In addition to these things, the existing batteries are hard to operate at internal human body, national defense arms or space environment. Since the development of semiconductor process and materials technology, the micro device is much more integrated. It is expected that, based on new semiconductor technology, the conversion device efficiency of betavoltaic battery will be highly increased. Furthermore, the radioactivity from the beta particle cannot penetrate a skin of human body, so it is safer than Li battery which has the probability to explosion. In the other words, the interest for radioisotope battery is increased because it can be applicable to an artificial internal organ power source without recharge and replacement, micro sensor applied to arctic and special environment, small size military equipment and space industry. However, there is not enough data for beta particle fluence from radioisotope source using nuclear battery. Beta particle fluence directly influences on battery efficiency and it is seriously affected by radioisotope source thickness because of self-absorption effect. Therefore, in this article, we present a basic design of Ni-63 nuclear battery and simulation data of beta particle fluence with various thickness of radioisotope source and design of battery

  17. Predictions of diffractive cross sections in proton-proton collisions

    Energy Technology Data Exchange (ETDEWEB)

    Goulianos, Konstantin [Rockefeller University, 1230 York Avenue, New York, NY 10065 (United States)

    2013-04-15

    We review our pre-LHC predictions of the total, elastic, total-inelastic, and diffractive components of proton-proton cross sections at high energies, expressed in the form of unitarized expressions based on a special parton-model approach to diffraction employing inclusive proton parton distribution functions and QCD color factors and compare with recent LHC results.

  18. A hospital-based proton linac for neutron therapy and radioisotope production

    International Nuclear Information System (INIS)

    Lennox, A.J.

    1988-10-01

    Fermilab's Alvarez proton linac has been used routinely for neutron therapy since 1976. The Neutron Therapy Facility (NTF) operates in a mode parasitic to the laboratory's high energy physics program, which uses the linac as an injector for a synchrotron. Parasitic operation is possible because the linac delivers /approximately/1.2 /times/ 10 13 protons per pulse at a 15 Hz rate, while the high energy physics program requires beam at a rate not greater than 0.5 Hz. Protons not needed for physics experiments strike a beryllium target to produce neutrons for neutron therapy. Encouraging clinical results from NTF have led to a study of the issues involved in providing hospitals with a neutron beam of the type available at Fermilab. This paper describes the issues addressed by that study. 12 refs., 1 fig., 1 tab

  19. Selection of voxel size and photon number in voxel-based Monte Carlo method: criteria and applications

    Science.gov (United States)

    Li, Dong; Chen, Bin; Ran, Wei Yu; Wang, Guo Xiang; Wu, Wen Juan

    2015-09-01

    The voxel-based Monte Carlo method (VMC) is now a gold standard in the simulation of light propagation in turbid media. For complex tissue structures, however, the computational cost will be higher when small voxels are used to improve smoothness of tissue interface and a large number of photons are used to obtain accurate results. To reduce computational cost, criteria were proposed to determine the voxel size and photon number in 3-dimensional VMC simulations with acceptable accuracy and computation time. The selection of the voxel size can be expressed as a function of tissue geometry and optical properties. The photon number should be at least 5 times the total voxel number. These criteria are further applied in developing a photon ray splitting scheme of local grid refinement technique to reduce computational cost of a nonuniform tissue structure with significantly varying optical properties. In the proposed technique, a nonuniform refined grid system is used, where fine grids are used for the tissue with high absorption and complex geometry, and coarse grids are used for the other part. In this technique, the total photon number is selected based on the voxel size of the coarse grid. Furthermore, the photon-splitting scheme is developed to satisfy the statistical accuracy requirement for the dense grid area. Result shows that local grid refinement technique photon ray splitting scheme can accelerate the computation by 7.6 times (reduce time consumption from 17.5