WorldWideScience

Sample records for multi-phase ramp2 computer

  1. Computer Based Porosity Design by Multi Phase Topology Optimization

    Science.gov (United States)

    Burblies, Andreas; Busse, Matthias

    2008-02-01

    A numerical simulation technique called Multi Phase Topology Optimization (MPTO) based on finite element method has been developed and refined by Fraunhofer IFAM during the last five years. MPTO is able to determine the optimum distribution of two or more different materials in components under thermal and mechanical loads. The objective of optimization is to minimize the component's elastic energy. Conventional topology optimization methods which simulate adaptive bone mineralization have got the disadvantage that there is a continuous change of mass by growth processes. MPTO keeps all initial material concentrations and uses methods adapted from molecular dynamics to find energy minimum. Applying MPTO to mechanically loaded components with a high number of different material densities, the optimization results show graded and sometimes anisotropic porosity distributions which are very similar to natural bone structures. Now it is possible to design the macro- and microstructure of a mechanical component in one step. Computer based porosity design structures can be manufactured by new Rapid Prototyping technologies. Fraunhofer IFAM has applied successfully 3D-Printing and Selective Laser Sintering methods in order to produce very stiff light weight components with graded porosities calculated by MPTO.

  2. Application of the level set method for multi-phase flow computation in fusion engineering

    International Nuclear Information System (INIS)

    Luo, X-Y.; Ni, M-J.; Ying, A.; Abdou, M.

    2006-01-01

    Numerical simulation of multi-phase flow is essential to evaluate the feasibility of a liquid protection scheme for the power plant chamber. The level set method is one of the best methods for computing and analyzing the motion of interface among the multi-phase flow. This paper presents a general formula for the second-order projection method combined with the level set method to simulate unsteady incompressible multi-phase flow with/out phase change flow encountered in fusion science and engineering. The third-order ENO scheme and second-order semi-implicit Crank-Nicholson scheme is used to update the convective and diffusion term. The numerical results show this method can handle the complex deformation of the interface and the effect of liquid-vapor phase change will be included in the future work

  3. Robust second-order scheme for multi-phase flow computations

    Science.gov (United States)

    Shahbazi, Khosro

    2017-06-01

    A robust high-order scheme for the multi-phase flow computations featuring jumps and discontinuities due to shock waves and phase interfaces is presented. The scheme is based on high-order weighted-essentially non-oscillatory (WENO) finite volume schemes and high-order limiters to ensure the maximum principle or positivity of the various field variables including the density, pressure, and order parameters identifying each phase. The two-phase flow model considered besides the Euler equations of gas dynamics consists of advection of two parameters of the stiffened-gas equation of states, characterizing each phase. The design of the high-order limiter is guided by the findings of Zhang and Shu (2011) [36], and is based on limiting the quadrature values of the density, pressure and order parameters reconstructed using a high-order WENO scheme. The proof of positivity-preserving and accuracy is given, and the convergence and the robustness of the scheme are illustrated using the smooth isentropic vortex problem with very small density and pressure. The effectiveness and robustness of the scheme in computing the challenging problem of shock wave interaction with a cluster of tightly packed air or helium bubbles placed in a body of liquid water is also demonstrated. The superior performance of the high-order schemes over the first-order Lax-Friedrichs scheme for computations of shock-bubble interaction is also shown. The scheme is implemented in two-dimensional space on parallel computers using message passing interface (MPI). The proposed scheme with limiter features approximately 50% higher number of inter-processor message communications compared to the corresponding scheme without limiter, but with only 10% higher total CPU time. The scheme is provably second-order accurate in regions requiring positivity enforcement and higher order in the rest of domain.

  4. Field-scale multi-phase LNAPL remediation: Validating a new computational framework against sequential field pilot trials.

    Science.gov (United States)

    Sookhak Lari, Kaveh; Johnston, Colin D; Rayner, John L; Davis, Greg B

    2018-03-05

    Remediation of subsurface systems, including groundwater, soil and soil gas, contaminated with light non-aqueous phase liquids (LNAPLs) is challenging. Field-scale pilot trials of multi-phase remediation were undertaken at a site to determine the effectiveness of recovery options. Sequential LNAPL skimming and vacuum-enhanced skimming, with and without water table drawdown were trialled over 78days; in total extracting over 5m 3 of LNAPL. For the first time, a multi-component simulation framework (including the multi-phase multi-component code TMVOC-MP and processing codes) was developed and applied to simulate the broad range of multi-phase remediation and recovery methods used in the field trials. This framework was validated against the sequential pilot trials by comparing predicted and measured LNAPL mass removal rates and compositional changes. The framework was tested on both a Cray supercomputer and a cluster. Simulations mimicked trends in LNAPL recovery rates (from 0.14 to 3mL/s) across all remediation techniques each operating over periods of 4-14days over the 78day trial. The code also approximated order of magnitude compositional changes of hazardous chemical concentrations in extracted gas during vacuum-enhanced recovery. The verified framework enables longer term prediction of the effectiveness of remediation approaches allowing better determination of remediation endpoints and long-term risks. Copyright © 2017 Commonwealth Scientific and Industrial Research Organisation. Published by Elsevier B.V. All rights reserved.

  5. Time-resolved computed tomography of the liver: retrospective, multi-phase image reconstruction derived from volumetric perfusion imaging

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, Michael A.; Kartalis, Nikolaos; Aspelin, Peter; Albiin, Nils; Brismar, Torkel B. [Karolinska University Hospital, Division of Medical Imaging and Technology, Department of Clinical Science, Intervention and Technology (CLINTEC), Karolinska Institutet, Stockholm (Sweden); Leidner, Bertil; Svensson, Anders [Karolinska University Hospital, Division of Medical Imaging and Technology, Department of Clinical Science, Intervention and Technology (CLINTEC), Karolinska Institutet, Stockholm (Sweden); Karolinska University Hospital Huddinge, Department of Radiology, Stockholm (Sweden)

    2014-01-15

    To assess feasibility and image quality (IQ) of a new post-processing algorithm for retrospective extraction of an optimised multi-phase CT (time-resolved CT) of the liver from volumetric perfusion imaging. Sixteen patients underwent clinically indicated perfusion CT using 4D spiral mode of dual-source 128-slice CT. Three image sets were reconstructed: motion-corrected and noise-reduced (MCNR) images derived from 4D raw data; maximum and average intensity projections (time MIP/AVG) of the arterial/portal/portal-venous phases and all phases (total MIP/ AVG) derived from retrospective fusion of dedicated MCNR split series. Two readers assessed the IQ, detection rate and evaluation time; one reader assessed image noise and lesion-to-liver contrast. Time-resolved CT was feasible in all patients. Each post-processing step yielded a significant reduction of image noise and evaluation time, maintaining lesion-to-liver contrast. Time MIPs/AVGs showed the highest overall IQ without relevant motion artefacts and best depiction of arterial and portal/portal-venous phases respectively. Time MIPs demonstrated a significantly higher detection rate for arterialised liver lesions than total MIPs/AVGs and the raw data series. Time-resolved CT allows data from volumetric perfusion imaging to be condensed into an optimised multi-phase liver CT, yielding a superior IQ and higher detection rate for arterialised liver lesions than the raw data series. (orig.)

  6. AOI 1— COMPUTATIONAL ENERGY SCIENCES:MULTIPHASE FLOW RESEARCH High-fidelity multi-phase radiation module for modern coal combustion systems

    Energy Technology Data Exchange (ETDEWEB)

    Modest, Michael

    2013-11-15

    The effects of radiation in particle-laden flows were the object of the present research. The presence of particles increases optical thickness substantially, making the use of the “optically thin” approximation in most cases a very poor assumption. However, since radiation fluxes peak at intermediate optical thicknesses, overall radiative effects may not necessarily be stronger than in gas combustion. Also, the spectral behavior of particle radiation properties is much more benign, making spectral models simpler (and making the assumption of a gray radiator halfway acceptable, at least for fluidized beds when gas radiation is not large). On the other hand, particles scatter radiation, making the radiative transfer equation (RTE) much more di fficult to solve. The research carried out in this project encompassed three general areas: (i) assessment of relevant radiation properties of particle clouds encountered in fluidized bed and pulverized coal combustors, (ii) development of proper spectral models for gas–particulate mixtures for various types of two-phase combustion flows, and (iii) development of a Radiative Transfer Equation (RTE) solution module for such applications. The resulting models were validated against artificial cases since open literature experimental data were not available. The final models are in modular form tailored toward maximum portability, and were incorporated into two research codes: (i) the open-source CFD code OpenFOAM, which we have extensively used in our previous work, and (ii) the open-source multi-phase flow code MFIX, which is maintained by NETL.

  7. Three-dimensional multi-phase flow computational fluid dynamics model for analysis of transport phenomena and thermal stresses in PEM fuel cells

    Energy Technology Data Exchange (ETDEWEB)

    Maher, A.R.; Al-Baghdadi, S. [International Technological Univ., London (United Kingdom). Dept. of Mechanical Engineering; Haroun, A.K.; Al-Janabi, S. [Babylon Univ., Babylon (Iraq). Dept. of Mechanical Engineering

    2007-07-01

    Fuel cell technology is expected to play an important role in meeting the growing demand for distributed generation because it can convert the chemical energy of a clean fuel directly into electrical energy. An operating fuel cell has varying local conditions of temperature, humidity, and power generation across the active area of the fuel cell in 3D. This paper presented a model that was developed to improve the basic understanding of the transport phenomena and thermal stresses in PEM fuel cells, and to investigate the behaviour of polymer membrane under hygro and thermal stresses during the cell operation. This comprehensive 3D, multiphase, non-isothermal model accounts for the major transport phenomena in a PEM fuel cell, notably convective and diffusive heat and mass transfer; electrode kinetics; transport and phase change mechanism of water; and potential fields. The model accounts for the liquid water flux inside the gas diffusion layers by viscous and capillary forces and can therefore predict the amount of liquid water inside the gas diffusion layers. This study also investigated the key parameters affecting fuel cell performance including geometry, materials and operating conditions. The model considers the many interacting, complex electrochemical, transport phenomena, thermal stresses and deformation that cannot be studied experimentally. It was concluded that the model can provide a computer-aided tool for the design and optimization of future fuel cells with much higher power density and lower cost. 21 refs., 2 tabs., 14 figs.

  8. Multi-phase chemistry in process simulation - MASIT04 (VISTA)

    Energy Technology Data Exchange (ETDEWEB)

    Brink, A.; Li Bingzhi; Hupa, M. (Aabo Akademi University, Combustion and Materials Chemistry, Turku (Finland)) (and others)

    2008-07-01

    A new generation of process models has been developed by using advanced multi-phase thermochemistry. The generality of the thermodynamic free energy concept enables use of common software tools for high and low temperature processes. Reactive multi-phase phenomena are integrated to advanced simulation procedures by using local equilibrium or constrained state free energy computation. The high-temperature applications include a process model for the heat recovery of copper flash smelting and coupled models for converter and bloom casting operations in steel-making. Wet suspension models are developed for boiler and desalination water chemistry, flash evaporation of black liquor and for selected fibre-line and paper-making processes. The simulation combines quantitative physical and chemical data from reactive flows to form their visual images, thus providing efficient tools for engineering design and industrial decision-making. Economic impacts are seen as both better process operations and improved end products. The software tools developed are internationally commercialised and being used to support Finnish process technology exports. (orig.)

  9. Multi-phase flow monitoring with electrical impedance tomography using level set based method

    International Nuclear Information System (INIS)

    Liu, Dong; Khambampati, Anil Kumar; Kim, Sin; Kim, Kyung Youn

    2015-01-01

    Highlights: • LSM has been used for shape reconstruction to monitor multi-phase flow using EIT. • Multi-phase level set model for conductivity is represented by two level set functions. • LSM handles topological merging and breaking naturally during evolution process. • To reduce the computational time, a narrowband technique was applied. • Use of narrowband and optimization approach results in efficient and fast method. - Abstract: In this paper, a level set-based reconstruction scheme is applied to multi-phase flow monitoring using electrical impedance tomography (EIT). The proposed scheme involves applying a narrowband level set method to solve the inverse problem of finding the interface between the regions having different conductivity values. The multi-phase level set model for the conductivity distribution inside the domain is represented by two level set functions. The key principle of the level set-based method is to implicitly represent the shape of interface as the zero level set of higher dimensional function and then solve a set of partial differential equations. The level set-based scheme handles topological merging and breaking naturally during the evolution process. It also offers several advantages compared to traditional pixel-based approach. Level set-based method for multi-phase flow is tested with numerical and experimental data. It is found that level set-based method has better reconstruction performance when compared to pixel-based method

  10. Optimization of multi-phase compressible lattice Boltzmann codes on massively parallel multi-core systems

    NARCIS (Netherlands)

    Biferale, L.; Mantovani, F.; Pivanti, M.; Pozzati, F.; Sbragaglia, M.; Schifano, S.F.; Toschi, F.; Tripiccione, R.

    2011-01-01

    We develop a Lattice Boltzmann code for computational fluid-dynamics and optimize it for massively parallel systems based on multi-core processors. Our code describes 2D multi-phase compressible flows. We analyze the performance bottlenecks that we find as we gradually expose a larger fraction of

  11. Advancement and Application of Multi-Phase CFD Modeling to High Speed Supercavitating Flows

    Science.gov (United States)

    2013-08-13

    October 2008 - December 2013 4. TITLE AND SUBTITLE Advancement and Application of Multi-Phase CFD Modeling to High Speed Supercavitating Flows...influence cavity hysteresis behavior. These observations are used to guide improved supercavitating -vehicle analyses including numerical predictions...experiments, and modeling 15. SUBJECT TERMS supercavitation , computational fluid dynamics, multiphase flow 16. SECURITY CLASSIFICATION OF: a

  12. Surface Tension of Multi-phase Flow with Multiple Junctions Governed by the Variational Principle

    International Nuclear Information System (INIS)

    Matsutani, Shigeki; Nakano, Kota; Shinjo, Katsuhiko

    2011-01-01

    We explore a computational model of an incompressible fluid with a multi-phase field in three-dimensional Euclidean space. By investigating an incompressible fluid with a two-phase field geometrically, we reformulate the expression of the surface tension for the two-phase field found by Lafaurie et al. (J Comput Phys 113:134–147, 1994) as a variational problem related to an infinite dimensional Lie group, the volume-preserving diffeomorphism. The variational principle to the action integral with the surface energy reproduces their Euler equation of the two-phase field with the surface tension. Since the surface energy of multiple interfaces even with singularities is not difficult to be evaluated in general and the variational formulation works for every action integral, the new formulation enables us to extend their expression to that of a multi-phase (N-phase, N ≥ 2) flow and to obtain a novel Euler equation with the surface tension of the multi-phase field. The obtained Euler equation governs the equation for motion of the multi-phase field with different surface tension coefficients without any difficulties for the singularities at multiple junctions. In other words, we unify the theory of multi-phase fields which express low dimensional interface geometry and the theory of the incompressible fluid dynamics on the infinite dimensional geometry as a variational problem. We apply the equation to the contact angle problems at triple junctions. We computed the fluid dynamics for a two-phase field with a wall numerically and show the numerical computational results that for given surface tension coefficients, the contact angles are generated by the surface tension as results of balances of the kinematic energy and the surface energy.

  13. Industrial applications of multi-functional, multi-phase reactors

    NARCIS (Netherlands)

    Harmsen, G.J.; Chewter, L.A.

    1999-01-01

    To reveal trends in the design and operation of multi-functional, multi-phase reactors, this paper describes, in historical sequence, three industrial applications of multi-functional, multi-phase reactors developed and operated by Shell Chemicals during the last five decades. For each case, we

  14. Is it possible to design universal multi-phase flow analyzer?

    International Nuclear Information System (INIS)

    Ivanov Kolev, N.

    2005-01-01

    Transient 3D-multiphase flows consisting of many chemical constituents in nature and technology (Figs. 1 and 2) are the common case of flows. In many technical applications we have to do with particular realization of the multi-phase flows like steady state flows, or single component flows or single phase flows etc. Engineers and scientists created hundreds of computer codes for description of more or less specific realizations of multi-phase flows. If one compares the structure of these codes one is astonished by the waste of the human resources for programming repeating model elements like equations of state, friction lows in variety of geometry, heat transfer coefficients, mathematical equation solvers, data handling procedures, graphical environment etc. It is hardly to expect, that the best solution for the specific sub-phenomenon is available in all codes. Looking in other branches of the technology like computer chips production we realize that the revolutionary idea of having common ''chips'' within complex applications is very far from its practical realization in the computational multi-phase flow dynamics. Following this line of arguments I expressed several times in my publications explicitly or implicitly the idea, that it is possible to create a universal multi-phase flow analyzer in the sense of computer architecture, that is capable to absorb the adequate multi-phase knowledge data base specified in Appendix 1. The subject of this paper is to summarize some of the main ideas, some of them already realized by this author, on the way of creating such computer code architecture, to illustrate haw they work, and to make an outlook regarding what are the challenges in the future developments. We confine deliberately our attention to the solution of the so called local volume and time averaged system of PDE's for a simple reason: Direct numerical resolution of interacting fluids is possible as demonstrated for small scales by many researchers, but for

  15. The comparison of MSCT multi-phase scan features between benign prostatic hyperplasia and prostate cancer

    International Nuclear Information System (INIS)

    Liu Jingang; Wang Xizhen; Niu Qingliang; Lu Hongkai; Wang Bin

    2009-01-01

    Objective: To investigate the multi-phase contrast-enhanced features of multi-slice computed tomography (MSCT) of benign prostatic hyperplasia (BPH) and prostate cancer (PCa). Methods: Thirty-five BPH and twenty- seven PCa were examined with multi-phase contrast-enhanced MSCT scan. The peak time, maximum attenuation value (MAV) and time density curve (TDC) were recorded, and the slope of the contrast media uptake curve was calculated. Result: Significant differences between BPH and PCa in the type of the curves and the peak time were observed (P<0.01). The slopes of BPH and PCa were 0.45+0.25 and 0.7 6+0.34 respectively, the slope of PCa was higher than that of BPH (P<0.05). MAVs of BPH and PCa were (44.057±10.261) HU and (46.778±11.140) HU respectively, and there was no significant difference between them (P>0.05). Conclusion: The multi-phase MSCT scan can reflect the blood supply and enhancement characters of BPH and PCa, which are important in detection and differential diagnosis of the prostate diseases. (authors)

  16. Multi-phase alternative current machine winding design | Khan ...

    African Journals Online (AJOL)

    ... single-phase to 18-phase excitation. Experimental results of a five-phase induction machine supplied from a static five-phase supply are provided to support the proposed design. Keywords: AC machine, Multi-phase machine, Stator winding, Five-phase. International Journal of Engineering, Science and Technology, Vol.

  17. Numerical simulation of 3-D incompressible, multi-phase flows over cavitating projectiles

    Energy Technology Data Exchange (ETDEWEB)

    Owis, F.M.; Nayfeh, A.H. [Blacksburg State University, Dept. of Engineering Science and Mechanics, MC 0219, Virginia Polytechnic Institute, VA (United States)

    2004-04-01

    The hydrodynamic cavitation over axisymmetric projectiles is computed using the unsteady incompressible Navier-Stokes equations for multi-fluid elements. The governing equations are discretized on a structured grid using an upwind difference scheme with flux limits. A preconditioning dual-time stepping method is used for the unsteady computations. The Eigen-system is derived for the Jacobian matrices. This Eigen-system is suitable for high-density ratio multi-fluid flows and it provides high numerical stability and fast convergence. This method can be used to compute single- as well as multi-phase flows. Cavitating flows over projectiles with different geometries are computed and the results are in good agreement with available experimental data and other published computations. (authors)

  18. Multi-phase physicochemical modeling of soil-cementitious material interaction

    International Nuclear Information System (INIS)

    Nakarai, Kenichiro; Ishida, Tetsuya; Maekawa, Koichi

    2005-01-01

    Multi-phase physicochemical modeling based on thermodynamic approach is studied on gel and capillary pores of nano-micrometers and large voids of micro-millimeters among soil foundation. A computational method about transportation of moisture and ions in pore structure for simulating concrete performance was extended for predicting time-dependent material properties of cemented soil. The proposed model was verified with experimental results of cement hydration, change of relative humidity and leaching of calcium ion from cement hydrate to underground water. (author)

  19. Energy-dissipation-model for metallurgical multi-phase-systems

    Energy Technology Data Exchange (ETDEWEB)

    Mavrommatis, K.T. [Rheinisch-Westfaelische Technische Hochschule Aachen, Aachen (Germany)

    1996-12-31

    Entropy production in real processes is directly associated with the dissipation of energy. Both are potential measures for the proceed of irreversible processes taking place in metallurgical systems. Many of these processes in multi-phase-systems could then be modelled on the basis of the energy-dissipation associated with. As this entity can often be estimated using very simple assumptions from first principles, the evolution of an overall measure of systems behaviour can be studied constructing an energy-dissipation -based model of the system. In this work a formulation of this concept, the Energy-Dissipation-Model (EDM), for metallurgical multi-phase-systems is given. Special examples are studied to illustrate the concept, and benefits as well as the range of validity are shown. This concept might be understood as complement to usual CFD-modelling of complex systems on a more abstract level but reproducing essential attributes of complex metallurgical systems. (author)

  20. Energy-dissipation-model for metallurgical multi-phase-systems

    Energy Technology Data Exchange (ETDEWEB)

    Mavrommatis, K T [Rheinisch-Westfaelische Technische Hochschule Aachen, Aachen (Germany)

    1997-12-31

    Entropy production in real processes is directly associated with the dissipation of energy. Both are potential measures for the proceed of irreversible processes taking place in metallurgical systems. Many of these processes in multi-phase-systems could then be modelled on the basis of the energy-dissipation associated with. As this entity can often be estimated using very simple assumptions from first principles, the evolution of an overall measure of systems behaviour can be studied constructing an energy-dissipation -based model of the system. In this work a formulation of this concept, the Energy-Dissipation-Model (EDM), for metallurgical multi-phase-systems is given. Special examples are studied to illustrate the concept, and benefits as well as the range of validity are shown. This concept might be understood as complement to usual CFD-modelling of complex systems on a more abstract level but reproducing essential attributes of complex metallurgical systems. (author)

  1. Multi-phase-field method for surface tension induced elasticity

    Science.gov (United States)

    Schiedung, Raphael; Steinbach, Ingo; Varnik, Fathollah

    2018-01-01

    A method, based on the multi-phase-field framework, is proposed that adequately accounts for the effects of a coupling between surface free energy and elastic deformation in solids. The method is validated via a number of analytically solvable problems. In addition to stress states at mechanical equilibrium in complex geometries, the underlying multi-phase-field framework naturally allows us to account for the influence of surface energy induced stresses on phase transformation kinetics. This issue, which is of fundamental importance on the nanoscale, is demonstrated in the limit of fast diffusion for a solid sphere, which melts due to the well-known Gibbs-Thompson effect. This melting process is slowed down when coupled to surface energy induced elastic deformation.

  2. Numerical simulation of complex multi-phase fluid of casting process and its applications

    Directory of Open Access Journals (Sweden)

    CHEN Li-liang

    2006-05-01

    Full Text Available The fluid of casting process is a typical kind of multi-phase flow. Actually, many casting phenomena have close relationship with the multi-phase flow, such as molten metal filling process, air entrapment, slag movement, venting process of die casting, gas escaping of lost foam casting and so on. Obviously, in order to analyze these phenomena accurately, numerical simulation of the multi-phase fluid is necessary. Unfortunately, so far, most of the commercial casting simulation systems do not have the ability of multi-phase flow modeling due to the difficulty in the multi-phase flow calculation. In the paper, Finite Different Method (FDM technique was adopt to solve the multi-phase fluid model. And a simple object of the muiti-phase fluid was analyzed to obtain the fluid rates of the liquid phase and the entrapped air phase.

  3. Design of multi-phase dynamic chemical networks

    Science.gov (United States)

    Chen, Chenrui; Tan, Junjun; Hsieh, Ming-Chien; Pan, Ting; Goodwin, Jay T.; Mehta, Anil K.; Grover, Martha A.; Lynn, David G.

    2017-08-01

    Template-directed polymerization reactions enable the accurate storage and processing of nature's biopolymer information. This mutualistic relationship of nucleic acids and proteins, a network known as life's central dogma, is now marvellously complex, and the progressive steps necessary for creating the initial sequence and chain-length-specific polymer templates are lost to time. Here we design and construct dynamic polymerization networks that exploit metastable prion cross-β phases. Mixed-phase environments have been used for constructing synthetic polymers, but these dynamic phases emerge naturally from the growing peptide oligomers and create environments suitable both to nucleate assembly and select for ordered templates. The resulting templates direct the amplification of a phase containing only chain-length-specific peptide-like oligomers. Such multi-phase biopolymer dynamics reveal pathways for the emergence, self-selection and amplification of chain-length- and possibly sequence-specific biopolymers.

  4. Universality Results for Multi-phase Hele-Shaw Flows

    Science.gov (United States)

    Daripa, Prabir

    2013-03-01

    Saffman-Taylor instability is a well known viscosity driven instability of an interface separating two immiscible fluids. We study linear stability of displacement processes in a Hele-Shaw cell involving an arbitrary number of immiscible fluid phases. This is a problem involving many interfaces. Universal stability results have been obtained for this multi-phase immiscible flow in the sense that the results hold for arbitrary number of interfaces. These stability results have been applied to design displacement processes that are considerably less unstable than the pure Saffman-Taylor case. In particular, we derive universal formula which gives specific values of the viscosities of the fluid layers corresponding to smallest unstable band. Other similar universal results will also be presented. The talk is based on the following paper. This work was supported by the Qatar National Research Fund (a member of The Qatar Foundation).

  5. Segmented motor drive - with multi-phase induction motor

    DEFF Research Database (Denmark)

    Bendixen, Flemming Buus

    of the induction motor is set up. The model is able to calculate dynamical electric, magnetic and mechanic state variables, but initially it is used to calculate static characteristics in motors with different number of phases and different voltage supply shapes. This analysis show i.e. that the efficiency....... The multi-phase motor is selected for further analysis. The project is limited to examine if increasing the number of phases can improve the characteristics for induction motor drives. In the literature it is demonstrated that torque production in a six-phase motor can be increased, if a 3rd harmonic......This PhD project commences in modulation of motor drives, i.e. having the advantage of reducing the number of variants and improves the system reliability at error situations. Four different motor drive topologies with modular construction as common denominator are compared on a general level...

  6. Multi-phase outflows as probes of AGN accretion history

    Science.gov (United States)

    Nardini, Emanuele; Zubovas, Kastytis

    2018-05-01

    Powerful outflows with a broad range of properties (such as velocity, ionization, radial scale and mass loss rate) represent a key feature of active galactic nuclei (AGN), even more so since they have been simultaneously revealed also in individual objects. Here we revisit in a simple analytical framework the recent remarkable cases of two ultraluminous infrared quasars, IRAS F11119+3257 and Mrk 231, which allow us to investigate the physical connection between multi-phase AGN outflows across the ladder of distance from the central supermassive black hole (SMBH). We argue that any major deviations from the standard outflow propagation models might encode unique information on the past SMBH accretion history, and briefly discuss how this could help address some controversial aspects of the current picture of AGN feedback.

  7. Microstructural modelling of nuclear graphite using multi-phase models

    International Nuclear Information System (INIS)

    Berre, C.; Fok, S.L.; Marsden, B.J.; Mummery, P.M.; Marrow, T.J.; Neighbour, G.B.

    2008-01-01

    This paper presents a new modelling technique using three-dimensional multi-phase finite element models in which meshes representing the microstructure of thermally oxidised nuclear graphite were generated from X-ray micro-tomography images. The density of the material was related to the image greyscale using Beer-Lambert's law, and multiple phases could thus be defined. The local elastic and non-linear properties of each phase were defined as a function of density and changes in Young's modulus, tensile and compressive strength with thermal oxidation were calculated. Numerical predictions compared well with experimental data and with other numerical results obtained using two-phase models. These models were found to be more representative of the actual microstructure of the scanned material than two-phase models and, possibly because of pore closure occurring during compression, compressive tests were also predicted to be less sensitive to the microstructure geometry than tensile tests

  8. Numerical Simulation of the Motion of Charged Suspended Particle in Multi-Phase Flow

    International Nuclear Information System (INIS)

    Abd-El Khalek, M.M.

    1998-01-01

    A method for computing Numerical simulation of the motion of charged suspended particle in multi-phase flow between two-long parallel plates is described in detail. The equation of motion of a suspended particle was suggested by Closkin. The equations of motion are reduced to ordinary differential equations by similarity transformations and solved numerically by using the Runge-Kutta method. The trajectories of particles are calculated by integrating the equation of motion of a single particle. Numerical solutions of the resulting ordinary differential equations provide velocity distributions for both fluid and solid phases and density distributions for the solid. The present simulation requires some empirical parameters concerning the collision of the particles with the wall. Some typical results for both fluid and particle phases and density distributions of the particles are presented graphically

  9. Numerical simulation of the motion of charged suspended particle in multi-phase flow

    Energy Technology Data Exchange (ETDEWEB)

    Abd Elkhalek, M M [Nuclear Research Center-Atomic Energy Authority, Cairo (Egypt)

    1997-12-31

    A method for computing numerical simulation of the motion of charged suspended particle in multi-phase flow between two-long parallel plates is described in detail. The equation of motion of a suspended particle was suggested by closkin. The equations of motion are reduced to ordinary differential equations by similarity transformations and solved numerically by using Runge-Kutta method. The trajectories of particles are calculated by integrating the equation of motion of a single particle. Numerical solutions of the resulting ordinary differential equations provide velocity distributions for both fluid and solid phases and density distributions for the solid. The present simulation requires some empirical parameters concerning the collision of the particles with the wall. Some typical results for both fluid and particle phases and density distributions of the particles are presented graphically. 4 figs.

  10. Multi-phase classification by a least-squares support vector machine approach in tomography images of geological samples

    Science.gov (United States)

    Khan, Faisal; Enzmann, Frieder; Kersten, Michael

    2016-03-01

    Image processing of X-ray-computed polychromatic cone-beam micro-tomography (μXCT) data of geological samples mainly involves artefact reduction and phase segmentation. For the former, the main beam-hardening (BH) artefact is removed by applying a best-fit quadratic surface algorithm to a given image data set (reconstructed slice), which minimizes the BH offsets of the attenuation data points from that surface. A Matlab code for this approach is provided in the Appendix. The final BH-corrected image is extracted from the residual data or from the difference between the surface elevation values and the original grey-scale values. For the segmentation, we propose a novel least-squares support vector machine (LS-SVM, an algorithm for pixel-based multi-phase classification) approach. A receiver operating characteristic (ROC) analysis was performed on BH-corrected and uncorrected samples to show that BH correction is in fact an important prerequisite for accurate multi-phase classification. The combination of the two approaches was thus used to classify successfully three different more or less complex multi-phase rock core samples.

  11. Dynamic dielectrophoresis model of multi-phase ionic fluids.

    Directory of Open Access Journals (Sweden)

    Ying Yan

    Full Text Available Ionic-based dielectrophoretic microchips have attracted significant attention due to their wide-ranging applications in electro kinetic and biological experiments. In this work, a numerical method is used to simulate the dynamic behaviors of ionic droplets in a microchannel under the effect of dielectrophoresis. When a discrete liquid dielectric is encompassed within a continuous fluid dielectric placed in an electric field, an electric force is produced due to the dielectrophoresis effect. If either or both of the fluids are ionic liquids, the magnitude and even the direction of the force will be changed because the net ionic charge induced by an electric field can affect the polarization degree of the dielectrics. However, using a dielectrophoresis model, assuming ideal dielectrics, results in significant errors. To avoid the inaccuracy caused by the model, this work incorporates the electrode kinetic equation and defines a relationship between the polarization charge and the net ionic charge. According to the simulation conditions presented herein, the electric force obtained in this work has an error exceeding 70% of the actual value if the false effect of net ionic charge is not accounted for, which would result in significant issues in the design and optimization of experimental parameters. Therefore, there is a clear motivation for developing a model adapted to ionic liquids to provide precise control for the dielectrophoresis of multi-phase ionic liquids.

  12. Dynamic dielectrophoresis model of multi-phase ionic fluids.

    Science.gov (United States)

    Yan, Ying; Luo, Jing; Guo, Dan; Wen, Shizhu

    2015-01-01

    Ionic-based dielectrophoretic microchips have attracted significant attention due to their wide-ranging applications in electro kinetic and biological experiments. In this work, a numerical method is used to simulate the dynamic behaviors of ionic droplets in a microchannel under the effect of dielectrophoresis. When a discrete liquid dielectric is encompassed within a continuous fluid dielectric placed in an electric field, an electric force is produced due to the dielectrophoresis effect. If either or both of the fluids are ionic liquids, the magnitude and even the direction of the force will be changed because the net ionic charge induced by an electric field can affect the polarization degree of the dielectrics. However, using a dielectrophoresis model, assuming ideal dielectrics, results in significant errors. To avoid the inaccuracy caused by the model, this work incorporates the electrode kinetic equation and defines a relationship between the polarization charge and the net ionic charge. According to the simulation conditions presented herein, the electric force obtained in this work has an error exceeding 70% of the actual value if the false effect of net ionic charge is not accounted for, which would result in significant issues in the design and optimization of experimental parameters. Therefore, there is a clear motivation for developing a model adapted to ionic liquids to provide precise control for the dielectrophoresis of multi-phase ionic liquids.

  13. Pore-scale Simulation and Imaging of Multi-phase Flow and Transport in Porous Media (Invited)

    Science.gov (United States)

    Crawshaw, J.; Welch, N.; Daher, I.; Yang, J.; Shah, S.; Grey, F.; Boek, E.

    2013-12-01

    We combine multi-scale imaging and computer simulation of multi-phase flow and reactive transport in rock samples to enhance our fundamental understanding of long term CO2 storage in rock formations. The imaging techniques include Confocal Laser Scanning Microscopy (CLSM), micro-CT and medical CT scanning, with spatial resolutions ranging from sub-micron to mm respectively. First, we report a new sample preparation technique to study micro-porosity in carbonates using CLSM in 3 dimensions. Second, we use micro-CT scanning to generate high resolution 3D pore space images of carbonate and cap rock samples. In addition, we employ micro-CT to image the processes of evaporation in fractures and cap rock degradation due to exposure to CO2 flow. Third, we use medical CT scanning to image spontaneous imbibition in carbonate rock samples. Our imaging studies are complemented by computer simulations of multi-phase flow and transport, using the 3D pore space images obtained from the scanning experiments. We have developed a massively parallel lattice-Boltzmann (LB) code to calculate the single phase flow field in these pore space images. The resulting flow fields are then used to calculate hydrodynamic dispersion using a novel scheme to predict probability distributions for molecular displacements using the LB method and a streamline algorithm, modified for optimal solid boundary conditions. We calculate solute transport on pore-space images of rock cores with increasing degree of heterogeneity: a bead pack, Bentheimer sandstone and Portland carbonate. We observe that for homogeneous rock samples, such as bead packs, the displacement distribution remains Gaussian with time increasing. In the more heterogeneous rocks, on the other hand, the displacement distribution develops a stagnant part. We observe that the fraction of trapped solute increases from the beadpack (0 %) to Bentheimer sandstone (1.5 %) to Portland carbonate (8.1 %), in excellent agreement with PFG

  14. Novel image reconstruction algorithm for multi-phase flow tomography system using γ ray method

    International Nuclear Information System (INIS)

    Hao Kuihong; Wang Huaxiang; Gao Mei

    2007-01-01

    After analyzing the reason of image reconstructed algorithm by using the conventional back projection (IBP) is prone to produce spurious line, and considering the characteristic of multi-phase flow tomography, a novel image reconstruction algorithm is proposed, which carries out the intersection calculation using back projection data. This algorithm can obtain a perfect system point spread function, and can eliminate spurious line better. Simulating results show that the algorithm is effective for identifying multi-phase flow pattern. (authors)

  15. Modelling and simulation of multi-phase effects on X-ray elasticity constants

    CERN Document Server

    Freour, S; Guillen, R; François, M X

    2003-01-01

    This paper deals with the calculation of X-ray Elasticity Constants (XEC) of phases embedded in multi-phase polycrystals. A three scales (macroscopic, pseudo-macroscopic, mesoscopic) model based on the classical self-consistent formalism is developed in order to analyse multi-phase effects on XEC values. Simulations are performed for cubic or hexagonal crystallographic structure phases embedded in several two-phases materials. In fact, it is demonstrated that XEC vary with the macroscopic stiffness of the whole polycrystal. In consequence, the constants of one particular phase depend on the elastic behaviour and the volume fraction of all the phases constituting the material. Now, XEC play a leading role in pseudo-macroscopic stresses determination by X-Ray Diffraction (XRD) methods. In this work, a quantitative analysis of the multi-phase effects on stresses determination by XRD methods was performed. Numerical results will be compared and discussed. (Abstract Copyright [2003], Wiley Periodicals, Inc.)

  16. A Multi-Phase Equation of State and Strength Model for Tin

    International Nuclear Information System (INIS)

    Cox, G. A.

    2006-01-01

    This paper considers a multi-phase equation of state and a multi-phase strength model for tin in the β, γ and liquid phases. At a phase transition there are changes in volume, energy, and properties of a material that should be included in an accurate model. The strength model will also be affected by a solid-solid phase transition. For many materials there is a lack of experimental data for strength at high pressures making the derivation of strength parameters for some phases difficult. In the case of tin there are longitudinal sound speed data on the Hugoniot available that have been used here in conjunction with a multi-phase equation of state to derive strength parameters for the γ phase, a phase which does not exist at room temperature and pressure

  17. A parallel solution-adaptive scheme for predicting multi-phase core flows in solid propellant rocket motors

    International Nuclear Information System (INIS)

    Sachdev, J.S.; Groth, C.P.T.; Gottlieb, J.J.

    2003-01-01

    The development of a parallel adaptive mesh refinement (AMR) scheme is described for solving the governing equations for multi-phase (gas-particle) core flows in solid propellant rocket motors (SRM). An Eulerian formulation is used to described the coupled motion between the gas and particle phases. A cell-centred upwind finite-volume discretization and the use of limited solution reconstruction, Riemann solver based flux functions for the gas and particle phases, and explicit multi-stage time-stepping allows for high solution accuracy and computational robustness. A Riemann problem is formulated for prescribing boundary data at the burning surface. Efficient and scalable parallel implementations are achieved with domain decomposition on distributed memory multiprocessor architectures. Numerical results are described to demonstrate the capabilities of the approach for predicting SRM core flows. (author)

  18. Segmentation of hepatic artery in multi-phase liver CT using directional dilation and connectivity analysis

    Science.gov (United States)

    Wang, Lei; Schnurr, Alena-Kathrin; Zidowitz, Stephan; Georgii, Joachim; Zhao, Yue; Razavi, Mohammad; Schwier, Michael; Hahn, Horst K.; Hansen, Christian

    2016-03-01

    Segmentation of hepatic arteries in multi-phase computed tomography (CT) images is indispensable in liver surgery planning. During image acquisition, the hepatic artery is enhanced by the injection of contrast agent. The enhanced signals are often not stably acquired due to non-optimal contrast timing. Other vascular structure, such as hepatic vein or portal vein, can be enhanced as well in the arterial phase, which can adversely affect the segmentation results. Furthermore, the arteries might suffer from partial volume effects due to their small diameter. To overcome these difficulties, we propose a framework for robust hepatic artery segmentation requiring a minimal amount of user interaction. First, an efficient multi-scale Hessian-based vesselness filter is applied on the artery phase CT image, aiming to enhance vessel structures with specified diameter range. Second, the vesselness response is processed using a Bayesian classifier to identify the most probable vessel structures. Considering the vesselness filter normally performs not ideally on the vessel bifurcations or the segments corrupted by noise, two vessel-reconnection techniques are proposed. The first technique uses a directional morphological operator to dilate vessel segments along their centerline directions, attempting to fill the gap between broken vascular segments. The second technique analyzes the connectivity of vessel segments and reconnects disconnected segments and branches. Finally, a 3D vessel tree is reconstructed. The algorithm has been evaluated using 18 CT images of the liver. To quantitatively measure the similarities between segmented and reference vessel trees, the skeleton coverage and mean symmetric distance are calculated to quantify the agreement between reference and segmented vessel skeletons, resulting in an average of 0:55+/-0:27 and 12:7+/-7:9 mm (mean standard deviation), respectively.

  19. Non-isothermal effects on multi-phase flow in porous medium

    DEFF Research Database (Denmark)

    Singh, Ashok; Wang, W; Park, C. H.

    2010-01-01

    In this paper a ppT -formulation for non-isothermal multi-phase flow is given including diffusion and latent heat effects. Temperature and pressure dependencies of governing parameters are considered, in particular surface tension variation on phase interfaces along with temperature changes. A we...

  20. Multi-phase flow modeling of soil contamination and soil remediation

    NARCIS (Netherlands)

    Dijke, van M.I.J.

    1997-01-01


    In this thesis multi-phase flow models are used to study the flow behavior of liquid contaminants in aquifers and of gases that are injected below the groundwater table for remediation purposes. Considered problems are redistribution of a lens of light nonaqueous phase

  1. The use of magnetic resonance imaging to quantify multi-phase flow patterns and transitions

    International Nuclear Information System (INIS)

    Reyes, J.N. Jr.; Lafi, A.Y.; Saloner, D.

    1998-01-01

    Conventional measurement techniques have given limited insights into the complex structure of multi-phase flows. This has led to highly subjective flow pattern classifications which have been cast in terms of flow regime maps. Rather than using static flow regime maps, some of the next generation of multi-phase flow analysis codes will implement interfacial area transport equations that would calculate the flow patterns that evolve spatially and temporally. To assess these new codes, a large data base needs to be established to quantify the essential characteristics of multi-phase flow structure. One such characteristic is the interfacial area concentration. In this paper, we discuss the current benefits and limitations of using Magnetic Resonance Imaging (MRI) to examine multi- phase flow patterns and transitions. Of particular interest, are the MRI measurements of interfacial area concentration for slug flow in an air-water system. These tests were performed at the University of California, San Francisco (UCSF) School of Medicine MRI Center as a collaborative research effort with Oregon State University (OSU). The special scanning sequences designed by UCSF were capable of imaging at repetition intervals as fast as 7 milliseconds. (author)

  2. The use of magnetic resonance imaging to quantify multi-phase flow patterns and transitions

    Energy Technology Data Exchange (ETDEWEB)

    Reyes, Jr, J N; Lafi, A Y [Department of Nuclear Engineering, Oregon State University, Corvallis, OR (United States); Saloner, D [University of California, San Francisco School of Medicine, Veterans Administration Medical Center, San Francisco, CA (United States)

    1998-09-01

    Conventional measurement techniques have given limited insights into the complex structure of multi-phase flows. This has led to highly subjective flow pattern classifications which have been cast in terms of flow regime maps. Rather than using static flow regime maps, some of the next generation of multi-phase flow analysis codes will implement interfacial area transport equations that would calculate the flow patterns that evolve spatially and temporally. To assess these new codes, a large data base needs to be established to quantify the essential characteristics of multi-phase flow structure. One such characteristic is the interfacial area concentration. In this paper, we discuss the current benefits and limitations of using Magnetic Resonance Imaging (MRI) to examine multi- phase flow patterns and transitions. Of particular interest, are the MRI measurements of interfacial area concentration for slug flow in an air-water system. These tests were performed at the University of California, San Francisco (UCSF) School of Medicine MRI Center as a collaborative research effort with Oregon State University (OSU). The special scanning sequences designed by UCSF were capable of imaging at repetition intervals as fast as 7 milliseconds. (author)

  3. Coherent Surface Clutter Suppression Techniques with Topography Estimation for Multi-Phase-Center Radar Ice Sounding

    DEFF Research Database (Denmark)

    Nielsen, Ulrik; Dall, Jørgen; Kristensen, Steen Savstrup

    2012-01-01

    Radar ice sounding enables measurement of the thickness and internal structures of the large ice sheets on Earth. Surface clutter masking the signal of interest is a major obstacle in ice sounding. Algorithms for surface clutter suppression based on multi-phase-center radars are presented. These ...

  4. A variational approach to multi-phase motion of gas, liquid and solid based on the level set method

    Science.gov (United States)

    Yokoi, Kensuke

    2009-07-01

    We propose a simple and robust numerical algorithm to deal with multi-phase motion of gas, liquid and solid based on the level set method [S. Osher, J.A. Sethian, Front propagating with curvature-dependent speed: Algorithms based on Hamilton-Jacobi formulation, J. Comput. Phys. 79 (1988) 12; M. Sussman, P. Smereka, S. Osher, A level set approach for capturing solution to incompressible two-phase flow, J. Comput. Phys. 114 (1994) 146; J.A. Sethian, Level Set Methods and Fast Marching Methods, Cambridge University Press, 1999; S. Osher, R. Fedkiw, Level Set Methods and Dynamics Implicit Surface, Applied Mathematical Sciences, vol. 153, Springer, 2003]. In Eulerian framework, to simulate interaction between a moving solid object and an interfacial flow, we need to define at least two functions (level set functions) to distinguish three materials. In such simulations, in general two functions overlap and/or disagree due to numerical errors such as numerical diffusion. In this paper, we resolved the problem using the idea of the active contour model [M. Kass, A. Witkin, D. Terzopoulos, Snakes: active contour models, International Journal of Computer Vision 1 (1988) 321; V. Caselles, R. Kimmel, G. Sapiro, Geodesic active contours, International Journal of Computer Vision 22 (1997) 61; G. Sapiro, Geometric Partial Differential Equations and Image Analysis, Cambridge University Press, 2001; R. Kimmel, Numerical Geometry of Images: Theory, Algorithms, and Applications, Springer-Verlag, 2003] introduced in the field of image processing.

  5. Speed-sensorless control strategy for multi-phase induction generator in wind energy conversion systems

    Directory of Open Access Journals (Sweden)

    Dumnić Boris P.

    2016-01-01

    Full Text Available Renewable energy sources, especially wind energy conversion systems (WECS, exhibit constant growth. Increase in power and installed capacity led to advances in WECS topologies. Multi-phase approach presents a new development direction, with several key advantages over three-phase systems. Paired with a sensorless control strategy, multi-phase machines are expected to take primacy over standard solutions. This paper presents speed sensorless vector control of an asymmetrical six-phase induction generator based on a model reference adaptive system (MRAS. Suggested topology and developed control algorithm show that sensorless control can yield appropriate dynamic characteristics for the use in WECS with increase in reliability and robustness. [Projekat Ministarstva nauke Republike Srbije, br. III 042004: Smart Electricity Distribution Grids Based on Distribution Management System and Distributed Generation

  6. Experimental analysis of influence of different lubricants types on the multi-phase ironing process

    Directory of Open Access Journals (Sweden)

    Milan Djordjević

    2013-05-01

    Full Text Available This paper is aimed at presenting results of an experimental analysis of the different types of lubricants influence on the multi-phase ironing process. Based on sliding of the metal strip between the two contact elements a special tribological model was adopted. The subject of experimental investigations was variations of the drawing force, contact pressure and the friction coefficient for each type of the applied lubricants. The ironing process was conducted in three-phases at the constant sliding velocity. The objective of this analysis was to compare all the applied lubricants in order to estimate their quality from the point of view of their applicability in the multi-phase ironing process.

  7. EXPERIMENTAL ANALYSIS OF INFLUENCE OF DIFFERENT LUBRICANTS TYPES ON THE MULTI-PHASE IRONING PROCESS

    Directory of Open Access Journals (Sweden)

    Milan Djordjević

    2013-09-01

    Full Text Available This paper is aimed at presenting results of an experimental analysis of the different types of lubricants influence on the multi-phase ironing process. Based on sliding of the metal strip between the two contact elements a special tribological model was adopted. The subject of experimental investigations was variations of the drawing force, contact pressure and the friction coefficient for each type of the applied lubricants. The ironing process was conducted in three-phases at the constant sliding velocity. The objective of this analysis was to compare all the applied lubricants in order to estimate their quality from the point of view of their applicability in the multi-phase ironing process.

  8. Well-posedness and stability characteristics of multi-phase models

    International Nuclear Information System (INIS)

    Ransom, V.H.; Trapp, J.A.

    1984-01-01

    The ill-posed characteristic associated with the basic two-fluid model for multi-phase flow is a natural consequence of the idealized physical model and the mean flow modeling approach. Two approaches are discussed whereby including added physics of the flow results in a well-posed system of partial differential equations. These models offer the possibility of improved accuracy and numerical efficiency compared to the numerical models used in the existing light water reactor safety analysis codes

  9. Generalized modeling of multi-component vaporization/condensation phenomena for multi-phase-flow analysis

    International Nuclear Information System (INIS)

    Morita, K.; Fukuda, K.; Tobita, Y.; Kondo, Sa.; Suzuki, T.; Maschek, W.

    2003-01-01

    A new multi-component vaporization/condensation (V/C) model was developed to provide a generalized model for safety analysis codes of liquid metal cooled reactors (LMRs). These codes simulate thermal-hydraulic phenomena of multi-phase, multi-component flows, which is essential to investigate core disruptive accidents of LMRs such as fast breeder reactors and accelerator driven systems. The developed model characterizes the V/C processes associated with phase transition by employing heat transfer and mass-diffusion limited models for analyses of relatively short-time-scale multi-phase, multi-component hydraulic problems, among which vaporization and condensation, or simultaneous heat and mass transfer, play an important role. The heat transfer limited model describes the non-equilibrium phase transition processes occurring at interfaces, while the mass-diffusion limited model is employed to represent effects of non-condensable gases and multi-component mixture on V/C processes. Verification of the model and method employed in the multi-component V/C model of a multi-phase flow code was performed successfully by analyzing a series of multi-bubble condensation experiments. The applicability of the model to the accident analysis of LMRs is also discussed by comparison between steam and metallic vapor systems. (orig.)

  10. Development of an Efficient Meso- scale Multi-phase Flow Solver in Nuclear Applications

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Taehun [City Univ. (CUNY), NY (United States)

    2015-10-20

    The proposed research aims at formulating a predictive high-order Lattice Boltzmann Equation for multi-phase flows relevant to nuclear energy related application - namely, saturated and sub-cooled boiling in reactors, and liquid- liquid mixing and extraction for fuel cycle separation. An efficient flow solver will be developed based on the Finite Element based Lattice Boltzmann Method (FE- LBM), accounting for phase-change heat transfer and capable of treating multiple phases over length scales from the submicron to the meter. A thermal LBM will be developed in order to handle adjustable Prandtl number, arbitrary specific heat ratio, a wide range of temperature variations, better numerical stability during liquid-vapor phase change, and full thermo-hydrodynamic consistency. Two-phase FE-LBM will be extended to liquid–liquid–gas multi-phase flows for application to high-fidelity simulations building up from the meso-scale up to the equipment sub-component scale. While several relevant applications exist, the initial applications for demonstration of the efficient methods to be developed as part of this project include numerical investigations of Critical Heat Flux (CHF) phenomena in nuclear reactor fuel bundles, and liquid-liquid mixing and interfacial area generation for liquid-liquid separations. In addition, targeted experiments will be conducted for validation of this advanced multi-phase model.

  11. Numerical study of Tallinn storm-water system flooding conditions using CFD simulations of multi-phase flow in a large-scale inverted siphon

    Science.gov (United States)

    Kaur, K.; Laanearu, J.; Annus, I.

    2017-10-01

    The numerical experiments are carried out for qualitative and quantitative interpretation of a multi-phase flow processes associated with malfunctioning of the Tallinn storm-water system during rain storms. The investigations are focused on the single-line inverted siphon, which is used as under-road connection of pipes of the storm-water system under interest. A multi-phase flow solver of Computational Fluid Dynamics software OpenFOAM is used for simulating the three-phase flow dynamics in the hydraulic system. The CFD simulations are performed with different inflow rates under same initial conditions. The computational results are compared essentially in two cases 1) design flow rate and 2) larger flow rate, for emptying the initially filled inverted siphon from a slurry-fluid. The larger flow-rate situations are under particular interest to detected possible flooding. In this regard, it is anticipated that the CFD solutions provide an important insight to functioning of inverted siphon under a restricted water-flow conditions at simultaneous presence of air and slurry-fluid.

  12. On the extension of multi-phase models to sub-residual saturations

    International Nuclear Information System (INIS)

    Lingineni, S.; Chen, Y.T.; Boehm, R.F.

    1995-01-01

    This paper focuses on the limitations of applying multi-phase flow and transport models to simulate the hydrothermal processes occurring when the liquid saturation falls below residual levels. A typical scenario of a heat-generating high-level waste package emplaced in a backfilled drift of a waste repository is presented. The hydrothermal conditions in the vicinity of the waste package as well as in the far-field are determined using multi-phase, non-isothermal codes such as TOUGH2 and FEHM. As the waste package temperature increases, heat-pipe effects are created and water is driven away from the package into colder regions where it condenses. The variations in the liquid saturations close to the waste package are determined using these models with extended capillary pressure-saturations relationships to sub-residual regime. The predictions indicate even at elevated temperatures, waste package surroundings are not completely dry. However, if transport based modeling is used to represent liquid saturation variations in the sub-residual regime, then complete dry conditions are predicted within the backfill for extended periods of time. The relative humidity conditions near the waste package are also found to be sensitive to the representation of capillary pressure-saturation relationship used for sub-residual regime. An experimental investigation is carried out to study the variations in liquid saturations and relative humidity conditions in sub-residual regimes. Experimental results indicated that extended multi-phase models without interphase transport can not predict dry-out conditions and the simulations underpredict the humidity conditions near the waste package

  13. Antiferromagnetic-ferromagnetic crossover in UO2-TiOx multi-phase systems

    International Nuclear Information System (INIS)

    Nakamura, Akio; Tsutsui, Satoshi; Yoshii, Kenji

    2001-01-01

    An antiferromagnetic (AF)-weakly ferromagnetic (WF) crossover has been found for UO 2 -TiO x multi-phase systems, (1-y)UO 2 +yTiO x (y=0.05-0.72, x=0, 1.0, 1.5 and 2.0), when these mixtures are heat treated at high temperature in vacuum. From the powder X-ray diffraction and electron-microprobe analyses, their phase assemblies were as follows: for x=0, 1.0 and 1.5, a heterogeneous two-phase mixture of UO 2 +TiO x ; for x=2.0, that of UO 2 +UTi 2 O 6 for y 0.67 that of UTi 2 O 6 +TiO 2 (plus residual minor UO 2 ). Magnetic susceptibility (χ) of the present UO 2 powder was confirmed to exhibit an antiferromagnetic sharp drop at T N (=30.5 K). In contrast, χ of these multi-phase systems was found to exhibit a sharp upturn at the respective T N , while their T N values remained almost constant with varying y. This χ upturn at T N is most pronounced for UO 2 +Ti-oxide (titania) systems (x=1.0, 1.5 and 2.0) over the wide mixture ratio above y∼0.10. These observations indicate that an AF-WF crossover is induced for these multi-phase systems, plausibly due to the interfacial magnetic modification of UO 2 in contact with the oxide partners

  14. Advanced high speed X-ray CT scanner for measurement and visualization of multi-phase flow

    International Nuclear Information System (INIS)

    Hori, Keiichi; Fujimoto, Tetsuro; Kawanishi, Kohei; Nishikawa, Hideo

    1998-01-01

    The development of an ultra-fast X-ray computed tomography (CT) scanner has been performed. The object of interest is in a transient or unsettled state, which makes the conventional CT scanner inappropriate. A concept of electrical switching of electron beam of X-ray generation unit is adopted to reduce the scanning time instead of a mechanical motion adopted by a conventional CT scanner. The mechanical motion is a major obstacle to improve the scanning speed. A prototype system with a scanning time of 3.6 milliseconds was developed at first. And, the feasibility was confirmed to measure the dynamic events of two-phase flow. However, faster scanning speed is generally required for the practical use in the thermalhydraulics research field. Therefore, the development of advanced type has been performed. This advanced type can operate under the scanning time of 0.5 milliseconds and is applicable for the measurement of the multi-phase flow with velocity up to 4-5 m/s. (author)

  15. Multi-organ segmentation from multi-phase abdominal CT via 4D graphs using enhancement, shape and location optimization.

    Science.gov (United States)

    Linguraru, Marius George; Pura, John A; Chowdhury, Ananda S; Summers, Ronald M

    2010-01-01

    The interpretation of medical images benefits from anatomical and physiological priors to optimize computer-aided diagnosis (CAD) applications. Diagnosis also relies on the comprehensive analysis of multiple organs and quantitative measures of soft tissue. An automated method optimized for medical image data is presented for the simultaneous segmentation of four abdominal organs from 4D CT data using graph cuts. Contrast-enhanced CT scans were obtained at two phases: non-contrast and portal venous. Intra-patient data were spatially normalized by non-linear registration. Then 4D erosion using population historic information of contrast-enhanced liver, spleen, and kidneys was applied to multi-phase data to initialize the 4D graph and adapt to patient specific data. CT enhancement information and constraints on shape, from Parzen windows, and location, from a probabilistic atlas, were input into a new formulation of a 4D graph. Comparative results demonstrate the effects of appearance and enhancement, and shape and location on organ segmentation.

  16. Sampling device for withdrawing a representative sample from single and multi-phase flows

    Science.gov (United States)

    Apley, Walter J.; Cliff, William C.; Creer, James M.

    1984-01-01

    A fluid stream sampling device has been developed for the purpose of obtaining a representative sample from a single or multi-phase fluid flow. This objective is carried out by means of a probe which may be inserted into the fluid stream. Individual samples are withdrawn from the fluid flow by sampling ports with particular spacings, and the sampling parts are coupled to various analytical systems for characterization of the physical, thermal, and chemical properties of the fluid flow as a whole and also individually.

  17. Multi-phase induced inflation in theories with non-minimal coupling to gravity

    Energy Technology Data Exchange (ETDEWEB)

    Artymowski, Michał [Institute of Physics, Jagiellonian University, Łojasiewicza 11, 30-348 Kraków (Poland); Lalak, Zygmunt; Lewicki, Marek, E-mail: Michal.Artymowski@uj.edu.pl, E-mail: Zygmunt.Lalak@fuw.edu.pl, E-mail: Marek.Lewicki@fuw.edu.pl [Institute of Theoretical Physics, Faculty of Physics, University of Warsaw ul. Hoża 69, 00-681 Warszawa (Poland)

    2017-01-01

    In this paper we investigate the induced inflation with two flat regions: one Starobinsky-like plateau in strong coupling regime and one shorter plateau around the saddle point of the Einstein frame potential. This multi-phase inflationary scenario can be used to solve problems of classical cosmology as well as the problem of initial conditions for inflation. The inflation at the saddle-point plateau is consistent with the data and can have arbitrarily low scale. The results can be useful in the context of the Higgs-Axion relaxation and in a certain limit they are equivalent to the α-attractors.

  18. Design and Analysis of Multi-Phase BLDC Motors for Electric Vehicles

    Directory of Open Access Journals (Sweden)

    G. Boztas

    2018-04-01

    Full Text Available This paper presents a design and analysis of multi-phase brushless direct current (BLDC motor for electric vehicles (EV. In this work, hub-wheels having 110Nm, 900rpm rated values have been designed for the proposed EV. This EV can produce 440 Nm without using transmission, differential and other mechanical components which have very high losses due to the mechanical fraction. The motors to be used in the EV have been designed as 3-, 5- and 7-phase by Infolytica/Motor Solve Software to compare their performances at the same load conditions. The same rotor geometry has been utilized for the motors. However, slot numbers and dimensions of the stator have been determined by considering the motor phase number. Performance curves of phase-currents, output powers, torques, efficiencies and power factors have been presented for these motors at the same operating conditions. It can be possible to use lower power switches in motor drive system thanks to the phase current reduction since the phase currents decrease proportionally to motor phase number. This work shows that the multi-phase BLDC motors are a good alternative in order to obtain lower torque and lower power inverter structure than the 3-phase BLDC motors which are used as standard.

  19. A Novel Multi-Phase Stochastic Model for Lithium-Ion Batteries’ Degradation with Regeneration Phenomena

    Directory of Open Access Journals (Sweden)

    Jianxun Zhang

    2017-10-01

    Full Text Available A lithium-Ion battery is a typical degradation product, and its performance will deteriorate over time. In its degradation process, regeneration phenomena have been frequently encountered, which affect both the degradation state and rate. In this paper, we focus on how to build the degradation model and estimate the lifetime. Toward this end, we first propose a multi-phase stochastic degradation model with random jumps based on the Wiener process, where the multi-phase model and random jumps at the changing point are used to describe the variation of degradation rate and state caused by regeneration phenomena accordingly. Owing to the complex structure and random variables, the traditional Maximum Likelihood Estimation (MLE is not suitable for the proposed model. In this case, we treat these random variables as latent parameters, and then develop an approach for model identification based on expectation conditional maximum (ECM algorithm. Moreover, depending on the proposed model, how to estimate the lifetime with fixed changing point is presented via the time-space transformation technique, and the approximate analytical solution is derived. Finally, a numerical simulation and a practical case are provided for illustration.

  20. Single-step reinitialization and extending algorithms for level-set based multi-phase flow simulations

    Science.gov (United States)

    Fu, Lin; Hu, Xiangyu Y.; Adams, Nikolaus A.

    2017-12-01

    We propose efficient single-step formulations for reinitialization and extending algorithms, which are critical components of level-set based interface-tracking methods. The level-set field is reinitialized with a single-step (non iterative) "forward tracing" algorithm. A minimum set of cells is defined that describes the interface, and reinitialization employs only data from these cells. Fluid states are extrapolated or extended across the interface by a single-step "backward tracing" algorithm. Both algorithms, which are motivated by analogy to ray-tracing, avoid multiple block-boundary data exchanges that are inevitable for iterative reinitialization and extending approaches within a parallel-computing environment. The single-step algorithms are combined with a multi-resolution conservative sharp-interface method and validated by a wide range of benchmark test cases. We demonstrate that the proposed reinitialization method achieves second-order accuracy in conserving the volume of each phase. The interface location is invariant to reapplication of the single-step reinitialization. Generally, we observe smaller absolute errors than for standard iterative reinitialization on the same grid. The computational efficiency is higher than for the standard and typical high-order iterative reinitialization methods. We observe a 2- to 6-times efficiency improvement over the standard method for serial execution. The proposed single-step extending algorithm, which is commonly employed for assigning data to ghost cells with ghost-fluid or conservative interface interaction methods, shows about 10-times efficiency improvement over the standard method while maintaining same accuracy. Despite their simplicity, the proposed algorithms offer an efficient and robust alternative to iterative reinitialization and extending methods for level-set based multi-phase simulations.

  1. A multi-phase equation of state for solid and liquid lead

    International Nuclear Information System (INIS)

    Robinson, C.M.

    2004-01-01

    This paper considers a multi-phase equation of state for solid and liquid lead. The thermodynamically consistent equation of state is constructed by calculating separate equations of state for the solid and liquid phases. The melt curve is the curve in the pressure, temperature plane where the Gibb's free energy of the solid and liquid phases are equal. In each phase a complete equation of state is obtained using the assumptions that the specific heat capacity is constant and that the Grueneisen parameter is proportional to the specific volume. The parameters for the equation of state are obtained from experimental data. In particular they are chosen to match melt curve and principal Hugoniot data. Predictions are made for the shock pressure required for melt to occur on shock and release

  2. Application of PNA-technique for the measurement of multi-phase flow

    International Nuclear Information System (INIS)

    Loevhoeiden, G.; Andersen, E.; Garder, K.; Rambaek, J.P.

    1986-09-01

    The pulsed neutron activation (PNA) technique is proposed for multi-phase flow monitoring of hydrocarbons. The reactions 12 C(n,p) 12 B and 12 C(n,n') 12 C both yeld 4.4 MeV in the form of gamma radiation as a measure of carbon content. Intensity measurement of the 4.4 MeV gamma line gives a measure of the carbon content in the irradiation zone. By use of a pulsed neutron source, an estimation of the carbon content time variation is possible. In the presence of sulphur in petroleum, the reaction 34 S(n,p) 34 P offers a better possibility for flow rate determination

  3. Energetics of the multi-phase fluid flow in a narrow kerf in laser cutting conditions

    Science.gov (United States)

    Golyshev, A. A.; Orishich, A. M.; Shulyatyev, V. B.

    2016-10-01

    The energy balance of the multi-phase medium flow is studied experimentally under the laser cutting. Experimental data are generalized due to the condition of minimal roughness of the created surface used as a quality criterion of the melt flow, and also due to the application of dimensionless parameters: Peclet number and dimensionless absorbed laser power. For the first time ever it is found that, regardless the assistant gas (oxygen or nitrogen), laser type (the fiber one with the wavelength of 1.07 µm or CO2-laser with the wavelength of 10.6 µm), the minimal roughness is provided at a certain energy input in a melt unit, about 26 J/mm3. With oxygen, 50% of this input is provided by the radiation, the other 50% - by the exothermic reaction of iron oxidation.

  4. A Fourier Collocation Approach for Transit-Time Ultrasonic Flowmeter Under Multi-Phase Flow Conditions

    DEFF Research Database (Denmark)

    Simurda, Matej; Lassen, Benny; Duggen, Lars

    2017-01-01

    A numerical model for a clamp-on transit-time ultrasonic flowmeter (TTUF) under multi-phase flow conditions is presented. The method solves equations of linear elasticity for isotropic heterogeneous materials with background flow where acoustic media are modeled by setting shear modulus to zero....... Spatial derivatives are calculated by a Fourier collocation method allowing the use of the fast Fourier transform (FFT) and time derivatives are approximated by a finite difference (FD) scheme. This approach is sometimes referred to as a pseudospectral time-domain method. Perfectly matched layers (PML......) are used to avoid wave-wrapping and staggered grids are implemented to improve stability and efficiency. The method is verified against exact analytical solutions and the effect of the time-staggering and associated lowest number of points per minimum wavelengths value is discussed. The method...

  5. Annealing treatments for producing cold-rolled low-silicon multi-phase steels for automotive applications

    Energy Technology Data Exchange (ETDEWEB)

    Vrieze, J.; Vortrefflich, W.; Winter, L. de [Applied Physical Metallurgy, Hoogovens Research and Development, Ijmuiden (Netherlands)

    2000-07-01

    Laboratory simulations of a hot-dip galvanizing line have been carried out on a 0.16%C-1.5%Mn-0.4%Si steel. It has been demonstrated that based on this steel composition materials with typical multi-phase properties can be produced on hot-dip galvanizing lines. By varying the process conditions a tensile strength between 570 and 650 MPa has been obtained. In addition, tensile tests at high strain rates have been carried out and the results have been compared with those of other non-multi-phase steel grades. (orig.)

  6. Multi-Phase Sub-Sampling Fractional-N PLL with soft loop switching for fast robust locking

    NARCIS (Netherlands)

    Liao, Dongyi; Dai, FA Foster; Nauta, Bram; Klumperink, Eric A.M.

    2017-01-01

    This paper presents a low phase noise sub-sampling PLL (SSPLL) with multi-phase outputs. Automatic soft switching between the sub-sampling phase loop and frequency loop is proposed to improve robustness against perturbations and interferences that may cause a traditional SSPLL to lose lock. A

  7. From damage to fracture, from micro to macro : a systematic study of ductile fracture in multi-phase microstructures

    NARCIS (Netherlands)

    de Geus, T.W.J.

    2016-01-01

    Multi-phase materials are of great importance for engineering applications, because of their favorable combination of strength and ductility. This unique combination of properties enables lightweight yet safe design for instance in the automotive industry. The in-depth understanding of the

  8. Numerical modeling of experimental observations on gas formation and multi-phase flow of carbon dioxide in subsurface formations

    Science.gov (United States)

    Pawar, R.; Dash, Z.; Sakaki, T.; Plampin, M. R.; Lassen, R. N.; Illangasekare, T. H.; Zyvoloski, G.

    2011-12-01

    One of the concerns related to geologic CO2 sequestration is potential leakage of CO2 and its subsequent migration to shallow groundwater resources leading to geochemical impacts. Developing approaches to monitor CO2 migration in shallow aquifer and mitigate leakage impacts will require improving our understanding of gas phase formation and multi-phase flow subsequent to CO2 leakage in shallow aquifers. We are utilizing an integrated approach combining laboratory experiments and numerical simulations to characterize the multi-phase flow of CO2 in shallow aquifers. The laboratory experiments involve a series of highly controlled experiments in which CO2 dissolved water is injected in homogeneous and heterogeneous soil columns and tanks. The experimental results are used to study the effects of soil properties, temperature, pressure gradients and heterogeneities on gas formation and migration. We utilize the Finite Element Heat and Mass (FEHM) simulator (Zyvoloski et al, 2010) to numerically model the experimental results. The numerical models capture the physics of CO2 exsolution, multi-phase fluid flow as well as sand heterogeneity. Experimental observations of pressure, temperature and gas saturations are used to develop and constrain conceptual models for CO2 gas-phase formation and multi-phase CO2 flow in porous media. This talk will provide details of development of conceptual models based on experimental observation, development of numerical models for laboratory experiments and modelling results.

  9. Is hepatotropic contrast enhanced MR a more effective method in differential diagnosis of hemangioma than multi-phase CT and unenhanced MR?

    Directory of Open Access Journals (Sweden)

    Markiet Karolina

    2011-04-01

    Full Text Available Abstract Background Cavernous hemangiomas are the most frequent neoplasms of the liver and in routine clinical practice they often need to be differentiated from malignant tumors and other benign focal lesions. The purpose of this study is to evaluate whether diagnostic accuracy of magnetic resonance imaging (MRI of hepatic hemangiomas, showing atypical pattern on US, improves with the use of Gd-BOPTA in comparison with contrast-enhanced multi-phase computed tomography (CT. Methods 178 consecutive patients with ambiguous hepatic masses showing atypical hyperechoic pattern on grey-scale US, underwent unenhanced and contrast-enhanced multi-phase multi-detector CT and MR (1.5T with the use of liver-specific contrast medium gadobenate dimeglumine (Gd-BOPTA. After intravenous contrast administration arterial (HAP, venous-portal (PVP, equilibrium phases (EP both in CT and MR and additionally hepatobiliary phase (HBP in MR were obtained. 398 lesions have been detected including 99 hemangiomas and 299 other lesions. Results In non-enhanced MDCT examination detection of hemangiomas was characterized by sensitivity of 76%, specificity of 90%, PPV of 71%, NPV of 92% and accuracy of 86%. Non-enhanced MR examination showed sensitivity of 98%, specificity of 99%, PPV of 99%, NPV of 99% and accuracy of 99%. After intravenous administration of contrast medium in MR the mentioned above parameters did not increase significantly. Conclusion Gd-BOPTA-enhanced MR in comparison with unenhanced MRI does not improve diagnostic accuracy in discriminating hemangiomas that show non-specific appearance in ultrasound examination. Unenhanced MR as a method of choice should directly follow US in course of diagnostic algorithm in differentiation of hemangiomas from other liver tumors.

  10. Multi-Phase Modular Drive System: A Case Study in Electrical Aircraft Applications

    Directory of Open Access Journals (Sweden)

    Charles Onambele

    2017-12-01

    Full Text Available In this article, an advanced multiphase modular power drive prototype is developed for More Electric Aircraft (MEA. The proposed drive is designed to supply a multi-phase permanent magnet (PM motor rating 120 kW with 24 slots and 11 pole pairs. The power converter of the drive system is based on Silicon Carbide Metal Oxide Semiconductor Field-Effect Transistor (SiC MOSFET technology to operate at high voltage, high frequency and low reverse recovery current. Firstly, an experimental characterization test is performed for the selected SiC power module in harsh conditions to evaluate the switching energy losses. Secondly, a finite element thermal analysis based on Ansys-Icepak is accomplished to validate the selected cooling system for the power converter. Thirdly, a co-simulation model is developed using Matlab-Simulink and LTspice® to evaluate the SiC power module impact on the performance of a multiphase drive system at different operating conditions. The results obtained show that the dynamic performance and efficiency of the power drive are significantly improved, which makes the proposed system an excellent candidate for future aircraft applications.

  11. A Multi-Phase Based Fluid-Structure-Microfluidic interaction sensor for Aerodynamic Shear Stress

    Science.gov (United States)

    Hughes, Christopher; Dutta, Diganta; Bashirzadeh, Yashar; Ahmed, Kareem; Qian, Shizhi

    2014-11-01

    A novel innovative microfluidic shear stress sensor is developed for measuring shear stress through multi-phase fluid-structure-microfluidic interaction. The device is composed of a microfluidic cavity filled with an electrolyte liquid. Inside the cavity, two electrodes make electrochemical velocimetry measurements of the induced convection. The cavity is sealed with a flexible superhydrophobic membrane. The membrane will dynamically stretch and flex as a result of direct shear cross-flow interaction with the seal structure, forming instability wave modes and inducing fluid motion within the microfluidic cavity. The shear stress on the membrane is measured by sensing the induced convection generated by membrane deflections. The advantages of the sensor over current MEMS based shear stress sensor technology are: a simplified design with no moving parts, optimum relationship between size and sensitivity, no gaps such as those created by micromachining sensors in MEMS processes. We present the findings of a feasibility study of the proposed sensor including wind-tunnel tests, microPIV measurements, electrochemical velocimetry, and simulation data results. The study investigates the sensor in the supersonic and subsonic flow regimes. Supported by a NASA SBIR phase 1 contract.

  12. Multifocus microscopy with precise color multi-phase diffractive optics applied in functional neuronal imaging.

    Science.gov (United States)

    Abrahamsson, Sara; Ilic, Rob; Wisniewski, Jan; Mehl, Brian; Yu, Liya; Chen, Lei; Davanco, Marcelo; Oudjedi, Laura; Fiche, Jean-Bernard; Hajj, Bassam; Jin, Xin; Pulupa, Joan; Cho, Christine; Mir, Mustafa; El Beheiry, Mohamed; Darzacq, Xavier; Nollmann, Marcelo; Dahan, Maxime; Wu, Carl; Lionnet, Timothée; Liddle, J Alexander; Bargmann, Cornelia I

    2016-03-01

    Multifocus microscopy (MFM) allows high-resolution instantaneous three-dimensional (3D) imaging and has been applied to study biological specimens ranging from single molecules inside cells nuclei to entire embryos. We here describe pattern designs and nanofabrication methods for diffractive optics that optimize the light-efficiency of the central optical component of MFM: the diffractive multifocus grating (MFG). We also implement a "precise color" MFM layout with MFGs tailored to individual fluorophores in separate optical arms. The reported advancements enable faster and brighter volumetric time-lapse imaging of biological samples. In live microscopy applications, photon budget is a critical parameter and light-efficiency must be optimized to obtain the fastest possible frame rate while minimizing photodamage. We provide comprehensive descriptions and code for designing diffractive optical devices, and a detailed methods description for nanofabrication of devices. Theoretical efficiencies of reported designs is ≈90% and we have obtained efficiencies of > 80% in MFGs of our own manufacture. We demonstrate the performance of a multi-phase MFG in 3D functional neuronal imaging in living C. elegans.

  13. Third harmonic current injection into highly saturated multi-phase machines

    Directory of Open Access Journals (Sweden)

    Klute Felix

    2017-03-01

    Full Text Available One advantage of multi-phase machines is the possibility to use the third harmonic of the rotor flux for additional torque generation. This effect can be maximised for Permanent Magnet Synchronous Machines (PMSM with a high third harmonic content in the magnet flux. This paper discusses the effects of third harmonic current injection (THCI on a five-phase PMSM with a conventional magnet shape depending on saturation. The effects of THCI in five-phase machines are shown in a 2D FEM model in Ansys Maxwell verified by measurement results. The results of the FEM model are analytically analysed using the Park model. It is shown in simulation and measurement that the torque improvement by THCI increases significantly with the saturation level, as the amplitude of the third harmonic flux linkage increases with the saturation level but the phase shift of the rotor flux linkage has to be considered. This paper gives a detailed analysis of saturation mechanisms of PMSM, which can be used for optimizing the efficiency in operating points of high saturations, without using special magnet shapes.

  14. New systematic methodology for incorporating dynamic heat transfer modelling in multi-phase biochemical reactors.

    Science.gov (United States)

    Fernández-Arévalo, T; Lizarralde, I; Grau, P; Ayesa, E

    2014-09-01

    This paper presents a new modelling methodology for dynamically predicting the heat produced or consumed in the transformations of any biological reactor using Hess's law. Starting from a complete description of model components stoichiometry and formation enthalpies, the proposed modelling methodology has integrated successfully the simultaneous calculation of both the conventional mass balances and the enthalpy change of reaction in an expandable multi-phase matrix structure, which facilitates a detailed prediction of the main heat fluxes in the biochemical reactors. The methodology has been implemented in a plant-wide modelling methodology in order to facilitate the dynamic description of mass and heat throughout the plant. After validation with literature data, as illustrative examples of the capability of the methodology, two case studies have been described. In the first one, a predenitrification-nitrification dynamic process has been analysed, with the aim of demonstrating the easy integration of the methodology in any system. In the second case study, the simulation of a thermal model for an ATAD has shown the potential of the proposed methodology for analysing the effect of ventilation and influent characterization. Copyright © 2014 Elsevier Ltd. All rights reserved.

  15. Multi-phase model development to assess RCIC system capabilities under severe accident conditions

    Energy Technology Data Exchange (ETDEWEB)

    Kirkland, Karen Vierow [Texas A & M Univ., College Station, TX (United States); Ross, Kyle [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Beeny, Bradley [Texas A & M Univ., College Station, TX (United States); Luthman, Nicholas [Texas A& M Engineering Experiment Station, College Station, TX (United States); Strater, Zachary [Texas A & M Univ., College Station, TX (United States)

    2017-12-23

    The Reactor Core Isolation Cooling (RCIC) System is a safety-related system that provides makeup water for core cooling of some Boiling Water Reactors (BWRs) with a Mark I containment. The RCIC System consists of a steam-driven Terry turbine that powers a centrifugal, multi-stage pump for providing water to the reactor pressure vessel. The Fukushima Dai-ichi accidents demonstrated that the RCIC System can play an important role under accident conditions in removing core decay heat. The unexpectedly sustained, good performance of the RCIC System in the Fukushima reactor demonstrates, firstly, that its capabilities are not well understood, and secondly, that the system has high potential for extended core cooling in accident scenarios. Better understanding and analysis tools would allow for more options to cope with a severe accident situation and to reduce the consequences. The objectives of this project were to develop physics-based models of the RCIC System, incorporate them into a multi-phase code and validate the models. This Final Technical Report details the progress throughout the project duration and the accomplishments.

  16. Modeling of multi-phase interactions of reactive nitrogen between snow and air in Antarctica

    Science.gov (United States)

    McCrystall, M.; Chan, H. G. V.; Frey, M. M.; King, M. D.

    2016-12-01

    In polar and snow-covered regions, the snowpack is an important link between atmospheric, terrestrial and oceanic systems. Trace gases, including nitrogen oxides, produced via photochemical reactions in snow are partially released to the lower atmosphere with considerable impact on its composition. However, the post-depositional processes that change the chemical composition and physical properties of the snowpack are still poorly understood. Most current snow chemistry models oversimplify as they assume air-liquid interactions and aqueous phase chemistry taking place at the interface between the snow grain and air. Here, we develop a novel temperature dependent multi-phase (gas-liquid-ice) physical exchange model for reactive nitrogen. The model is validated with existing year-round observations of nitrate in the top 0.5-2 cm of snow and the overlying atmosphere at two very different Antarctic locations: Dome C on the East Antarctic Plateau with very low annual mean temperature (-54ºC) and accumulation rate (rate and high background level of sea salt aerosol. We find that below the eutectic temperature of the H2O/dominant ion mixture the surface snow nitrate is controlled by kinetic adsorption onto the surface of snow grains followed by grain diffusion. Above the eutectic temperature, in addition to the former two processes, thermodynamic equilibrium of HNO3 between interstitial air and liquid water pockets, possibly present at triple junctions or grooves at grain boundaries, greatly enhances the nitrate uptake by snow in agreement with the concentration peak observed in summer.

  17. A multi-phase ferrofluid flow model with equation of state for thermomagnetic pumping and heat transfer

    Energy Technology Data Exchange (ETDEWEB)

    Aursand, Eskil, E-mail: eskil.aursand@sintef.no; Gjennestad, Magnus Aa.; Yngve Lervåg, Karl; Lund, Halvor

    2016-03-15

    A one-dimensional multi-phase flow model for thermomagnetically pumped ferrofluid with heat transfer is proposed. The thermodynamic model is a combination of a simplified particle model and thermodynamic equations of state for the base fluid. The magnetization model is based on statistical mechanics, taking into account non-uniform particle size distributions. An implementation of the proposed model is validated against experiments from the literature, and found to give good predictions for the thermomagnetic pumping performance. However, the results reveal a very large sensitivity to uncertainties in heat transfer coefficient predictions. - Highlights: • A multi-phase flow model for thermomagnetically pumped ferrofluid is proposed. • An implementation is validated against experiments from the literature. • Predicted thermomagnetic pumping effect agrees with experiments. • However, a very large sensitivity to heat transfer coefficient is revealed.

  18. The experimental study on bowel ischemia in closed loop obstruction by using multi-phase spiral CT

    International Nuclear Information System (INIS)

    Zhang Xiaoming; Yang Hanfeng; Huang Xiaohua; Tang Xianying; Jian Pu; Yang Zhengwei; Zhou Jiyong; Zhao Zongwen

    2005-01-01

    Objective: To evaluate the bowel ischemia in experimental closed loop obstruction by using multi-phase spiral CT. Methods: Twenty-four New Zealand rabbits of both sexes (mean age, 4 months, and mean body weight, 2.5-3.0 kg) were divided randomly into three groups with each group containing 8 rabbits. After clamping 10-15 cm segments of small bowel and their veins for 0.5 hours (Group A), 1-2 hours (Group B), and 3-5 hours (Group C), respectively, multi-phase spiral CT was performed at baseline, and at arterial, venous, and delayed phases after intravenous contrast administration. Then the rabbits were sacrificed to observe their surgical and histological changes. Two radiologists, blinded to the animal model classification and their histological results, individually reviewed the CT images to observe the CT appearances of the closed loop. Statistical significance criteria was determined by P 0.05) at baseline, however, they were significantly different (P<0.05) at all phases after enhancement. Among rabbits without necrotic closed loop, 11 of 13 had continuous enhancement at all phases, while only 1 of 11 rabbits with necrotic closed loop showed continuous enhancement (P<0.05). Conclusion: The ischemia of bowel wall in different phases after clamping small bowel and their veins can be evaluated by using enhanced multi-phase spiral CT. Continuous enhancement of bowel wall in multi-phase spiral CT can be seen prominently in the early bowel ischemia, but necrotic bowel shows no enhancement. (authors)

  19. Microstructure characterization of multi-phase composites and utilization of phase change materials and recycled rubbers in cementitious materials

    Science.gov (United States)

    Meshgin, Pania

    2011-12-01

    This research focuses on two important subjects: (1) Characterization of heterogeneous microstructure of multi-phase composites and the effect of microstructural features on effective properties of the material. (2) Utilizations of phase change materials and recycled rubber particles from waste tires to improve thermal properties of insulation materials used in building envelopes. Spatial pattern of multi-phase and multidimensional internal structures of most composite materials are highly random. Quantitative description of the spatial distribution should be developed based on proper statistical models, which characterize the morphological features. For a composite material with multi-phases, the volume fraction of the phases as well as the morphological parameters of the phases have very strong influences on the effective property of the composite. These morphological parameters depend on the microstructure of each phase. This study intends to include the effect of higher order morphological details of the microstructure in the composite models. The higher order statistics, called two-point correlation functions characterize various behaviors of the composite at any two points in a stochastic field. Specifically, correlation functions of mosaic patterns are used in the study for characterizing transport properties of composite materials. One of the most effective methods to improve energy efficiency of buildings is to enhance thermal properties of insulation materials. The idea of using phase change materials and recycled rubber particles such as scrap tires in insulation materials for building envelopes has been studied.

  20. Fast Streaming 3D Level set Segmentation on the GPU for Smooth Multi-phase Segmentation

    DEFF Research Database (Denmark)

    Sharma, Ojaswa; Zhang, Qin; Anton, François

    2011-01-01

    Level set method based segmentation provides an efficient tool for topological and geometrical shape handling, but it is slow due to high computational burden. In this work, we provide a framework for streaming computations on large volumetric images on the GPU. A streaming computational model...

  1. Micro-Ct Imaging of Multi-Phase Flow in Carbonates and Sandstones

    Science.gov (United States)

    Andrew, M. G.; Bijeljic, B.; Blunt, M. J.

    2013-12-01

    One of the most important mechanisms that limits the escape of CO2 when injected into the subsurface for the purposes of carbon storage is capillary trapping, where CO2 is stranded as pore-scale droplets (ganglia). Prospective storage sites are aquifers or reservoirs that tend to be at conditions where CO2 will reside as a super-critical phase. In order to fully describe physical mechanisms characterising multi-phase flow during and post CO2 injection, experiments need to be conducted at these elevated aquifer/reservoir conditions - this poses a considerable experimental challenge. A novel experimental apparatus has been developed which uses μCT scanning for the non-invasive imaging of the distribution of CO2 in the pore space of rock with resolutions of 7μm at temperatures and pressures representative of the conditions present in prospective saline aquifer CO2 storage sites. The fluids are kept in chemical equilibrium with one-another and with the rock into which they are injected. This is done to prevent the dissolution of the CO2 in the brine to form carbonic acid, which can then react with the rock, particularly carbonates. By eliminating reaction we study the fundamental mechanisms of capillary trapping for an unchanging pore structure. In this study we present a suite of results from three carbonate and two sandstone rock types, showing that, for both cases the CO2 acts as the non-wetting phase and significant quantities of CO2 is trapped. The carbonate examined represent a wide variety of pore topologies with one rock with a very well connected, high porosity pore space (Mt Gambier), one with a lower porosity, poorly connected pore space (Estaillades) and one with a cemented bead pack type pore space (Ketton). Both sandstones (Doddington and Bentheimer) were high permeability granular quartzites. CO2 was injected into each rock, followed by brine injection. After brine injection the entire length of the rock core was scanned, processed and segmented into

  2. Choice ofoptimal phase for liver angiography and multi-phase scanning with multi-slice spiral CT

    International Nuclear Information System (INIS)

    Fang Hong; Song Yunlong; Bi Yongmin; Wang Dong; Shi Huiping; Zhang Wanshi; Zhu Hongxian; Yang Hua; Ji Xudong; Fan Hongxia

    2008-01-01

    Objective: To evaluate the efficacy of test bolus technique with multi-slice spiral CT (MSCT) for determining the optimal scan delay time in CT Hepatic artery (HA)-portal vein (PV) angiography and multi-phase scanning. Methods: MSCT liver angiography and multi-phase scanning were performed in 187 patients divided randomly into two groups. In group A (n=59), the scan delay time was set according to the subjective experiences of operators; in group B (n=128), the scan delay time was determined by test bolus technique. Abdominal aorta and superior mesenteric, vein were selected as target blood vessels, and 50 HU was set as enhancement threshold value. 20 ml contrast agent was injected intravenously and time-density curve of target blood vessels were obtained, then HA-PV scanning delay time were calculated respectively. The quality of CTA images obtained by using these 2 methods were compared and statistically analysed using Chi-square criterion. Results: For hepatic artery phase, the images of group A are: excellent in 34 (58%), good in 17 (29%), and poor in 8 (13%), while those of group B are excellent in 128(100%), good in 0(0%), and poor in 0(0%). For portal vein phase, the images of group A are: excellent in 23 (39%), good in 27 (46%), and poor in 9 (15%), while those of group B are excellent in 96 (75%), good in 28 (22%), and poor in 4 (3%) respectively. There was statistically significant difference between the ratios of image quality in group A and group B (χ 2 =14.97, 9.18, P< 0.05). Conclusion: Accurate scan delay time was best determined by using test bolus technique, which can improve the image quality of liver angiography and multi-phase scanning. (authors)

  3. Multi-Phase Equilibrium and Solubilities of Aromatic Compounds and Inorganic Compounds in Sub- and Supercritical Water: A Review.

    Science.gov (United States)

    Liu, Qinli; Ding, Xin; Du, Bowen; Fang, Tao

    2017-11-02

    Supercritical water oxidation (SCWO), as a novel and efficient technology, has been applied to wastewater treatment processes. The use of phase equilibrium data to optimize process parameters can offer a theoretical guidance for designing SCWO processes and reducing the equipment and operating costs. In this work, high-pressure phase equilibrium data for aromatic compounds+water systems and inorganic compounds+water systems are given. Moreover, thermodynamic models, equations of state (EOS) and empirical and semi-empirical approaches are summarized and evaluated. This paper also lists the existing problems of multi-phase equilibria and solubility studies on aromatic compounds and inorganic compounds in sub- and supercritical water.

  4. A multi-phase algorithm for a joint lot-sizing and pricing problem with stochastic demands

    DEFF Research Database (Denmark)

    Jenny Li, Hongyan; Thorstenson, Anders

    2014-01-01

    to a practically viable approach to decision-making. In addition to incorporating market uncertainty and pricing decisions in the traditional production and inventory planning process, our approach also accommodates the complexity of time-varying cost and capacity constraints. Finally, our numerical results show......Stochastic lot-sizing problems have been addressed quite extensively, but relatively few studies also consider marketing factors, such as pricing. In this paper, we address a joint stochastic lot-sizing and pricing problem with capacity constraints and backlogging for a firm that produces a single...... that the multi-phase heuristic algorithm solves the example problems effectively....

  5. Clean Grain Boundary Found in C14/Body-Center-Cubic Multi-Phase Metal Hydride Alloys

    Directory of Open Access Journals (Sweden)

    Hao-Ting Shen

    2016-06-01

    Full Text Available The grain boundaries of three Laves phase-related body-center-cubic (bcc solid-solution, metal hydride (MH alloys with different phase abundances were closely examined by scanning electron microscopy (SEM, transmission electron microscopy (TEM, and more importantly, electron backscatter diffraction (EBSD techniques. By using EBSD, we were able to identify the alignment of the crystallographic orientations of the three major phases in the alloys (C14, bcc, and B2 structures. This finding confirms the presence of crystallographically sharp interfaces between neighboring phases, which is a basic assumption for synergetic effects in a multi-phase MH system.

  6. Modeling optimizes PEM fuel cell durability using three-dimensional multi-phase computational fluid dynamics model

    OpenAIRE

    Maher A.R. Sadiq Al-Baghdadi

    2010-01-01

    Damage mechanisms in a proton exchange membrane (PEM) fuel cell are accelerated by mechanical stresses arising during fuel cell assembly (bolt assembling), and the stresses arise during fuel cell running, because it consists of the materials with different thermal expansion and swelling coefficients. Therefore, in order to acquire a complete understanding of the damage mechanisms in the membrane and gas diffusion layers, mechanical response under steady-state hygro-thermal stresses should be ...

  7. High-Speed Visualization of Evaporation Phenomena from Tungsten Based Electrode in Multi-Phase AC Arc

    Science.gov (United States)

    Tanaka, Manabu; Hashizume, Taro; Imatsuji, Tomoyuki; Nawata, Yushi; Watanabe, Takayuki

    2015-09-01

    A multi-phase AC arc has been developed for applications in various fields of engineering because it possesses unique advantages such as high energy efficiency. However, understanding of fundamental phenomena in the multi-phase AC arc is still insufficient for practical use. Purpose of this study is to investigate electrode erosion mechanism by high-speed visualization of the electrode metal vapor in the arc. Results indicated that the electrode mainly evaporated at anodic period, leading to the arc constriction. Moreover, evaporation of W electrode with 2wt% La2O3 at the anodic period was much higher than that with 2wt% ThO2. This can be explained by different properties of these oxide additives. Evaporation of the oxide additive resulted in the arc constriction, which accelerated the evaporation of W electrode. Therefore, addition of La2O3 with lower melting and boiling point than ThO2 lead to stronger arc constriction, resulting in severer evaporation of W electrode.

  8. Inference in Complex Systems Using Multi-Phase MCMC Sampling With Gradient Matching Burn-in

    OpenAIRE

    Lazarus, Alan; Husmeier, Dirk; Papamarkou, Theodore

    2017-01-01

    We propose a novel method for parameter inference that builds on the current research in gradient matching surrogate likelihood spaces. Adopting a three phase technique, we demonstrate that it is possible to obtain parameter estimates of limited bias whilst still adopting the paradigm of the computationally cheap surrogate approximation.

  9. Stability of retained austenite in multi-phase microstructure during austempering and its effect on the ductility of a low carbon steel

    Energy Technology Data Exchange (ETDEWEB)

    Xie, Z.J.; Ren, Y.Q.; Zhou, W.H. [School of Materials Science and Engineering, University of Science and Technology, Beijing (China); Yang, J.R. [Department of Materials Science and Engineering, National Taiwan University, Taipei, Taiwan (China); Shang, C.J., E-mail: cjshang@ustb.edu.cn [School of Materials Science and Engineering, University of Science and Technology, Beijing (China); Misra, R.D.K. [Laboratory for Excellence in Advanced Steel Research, Center for Structural and Functional Materials, Institute for Material Research and Innovation, University of Louisiana at Lafayette, P.O. Box 44130, Lafayette, LA 70503 (United States)

    2014-05-01

    The contribution of multi-phase microstructure and retained austenite on mechanical properties of austempered and intercritical annealed Fe–0.23C–1.8Mn–1.35Si (wt%) steel was studied. The multi-phase microstructure comprised of intercritical ferrite (IF), bainite/martensite, and retained austenite. During austempering, the retained austenite was stabilized, which was studied using a combination of experimental (XRD, TEM) and thermodynamic analysis. The termination of bainitic transformation combined with carbon rejection into residual austenite during the second step austempering treatment is believed to be the underlying basis for stabilization of retained austenite. This led to significant increase in uniform and total elongation (25% and 36%, respectively) and the product of tensile strength and % elongation was 33 GPa%. The work hardening behavior of retained austenite exhibited a three-stage process such that necking was delayed. The increased work hardening rate is attributed to the multi-phase microstructure and TRIP effect.

  10. Contribution of the study of thermal interaction: modelling of a thermal blast in a multi-phase medium

    International Nuclear Information System (INIS)

    Scott, Edouard

    1978-01-01

    This research thesis aims at being a contribution to the safety of nuclear facilities by reporting the study of the interaction between nuclear fuel and coolant in simplified conditions. It focuses on the thermal aspect of this interaction between a very hot body and an easily vaporized cold body, which could produce a blast. Thus, this author addresses the field of existence of a thermal blast, and reports the development of a hydrodynamic model which takes the heterogeneous nature of the interacting medium into account, in order to precisely describe the conditions of fuel fragmentation. This model includes the propagation of a shock in a mixture, and the calculation of a multi-phase flow in the reaction zone, and proposes criteria for a self-sustained shock wave propagation in the reactive medium. Results are compared with those obtained with the Bankoff model [fr

  11. Theory-Guided Materials Design of Multi-Phase Ti-Nb Alloys with Bone-Matching Elastic Properties

    Directory of Open Access Journals (Sweden)

    Jörg Neugebauer

    2012-10-01

    Full Text Available We present a scale-bridging approach for modeling the integral elasticresponse of polycrystalline composite that is based on a multi-disciplinary combination of(i parameter-free first-principles calculations of thermodynamic phase stability andsingle-crystal elastic stiffness; and (ii homogenization schemes developed forpolycrystalline aggregates and composites. The modeling is used as a theory-guidedbottom-up materials design strategy and applied to Ti-Nb alloys as promising candidatesfor biomedical implant applications. The theoretical results (i show an excellent agreementwith experimental data and (ii reveal a decisive influence of the multi-phase character ofthe polycrystalline composites on their integral elastic properties. The study shows thatthe results based on the density functional theory calculations at the atomistic level canbe directly used for predictions at the macroscopic scale, effectively scale-jumping severalorders of magnitude without using any empirical parameters.

  12. A new general methodology for incorporating physico-chemical transformations into multi-phase wastewater treatment process models.

    Science.gov (United States)

    Lizarralde, I; Fernández-Arévalo, T; Brouckaert, C; Vanrolleghem, P; Ikumi, D S; Ekama, G A; Ayesa, E; Grau, P

    2015-05-01

    This paper introduces a new general methodology for incorporating physico-chemical and chemical transformations into multi-phase wastewater treatment process models in a systematic and rigorous way under a Plant-Wide modelling (PWM) framework. The methodology presented in this paper requires the selection of the relevant biochemical, chemical and physico-chemical transformations taking place and the definition of the mass transport for the co-existing phases. As an example a mathematical model has been constructed to describe a system for biological COD, nitrogen and phosphorus removal, liquid-gas transfer, precipitation processes, and chemical reactions. The capability of the model has been tested by comparing simulated and experimental results for a nutrient removal system with sludge digestion. Finally, a scenario analysis has been undertaken to show the potential of the obtained mathematical model to study phosphorus recovery. Copyright © 2015 Elsevier Ltd. All rights reserved.

  13. Comparison of structure, morphology, and leach characteristics of multi-phase ceramics produced via melt processing and hot isostatic pressing

    Science.gov (United States)

    Dandeneau, Christopher S.; Hong, Tao; Brinkman, Kyle S.; Vance, Eric R.; Amoroso, Jake W.

    2018-04-01

    Melt processing of multi-phase ceramic waste forms offers potential advantages over traditional solid-state synthesis methods given both the prevalence of melters currently in use and the ability to reduce the possibility of airborne radionuclide contamination. In this work, multi-phase ceramics with a targeted hollandite composition of Ba1.0Cs0.3Cr1.0Al0.3Fe1.0Ti5.7O16 were fabricated by melt processing at 1675 °C and hot isostatic pressing (HIP) at 1250 and 1300 °C. X-ray diffraction analysis (XRD) confirmed hollandite as the major phase in all specimens. Zirconolite/pyrochlore peaks and weaker perovskite reflections were observed after melt processing, while HIP samples displayed prominent perovskite peaks and low-intensity zirconolite reflections. Melt processing produced specimens with large (>50 μm) well-defined hollandite grains, while HIP yielded samples with a more fine-grained morphology. Elemental analysis showed "islands" rich in Cs and Ti across the surface of the 1300 °C HIP sample, suggesting partial melting and partitioning of Cs into multiple phases. Photoemission data revealed multiple Cs 3d spin-orbit pairs for the HIP samples, with the lower binding energy doublets likely corresponding to Cs located in more leachable phases. Among all specimens examined, the melt-processed sample exhibited the lowest fractional release rates for Rb and Cs. However, the retention of Sr and Mo was greater in the HIP specimens.

  14. Modeling of Multi Phase Flow in Porous Media: Operator Splitting, Front Tracking, Interfacial Area and Network Models

    Energy Technology Data Exchange (ETDEWEB)

    Nordhaug, Hans Fredrik

    2001-07-01

    In reservoir problems we consider some or all of the following phases: Oil, gas, water and solid. The solid phase is normally assumed to be immobile and non-deforming, but in general this does not need to be the case. By multi phase flow we will mean the flow of oil, gas and water. The phases are categorized according to their different physical quantities. A hydrocarbon phase, may consist of different hydrocarbon components, e.g., the oil phase can contain several oil and gas types. In this work the components are neglected and only the phases are considered. A porous medium is any solid phase, e.g. sand stone, that is permeable. The flow in a porous medium takes place through connected pores in the rock. Regions on a larger scale that contain oil or gas are called reservoirs. The typical size of a reservoir is kilometers in each direction while the pore scale size is millimeters or less. Solving the Navier-Stokes equation at the pore scale to obtain the transport on a larger scale is not numerically feasible because of the huge difference in scales. Therefore, some averaging is necessary to go from the pore scale (micro scale) to the reservoir scale (macro scale). In this process the Navier-Stokes equations are replaced by macro scale equations that are solved for macro scale variables. The papers presented herein cover several topics in multi phase flow in porous media, and they address some central problems both on the micro scale as well as on the macro scale. In addition, operator splitting techniques have been developed for convection dominated non-linear transport equations.

  15. Non-linear behaviour of multi-phase MOX fuels: a micro-mechanical approach

    International Nuclear Information System (INIS)

    Rousette, S.; Gatt, J.M.; Michel, J.C.

    2005-01-01

    The modelling of mechanical pellet-clad interaction requires knowledge of the thermo-mechanical behaviour of nuclear fuels. Some nuclear fuels such as MOX are composed of several phases. The mechanical properties of these phases, which are elasto-visco-plastic in-pile, are changing in-pile. The objective is to formulate a mechanical behaviour law taking all the physical phenomena into account in the different phases, which can easily be introduced into a fuel rod modelling code. Consequently, Non-uniform Transformation Field Analysis (NTFA) is used on the one hand, to correctly capture the heterogeneity of the anelastic strain in the different phases and, on the other hand, to provide a simple overall constitutive law for computational codes. This method is a good way to describe the behaviour of MOX fuel. Transformation Field Analysis (TFA), which corresponds to piecewise uniform transformation fields, is used to perform a sensitivity study. (authors)

  16. Complexity reduction of multi-phase flows in heterogeneous porous media

    KAUST Repository

    Ghommem, Mehdi

    2013-01-01

    In this paper, we apply mode decomposition and interpolatory projection methods to speed up simulations of two-phase flows in highly heterogeneous porous media. We propose intrusive and non-intrusive model reduction approaches that enable a significant reduction in the dimension of the flow problem size while capturing the behavior of the fully-resolved solutions. In one approach, we employ the dynamic mode decomposition (DMD) and the discrete empirical interpolation method (DEIM). This approach does not require any modification of the reservoir simulation code but rather postprocesses a set of global snapshots to identify the dynamically-relevant structures associated with the flow behavior. In a second approach, we project the governing equations of the velocity and the pressure fields on the subspace spanned by their proper orthogonal decomposition (POD) modes. Furthermore, we use DEIM to approximate the mobility related term in the global system assembly and then reduce the online computational cost and make it independent of the fine grid. To show the effectiveness and usefulness of the aforementioned approaches, we consider the SPE 10 benchmark permeability field and present a variety of numerical examples of two-phase flow and transport. The proposed model reduction methods can be efficiently used when performing uncertainty quantification or optimization studies and history matching.

  17. Practical aspects and applications of the biological effective dose three-dimensional calculation for multi-phase radiotherapy treatment plans

    Science.gov (United States)

    Kauweloa, Kevin Ikaika

    The approximate BED (BEDA) is calculated for multi-phase cases due to current treatment planning systems (TPSs) being incapable of performing BED calculations. There has been no study on the mathematical accuracy and precision of BEDA relative to the true BED (BEDT), and how that might negatively impact patient care. The purpose of the first aim was to study the mathematical accuracy and precision in both hypothetical and clinical situations, while the next two aims were to create multi-phase BED optimization ideas for both multi-target liver stereotactic body radiation therapy (SBRT) cases, and gynecological cases where patients are treated with high-dose rate (HDR) brachytherapy along with external beam radiotherapy (EBRT). MATLAB algorithms created for this work were used to mathematically analyze the accuracy and precision of BEDA relative to BEDT in both hypothetical and clinical situations on a 3D basis. The organs-at-risk (OARs) of ten head & neck and ten prostate cancer patients were studied for the clinical situations. The accuracy of BEDA was shown to vary between OARs as well as between patients. The percentage of patients with an overall BEDA percent error less than 1% were, 50% for the Optic Chiasm and Brainstem, 70% for the Left and Right Optic Nerves, as well as the Rectum and Bladder, and 80% for the Normal Brain and Spinal Cord. As seen for each OAR among different patients, there were always cases where the percent error was greater than 1%. This is a cause for concern since the goal of radiation therapy is to reduce the overall uncertainty of treatment, and calculating BEDA distributions increases the treatment uncertainty with percent errors greater than 1%. The revealed inaccuracy and imprecision of BEDA supports the argument to use BEDT. The multi-target liver study involved applying BEDT in order to reduce the number of dose limits to one rather than have one for each fractionation scheme in multi-target liver SBRT treatments. A BEDT limit

  18. Predictions of wet natural gases condensation rates via multi-component and multi-phase simulation of supersonic separators

    International Nuclear Information System (INIS)

    Shooshtari, Seyed Heydar Rajaee; Shahsavand, Akbar

    2014-01-01

    Proper correction of water and heavy hydrocarbon dew points of sweet natural gases is essential from various technical and economical standpoints. Supersonic separators (3S) are proved to be capable of achieving these tasks with maximum reliability and minimal expenses. The majority of the previous articles have focused on the flow behavior of pure fluids across a 3S unit. Multicomponent fluid flow inside 3S accompanied with condensation phenomenon will drastically increase the complexity of the simulation process. We tackle this issue by considering a proper combination of fundamental governing equations and phase equilibrium calculations to predict various operating conditions and composition profiles across two multi-component and multi-phase 3S units. Various Iranian sweet gases are used as real case studies to demonstrate the importance of 3S unit practical applications. Simulation results clearly illustrate the effectiveness of 3S units for faithful dehydration of various natural gases, while successfully controlling its dew point, suitable for any practical applications. Conventional HYSYS simulation software is used to validate the simulation results

  19. COMPUTING

    CERN Multimedia

    M. Kasemann

    Overview In autumn the main focus was to process and handle CRAFT data and to perform the Summer08 MC production. The operational aspects were well covered by regular Computing Shifts, experts on duty and Computing Run Coordination. At the Computing Resource Board (CRB) in October a model to account for service work at Tier 2s was approved. The computing resources for 2009 were reviewed for presentation at the C-RRB. The quarterly resource monitoring is continuing. Facilities/Infrastructure operations Operations during CRAFT data taking ran fine. This proved to be a very valuable experience for T0 workflows and operations. The transfers of custodial data to most T1s went smoothly. A first round of reprocessing started at the Tier-1 centers end of November; it will take about two weeks. The Computing Shifts procedure was tested full scale during this period and proved to be very efficient: 30 Computing Shifts Persons (CSP) and 10 Computing Resources Coordinators (CRC). The shift program for the shut down w...

  20. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction CMS distributed computing system performed well during the 2011 start-up. The events in 2011 have more pile-up and are more complex than last year; this results in longer reconstruction times and harder events to simulate. Significant increases in computing capacity were delivered in April for all computing tiers, and the utilisation and load is close to the planning predictions. All computing centre tiers performed their expected functionalities. Heavy-Ion Programme The CMS Heavy-Ion Programme had a very strong showing at the Quark Matter conference. A large number of analyses were shown. The dedicated heavy-ion reconstruction facility at the Vanderbilt Tier-2 is still involved in some commissioning activities, but is available for processing and analysis. Facilities and Infrastructure Operations Facility and Infrastructure operations have been active with operations and several important deployment tasks. Facilities participated in the testing and deployment of WMAgent and WorkQueue+Request...

  1. COMPUTING

    CERN Multimedia

    P. McBride

    The Computing Project is preparing for a busy year where the primary emphasis of the project moves towards steady operations. Following the very successful completion of Computing Software and Analysis challenge, CSA06, last fall, we have reorganized and established four groups in computing area: Commissioning, User Support, Facility/Infrastructure Operations and Data Operations. These groups work closely together with groups from the Offline Project in planning for data processing and operations. Monte Carlo production has continued since CSA06, with about 30M events produced each month to be used for HLT studies and physics validation. Monte Carlo production will continue throughout the year in the preparation of large samples for physics and detector studies ramping to 50 M events/month for CSA07. Commissioning of the full CMS computing system is a major goal for 2007. Site monitoring is an important commissioning component and work is ongoing to devise CMS specific tests to be included in Service Availa...

  2. COMPUTING

    CERN Multimedia

    M. Kasemann

    Overview During the past three months activities were focused on data operations, testing and re-enforcing shift and operational procedures for data production and transfer, MC production and on user support. Planning of the computing resources in view of the new LHC calendar in ongoing. Two new task forces were created for supporting the integration work: Site Commissioning, which develops tools helping distributed sites to monitor job and data workflows, and Analysis Support, collecting the user experience and feedback during analysis activities and developing tools to increase efficiency. The development plan for DMWM for 2009/2011 was developed at the beginning of the year, based on the requirements from the Physics, Computing and Offline groups (see Offline section). The Computing management meeting at FermiLab on February 19th and 20th was an excellent opportunity discussing the impact and for addressing issues and solutions to the main challenges facing CMS computing. The lack of manpower is particul...

  3. Workshop `Measurement technology for steady state and transient multi phase flows`; Workshop `Messtechnik fuer stationaere und transiente Mehrphasenstroemungen`

    Energy Technology Data Exchange (ETDEWEB)

    Prasser, H.M. [ed.

    1997-12-01

    There is hardly another area of physics which has a comparable multiplicity of phenomena, like flow in multi-phase mixtures. The wishes of experimenters regarding measurement technique are correspondingly great: Apart from the conventional parameters of pressure, temperature and speed of flow, as great a collection with resolution of the instantaneous phase distribution is required. Also, the phases themselves frequently consists of several components, whose concentration should also be measured. The enormous progress which has recently been made with laser optics and tomographic processes, must be compared with a long list of unsolved problems, above all where non-contact measurement is concerned. The attempts at solutions are multifarious, the need for the exchange of experience is great and the comparson of measurement processes with one another must be strengthened. The workshop has set itself these targets. (orig.) [Deutsch] Es gibt kaum ein anderes Gebiet der Physik, das eine vergleichbare Vielfalt der Erscheinungen aufweist wie Stroemungen von Mehrphasengemischen. Entsprechend gross sind die Wuensche der Experimentatoren hinsichtlich der Messtechnik: Neben den klassischen Parametern Druck, Temperatur und Stroemungsgeschwindigkeit wird eine moeglichst hoch aufloesende Erfassung der momentanen Phasenverteilung benoetigt. Ausserdem bestehen die Phasen selbst haeufig aus mehreren Komponenten, deren Konzentration ebenfalls gemessen werden soll. Den enormen Fortschritten, ie mit laseroptischen und tomographischen Verfahren in letzter Zeit gemacht wurden, steht nach wie vor eine lange Liste bisher ungeloester Aufgaben gegenueber, vor allen Dingen, wenn beruehrungslos gemessen werden soll. Die Loesungsansaetze sind vielfaeltig, der Bedarf an Erfahrungsaustausch ist gross, der Vergleich der Messverfahren untereinander muss verstaerkt werden. Diesen Zielen hatte sich der Workshop ``Messtechnik fuer tationaere und transiente Mehrphasenstroemungen`` verschrieben.

  4. Multi-phase partitioning, ecological risk and fate of acidic pharmaceuticals in a wastewater receiving river: The role of colloids

    International Nuclear Information System (INIS)

    Duan, Yan-Ping; Meng, Xiang-Zhou; Wen, Zhi-Hao; Ke, Run-Hui; Chen, Ling

    2013-01-01

    The occurrence and multi-phase distribution of five pharmaceutical compounds were investigated in an urban wastewater treatment plant (WWTP) receiving river by analysis of pharmaceuticals in sediment, particulate matter, conventional dissolved phase (> 0.7 μm), colloidal phase (5 kDa to 0.7 μm), and truly dissolved phase (< 5 kDa) water. Diclofenac was found in all samples, followed by clofibric acid, ibuprofen, ketoprofen, and naproxen with the decreasing detection frequency. All targets in WWTP outfall site were higher than those in the upstream and downstream, indicating that the WWTP is an important input source of pharmaceuticals in the river. The colloidal phase contributed 10–14% of ketoprofen, 8–26% of naproxen, 17–36% of clofibric acid, 22–33% of diclofenac, and 9–28% of ibuprofen in the aquatic system, suggesting the colloids will play an important role as carrier to contaminants in the aquatic environment. Based on truly dissolved concentrations of pharmaceuticals in water, only the risk quotient (RQ) value for diclofenac towards fish was higher than 1, indicating it poses a potential risk to aquatic organisms. Finally, a Level III fugacity model was used to further assess the environmental fate of the selected pharmaceuticals (exemplified for clofibric acid and diclofenac). Both clofibric acid and diclofenac tend to accumulate in water compartment with the percentage of 99.7% and 60.6%, respectively. Advection in river is a significant loss process for clofibric acid (56.4%) and diclofenac (54.4%). - Highlights: ► WWTP is the main source of pharmaceuticals to the receiving river in Shanghai. ► The colloids contribute 9–36% to the total pharmaceutical concentration in water. ► Truly dissolved diclofenac poses a potential risk to aquatic organisms. ► Clofibric acid and diclofenac tend to accumulate in water compartment

  5. Violent flows in aqueous foams III: physical multi-phase model comparison with aqueous foam shock tube experiments

    Science.gov (United States)

    Redford, J. A.; Ghidaglia, J.-M.; Faure, S.

    2018-06-01

    Mitigation of blast waves in aqueous foams is a problem that has a strong dependence on multi-phase effects. Here, a simplified model is developed from the previous articles treating violent flows (D'Alesio et al. in Eur J Mech B Fluids 54:105-124, 2015; Faure and Ghidaglia in Eur J Mech B Fluids 30:341-359, 2011) to capture the essential phenomena. The key is to have two fluids with separate velocities to represent the liquid and gas phases. This allows for the interaction between the two phases, which may include terms for drag, heat transfer, mass transfer due to phase change, added mass effects, to be included explicitly in the model. A good test for the proposed model is provided by two experimental data sets that use a specially designed shock tube. The first experiment has a test section filled with spray droplets, and the second has a range of aqueous foams in the test section. A substantial attenuation of the shock wave is seen in both cases, but a large difference is observed in the sound speeds. The droplets cause no observable change from the air sound speed, while the foams have a reduced sound speed of approximately 50-75 m/s . In the model given here, an added mass term is introduced in the governing equations to capture the low sound speed. The match between simulation and experiment is found to be satisfactory for both droplets and the foam. This is especially good when considering the complexity of the physics and the effects that are unaccounted for, such as three-dimensionality and droplet atomisation. The resulting statistics illuminate the processes occurring in such flows.

  6. Multi-phase partitioning, ecological risk and fate of acidic pharmaceuticals in a wastewater receiving river: The role of colloids

    Energy Technology Data Exchange (ETDEWEB)

    Duan, Yan-Ping [State Key Laboratory of Pollution Control and Resources Reuse, College of Environmental Science and Engineering, Tongji University, Shanghai 200092 (China); Meng, Xiang-Zhou, E-mail: xzmeng@tongji.edu.cn [State Key Laboratory of Pollution Control and Resources Reuse, College of Environmental Science and Engineering, Tongji University, Shanghai 200092 (China); Wen, Zhi-Hao [State Key Laboratory of Pollution Control and Resources Reuse, College of Environmental Science and Engineering, Tongji University, Shanghai 200092 (China); Ke, Run-Hui [China National Research Institute of Food and Fermentation Industries, Beijing 100027 (China); Chen, Ling [State Key Laboratory of Pollution Control and Resources Reuse, College of Environmental Science and Engineering, Tongji University, Shanghai 200092 (China)

    2013-03-01

    The occurrence and multi-phase distribution of five pharmaceutical compounds were investigated in an urban wastewater treatment plant (WWTP) receiving river by analysis of pharmaceuticals in sediment, particulate matter, conventional dissolved phase (> 0.7 μm), colloidal phase (5 kDa to 0.7 μm), and truly dissolved phase (< 5 kDa) water. Diclofenac was found in all samples, followed by clofibric acid, ibuprofen, ketoprofen, and naproxen with the decreasing detection frequency. All targets in WWTP outfall site were higher than those in the upstream and downstream, indicating that the WWTP is an important input source of pharmaceuticals in the river. The colloidal phase contributed 10–14% of ketoprofen, 8–26% of naproxen, 17–36% of clofibric acid, 22–33% of diclofenac, and 9–28% of ibuprofen in the aquatic system, suggesting the colloids will play an important role as carrier to contaminants in the aquatic environment. Based on truly dissolved concentrations of pharmaceuticals in water, only the risk quotient (RQ) value for diclofenac towards fish was higher than 1, indicating it poses a potential risk to aquatic organisms. Finally, a Level III fugacity model was used to further assess the environmental fate of the selected pharmaceuticals (exemplified for clofibric acid and diclofenac). Both clofibric acid and diclofenac tend to accumulate in water compartment with the percentage of 99.7% and 60.6%, respectively. Advection in river is a significant loss process for clofibric acid (56.4%) and diclofenac (54.4%). - Highlights: ► WWTP is the main source of pharmaceuticals to the receiving river in Shanghai. ► The colloids contribute 9–36% to the total pharmaceutical concentration in water. ► Truly dissolved diclofenac poses a potential risk to aquatic organisms. ► Clofibric acid and diclofenac tend to accumulate in water compartment.

  7. Simulating the formation and evolution of galaxies: multi-phase description of the interstellar medium, star formation, and energy feedback

    Science.gov (United States)

    Merlin, E.; Chiosi, C.

    2007-10-01

    Context: Modelling the gaseous component of the interstellar medium (ISM) by Smoothed Particles Hydrodynamics in N-Body simulations (NB-TSPH) is still very crude when compared to the complex real situation. In the real ISM, many different and almost physically decoupled components (phases) coexist for long periods of time, and since they spread over wide ranges of density and temperature, they cannot be correctly represented by a unique continuous fluid. This would influence star formation which is thought to take place in clumps of cold, dense, molecular clouds, embedded in a warmer, neutral medium, that are almost freely moving throughout the tenuous hot ISM. Therefore, assuming that star formation is simply related to the gas content without specifying the component in which this is both observed and expected to occur may not be physically sound. Aims: We consider a multi-phase representation of the ISM in NB-TSPH simulations of galaxy formation and evolution with particular attention to the case of early-type galaxies. Methods: Cold gas clouds are described by the so-called sticky particles algorithm. They can freely move throughout the hot ISM medium; stars form within these clouds and the mass exchange among the three baryonic phases (hot gas, cold clouds, stars) is governed by radiative and Compton cooling and energy feedback by supernova (SN) explosions, stellar winds, and UV radiation. We also consider thermal conduction, cloud-cloud collisions, and chemical enrichment. Results: Our model agrees with and improves upon previous studies on the same subject. The results for the star formation rate agree with recent observational data on early-type galaxies. Conclusions: These models lend further support to the revised monolithic scheme of galaxy formation, which has recently been strengthened by high redshift data leading to the so-called downsizing and top-down scenarios.

  8. COMPUTING

    CERN Multimedia

    I. Fisk

    2013-01-01

    Computing activity had ramped down after the completion of the reprocessing of the 2012 data and parked data, but is increasing with new simulation samples for analysis and upgrade studies. Much of the Computing effort is currently involved in activities to improve the computing system in preparation for 2015. Operations Office Since the beginning of 2013, the Computing Operations team successfully re-processed the 2012 data in record time, not only by using opportunistic resources like the San Diego Supercomputer Center which was accessible, to re-process the primary datasets HTMHT and MultiJet in Run2012D much earlier than planned. The Heavy-Ion data-taking period was successfully concluded in February collecting almost 500 T. Figure 3: Number of events per month (data) In LS1, our emphasis is to increase efficiency and flexibility of the infrastructure and operation. Computing Operations is working on separating disk and tape at the Tier-1 sites and the full implementation of the xrootd federation ...

  9. COMPUTING

    CERN Multimedia

    I. Fisk

    2010-01-01

    Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion An activity that is still in progress is computing for the heavy-ion program. The heavy-ion events are collected without zero suppression, so the event size is much large at roughly 11 MB per event of RAW. The central collisions are more complex and...

  10. COMPUTING

    CERN Multimedia

    M. Kasemann P. McBride Edited by M-C. Sawley with contributions from: P. Kreuzer D. Bonacorsi S. Belforte F. Wuerthwein L. Bauerdick K. Lassila-Perini M-C. Sawley

    Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the comput...

  11. Energy and Resource Saving of Steelmaking Process: Utilization of Innovative Multi-phase Flux During Dephosphorization Process

    Science.gov (United States)

    Matsuura, Hiroyuki; Hamano, Tasuku; Zhong, Ming; Gao, Xu; Yang, Xiao; Tsukihashi, Fumitaka

    2014-09-01

    An increase in the utilization efficiency of CaO, one of the major fluxing agents used in various steelmaking processes, is required to reduce the amount of discharged slag and energy consumption of the process. The authors have intensively focused on the development of innovative dephosphorization process by using so called "multi-phase flux" composed of solid and liquid phases. This article summarizes the research on the above topic done by the authors, in which the formation mechanisms of P2O5-containing phase during CaO or 2CaO·SiO2 dissolution into molten slag, the phase relationship between solid and liquid phases at equilibrium, and thermodynamic properties of P2O5-containing phase have been clarified. The reactions between solid CaO or 2CaO·SiO2 and molten CaO-FeO x -SiO2-P2O5 slag were observed by dipping solid specimen in the synthesized slag at 1573 K or 1673 K. The formation of the CaO-FeO layer and dual-phase layer of solid 2CaO·SiO2 and FeO x -rich liquid phase was observed around the interface from the solid CaO side toward the bulk slag phase side. Condensation of P2O5 into 2CaO·SiO2 phase as 2CaO·SiO2-3CaO·P2O5 solid solution was observed in both cases of CaO and 2CaO·SiO2 as solid specimens. Measurement of the phase relationship for the CaO-FeO x -SiO2-P2O5 system confirmed the condensation of P2O5 in solid phase at low oxygen partial pressure. The thermodynamics of 2CaO·SiO2-3CaO·P2O5 solid solution are to be clarified to quantitatively simulate the dephosphorization process, and the current results are also introduced. Based on the above results, the reduction of CaO consumption, the discharged slag curtailment, and energy-saving effects have been discussed.

  12. Hot-working behavior of an advanced intermetallic multi-phase γ-TiAl based alloy

    Energy Technology Data Exchange (ETDEWEB)

    Schwaighofer, Emanuel, E-mail: emanuel.schwaighofer@unileoben.ac.at [Department of Physical Metallurgy and Materials Testing, Montanuniversität Leoben, Roseggerstr. 12, A-8700 Leoben (Austria); Clemens, Helmut [Department of Physical Metallurgy and Materials Testing, Montanuniversität Leoben, Roseggerstr. 12, A-8700 Leoben (Austria); Lindemann, Janny [Chair of Physical Metallurgy and Materials Technology, Brandenburg University of Technology, Konrad-Wachsmann-Allee 17, D-03046 Cottbus (Germany); GfE Fremat GmbH, Lessingstr. 41, D-09599 Freiberg (Germany); Stark, Andreas [Institute of Materials Research, Helmholtz-Zentrum Geesthacht, Max-Planck-Str. 1, D-21502 Geesthacht (Germany); Mayer, Svea [Department of Physical Metallurgy and Materials Testing, Montanuniversität Leoben, Roseggerstr. 12, A-8700 Leoben (Austria)

    2014-09-22

    deformation within the (α+β) phase field region, leading to refined and nearly texture-free α/α{sub 2}-grains. In conclusion, robust deformation parameters for the refinement of critical microstructural defects could be defined for the investigated multi-phase γ-TiAl based alloy.

  13. Hot-working behavior of an advanced intermetallic multi-phase γ-TiAl based alloy

    International Nuclear Information System (INIS)

    Schwaighofer, Emanuel; Clemens, Helmut; Lindemann, Janny; Stark, Andreas; Mayer, Svea

    2014-01-01

    the (α+β) phase field region, leading to refined and nearly texture-free α/α 2 -grains. In conclusion, robust deformation parameters for the refinement of critical microstructural defects could be defined for the investigated multi-phase γ-TiAl based alloy

  14. COMPUTING

    CERN Multimedia

    P. McBride

    It has been a very active year for the computing project with strong contributions from members of the global community. The project has focused on site preparation and Monte Carlo production. The operations group has begun processing data from P5 as part of the global data commissioning. Improvements in transfer rates and site availability have been seen as computing sites across the globe prepare for large scale production and analysis as part of CSA07. Preparations for the upcoming Computing Software and Analysis Challenge CSA07 are progressing. Ian Fisk and Neil Geddes have been appointed as coordinators for the challenge. CSA07 will include production tests of the Tier-0 production system, reprocessing at the Tier-1 sites and Monte Carlo production at the Tier-2 sites. At the same time there will be a large analysis exercise at the Tier-2 centres. Pre-production simulation of the Monte Carlo events for the challenge is beginning. Scale tests of the Tier-0 will begin in mid-July and the challenge it...

  15. COMPUTING

    CERN Multimedia

    M. Kasemann

    Introduction During the past six months, Computing participated in the STEP09 exercise, had a major involvement in the October exercise and has been working with CMS sites on improving open issues relevant for data taking. At the same time operations for MC production, real data reconstruction and re-reconstructions and data transfers at large scales were performed. STEP09 was successfully conducted in June as a joint exercise with ATLAS and the other experiments. It gave good indication about the readiness of the WLCG infrastructure with the two major LHC experiments stressing the reading, writing and processing of physics data. The October Exercise, in contrast, was conducted as an all-CMS exercise, where Physics, Computing and Offline worked on a common plan to exercise all steps to efficiently access and analyze data. As one of the major results, the CMS Tier-2s demonstrated to be fully capable for performing data analysis. In recent weeks, efforts were devoted to CMS Computing readiness. All th...

  16. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion The Tier 0 infrastructure was able to repack and promptly reconstruct heavy-ion collision data. Two copies were made of the data at CERN using a large CASTOR disk pool, and the core physics sample was replicated ...

  17. COMPUTING

    CERN Multimedia

    I. Fisk

    2012-01-01

    Introduction Computing continued with a high level of activity over the winter in preparation for conferences and the start of the 2012 run. 2012 brings new challenges with a new energy, more complex events, and the need to make the best use of the available time before the Long Shutdown. We expect to be resource constrained on all tiers of the computing system in 2012 and are working to ensure the high-priority goals of CMS are not impacted. Heavy ions After a successful 2011 heavy-ion run, the programme is moving to analysis. During the run, the CAF resources were well used for prompt analysis. Since then in 2012 on average 200 job slots have been used continuously at Vanderbilt for analysis workflows. Operations Office As of 2012, the Computing Project emphasis has moved from commissioning to operation of the various systems. This is reflected in the new organisation structure where the Facilities and Data Operations tasks have been merged into a common Operations Office, which now covers everything ...

  18. COMPUTING

    CERN Multimedia

    M. Kasemann

    CCRC’08 challenges and CSA08 During the February campaign of the Common Computing readiness challenges (CCRC’08), the CMS computing team had achieved very good results. The link between the detector site and the Tier0 was tested by gradually increasing the number of parallel transfer streams well beyond the target. Tests covered the global robustness at the Tier0, processing a massive number of very large files and with a high writing speed to tapes.  Other tests covered the links between the different Tiers of the distributed infrastructure and the pre-staging and reprocessing capacity of the Tier1’s: response time, data transfer rate and success rate for Tape to Buffer staging of files kept exclusively on Tape were measured. In all cases, coordination with the sites was efficient and no serious problem was found. These successful preparations prepared the ground for the second phase of the CCRC’08 campaign, in May. The Computing Software and Analysis challen...

  19. COMPUTING

    CERN Multimedia

    I. Fisk

    2010-01-01

    Introduction The first data taking period of November produced a first scientific paper, and this is a very satisfactory step for Computing. It also gave the invaluable opportunity to learn and debrief from this first, intense period, and make the necessary adaptations. The alarm procedures between different groups (DAQ, Physics, T0 processing, Alignment/calibration, T1 and T2 communications) have been reinforced. A major effort has also been invested into remodeling and optimizing operator tasks in all activities in Computing, in parallel with the recruitment of new Cat A operators. The teams are being completed and by mid year the new tasks will have been assigned. CRB (Computing Resource Board) The Board met twice since last CMS week. In December it reviewed the experience of the November data-taking period and could measure the positive improvements made for the site readiness. It also reviewed the policy under which Tier-2 are associated with Physics Groups. Such associations are decided twice per ye...

  20. COMPUTING

    CERN Multimedia

    M. Kasemann

    Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the co...

  1. COMPUTING

    CERN Multimedia

    2010-01-01

    Introduction Just two months after the “LHC First Physics” event of 30th March, the analysis of the O(200) million 7 TeV collision events in CMS accumulated during the first 60 days is well under way. The consistency of the CMS computing model has been confirmed during these first weeks of data taking. This model is based on a hierarchy of use-cases deployed between the different tiers and, in particular, the distribution of RECO data to T1s, who then serve data on request to T2s, along a topology known as “fat tree”. Indeed, during this period this model was further extended by almost full “mesh” commissioning, meaning that RECO data were shipped to T2s whenever possible, enabling additional physics analyses compared with the “fat tree” model. Computing activities at the CMS Analysis Facility (CAF) have been marked by a good time response for a load almost evenly shared between ALCA (Alignment and Calibration tasks - highest p...

  2. COMPUTING

    CERN Multimedia

    Contributions from I. Fisk

    2012-01-01

    Introduction The start of the 2012 run has been busy for Computing. We have reconstructed, archived, and served a larger sample of new data than in 2011, and we are in the process of producing an even larger new sample of simulations at 8 TeV. The running conditions and system performance are largely what was anticipated in the plan, thanks to the hard work and preparation of many people. Heavy ions Heavy Ions has been actively analysing data and preparing for conferences.  Operations Office Figure 6: Transfers from all sites in the last 90 days For ICHEP and the Upgrade efforts, we needed to produce and process record amounts of MC samples while supporting the very successful data-taking. This was a large burden, especially on the team members. Nevertheless the last three months were very successful and the total output was phenomenal, thanks to our dedicated site admins who keep the sites operational and the computing project members who spend countless hours nursing the...

  3. COMPUTING

    CERN Multimedia

    M. Kasemann

    Introduction A large fraction of the effort was focused during the last period into the preparation and monitoring of the February tests of Common VO Computing Readiness Challenge 08. CCRC08 is being run by the WLCG collaboration in two phases, between the centres and all experiments. The February test is dedicated to functionality tests, while the May challenge will consist of running at all centres and with full workflows. For this first period, a number of functionality checks of the computing power, data repositories and archives as well as network links are planned. This will help assess the reliability of the systems under a variety of loads, and identifying possible bottlenecks. Many tests are scheduled together with other VOs, allowing the full scale stress test. The data rates (writing, accessing and transfer¬ring) are being checked under a variety of loads and operating conditions, as well as the reliability and transfer rates of the links between Tier-0 and Tier-1s. In addition, the capa...

  4. COMPUTING

    CERN Multimedia

    Matthias Kasemann

    Overview The main focus during the summer was to handle data coming from the detector and to perform Monte Carlo production. The lessons learned during the CCRC and CSA08 challenges in May were addressed by dedicated PADA campaigns lead by the Integration team. Big improvements were achieved in the stability and reliability of the CMS Tier1 and Tier2 centres by regular and systematic follow-up of faults and errors with the help of the Savannah bug tracking system. In preparation for data taking the roles of a Computing Run Coordinator and regular computing shifts monitoring the services and infrastructure as well as interfacing to the data operations tasks are being defined. The shift plan until the end of 2008 is being put together. User support worked on documentation and organized several training sessions. The ECoM task force delivered the report on “Use Cases for Start-up of pp Data-Taking” with recommendations and a set of tests to be performed for trigger rates much higher than the ...

  5. COMPUTING

    CERN Multimedia

    P. MacBride

    The Computing Software and Analysis Challenge CSA07 has been the main focus of the Computing Project for the past few months. Activities began over the summer with the preparation of the Monte Carlo data sets for the challenge and tests of the new production system at the Tier-0 at CERN. The pre-challenge Monte Carlo production was done in several steps: physics generation, detector simulation, digitization, conversion to RAW format and the samples were run through the High Level Trigger (HLT). The data was then merged into three "Soups": Chowder (ALPGEN), Stew (Filtered Pythia) and Gumbo (Pythia). The challenge officially started when the first Chowder events were reconstructed on the Tier-0 on October 3rd. The data operations teams were very busy during the the challenge period. The MC production teams continued with signal production and processing while the Tier-0 and Tier-1 teams worked on splitting the Soups into Primary Data Sets (PDS), reconstruction and skimming. The storage sys...

  6. COMPUTING

    CERN Multimedia

    I. Fisk

    2013-01-01

    Computing operation has been lower as the Run 1 samples are completing and smaller samples for upgrades and preparations are ramping up. Much of the computing activity is focusing on preparations for Run 2 and improvements in data access and flexibility of using resources. Operations Office Data processing was slow in the second half of 2013 with only the legacy re-reconstruction pass of 2011 data being processed at the sites.   Figure 1: MC production and processing was more in demand with a peak of over 750 Million GEN-SIM events in a single month.   Figure 2: The transfer system worked reliably and efficiently and transferred on average close to 520 TB per week with peaks at close to 1.2 PB.   Figure 3: The volume of data moved between CMS sites in the last six months   The tape utilisation was a focus for the operation teams with frequent deletion campaigns from deprecated 7 TeV MC GEN-SIM samples to INVALID datasets, which could be cleaned up...

  7. COMPUTING

    CERN Multimedia

    I. Fisk

    2012-01-01

      Introduction Computing activity has been running at a sustained, high rate as we collect data at high luminosity, process simulation, and begin to process the parked data. The system is functional, though a number of improvements are planned during LS1. Many of the changes will impact users, we hope only in positive ways. We are trying to improve the distributed analysis tools as well as the ability to access more data samples more transparently.  Operations Office Figure 2: Number of events per month, for 2012 Since the June CMS Week, Computing Operations teams successfully completed data re-reconstruction passes and finished the CMSSW_53X MC campaign with over three billion events available in AOD format. Recorded data was successfully processed in parallel, exceeding 1.2 billion raw physics events per month for the first time in October 2012 due to the increase in data-parking rate. In parallel, large efforts were dedicated to WMAgent development and integrati...

  8. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction The Computing Team successfully completed the storage, initial processing, and distribution for analysis of proton-proton data in 2011. There are still a variety of activities ongoing to support winter conference activities and preparations for 2012. Heavy ions The heavy-ion run for 2011 started in early November and has already demonstrated good machine performance and success of some of the more advanced workflows planned for 2011. Data collection will continue until early December. Facilities and Infrastructure Operations Operational and deployment support for WMAgent and WorkQueue+Request Manager components, routinely used in production by Data Operations, are provided. The GlideInWMS and components installation are now deployed at CERN, which is added to the GlideInWMS factory placed in the US. There has been new operational collaboration between the CERN team and the UCSD GlideIn factory operators, covering each others time zones by monitoring/debugging pilot jobs sent from the facto...

  9. Multi-Phased, Post-Accident Support of the Fukushima Dai-Ichi Nuclear Power Plant - 12246

    Energy Technology Data Exchange (ETDEWEB)

    Gay, Arnaud; Gillet, Philippe; Ytournel, Bertrand; Varet, Thierry; David, Laurent; Prevost, Thierry; Redonnet, Carol; Piot, Gregoire; Jouaville, Stephane; Pagis, Georges [AREVA NC (France)

    2012-07-01

    operation results to date. AREVA's response to the Fukushima Dai-Ichi crisis was multi-phased: emergency aid and relief supply was sent within days after the accident; AREVA-Veolia engineering teams designed and implemented a water treatment solution in record time, only 3 months; and AREVA continues to support TEPCO and propose solutions for waste management, soil remediation and decontamination of the Fukushima Dai-Ichi site. Despite the huge challenges, the Actiflo{sup TM}-Rad project has been a success: the water treatment unit started on time and performed as expected. The performance is the result of many key elements: AREVA expertise in radioactive effluents decontamination, Veolia know-how in water treatment equipments in crisis environment, and of course AREVA and Veolia teams' creativity. The project success is also due to AREVA and Veolia teams' reactivity and high level of commitment with engineering teams working 24/7 in Japan, France and Germany. AREVA and Veolia deep knowledge of the Japanese industry ensured that the multi-cultural exchanges were not an issue. Finally the excellent overall project management and execution by TEPCO and other Japanese stakeholders was very efficient. The emergency water treatment was a key step of the roadmap towards restoration from the accident at Fukushima Dai-Ichi that TEPCO designed and keeps executing with success. (authors)

  10. A Study on the Role of Reaction Modeling in Multi-phase CFD-based Simulations of Chemical Looping Combustion; Impact du modele de reaction sur les simulations CFD de la combustion en boucle chimique

    Energy Technology Data Exchange (ETDEWEB)

    Kruggel-Emden, H.; Stepanek, F. [Department of Chemical Engineering, South Kensington Campus, Imperial College London, SW7 2AZ, London (United Kingdom); Kruggel-Emden, H.; Munjiza, A. [Department of Engineering, Queen Mary, University of London, Mile End Road, E1 4NS, London (United Kingdom)

    2011-03-15

    Chemical Looping Combustion is an energy efficient combustion technology for the inherent separation of carbon dioxide for both gaseous and solid fuels. For scale up and further development of this process multi-phase CFD-based simulations have a strong potential which rely on kinetic models for the solid/gaseous reactions. Reaction models are usually simple in structure in order to keep the computational cost low. They are commonly derived from thermogravimetric experiments. With only few CFD-based simulations performed on chemical looping combustion, there is a lack in understanding of the role and of the sensitivity of the applied chemical reaction model on the outcome of a simulation. The aim of this investigation is therefore the study of three different carrier materials CaSO{sub 4}, Mn{sub 3}O{sub 4} and NiO with the gaseous fuels H{sub 2} and CH{sub 4} in a batch type reaction vessel. Four reaction models namely the linear shrinking core, the spherical shrinking core, the Avrami-Erofeev and a recently proposed multi parameter model are applied and compared on a case by case basis. (authors)

  11. COMPUTING

    CERN Multimedia

    M. Kasemann

    CMS relies on a well functioning, distributed computing infrastructure. The Site Availability Monitoring (SAM) and the Job Robot submission have been very instrumental for site commissioning in order to increase availability of more sites such that they are available to participate in CSA07 and are ready to be used for analysis. The commissioning process has been further developed, including "lessons learned" documentation via the CMS twiki. Recently the visualization, presentation and summarizing of SAM tests for sites has been redesigned, it is now developed by the central ARDA project of WLCG. Work to test the new gLite Workload Management System was performed; a 4 times increase in throughput with respect to LCG Resource Broker is observed. CMS has designed and launched a new-generation traffic load generator called "LoadTest" to commission and to keep exercised all data transfer routes in the CMS PhE-DEx topology. Since mid-February, a transfer volume of about 12 P...

  12. Multi-phase simulation of fast ion profile flattening due to Alfvén eigenmodes in a DIII-D experiment

    International Nuclear Information System (INIS)

    Todo, Y.; Van Zeeland, M.A.; Bierwage, A.; Heidbrink, W.W.

    2014-01-01

    A multi-phase simulation that is a combination of classical simulation and hybrid simulation for energetic particles interacting with a magnetohydrodynamic (MHD) fluid is developed to simulate the nonlinear dynamics on the slowing down time scale of the energetic particles. The hybrid simulation code is extended with realistic beam deposition profile, collisions and losses, and is used for both the classical and hybrid phases. The code is run without MHD perturbations in the classical phase, while the interaction between the energetic particles and the MHD fluid is simulated in the hybrid phase. In a multi-phase simulation of DIII-D discharge #142111, the stored beam ion energy is saturated due to Alfvén eigenmodes (AE modes) at a level lower than in the classical simulation. After the stored fast ion energy is saturated, the hybrid simulation is run continuously. It is demonstrated that the fast ion spatial profile is significantly flattened due to the interaction with the multiple AE modes with amplitude v/v A  ∼ δB/B ∼ O(10 −4 ). The dominant AE modes are toroidal Alfvén eigenmodes (TAE modes), which is consistent with the experimental observation at the simulated moment. The amplitude of the temperature fluctuations brought about by the TAE modes is of the order of 1% of the equilibrium temperature. This is also comparable with electron cyclotron emission measurements in the experiment. (paper)

  13. Robust iterative learning control for multi-phase batch processes: an average dwell-time method with 2D convergence indexes

    Science.gov (United States)

    Wang, Limin; Shen, Yiteng; Yu, Jingxian; Li, Ping; Zhang, Ridong; Gao, Furong

    2018-01-01

    In order to cope with system disturbances in multi-phase batch processes with different dimensions, a hybrid robust control scheme of iterative learning control combined with feedback control is proposed in this paper. First, with a hybrid iterative learning control law designed by introducing the state error, the tracking error and the extended information, the multi-phase batch process is converted into a two-dimensional Fornasini-Marchesini (2D-FM) switched system with different dimensions. Second, a switching signal is designed using the average dwell-time method integrated with the related switching conditions to give sufficient conditions ensuring stable running for the system. Finally, the minimum running time of the subsystems and the control law gains are calculated by solving the linear matrix inequalities. Meanwhile, a compound 2D controller with robust performance is obtained, which includes a robust extended feedback control for ensuring the steady-state tracking error to converge rapidly. The application on an injection molding process displays the effectiveness and superiority of the proposed strategy.

  14. A graphical user interface (GUI) toolkit for the calculation of three-dimensional (3D) multi-phase biological effective dose (BED) distributions including statistical analyses.

    Science.gov (United States)

    Kauweloa, Kevin I; Gutierrez, Alonso N; Stathakis, Sotirios; Papanikolaou, Niko; Mavroidis, Panayiotis

    2016-07-01

    A toolkit has been developed for calculating the 3-dimensional biological effective dose (BED) distributions in multi-phase, external beam radiotherapy treatments such as those applied in liver stereotactic body radiation therapy (SBRT) and in multi-prescription treatments. This toolkit also provides a wide range of statistical results related to dose and BED distributions. MATLAB 2010a, version 7.10 was used to create this GUI toolkit. The input data consist of the dose distribution matrices, organ contour coordinates, and treatment planning parameters from the treatment planning system (TPS). The toolkit has the capability of calculating the multi-phase BED distributions using different formulas (denoted as true and approximate). Following the calculations of the BED distributions, the dose and BED distributions can be viewed in different projections (e.g. coronal, sagittal and transverse). The different elements of this toolkit are presented and the important steps for the execution of its calculations are illustrated. The toolkit is applied on brain, head & neck and prostate cancer patients, who received primary and boost phases in order to demonstrate its capability in calculating BED distributions, as well as measuring the inaccuracy and imprecision of the approximate BED distributions. Finally, the clinical situations in which the use of the present toolkit would have a significant clinical impact are indicated. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  15. A study of multi-phase flow through the cathode side of an interdigitated flow field using a multi-fluid model

    DEFF Research Database (Denmark)

    Berning, Torsten; Odgaard, Madeleine; Kær, Søren Knudsen

    2010-01-01

    This work presents a study of multi-phase flow through the cathode side of a polymer electrolyte membrane fuel cell employing an interdigitated flow field plate. A previously published model has been extended in order to account for phase change kinetics, and a comparison between the interdigitated...... flow field design and a conventional straight channel design has been conducted. It is found that the parasitic pressure drop in the interdigitated design is in the range of a few thousand Pa and could be reduced to a few hundred Pa by choosing diffusion media with high in-plane permeability....... In the interdigitated design more product water is carried out of the cell in the vapor phase compared to the straight channel design which indicates that liquid water management might be less problematic. This effect also leads to the finding that in the interdigitated design more waste heat is carried out of the cell...

  16. Beyond the Black Box: Coupling x-ray tomographic imaging of multi-phase flow processes to numerical models and traditional laboratory measurements

    DEFF Research Database (Denmark)

    Wildenschild, Dorthe; Porter, M.L.; Schaap, M.G.

    Quantitative non-invasive imaging has evolved rapidly in the last decade, and is now being used to assess a variety of problems in vadose zone research, including unsaturated flow and transport of water and contaminants, macropore-dominated processes, soil-water-root interactions, more recent work...... on colloidal processes, and significant work on NAPL-water interactions . We are now able to use non-invasive imaging to probe processes that could not previously be quantified because of lack of opacity, resolution, or accurate techniques for quantitative measurement. This work presents an overview of recent...... advances in x-ray microtomography techniques that can generate high-resolution image-based data for (1) validation of pore-scale multi-phase flow models such as the lattice-Boltzmann technique and pore network models (with respect to fluid saturations, fluid distribution, and relationships among capillary...

  17. Towards a Usability and Error "Safety Net": A Multi-Phased Multi-Method Approach to Ensuring System Usability and Safety.

    Science.gov (United States)

    Kushniruk, Andre; Senathirajah, Yalini; Borycki, Elizabeth

    2017-01-01

    The usability and safety of health information systems have become major issues in the design and implementation of useful healthcare IT. In this paper we describe a multi-phased multi-method approach to integrating usability engineering methods into system testing to ensure both usability and safety of healthcare IT upon widespread deployment. The approach involves usability testing followed by clinical simulation (conducted in-situ) and "near-live" recording of user interactions with systems. At key stages in this process, usability problems are identified and rectified forming a usability and technology-induced error "safety net" that catches different types of usability and safety problems prior to releasing systems widely in healthcare settings.

  18. Modelling multi-phase liquid-sediment scour and resuspension induced by rapid flows using Smoothed Particle Hydrodynamics (SPH) accelerated with a Graphics Processing Unit (GPU)

    Science.gov (United States)

    Fourtakas, G.; Rogers, B. D.

    2016-06-01

    A two-phase numerical model using Smoothed Particle Hydrodynamics (SPH) is applied to two-phase liquid-sediments flows. The absence of a mesh in SPH is ideal for interfacial and highly non-linear flows with changing fragmentation of the interface, mixing and resuspension. The rheology of sediment induced under rapid flows undergoes several states which are only partially described by previous research in SPH. This paper attempts to bridge the gap between the geotechnics, non-Newtonian and Newtonian flows by proposing a model that combines the yielding, shear and suspension layer which are needed to predict accurately the global erosion phenomena, from a hydrodynamics prospective. The numerical SPH scheme is based on the explicit treatment of both phases using Newtonian and the non-Newtonian Bingham-type Herschel-Bulkley-Papanastasiou constitutive model. This is supplemented by the Drucker-Prager yield criterion to predict the onset of yielding of the sediment surface and a concentration suspension model. The multi-phase model has been compared with experimental and 2-D reference numerical models for scour following a dry-bed dam break yielding satisfactory results and improvements over well-known SPH multi-phase models. With 3-D simulations requiring a large number of particles, the code is accelerated with a graphics processing unit (GPU) in the open-source DualSPHysics code. The implementation and optimisation of the code achieved a speed up of x58 over an optimised single thread serial code. A 3-D dam break over a non-cohesive erodible bed simulation with over 4 million particles yields close agreement with experimental scour and water surface profiles.

  19. Interactive computer graphics applications for compressible aerodynamics

    Science.gov (United States)

    Benson, Thomas J.

    1994-01-01

    Three computer applications have been developed to solve inviscid compressible fluids problems using interactive computer graphics. The first application is a compressible flow calculator which solves for isentropic flow, normal shocks, and oblique shocks or centered expansions produced by two dimensional ramps. The second application couples the solutions generated by the first application to a more graphical presentation of the results to produce a desk top simulator of three compressible flow problems: 1) flow past a single compression ramp; 2) flow past two ramps in series; and 3) flow past two opposed ramps. The third application extends the results of the second to produce a design tool which solves for the flow through supersonic external or mixed compression inlets. The applications were originally developed to run on SGI or IBM workstations running GL graphics. They are currently being extended to solve additional types of flow problems and modified to operate on any X-based workstation.

  20. Multi-phase simultaneous segmentation of tumor in lung 4D-CT data with context information.

    Directory of Open Access Journals (Sweden)

    Zhengwen Shen

    Full Text Available Lung 4D computed tomography (4D-CT plays an important role in high-precision radiotherapy because it characterizes respiratory motion, which is crucial for accurate target definition. However, the manual segmentation of a lung tumor is a heavy workload for doctors because of the large number of lung 4D-CT data slices. Meanwhile, tumor segmentation is still a notoriously challenging problem in computer-aided diagnosis. In this paper, we propose a new method based on an improved graph cut algorithm with context information constraint to find a convenient and robust approach of lung 4D-CT tumor segmentation. We combine all phases of the lung 4D-CT into a global graph, and construct a global energy function accordingly. The sub-graph is first constructed for each phase. A context cost term is enforced to achieve segmentation results in every phase by adding a context constraint between neighboring phases. A global energy function is finally constructed by combining all cost terms. The optimization is achieved by solving a max-flow/min-cut problem, which leads to simultaneous and robust segmentation of the tumor in all the lung 4D-CT phases. The effectiveness of our approach is validated through experiments on 10 different lung 4D-CT cases. The comparison with the graph cut without context constraint, the level set method and the graph cut with star shape prior demonstrates that the proposed method obtains more accurate and robust segmentation results.

  1. Computer applications in thermochemistry

    International Nuclear Information System (INIS)

    Vana Varamban, S.

    1996-01-01

    Knowledge of equilibrium is needed under many practical situations. Simple stoichiometric calculations can be performed by the use of hand calculators. Multi-component, multi-phase gas - solid chemical equilibrium calculations are far beyond the conventional devices and methods. Iterative techniques have to be resorted. Such problems are most elegantly handled by the use of modern computers. This report demonstrates the possible use of computers for chemical equilibrium calculations in the field of thermochemistry and chemical metallurgy. Four modules are explained. To fit the experimental C p data and to generate the thermal functions, to perform equilibrium calculations to the defined conditions, to prepare the elaborate input to the equilibrium and to analyse the calculated results graphically. The principles of thermochemical calculations are briefly described. An extensive input guide is given. Several illustrations are included to help the understanding and usage. (author)

  2. Studies on Single-phase and Multi-phase Heat Pipe for LED Panel for Efficient Heat Dissipation

    Science.gov (United States)

    Vyshnave, K. C.; Rohit, G.; Maithreya, D. V. N. S.; Rakesh, S. G.

    2017-08-01

    The popularity of LED panel as a source of illumination has soared recently due to its high efficiency. However, the removal of heat that is produced in the chip is still a major challenge in its design since this has an adverse effect on its reliability. If high junction temperature develops, the colour of the emitted light may diminish over prolonged usage or even a colour shift may occur. In this paper, a solution has been developed to address this problem by using a combination of heat pipe and heat fin technology. A single-phase and a two-phase heat pipes have been designed theoretically and computational simulations carried out using ANSYS FLUENT. The results of the theoretical calculations and those obtained from the simulations are found to be in agreement with each other.

  3. Location capability of a sparse regional network (RSTN) using a multi-phase earthquake location algorithm (REGLOC)

    Energy Technology Data Exchange (ETDEWEB)

    Hutchings, L.

    1994-01-01

    The Regional Seismic Test Network (RSTN) was deployed by the US Department of Energy (DOE) to determine whether data recorded by a regional network could be used to detect and accurately locate seismic events that might be clandestine nuclear tests. The purpose of this paper is to evaluate the location capability of the RSTN. A major part of this project was the development of the location algorithm REGLOC and application of Basian a prior statistics for determining the accuracy of the location estimates. REGLOC utilizes all identifiable phases, including backazimuth, in the location. Ninty-four events, distributed throughout the network area, detected by both the RSTN and located by local networks were used in the study. The location capability of the RSTN was evaluated by estimating the location accuracy, error ellipse accuracy, and the percentage of events that could be located, as a function of magnitude. The location accuracy was verified by comparing the RSTN results for the 94 events with published locations based on data from the local networks. The error ellipse accuracy was evaluated by determining whether the error ellipse includes the actual location. The percentage of events located was assessed by combining detection capability with location capability to determine the percentage of events that could be located within the study area. Events were located with both an average crustal model for the entire region, and with regional velocity models along with station corrections obtained from master events. Most events with a magnitude <3.0 can only be located with arrivals from one station. Their average location errors are 453 and 414 km for the average- and regional-velocity model locations, respectively. Single station locations are very unreliable because they depend on accurate backazimuth estimates, and backazimuth proved to be a very unreliable computation.

  4. A New and General Formulation of the Parametric HFGMC Micromechanical Method for Three-Dimensional Multi-Phase Composites

    Science.gov (United States)

    Haj-Ali, Rami; Aboudi, Jacob

    2012-01-01

    The recent two-dimensional (2-D) parametric formulation of the high fidelity generalized method of cells (HFGMC) reported by the authors is generalized for the micromechanical analysis of three-dimensional (3-D) multiphase composites with periodic microstructure. Arbitrary hexahedral subcell geometry is developed to discretize a triply periodic repeating unit-cell (RUC). Linear parametric-geometric mapping is employed to transform the arbitrary hexahedral subcell shapes from the physical space to an auxiliary orthogonal shape, where a complete quadratic displacement expansion is performed. Previously in the 2-D case, additional three equations are needed in the form of average moments of equilibrium as a result of the inclusion of the bilinear terms. However, the present 3-D parametric HFGMC formulation eliminates the need for such additional equations. This is achieved by expressing the coefficients of the full quadratic polynomial expansion of the subcell in terms of the side or face average-displacement vectors. The 2-D parametric and orthogonal HFGMC are special cases of the present 3-D formulation. The continuity of displacements and tractions, as well as the equilibrium equations, are imposed in the average (integral) sense as in the original HFGMC formulation. Each of the six sides (faces) of a subcell has an independent average displacement micro-variable vector which forms an energy-conjugate pair with the transformed average-traction vector. This allows generating symmetric stiffness matrices along with internal resisting vectors for the subcells which enhances the computational efficiency. The established new parametric 3-D HFGMC equations are formulated and solution implementations are addressed. Several applications for triply periodic 3-D composites are presented to demonstrate the general capability and varsity of the present parametric HFGMC method for refined micromechanical analysis by generating the spatial distributions of local stress fields

  5. Formation mechanism of channel segregation in carbon steels by inclusion flotation: X-ray microtomography characterization and multi-phase flow modeling

    International Nuclear Information System (INIS)

    Cao, Yan Fei; Chen, Yun; Li, Dian Zhong

    2016-01-01

    Recent experimental dissections of steel ingots and multi-scale simulations have led to the discovery of a potential driving force for channel segregation: the flotation of oxide-based inclusion (D. Li et al., Nat. Commun. 5:5572 (2014)). Further experimental analysis and numerical modeling are necessary to clarify this mechanism in detail. In this work, the inclusions in a carbon steel ingot that exhibits severe channel segregations were characterized by the 3D X-ray microtomography, which revealed a significant enrichment and growth of inclusions in the channels. Based on above microtomography characterization, a 2D macrosegregation model encompassing the inclusion flotation was established. In the model, the motions of solid inclusions and liquid were described using the multi-phase flow scheme within the Euler-Lagrange framework. The benchmark simulations showed that sufficient inclusion populations with appropriate sizes are capable of altering the local flow patterns and destabilize the mushy zone, initiating the subsequent channel segregation. The continuous interplay between melt convection, inclusion flotation and solidification eventually causes the formation of macroscale channel. The predicted sizes and volume fraction of inclusions that are able to trigger the channel segregation effectively are consistent with the data obtained via microtomography characterization. The macrosegregation model was then applied to predict the channel segregations in an industrial carbon steel ingot. A rather good agreement of A-segregates was achieved between the simulation and the dissected ingot.

  6. A new technique for noise reduction at coronary CT angiography with multi-phase data-averaging and non-rigid image registration

    Energy Technology Data Exchange (ETDEWEB)

    Tatsugami, Fuminari; Higaki, Toru; Nakamura, Yuko; Yamagami, Takuji; Date, Shuji; Awai, Kazuo [Hiroshima University, Department of Diagnostic Radiology, Minami-ku, Hiroshima (Japan); Fujioka, Chikako; Kiguchi, Masao [Hiroshima University, Department of Radiology, Minami-ku, Hiroshima (Japan); Kihara, Yasuki [Hiroshima University, Department of Cardiovascular Medicine, Minami-ku, Hiroshima (Japan)

    2015-01-15

    To investigate the feasibility of a newly developed noise reduction technique at coronary CT angiography (CTA) that uses multi-phase data-averaging and non-rigid image registration. Sixty-five patients underwent coronary CTA with prospective ECG-triggering. The range of the phase window was set at 70-80 % of the R-R interval. First, three sets of consecutive volume data at 70 %, 75 % and 80 % of the R-R interval were prepared. Second, we applied non-rigid registration to align the 70 % and 80 % images to the 75 % image. Finally, we performed weighted averaging of the three images and generated a de-noised image. The image noise and contrast-to-noise ratio (CNR) in the proximal coronary arteries between the conventional 75 % and the de-noised images were compared. Two radiologists evaluated the image quality using a 5-point scale (1, poor; 5, excellent). On de-noised images, mean image noise was significantly lower than on conventional 75 % images (18.3 HU ± 2.6 vs. 23.0 HU ± 3.3, P < 0.01) and the CNR was significantly higher (P < 0.01). The mean image quality score for conventional 75 % and de-noised images was 3.9 and 4.4, respectively (P < 0.01). Our method reduces image noise and improves image quality at coronary CTA. (orig.)

  7. Initiating the Validation of CCIM Processability for Multi-phase all Ceramic (SYNROC) HLW Form: Plan for Test BFY14CCIM-C

    Energy Technology Data Exchange (ETDEWEB)

    Maio, Vince [Idaho National Lab. (INL), Idaho Falls, ID (United States)

    2014-08-01

    This plan covers test BFY14CCIM-C which will be a first–of–its-kind demonstration for the complete non-radioactive surrogate production of multi-phase ceramic (SYNROC) High Level Waste Forms (HLW) using Cold Crucible Induction Melting (CCIM) Technology. The test will occur in the Idaho National Laboratory’s (INL) CCIM Pilot Plant and is tentatively scheduled for the week of September 15, 2014. The purpose of the test is to begin collecting qualitative data for validating the ceramic HLW form processability advantages using CCIM technology- as opposed to existing ceramic–lined Joule Heated Melters (JHM) currently producing BSG HLW forms. The major objectives of BFY14CCIM-C are to complete crystalline melt initiation with a new joule-heated resistive starter ring, sustain inductive melting at temperatures between 1600 to 1700°C for two different relatively high conductive materials representative of the SYNROC ceramic formation inclusive of a HLW surrogate, complete melter tapping and pouring of molten ceramic material in to a preheated 4 inch graphite canister and a similar canister at room temperature. Other goals include assessing the performance of a new crucible specially designed to accommodate the tapping and pouring of pure crystalline forms in contrast to less recalcitrant amorphous glass, assessing the overall operational effectiveness of melt initiation using a resistive starter ring with a dedicated power source, and observing the tapped molten flow and subsequent relatively quick crystallization behavior in pans with areas identical to standard HLW disposal canisters. Surrogate waste compositions with ceramic SYNROC forming additives and their measured properties for inductive melting, testing parameters, pre-test conditions and modifications, data collection requirements, and sampling/post-demonstration analysis requirements for the produced forms are provided and defined.

  8. Long residence times of rapidly decomposable soil organic matter: application of a multi-phase, multi-component, and vertically resolved model (BAMS1) to soil carbon dynamics

    Science.gov (United States)

    Riley, W. J.; Maggi, F.; Kleber, M.; Torn, M. S.; Tang, J. Y.; Dwivedi, D.; Guerry, N.

    2014-07-01

    Accurate representation of soil organic matter (SOM) dynamics in Earth system models is critical for future climate prediction, yet large uncertainties exist regarding how, and to what extent, the suite of proposed relevant mechanisms should be included. To investigate how various mechanisms interact to influence SOM storage and dynamics, we developed an SOM reaction network integrated in a one-dimensional, multi-phase, and multi-component reactive transport solver. The model includes representations of bacterial and fungal activity, multiple archetypal polymeric and monomeric carbon substrate groups, aqueous chemistry, aqueous advection and diffusion, gaseous diffusion, and adsorption (and protection) and desorption from the soil mineral phase. The model predictions reasonably matched observed depth-resolved SOM and dissolved organic matter (DOM) stocks and fluxes, lignin content, and fungi to aerobic bacteria ratios. We performed a suite of sensitivity analyses under equilibrium and dynamic conditions to examine the role of dynamic sorption, microbial assimilation rates, and carbon inputs. To our knowledge, observations do not exist to fully test such a complicated model structure or to test the hypotheses used to explain observations of substantial storage of very old SOM below the rooting depth. Nevertheless, we demonstrated that a reasonable combination of sorption parameters, microbial biomass and necromass dynamics, and advective transport can match observations without resorting to an arbitrary depth-dependent decline in SOM turnover rates, as is often done. We conclude that, contrary to assertions derived from existing turnover time based model formulations, observed carbon content and Δ14C vertical profiles are consistent with a representation of SOM consisting of carbon compounds with relatively fast reaction rates, vertical aqueous transport, and dynamic protection on mineral surfaces.

  9. New insights on lithofacies architecture, sedimentological characteristics and volcanological evolution of pre-caldera (> 22 ka), multi-phase, scoria- and spatter-cones at Somma-Vesuvius

    Science.gov (United States)

    Sparice, Domenico; Scarpati, Claudio; Perrotta, Annamaria; Mazzeo, Fabio Carmine; Calvert, Andrew T.; Lanphere, Marvin A.

    2017-11-01

    Pre-caldera (> 22 ka) lateral activity at Somma-Vesuvius is related to scoria- and spatter-cone forming events of monogenetic or polygenetic nature. A new stratigraphic, sedimentological, textural and lithofacies investigation was performed on five parasitic cones (Pollena cones, Traianello cone, S. Maria a Castello cone and the recently found Terzigno cone) occurring below the Pomici di Base (22 ka) Plinian products emplaced during the first caldera collapse at Somma-Vesuvius. A new Ar/Ar age of 23.6 ± 0.3 ka obtained for the Traianello cone as well as the absence of a paleosol or reworked material between the S. Maria a Castello cone and the Pomici di Base deposits suggest that such cone-forming eruptions occurred near the upper limit of the pre-caldera period (22-39 ky). The stratigraphy of three of these eccentric cones (Pollena cones and Traianello cone) exhibits erosion surfaces, exotic tephras, volcaniclastic layers, paleosols, unconformity and paraconformity between superimposed eruptive units revealing their multi-phase, polygenetic evolution related to activation of separate vents and periods of quiescence. Such eccentric cones have been described as composed of scoria deposits and pure effusive lavas by previous authors. Lavas are here re-interpreted as welded horizons (lava-like) composed of coalesced spatter fragments whose pyroclastic nature is locally revealed by relicts of original fragments and remnants of clast outlines. These welded horizons show, locally, rheomorphic structures allowing to define them as emplaced as clastogenic lava flows. The lava-like facies is transitional, upward and downward, to less welded facies composed of agglutinated to unwelded spatter horizons in which clasts outlines are increasingly discernible. Such textural characteristics and facies variation are consistent with a continuous fall deposition of Hawaiian fire-fountains episodes alternated with Strombolian phases emplacing loose scoria deposits. High enrichment

  10. Ensemble averaged multi-phase Eulerian model for columnar/equiaxed solidification of a binary alloy: II. Simulation of the columnar-to-equiaxed transition (CET)

    International Nuclear Information System (INIS)

    Ciobanas, A I; Fautrelle, Y

    2007-01-01

    A new multi-phase Eulerian model for the columnar and equiaxed dendritic solidification has been developed. In this paper we first focus on the numerical simulation of quasi-steady solidification experiments in order to obtain corresponding CET maps. We have identified three main zones on the CET map: the pure columnar, the pure equiaxed zone and finally the mixed columnar+equiaxed zone. The mixed c/e zone was further quantified by means of a columnar fraction ε c which quantifies in a rigorous way the two coexisting structures. Since it intrinsically includes the solutal and the mechanical blocking effects, the new ensemble model unifies the semi-empirical Hunt's approach (pure mechanical blocking mechanism) and the Martorano et al approach (pure solutal blocking mechanism). Secondly the present model was used to simulate unidirectional solidification experiments. It was found that the columnar front evolved in a quasi-steady state until a time very close to the critical CET moment. It is also found that the equiaxed nucleation undercooling is close to the maximum columnar dendrite tip undercooling and that the CET is virtually independent of the equiaxed zone ahead of the columnar front. If the equiaxed zone is not taken into account it is observed that the columnar front velocity exhibits a sudden increase at the beginning of the solidification followed by a quasi-plateau corresponding to a quasi-state at the columnar tips and finally, above a critical time, an oscillatory evolution. The beginning of the oscillatory evolution of the columnar front was well correlated with the CET position measured in the experiments. We also find that this oscillatory evolution of the columnar front is very favourable for the fragmentation of the columnar dendrites and thus for the CET. In this respect, it seems that the unsteady regime of the columnar front with respect to the local cooling conditions represents the main cause for the CET phenomena, at least for the non

  11. Multi-phase distribution and comprehensive ecological risk assessment of heavy metal pollutants in a river affected by acid mine drainage.

    Science.gov (United States)

    Liao, Jianbo; Ru, Xuan; Xie, Binbin; Zhang, Wanhui; Wu, Haizhen; Wu, Chaofei; Wei, Chaohai

    2017-07-01

    To date, there is a lack of a comprehensive research on heavy metals detection and ecological risk assessment in river water, sediments, pore water (PW) and suspended solids (SS). Here, the concentrations of heavy metals, including Cu, Zn, Mn, Cd, Pb and As, and their distribution between the four phases was studied. Samples for analysis were taken from twelve sites of the Hengshi River, Guangdong Province, China, during the rainy and dry seasons. A new comprehensive ecological risk index (CERI) based on considering metal contents, pollution indices, toxicity coefficients and water categories is offered for prediction of potential risk on aquatic organisms. The results of comprehensive analysis showed that the highest concentrations of Cu, Zn and Mn of 6.42, 87.17 and 98.74mg/L, respectively, in PW were comparable with those in water, while concentrations of Cd, Pb and As of 609.5, 2757 and 96.38μg/L, respectively, were 2-5 times higher. The sum of the exchangeable and carbonate fractions of target metals in sediments followed the order of Cd > Mn > Zn > Pb > Cu > As. The distribution of heavy metals in phases followed the order of sediment > SS > water > PW, having the sum content in water and PW lower than 2% of total. The elevated ecological risk for a single metal and the phase were 34,585 for Cd and 1160 for water, respectively, implied Cd as a priority pollutant in the considered area. According to the CERI, the maximum risk value of 769.3 was smaller than 1160 in water, but higher than those in other phases. Out of considering the water categories and contribution coefficients, the CERI was proved to be more reliable for assessing the pollution of rivers with heavy metals. These results imply that the CERI has a potential of adequate assessment of multi-phase composite metals pollution. Copyright © 2017 Elsevier Inc. All rights reserved.

  12. Modeling the Physical Multi-Phase Interactions of HNO3 Between Snow and Air on the Antarctic Plateau (Dome C) and coast (Halley)

    Science.gov (United States)

    Chan, Hoi Ga; Frey, Markus M.; King, Martin D.

    2017-04-01

    Nitrogen oxides (NOx = NO + NO2) emissions from nitrate (NO3-) photolysis in snow affect the oxidising capacity of the lower troposphere especially in remote regions of the high latitudes with low pollution levels. The porous structure of snowpack allows the exchange of gases with the atmosphere driven by physicochemical processes, and hence, snow can act as both source and sink of atmospheric chemical trace gases. Current models are limited by poor process understanding and often require tuning parameters. Here, two multi-phase physical models were developed from first principles constrained by observed atmospheric nitrate, HNO3, to describe the air-snow interaction of nitrate. Similar to most of the previous approaches, the first model assumes that below a threshold temperature, To, the air-snow grain interface is pure ice and above To, a disordered interface (DI) emerges assumed to be covering the entire grain surface. The second model assumes that Air-Ice interactions dominate over the entire temperature range below melting and that only above the eutectic temperature, liquid is present in the form of micropockets in grooves. The models are validated with available year-round observations of nitrate in snow and air at a cold site on the Antarctica Plateau (Dome C, 75°06'S, 123°33'E, 3233 m a.s.l.) and at a relatively warm site on the Antarctica coast (Halley, 75°35'S, 26°39'E, 35 m a.s.l). The first model agrees reasonably well with observations at Dome C (Cv(RMSE) = 1.34), but performs poorly at Halley (Cv(RMSE) = 89.28) while the second model reproduces with good agreement observations at both sites without any tuning (Cv(RMSE) = 0.84 at both sites). It is therefore suggested that air-snow interactions of nitrate in the winter are determined by non-equilibrium surface adsorption and co-condensation on ice coupled with solid-state diffusion inside the grain. In summer, however, the air-snow exchange of nitrate is mainly driven by solvation into liquid

  13. Using pore-scale imaging and modeling to provide new insights in multi-phase flow, transport and reaction phenomena in porous media (Invited)

    Science.gov (United States)

    Bijeljic, B.; Andrew, M. G.; Menke, H. P.; Blunt, M. J.

    2013-12-01

    Advances in X ray imaging techniques made it possible not only to accurately describe solid and fluid(s) distributions in the pore space but also to study dynamics of multi-phase flow and reactive transport in-situ. This has opened up a range of new opportunities to better understand fundamental physics at the pore scale by experiment, and test and validate theoretical models in order to develop predictive tools at the pore scale and use it for upscaling. Firstly, we illustrate this concept by describing a new methodology for predicting non-Fickian transport in millimeter-sized three-dimensional micro-CT images of a beadpack, a sandstone, and a carbonate, representing porous media with an increasing degree of pore-scale complexity. The key strategy is to retain the full information on flow and transport signature of a porous medium by using probability distribution functions (PDFs) of voxel velocities for flow, and both PDFs of particle displacements and PDFs of particle transit times between voxels for transport. For this purpose, direct-simulation flow and transport model is used to analyse the relationship between pore structure, velocity, and the dynamics of the evolving plume. The model predictions for PDFs of particle displacements obtained by the model are in excellent agreement with those measured on similar cores in nuclear magnetic resonance experiments. A key determinant for non-Fickian transport is the spread in velocity distribution in the pore space. Further, we present micro-CT imaging of capillary trapping of scCO2 at reservoir conditions in a range of carbonates and sandstones having different pore structure and demonstrate that substantial quantities of scCO2 can be trapped in the pore space. Higher residual scCO2 saturations are found in sandstones compared to carbonates. The trapped ganglia exhibit different distribution of size, related to the inherent structure of pore space. Pore structures with large, open pores that are well connected lead

  14. The multi-phase winds of Markarian 231: from the hot, nuclear, ultra-fast wind to the galaxy-scale, molecular outflow

    Science.gov (United States)

    Feruglio, C.; Fiore, F.; Carniani, S.; Piconcelli, E.; Zappacosta, L.; Bongiorno, A.; Cicone, C.; Maiolino, R.; Marconi, A.; Menci, N.; Puccetti, S.; Veilleux, S.

    2015-11-01

    Mrk 231 is a nearby ultra-luminous IR galaxy exhibiting a kpc-scale, multi-phase AGN-driven outflow. This galaxy represents the best target to investigate in detail the morphology and energetics of powerful outflows, as well as their still poorly-understood expansion mechanism and impact on the host galaxy. In this work, we present the best sensitivity and angular resolution maps of the molecular disk and outflow of Mrk 231, as traced by CO(2-1) and (3-2) observations obtained with the IRAM/PdBI. In addition, we analyze archival deep Chandra and NuSTAR X-ray observations. We use this unprecedented combination of multi-wavelength data sets to constrain the physical properties of both the molecular disk and outflow, the presence of a highly-ionized ultra-fast nuclear wind, and their connection. The molecular CO(2-1) outflow has a size of 1 kpc, and extends in all directions around the nucleus, being more prominent along the south-west to north-east direction, suggesting a wide-angle biconical geometry. The maximum projected velocity of the outflow is nearly constant out to 1 kpc, thus implying that the density of the outflowing material must decrease from the nucleus outwards as r-2. This suggests that either a large part of the gas leaves the flow during its expansion or that the bulk of the outflow has not yet reached out to 1 kpc, thus implying a limit on its age of 1 Myr. Mapping the mass and energy rates of the molecular outflow yields dot {M} OF = [500-1000] M⊙ yr-1 and Ėkin,OF = [7-10] × 1043 erg s-1. The total kinetic energy of the outflow is Ekin,OF is of the same order of the total energy of the molecular disk, Edisk. Remarkably, our analysis of the X-ray data reveals a nuclear ultra-fast outflow (UFO) with velocity -20 000 km s-1, dot {M}UFO = [0.3-2.1] M⊙ yr-1, and momentum load dot {P}UFO/ dot {P}rad = [0.2-1.6]. We find Ėkin,UFO Ėkin,OF as predicted for outflows undergoing an energy conserving expansion. This suggests that most of the UFO

  15. Computational advances in transition phase analysis

    International Nuclear Information System (INIS)

    Morita, K.; Kondo, S.; Tobita, Y.; Shirakawa, N.; Brear, D.J.; Fischer, E.A.

    1994-01-01

    In this paper, historical perspective and recent advances are reviewed on computational technologies to evaluate a transition phase of core disruptive accidents in liquid-metal fast reactors. An analysis of the transition phase requires treatment of multi-phase multi-component thermohydraulics coupled with space- and energy-dependent neutron kinetics. Such a comprehensive modeling effort was initiated when the program of SIMMER-series computer code development was initiated in the late 1970s in the USA. Successful application of the latest SIMMER-II in USA, western Europe and Japan have proved its effectiveness, but, at the same time, several areas that require further research have been identified. Based on the experience and lessons learned during the SIMMER-II application through 1980s, a new project of SIMMER-III development is underway at the Power Reactor and Nuclear Fuel Development Corporation (PNC), Japan. The models and methods of SIMMER-III are briefly described with emphasis on recent advances in multi-phase multi-component fluid dynamics technologies and their expected implication on a future reliable transition phase analysis. (author)

  16. HAI: A novel airborne multi-channel hygrometer for fast multi-phase H2O quantification: Performance of the HAI instrument during the first flights on the German HALO aircraft

    Science.gov (United States)

    Buchholz, B.; Ebert, V.; Kraemer, M.; Afchine, A.

    2014-12-01

    Common gas phase H2O measurements on fast airborne platforms e.g. using backward facing or "Rosemount"-inlets can lead to a high risk of ice and droplets contamination. In addition, currently no single hygrometer exists that allows a simultaneous, high-speed measurement of all phases (gas, liquid, ice) with the same detection principle. In the rare occasions multi-phase measurements are realized, gas-and condensed-phase observations rely on different methods, instruments and calibration strategies so that precision and accuracy levels are quite difficult to quantify. This is effectively avoided by the novel TDLAS instrument, HAI, Hygrometer for Atmospheric Investigation, which allows a simultaneous, high speed, multi-phase detection without any sensor calibration in a unique "2+2" channel concept. Hai combines two independent wavelength channels, at 1.4 µm and at 2.6 µm, for a wide dynamic range from 1 to 30 000 ppmv, with a simultaneous closed path (extractive) and open path detection. Thus, "Total", i.e. gas-phase plus condensed-phase water is measured by sampling via a forward facing inlet into "closed-path" extractive cells. A selective, sampling-free, high speed gas phase detection is realized via a dual-wavelength "open-path" cell placed outside of the aircraft fuselage. All channels can be sampled with 120 Hz (measurement cycle time Dt=1.6 ms) allowing an unprecedented spatial resolution of 30 cm at 900 km/h. The evaluation of the individual multi-channel raw-data is done post flight, without any channel interdependencies, in calibration-free mode, thus allowing fast, accurate and precise multi-phase water detection in flight. The performance could be shown in more than 200 net flights hours in three scientific flight campaigns (TACTS, ESMVal, ML-CIRRUS) on the new German HALO aircraft. In addition the level of the accuracy of the calibration free evaluation was evaluated at the German national primary water vapor standard.

  17. 2nd International Conference on Multiscale Computational Methods for Solids and Fluids

    CERN Document Server

    2016-01-01

    This volume contains the best papers presented at the 2nd ECCOMAS International Conference on Multiscale Computations for Solids and Fluids, held June 10-12, 2015. Topics dealt with include multiscale strategy for efficient development of scientific software for large-scale computations, coupled probability-nonlinear-mechanics problems and solution methods, and modern mathematical and computational setting for multi-phase flows and fluid-structure interaction. The papers consist of contributions by six experts who taught short courses prior to the conference, along with several selected articles from other participants dealing with complementary issues, covering both solid mechanics and applied mathematics. .

  18. Study of retained austenite and nano-scale precipitation and their effects on properties of a low alloyed multi-phase steel by the two-step intercritical treatment

    Energy Technology Data Exchange (ETDEWEB)

    Xie, Z.J.; Han, G., E-mail: hangang@mater.ustb.edu.cn; Zhou, W.H.; Zeng, C.Y.; Shang, C.J., E-mail: cjshang@ustb.edu.cn

    2016-03-15

    Microstructure evolution and properties were studied in a low carbon low alloyed hot-rolled bainitic steel by annealing and annealing plus tempering. Microstructure of the hot-rolled steel consists of lath bainite and martensite. By annealing at 720 °C for 30 min and water quenching, multi-phase microstructure consisting of intercritical ferrite, tempered bainite/martensite, retained austenite and fresh martensite was obtained. With increasing annealing temperature to 760 °C, microstructure of the steel consisted of intercritical ferrite, fresh martensite without retained austenite. After the second step of tempering at 680 °C for samples annealed both at 720 °C and 760 °C, ~ 8–9% volume fraction of retained austenite was obtained in the multi-phase microstructure. Moreover, fine precipitates of VC with size smaller than 10 nm and copper precipitates with size of ~ 10–50 nm were obtained after tempering. Results from scanning transmission electron microscopy (STEM) give evidence to support that the partitioning of Mn, Ni and Cu is of significance for retained austenite stabilization. Due to the combined contribution of multiphase microstructure, the transformation-induced-plasticity effect of retained austenite and strengthening effect of nanometer-sized precipitates, yield strength greater than 800 MPa, yield to tensile ratio of 0.9, uniform elongation of ~ 9% and good low temperature impact toughness of 147 J at − 40 °C were achieved. - Highlights: • Stable retained austenite was produced in a low alloyed steel. • Partition of Mn, Ni and Cu was confirmed by STEM for austenite stabilization. • Nano-sized VC and Cu precipitates were achieved by second tempering. • High strength–high toughness with low Y/T ratio was obtained.

  19. Application of GPU to computational multiphase fluid dynamics

    International Nuclear Information System (INIS)

    Nagatake, T; Kunugi, T

    2010-01-01

    The MARS (Multi-interfaces Advection and Reconstruction Solver) [1] is one of the surface volume tracking methods for multi-phase flows. Nowadays, the performance of GPU (Graphics Processing Unit) is much higher than the CPU (Central Processing Unit). In this study, the GPU was applied to the MARS in order to accelerate the computation of multi-phase flows (GPU-MARS), and the performance of the GPU-MARS was discussed. From the performance of the interface tracking method for the analyses of one-directional advection problem, it is found that the computing time of GPU(single GTX280) was around 4 times faster than that of the CPU (Xeon 5040, 4 threads parallelized). From the performance of Poisson Solver by using the algorithm developed in this study, it is found that the performance of the GPU showed around 30 times faster than that of the CPU. Finally, it is confirmed that the GPU showed the large acceleration of the fluid flow computation (GPU-MARS) compared to the CPU. However, it is also found that the double-precision computation of the GPU must perform with very high precision.

  20. Multidetector helical CT plus superparamagnetic iron oxide-enhanced MR imaging for focal hepatic lesions in cirrhotic liver: A comparison with multi-phase CT during hepatic arteriography

    International Nuclear Information System (INIS)

    Yukisawa, Seigo; Okugawa, Hidehiro; Masuya, Yoshio; Okabe, Shinichirou; Fukuda, Hiroyuki; Yoshikawa, Masaharu; Ebara, Masaaki; Saisho, Hiromitsu

    2007-01-01

    The aim of this study was to evaluate multidetector helical computed tomography (MDCT), superparamagnetic iron oxide (SPIO)-enhanced magnetic resonance (MR) imaging, and CT arterial portography (CTAP) and CT during hepatic arteriography (CTHA) for the detection and diagnosis of hepatocellular carcinomas (HCC). This included visual correlations of MDCT and SPIO-MR imaging in the detection of HCC using receiver operating characteristic (ROC) analysis. Twenty-five patients with 57 nodular HCCs were retrospectively analyzed. A total of 200 segments, including 49 segments with 57 HCCs, were reviewed independently by three observers. Each observer read four sets of images (set 1, MDCT; set 2, unenhanced and SPIO-enhanced MR images; set 3, combined MDCT and SPIO-enhanced MR images; set 4, combined CTAP and CTHA). The mean Az values representing the diagnostic accuracy for HCCs of sets 1, 2, 3, and 4 were 0.777, 0.814, 0.849, and 0.911, respectively, and there was no significant difference between sets 3 and 4. The sensitivity of set 4 was significantly higher than those of set 3 for all the lesions and for lesions 10 mm or smaller (p < 0.05); however, for lesions larger than 10 mm, the sensitivities of the two sets were similar. No significant difference in positive predictive value and specificity was observed between set 3 and set 4. Combined MDCT and SPIO-enhanced MR imaging may obviate the need for more invasive CTAP and CTHA for the pre-therapeutic evaluation of patients with HCC more than 10 mm

  1. Trade-off analysis of discharge-desiltation-turbidity and ANN analysis on sedimentation of a combined reservoir-reach system under multi-phase and multi-layer conjunctive releasing operation

    Science.gov (United States)

    Huang, Chien-Lin; Hsu, Nien-Sheng; Wei, Chih-Chiang; Yao, Chun-Hao

    2017-10-01

    Multi-objective reservoir operation considering the trade-off of discharge-desiltation-turbidity during typhoons and sediment concentration (SC) simulation modeling are the vital components for sustainable reservoir management. The purposes of this study were (1) to analyze the multi-layer release trade-offs between reservoir desiltation and intake turbidity of downstream purification plants and thus propose a superior conjunctive operation strategy and (2) to develop ANFIS-based (adaptive network-based fuzzy inference system) and RTRLNN-based (real-time recurrent learning neural networks) substitute SC simulation models. To this end, this study proposed a methodology to develop (1) a series of multi-phase and multi-layer sediment-flood conjunctive release modes and (2) a specialized SC numerical model for a combined reservoir-reach system. The conjunctive release modes involve (1) an optimization model where the decision variables are multi-phase reduction/scaling ratios and the timings to generate a superior total release hydrograph for flood control (Phase I: phase prior to flood arrival, Phase II/III: phase prior to/subsequent to peak flow) and (2) a combination method with physical limitations regarding separation of the singular hydrograph into multi-layer release hydrographs for sediment control. This study employed the featured signals obtained from statistical quartiles/sediment duration curve in mesh segmentation, and an iterative optimization model with a sediment unit response matrix and corresponding geophysical-based acceleration factors, for efficient parameter calibration. This research applied the developed methodology to the Shihmen Reservoir basin in Taiwan. The trade-off analytical results using Typhoons Sinlaku and Jangmi as case examples revealed that owing to gravity current and re-suspension effects, Phase I + II can de-silt safely without violating the intake's turbidity limitation before reservoir discharge reaches 2238 m3/s; however

  2. Multi-phase reactive transport theory

    International Nuclear Information System (INIS)

    Lichtner, P.C.

    1995-07-01

    Physicochemical processes in the near-field region of a high-level waste repository may involve a diverse set of phenomena including flow of liquid and gas, gaseous diffusion, and chemical reaction of the host rock with aqueous solutions at elevated temperatures. This report develops some of the formalism for describing simultaneous multicomponent solute and heat transport in a two-phase system for partially saturated porous media. Diffusion of gaseous species is described using the Dusty Gas Model which provides for simultaneous Knudsen and Fickian diffusion in addition to Darcy flow. A new form of the Dusty Gas Model equations is derived for binary diffusion which separates the total diffusive flux into segregative and nonsegregative components. Migration of a wetting front is analyzed using the quasi-stationary state approximation to the Richards' equation. Heat-pipe phenomena are investigated for both gravity- and capillary-driven reflux of liquid water. An expression for the burnout permeability is derived for a gravity-driven heat-pipe. Finally an estimate is given for the change in porosity and permeability due to mineral dissolution which could occur in the region of condensate formation in a heat-pipe

  3. Uncertainty Quantification of Multi-Phase Closures

    Energy Technology Data Exchange (ETDEWEB)

    Nadiga, Balasubramanya T. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Baglietto, Emilio [Massachusetts Inst. of Technology (MIT), Cambridge, MA (United States)

    2017-10-27

    In the ensemble-averaged dispersed phase formulation used for CFD of multiphase ows in nuclear reactor thermohydraulics, closures of interphase transfer of mass, momentum, and energy constitute, by far, the biggest source of error and uncertainty. Reliable estimators of this source of error and uncertainty are currently non-existent. Here, we report on how modern Validation and Uncertainty Quanti cation (VUQ) techniques can be leveraged to not only quantify such errors and uncertainties, but also to uncover (unintended) interactions between closures of di erent phenomena. As such this approach serves as a valuable aide in the research and development of multiphase closures. The joint modeling of lift, drag, wall lubrication, and turbulent dispersion|forces that lead to tranfer of momentum between the liquid and gas phases|is examined in the frame- work of validation of the adiabatic but turbulent experiments of Liu and Banko , 1993. An extensive calibration study is undertaken with a popular combination of closure relations and the popular k-ϵ turbulence model in a Bayesian framework. When a wide range of super cial liquid and gas velocities and void fractions is considered, it is found that this set of closures can be validated against the experimental data only by allowing large variations in the coe cients associated with the closures. We argue that such an extent of variation is a measure of uncertainty induced by the chosen set of closures. We also nd that while mean uid velocity and void fraction pro les are properly t, uctuating uid velocity may or may not be properly t. This aspect needs to be investigated further. The popular set of closures considered contains ad-hoc components and are undesirable from a predictive modeling point of view. Consequently, we next consider improvements that are being developed by the MIT group under CASL and which remove the ad-hoc elements. We use non-intrusive methodologies for sensitivity analysis and calibration (using Dakota) to study sensitivities of the CFD representation (STARCCM+) of uid velocity pro les and void fraction pro les in the context of Shaver and Podowski, 2015 correction to lift, and the Lubchenko et al., 2017 formulation of wall lubrication.

  4. Heat transfer in multi-phase materials

    CERN Document Server

    Öchsner, Andreas

    2011-01-01

    This book provides a profound understanding, which physical processes and mechanisms cause the heat transfer in composite and cellular materials. It shows models for all important classes of composite materials and introduces into the latest advances. In three parts, the book covers Composite Materials (Part A), Porous and Cellular Materials (Part B) and the appearance of a conjoint solid phase and fluid aggregate (Part C).

  5. Long residence times of rapidly decomposable soil organic matter: application of a multi-phase, multi-component, and vertically-resolved model (TOUGHREACTv1) to soil carbon dynamics

    Science.gov (United States)

    Riley, W. J.; Maggi, F. M.; Kleber, M.; Torn, M. S.; Tang, J. Y.; Dwivedi, D.; Guerry, N.

    2014-01-01

    Accurate representation of soil organic matter (SOM) dynamics in Earth System Models is critical for future climate prediction, yet large uncertainties exist regarding how, and to what extent, the suite of proposed relevant mechanisms should be included. To investigate how various mechanisms interact to influence SOM storage and dynamics, we developed a SOM reaction network integrated in a one-dimensional, multi-phase, and multi-component reactive transport solver. The model includes representations of bacterial and fungal activity, multiple archetypal polymeric and monomeric carbon substrate groups, aqueous chemistry, aqueous advection and diffusion, gaseous diffusion, and adsorption (and protection) and desorption from the soil mineral phase. The model predictions reasonably matched observed depth-resolved SOM and dissolved organic carbon (DOC) stocks in grassland ecosystems as well as lignin content and fungi to aerobic bacteria ratios. We performed a suite of sensitivity analyses under equilibrium and dynamic conditions to examine the role of dynamic sorption, microbial assimilation rates, and carbon inputs. To our knowledge, observations do not exist to fully test such a complicated model structure or to test the hypotheses used to explain observations of substantial storage of very old SOM below the rooting depth. Nevertheless, we demonstrated that a reasonable combination of sorption parameters, microbial biomass and necromass dynamics, and advective transport can match observations without resorting to an arbitrary depth-dependent decline in SOM turnover rates, as is often done. We conclude that, contrary to assertions derived from existing turnover time based model formulations, observed carbon content and δ14C vertical profiles are consistent with a representation of SOM dynamics consisting of (1) carbon compounds without designated intrinsic turnover times, (2) vertical aqueous transport, and (3) dynamic protection on mineral surfaces.

  6. Simulating subsurface flow and transport on ultrascale computers using PFLOTRAN

    International Nuclear Information System (INIS)

    Mills, Richard Tran; Lu, Chuan; Lichtner, Peter C; Hammond, Glenn E

    2007-01-01

    We describe PFLOTRAN, a recently developed code for modeling multi-phase, multi-component subsurface flow and reactive transport using massively parallel computers. PFLOTRAN is built on top of PETSc, the Portable, Extensible Toolkit for Scientific Computation. Leveraging PETSc has allowed us to develop-with a relatively modest investment in development effort-a code that exhibits excellent performance on the largest-scale supercomputers. Very significant enhancements to the code are planned during our SciDAC-2 project. Here we describe the current state of the code, present an example of its use on Jaguar, the Cray XT3/4 system at Oak Ridge National Laboratory consisting of 11706 dual-core Opteron processor nodes, and briefly outline our future plans for the code

  7. Simulating subsurface flow and transport on ultrascale computers using PFLOTRAN

    Energy Technology Data Exchange (ETDEWEB)

    Mills, Richard Tran [Computational Earth Sciences Group, Computer Science and Mathematics Division, Oak Ridge National Laboratory, Oak Ridge, TN 37831-6015 (United States); Lu, Chuan [Hydrology, Geochemistry, and Geology Group, Earth and Environmental Sciences Division, Los Alamos National Laboratory, Los Alamos, NM 87545 (United States); Lichtner, Peter C [Hydrology, Geochemistry, and Geology Group, Earth and Environmental Sciences Division, Los Alamos National Laboratory, Los Alamos, NM 87545 (United States); Hammond, Glenn E [Hydrology Group, Environmental Technology Division, Pacific Northwest National Laboratory, Richland, WA 99352 (United States)

    2007-07-15

    We describe PFLOTRAN, a recently developed code for modeling multi-phase, multi-component subsurface flow and reactive transport using massively parallel computers. PFLOTRAN is built on top of PETSc, the Portable, Extensible Toolkit for Scientific Computation. Leveraging PETSc has allowed us to develop-with a relatively modest investment in development effort-a code that exhibits excellent performance on the largest-scale supercomputers. Very significant enhancements to the code are planned during our SciDAC-2 project. Here we describe the current state of the code, present an example of its use on Jaguar, the Cray XT3/4 system at Oak Ridge National Laboratory consisting of 11706 dual-core Opteron processor nodes, and briefly outline our future plans for the code.

  8. Optical Computing

    OpenAIRE

    Woods, Damien; Naughton, Thomas J.

    2008-01-01

    We consider optical computers that encode data using images and compute by transforming such images. We give an overview of a number of such optical computing architectures, including descriptions of the type of hardware commonly used in optical computing, as well as some of the computational efficiencies of optical devices. We go on to discuss optical computing from the point of view of computational complexity theory, with the aim of putting some old, and some very recent, re...

  9. Making eco-friendly transportation safer: developing computer-based simulations to assess of the impacts of bicycle accident prevention interventions on healthcare utilization.

    Science.gov (United States)

    Juhra, Christian; Borycki, Elizabeth M; Kushniruk, Andre W; Anderson, Jim; Anderson, Marilyn

    2011-01-01

    Computer-based modeling and simulations are becoming increasingly used for applications in health and safety. In this paper we describe a multi-phase project aimed at modeling bicycle accidents in Munster, Germany. The work involved a first phase of collecting empirical data on accident rates and severity. In the second phase a computer-based simulation model of bicycle accidents was created, using data from phase one to identify relevant parameters in the model. Finally, initial results from running the model are described that will be used to inform decision making regarding safety initiatives.

  10. Energy law preserving C0 finite element schemes for phase field models in two-phase flow computations

    International Nuclear Information System (INIS)

    Hua Jinsong; Lin Ping; Liu Chun; Wang Qi

    2011-01-01

    Highlights: → We study phase-field models for multi-phase flow computation. → We develop an energy-law preserving C0 FEM. → We show that the energy-law preserving method work better. → We overcome unphysical oscillation associated with the Cahn-Hilliard model. - Abstract: We use the idea in to develop the energy law preserving method and compute the diffusive interface (phase-field) models of Allen-Cahn and Cahn-Hilliard type, respectively, governing the motion of two-phase incompressible flows. We discretize these two models using a C 0 finite element in space and a modified midpoint scheme in time. To increase the stability in the pressure variable we treat the divergence free condition by a penalty formulation, under which the discrete energy law can still be derived for these diffusive interface models. Through an example we demonstrate that the energy law preserving method is beneficial for computing these multi-phase flow models. We also demonstrate that when applying the energy law preserving method to the model of Cahn-Hilliard type, un-physical interfacial oscillations may occur. We examine the source of such oscillations and a remedy is presented to eliminate the oscillations. A few two-phase incompressible flow examples are computed to show the good performance of our method.

  11. Computer group

    International Nuclear Information System (INIS)

    Bauer, H.; Black, I.; Heusler, A.; Hoeptner, G.; Krafft, F.; Lang, R.; Moellenkamp, R.; Mueller, W.; Mueller, W.F.; Schati, C.; Schmidt, A.; Schwind, D.; Weber, G.

    1983-01-01

    The computer groups has been reorganized to take charge for the general purpose computers DEC10 and VAX and the computer network (Dataswitch, DECnet, IBM - connections to GSI and IPP, preparation for Datex-P). (orig.)

  12. Computer Engineers.

    Science.gov (United States)

    Moncarz, Roger

    2000-01-01

    Looks at computer engineers and describes their job, employment outlook, earnings, and training and qualifications. Provides a list of resources related to computer engineering careers and the computer industry. (JOW)

  13. Computer Music

    Science.gov (United States)

    Cook, Perry R.

    This chapter covers algorithms, technologies, computer languages, and systems for computer music. Computer music involves the application of computers and other digital/electronic technologies to music composition, performance, theory, history, and the study of perception. The field combines digital signal processing, computational algorithms, computer languages, hardware and software systems, acoustics, psychoacoustics (low-level perception of sounds from the raw acoustic signal), and music cognition (higher-level perception of musical style, form, emotion, etc.).

  14. Analog computing

    CERN Document Server

    Ulmann, Bernd

    2013-01-01

    This book is a comprehensive introduction to analog computing. As most textbooks about this powerful computing paradigm date back to the 1960s and 1970s, it fills a void and forges a bridge from the early days of analog computing to future applications. The idea of analog computing is not new. In fact, this computing paradigm is nearly forgotten, although it offers a path to both high-speed and low-power computing, which are in even more demand now than they were back in the heyday of electronic analog computers.

  15. Computational composites

    DEFF Research Database (Denmark)

    Vallgårda, Anna K. A.; Redström, Johan

    2007-01-01

    Computational composite is introduced as a new type of composite material. Arguing that this is not just a metaphorical maneuver, we provide an analysis of computational technology as material in design, which shows how computers share important characteristics with other materials used in design...... and architecture. We argue that the notion of computational composites provides a precise understanding of the computer as material, and of how computations need to be combined with other materials to come to expression as material. Besides working as an analysis of computers from a designer’s point of view......, the notion of computational composites may also provide a link for computer science and human-computer interaction to an increasingly rapid development and use of new materials in design and architecture....

  16. Quantum Computing

    OpenAIRE

    Scarani, Valerio

    1998-01-01

    The aim of this thesis was to explain what quantum computing is. The information for the thesis was gathered from books, scientific publications, and news articles. The analysis of the information revealed that quantum computing can be broken down to three areas: theories behind quantum computing explaining the structure of a quantum computer, known quantum algorithms, and the actual physical realizations of a quantum computer. The thesis reveals that moving from classical memor...

  17. Reference Computational Meshing Strategy for Computational Fluid Dynamics Simulation of Departure from Nucleate BoilingReference Computational Meshing Strategy for Computational Fluid Dynamics Simulation of Departure from Nucleate Boiling

    Energy Technology Data Exchange (ETDEWEB)

    Pointer, William David [ORNL

    2017-08-01

    The objective of this effort is to establish a strategy and process for generation of suitable computational mesh for computational fluid dynamics simulations of departure from nucleate boiling in a 5 by 5 fuel rod assembly held in place by PWR mixing vane spacer grids. This mesh generation process will support ongoing efforts to develop, demonstrate and validate advanced multi-phase computational fluid dynamics methods that enable more robust identification of dryout conditions and DNB occurrence.Building upon prior efforts and experience, multiple computational meshes were developed using the native mesh generation capabilities of the commercial CFD code STAR-CCM+. These meshes were used to simulate two test cases from the Westinghouse 5 by 5 rod bundle facility. The sensitivity of predicted quantities of interest to the mesh resolution was then established using two evaluation methods, the Grid Convergence Index method and the Least Squares method. This evaluation suggests that the Least Squares method can reliably establish the uncertainty associated with local parameters such as vector velocity components at a point in the domain or surface averaged quantities such as outlet velocity magnitude. However, neither method is suitable for characterization of uncertainty in global extrema such as peak fuel surface temperature, primarily because such parameters are not necessarily associated with a fixed point in space. This shortcoming is significant because the current generation algorithm for identification of DNB event conditions relies on identification of such global extrema. Ongoing efforts to identify DNB based on local surface conditions will address this challenge

  18. Computational Medicine

    DEFF Research Database (Denmark)

    Nygaard, Jens Vinge

    2017-01-01

    The Health Technology Program at Aarhus University applies computational biology to investigate the heterogeneity of tumours......The Health Technology Program at Aarhus University applies computational biology to investigate the heterogeneity of tumours...

  19. Grid Computing

    Indian Academy of Sciences (India)

    A computing grid interconnects resources such as high performancecomputers, scientific databases, and computercontrolledscientific instruments of cooperating organizationseach of which is autonomous. It precedes and is quitedifferent from cloud computing, which provides computingresources by vendors to customers ...

  20. Green Computing

    Directory of Open Access Journals (Sweden)

    K. Shalini

    2013-01-01

    Full Text Available Green computing is all about using computers in a smarter and eco-friendly way. It is the environmentally responsible use of computers and related resources which includes the implementation of energy-efficient central processing units, servers and peripherals as well as reduced resource consumption and proper disposal of electronic waste .Computers certainly make up a large part of many people lives and traditionally are extremely damaging to the environment. Manufacturers of computer and its parts have been espousing the green cause to help protect environment from computers and electronic waste in any way.Research continues into key areas such as making the use of computers as energy-efficient as Possible, and designing algorithms and systems for efficiency-related computer technologies.

  1. Computer Aided Design and Analysis of Separation Processes with Electrolyte Systems

    DEFF Research Database (Denmark)

    A methodology for computer aided design and analysis of separation processes involving electrolyte systems is presented. The methodology consists of three main parts. The thermodynamic part "creates" the problem specific property model package, which is a collection of pure component and mixture...... property models. The design and analysis part generates process (flowsheet) alternatives, evaluates/analyses feasibility of separation and provides a visual operation path for the desired separation. The simulation part consists of a simulation/calculation engine that allows the screening and validation...... of process alternatives. For the simulation part, a general multi-purpose, multi-phase separation model has been developed and integrated to an existing computer aided system. Application of the design and analysis methodology is highlighted through two illustrative case studies....

  2. Computer Aided Design and Analysis of Separation Processes with Electrolyte Systems

    DEFF Research Database (Denmark)

    Takano, Kiyoteru; Gani, Rafiqul; Kolar, P.

    2000-01-01

    A methodology for computer aided design and analysis of separation processes involving electrolyte systems is presented. The methodology consists of three main parts. The thermodynamic part 'creates' the problem specific property model package, which is a collection of pure component and mixture...... property models. The design and analysis part generates process (flowsheet) alternatives, evaluates/analyses feasibility of separation and provides a visual operation path for the desired separation. The simulation part consists of a simulation/calculation engine that allows the screening and validation...... of process alternatives. For the simulation part, a general multi-purpose, multi-phase separation model has been developed and integrated to an existing computer aided system. Application of the design and analysis methodology is highlighted through two illustrative case studies, (C) 2000 Elsevier Science...

  3. Quantum computers and quantum computations

    International Nuclear Information System (INIS)

    Valiev, Kamil' A

    2005-01-01

    This review outlines the principles of operation of quantum computers and their elements. The theory of ideal computers that do not interact with the environment and are immune to quantum decohering processes is presented. Decohering processes in quantum computers are investigated. The review considers methods for correcting quantum computing errors arising from the decoherence of the state of the quantum computer, as well as possible methods for the suppression of the decohering processes. A brief enumeration of proposed quantum computer realizations concludes the review. (reviews of topical problems)

  4. Quantum Computing for Computer Architects

    CERN Document Server

    Metodi, Tzvetan

    2011-01-01

    Quantum computers can (in theory) solve certain problems far faster than a classical computer running any known classical algorithm. While existing technologies for building quantum computers are in their infancy, it is not too early to consider their scalability and reliability in the context of the design of large-scale quantum computers. To architect such systems, one must understand what it takes to design and model a balanced, fault-tolerant quantum computer architecture. The goal of this lecture is to provide architectural abstractions for the design of a quantum computer and to explore

  5. Pervasive Computing

    NARCIS (Netherlands)

    Silvis-Cividjian, N.

    This book provides a concise introduction to Pervasive Computing, otherwise known as Internet of Things (IoT) and Ubiquitous Computing (Ubicomp) which addresses the seamless integration of computing systems within everyday objects. By introducing the core topics and exploring assistive pervasive

  6. Computational vision

    CERN Document Server

    Wechsler, Harry

    1990-01-01

    The book is suitable for advanced courses in computer vision and image processing. In addition to providing an overall view of computational vision, it contains extensive material on topics that are not usually covered in computer vision texts (including parallel distributed processing and neural networks) and considers many real applications.

  7. Spatial Computation

    Science.gov (United States)

    2003-12-01

    Computation and today’s microprocessors with the approach to operating system architecture, and the controversy between microkernels and monolithic kernels...Both Spatial Computation and microkernels break away a relatively monolithic architecture into in- dividual lightweight pieces, well specialized...for their particular functionality. Spatial Computation removes global signals and control, in the same way microkernels remove the global address

  8. Parallel computations

    CERN Document Server

    1982-01-01

    Parallel Computations focuses on parallel computation, with emphasis on algorithms used in a variety of numerical and physical applications and for many different types of parallel computers. Topics covered range from vectorization of fast Fourier transforms (FFTs) and of the incomplete Cholesky conjugate gradient (ICCG) algorithm on the Cray-1 to calculation of table lookups and piecewise functions. Single tridiagonal linear systems and vectorized computation of reactive flow are also discussed.Comprised of 13 chapters, this volume begins by classifying parallel computers and describing techn

  9. Human Computation

    CERN Multimedia

    CERN. Geneva

    2008-01-01

    What if people could play computer games and accomplish work without even realizing it? What if billions of people collaborated to solve important problems for humanity or generate training data for computers? My work aims at a general paradigm for doing exactly that: utilizing human processing power to solve computational problems in a distributed manner. In particular, I focus on harnessing human time and energy for addressing problems that computers cannot yet solve. Although computers have advanced dramatically in many respects over the last 50 years, they still do not possess the basic conceptual intelligence or perceptual capabilities...

  10. Quantum computation

    International Nuclear Information System (INIS)

    Deutsch, D.

    1992-01-01

    As computers become ever more complex, they inevitably become smaller. This leads to a need for components which are fabricated and operate on increasingly smaller size scales. Quantum theory is already taken into account in microelectronics design. This article explores how quantum theory will need to be incorporated into computers in future in order to give them their components functionality. Computation tasks which depend on quantum effects will become possible. Physicists may have to reconsider their perspective on computation in the light of understanding developed in connection with universal quantum computers. (UK)

  11. Computer software.

    Science.gov (United States)

    Rosenthal, L E

    1986-10-01

    Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips.

  12. Computer sciences

    Science.gov (United States)

    Smith, Paul H.

    1988-01-01

    The Computer Science Program provides advanced concepts, techniques, system architectures, algorithms, and software for both space and aeronautics information sciences and computer systems. The overall goal is to provide the technical foundation within NASA for the advancement of computing technology in aerospace applications. The research program is improving the state of knowledge of fundamental aerospace computing principles and advancing computing technology in space applications such as software engineering and information extraction from data collected by scientific instruments in space. The program includes the development of special algorithms and techniques to exploit the computing power provided by high performance parallel processors and special purpose architectures. Research is being conducted in the fundamentals of data base logic and improvement techniques for producing reliable computing systems.

  13. Computer programming and computer systems

    CERN Document Server

    Hassitt, Anthony

    1966-01-01

    Computer Programming and Computer Systems imparts a "reading knowledge? of computer systems.This book describes the aspects of machine-language programming, monitor systems, computer hardware, and advanced programming that every thorough programmer should be acquainted with. This text discusses the automatic electronic digital computers, symbolic language, Reverse Polish Notation, and Fortran into assembly language. The routine for reading blocked tapes, dimension statements in subroutines, general-purpose input routine, and efficient use of memory are also elaborated.This publication is inten

  14. Organic Computing

    CERN Document Server

    Würtz, Rolf P

    2008-01-01

    Organic Computing is a research field emerging around the conviction that problems of organization in complex systems in computer science, telecommunications, neurobiology, molecular biology, ethology, and possibly even sociology can be tackled scientifically in a unified way. From the computer science point of view, the apparent ease in which living systems solve computationally difficult problems makes it inevitable to adopt strategies observed in nature for creating information processing machinery. In this book, the major ideas behind Organic Computing are delineated, together with a sparse sample of computational projects undertaken in this new field. Biological metaphors include evolution, neural networks, gene-regulatory networks, networks of brain modules, hormone system, insect swarms, and ant colonies. Applications are as diverse as system design, optimization, artificial growth, task allocation, clustering, routing, face recognition, and sign language understanding.

  15. Computational biomechanics

    International Nuclear Information System (INIS)

    Ethier, C.R.

    2004-01-01

    Computational biomechanics is a fast-growing field that integrates modern biological techniques and computer modelling to solve problems of medical and biological interest. Modelling of blood flow in the large arteries is the best-known application of computational biomechanics, but there are many others. Described here is work being carried out in the laboratory on the modelling of blood flow in the coronary arteries and on the transport of viral particles in the eye. (author)

  16. Computational Composites

    DEFF Research Database (Denmark)

    Vallgårda, Anna K. A.

    to understand the computer as a material like any other material we would use for design, like wood, aluminum, or plastic. That as soon as the computer forms a composition with other materials it becomes just as approachable and inspiring as other smart materials. I present a series of investigations of what...... Computational Composite, and Telltale). Through the investigations, I show how the computer can be understood as a material and how it partakes in a new strand of materials whose expressions come to be in context. I uncover some of their essential material properties and potential expressions. I develop a way...

  17. GPGPU COMPUTING

    Directory of Open Access Journals (Sweden)

    BOGDAN OANCEA

    2012-05-01

    Full Text Available Since the first idea of using GPU to general purpose computing, things have evolved over the years and now there are several approaches to GPU programming. GPU computing practically began with the introduction of CUDA (Compute Unified Device Architecture by NVIDIA and Stream by AMD. These are APIs designed by the GPU vendors to be used together with the hardware that they provide. A new emerging standard, OpenCL (Open Computing Language tries to unify different GPU general computing API implementations and provides a framework for writing programs executed across heterogeneous platforms consisting of both CPUs and GPUs. OpenCL provides parallel computing using task-based and data-based parallelism. In this paper we will focus on the CUDA parallel computing architecture and programming model introduced by NVIDIA. We will present the benefits of the CUDA programming model. We will also compare the two main approaches, CUDA and AMD APP (STREAM and the new framwork, OpenCL that tries to unify the GPGPU computing models.

  18. Quantum Computing

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 5; Issue 9. Quantum Computing - Building Blocks of a Quantum Computer. C S Vijay Vishal Gupta. General Article Volume 5 Issue 9 September 2000 pp 69-81. Fulltext. Click here to view fulltext PDF. Permanent link:

  19. Platform computing

    CERN Multimedia

    2002-01-01

    "Platform Computing releases first grid-enabled workload management solution for IBM eServer Intel and UNIX high performance computing clusters. This Out-of-the-box solution maximizes the performance and capability of applications on IBM HPC clusters" (1/2 page) .

  20. Quantum Computing

    Indian Academy of Sciences (India)

    In the first part of this article, we had looked at how quantum physics can be harnessed to make the building blocks of a quantum computer. In this concluding part, we look at algorithms which can exploit the power of this computational device, and some practical difficulties in building such a device. Quantum Algorithms.

  1. Quantum computing

    OpenAIRE

    Burba, M.; Lapitskaya, T.

    2017-01-01

    This article gives an elementary introduction to quantum computing. It is a draft for a book chapter of the "Handbook of Nature-Inspired and Innovative Computing", Eds. A. Zomaya, G.J. Milburn, J. Dongarra, D. Bader, R. Brent, M. Eshaghian-Wilner, F. Seredynski (Springer, Berlin Heidelberg New York, 2006).

  2. Computational Pathology

    Science.gov (United States)

    Louis, David N.; Feldman, Michael; Carter, Alexis B.; Dighe, Anand S.; Pfeifer, John D.; Bry, Lynn; Almeida, Jonas S.; Saltz, Joel; Braun, Jonathan; Tomaszewski, John E.; Gilbertson, John R.; Sinard, John H.; Gerber, Georg K.; Galli, Stephen J.; Golden, Jeffrey A.; Becich, Michael J.

    2016-01-01

    Context We define the scope and needs within the new discipline of computational pathology, a discipline critical to the future of both the practice of pathology and, more broadly, medical practice in general. Objective To define the scope and needs of computational pathology. Data Sources A meeting was convened in Boston, Massachusetts, in July 2014 prior to the annual Association of Pathology Chairs meeting, and it was attended by a variety of pathologists, including individuals highly invested in pathology informatics as well as chairs of pathology departments. Conclusions The meeting made recommendations to promote computational pathology, including clearly defining the field and articulating its value propositions; asserting that the value propositions for health care systems must include means to incorporate robust computational approaches to implement data-driven methods that aid in guiding individual and population health care; leveraging computational pathology as a center for data interpretation in modern health care systems; stating that realizing the value proposition will require working with institutional administrations, other departments, and pathology colleagues; declaring that a robust pipeline should be fostered that trains and develops future computational pathologists, for those with both pathology and non-pathology backgrounds; and deciding that computational pathology should serve as a hub for data-related research in health care systems. The dissemination of these recommendations to pathology and bioinformatics departments should help facilitate the development of computational pathology. PMID:26098131

  3. Cloud Computing

    DEFF Research Database (Denmark)

    Krogh, Simon

    2013-01-01

    with technological changes, the paradigmatic pendulum has swung between increased centralization on one side and a focus on distributed computing that pushes IT power out to end users on the other. With the introduction of outsourcing and cloud computing, centralization in large data centers is again dominating...... the IT scene. In line with the views presented by Nicolas Carr in 2003 (Carr, 2003), it is a popular assumption that cloud computing will be the next utility (like water, electricity and gas) (Buyya, Yeo, Venugopal, Broberg, & Brandic, 2009). However, this assumption disregards the fact that most IT production......), for instance, in establishing and maintaining trust between the involved parties (Sabherwal, 1999). So far, research in cloud computing has neglected this perspective and focused entirely on aspects relating to technology, economy, security and legal questions. While the core technologies of cloud computing (e...

  4. Computability theory

    CERN Document Server

    Weber, Rebecca

    2012-01-01

    What can we compute--even with unlimited resources? Is everything within reach? Or are computations necessarily drastically limited, not just in practice, but theoretically? These questions are at the heart of computability theory. The goal of this book is to give the reader a firm grounding in the fundamentals of computability theory and an overview of currently active areas of research, such as reverse mathematics and algorithmic randomness. Turing machines and partial recursive functions are explored in detail, and vital tools and concepts including coding, uniformity, and diagonalization are described explicitly. From there the material continues with universal machines, the halting problem, parametrization and the recursion theorem, and thence to computability for sets, enumerability, and Turing reduction and degrees. A few more advanced topics round out the book before the chapter on areas of research. The text is designed to be self-contained, with an entire chapter of preliminary material including re...

  5. Computational Streetscapes

    Directory of Open Access Journals (Sweden)

    Paul M. Torrens

    2016-09-01

    Full Text Available Streetscapes have presented a long-standing interest in many fields. Recently, there has been a resurgence of attention on streetscape issues, catalyzed in large part by computing. Because of computing, there is more understanding, vistas, data, and analysis of and on streetscape phenomena than ever before. This diversity of lenses trained on streetscapes permits us to address long-standing questions, such as how people use information while mobile, how interactions with people and things occur on streets, how we might safeguard crowds, how we can design services to assist pedestrians, and how we could better support special populations as they traverse cities. Amid each of these avenues of inquiry, computing is facilitating new ways of posing these questions, particularly by expanding the scope of what-if exploration that is possible. With assistance from computing, consideration of streetscapes now reaches across scales, from the neurological interactions that form among place cells in the brain up to informatics that afford real-time views of activity over whole urban spaces. For some streetscape phenomena, computing allows us to build realistic but synthetic facsimiles in computation, which can function as artificial laboratories for testing ideas. In this paper, I review the domain science for studying streetscapes from vantages in physics, urban studies, animation and the visual arts, psychology, biology, and behavioral geography. I also review the computational developments shaping streetscape science, with particular emphasis on modeling and simulation as informed by data acquisition and generation, data models, path-planning heuristics, artificial intelligence for navigation and way-finding, timing, synthetic vision, steering routines, kinematics, and geometrical treatment of collision detection and avoidance. I also discuss the implications that the advances in computing streetscapes might have on emerging developments in cyber

  6. COMPUTATIONAL THINKING

    Directory of Open Access Journals (Sweden)

    Evgeniy K. Khenner

    2016-01-01

    Full Text Available Abstract. The aim of the research is to draw attention of the educational community to the phenomenon of computational thinking which actively discussed in the last decade in the foreign scientific and educational literature, to substantiate of its importance, practical utility and the right on affirmation in Russian education.Methods. The research is based on the analysis of foreign studies of the phenomenon of computational thinking and the ways of its formation in the process of education; on comparing the notion of «computational thinking» with related concepts used in the Russian scientific and pedagogical literature.Results. The concept «computational thinking» is analyzed from the point of view of intuitive understanding and scientific and applied aspects. It is shown as computational thinking has evolved in the process of development of computers hardware and software. The practice-oriented interpretation of computational thinking which dominant among educators is described along with some ways of its formation. It is shown that computational thinking is a metasubject result of general education as well as its tool. From the point of view of the author, purposeful development of computational thinking should be one of the tasks of the Russian education.Scientific novelty. The author gives a theoretical justification of the role of computational thinking schemes as metasubject results of learning. The dynamics of the development of this concept is described. This process is connected with the evolution of computer and information technologies as well as increase of number of the tasks for effective solutions of which computational thinking is required. Author substantiated the affirmation that including «computational thinking » in the set of pedagogical concepts which are used in the national education system fills an existing gap.Practical significance. New metasubject result of education associated with

  7. Computer interfacing

    CERN Document Server

    Dixey, Graham

    1994-01-01

    This book explains how computers interact with the world around them and therefore how to make them a useful tool. Topics covered include descriptions of all the components that make up a computer, principles of data exchange, interaction with peripherals, serial communication, input devices, recording methods, computer-controlled motors, and printers.In an informative and straightforward manner, Graham Dixey describes how to turn what might seem an incomprehensible 'black box' PC into a powerful and enjoyable tool that can help you in all areas of your work and leisure. With plenty of handy

  8. Computational physics

    CERN Document Server

    Newman, Mark

    2013-01-01

    A complete introduction to the field of computational physics, with examples and exercises in the Python programming language. Computers play a central role in virtually every major physics discovery today, from astrophysics and particle physics to biophysics and condensed matter. This book explains the fundamentals of computational physics and describes in simple terms the techniques that every physicist should know, such as finite difference methods, numerical quadrature, and the fast Fourier transform. The book offers a complete introduction to the topic at the undergraduate level, and is also suitable for the advanced student or researcher who wants to learn the foundational elements of this important field.

  9. Computational physics

    Energy Technology Data Exchange (ETDEWEB)

    Anon.

    1987-01-15

    Computers have for many years played a vital role in the acquisition and treatment of experimental data, but they have more recently taken up a much more extended role in physics research. The numerical and algebraic calculations now performed on modern computers make it possible to explore consequences of basic theories in a way which goes beyond the limits of both analytic insight and experimental investigation. This was brought out clearly at the Conference on Perspectives in Computational Physics, held at the International Centre for Theoretical Physics, Trieste, Italy, from 29-31 October.

  10. Cloud Computing

    CERN Document Server

    Baun, Christian; Nimis, Jens; Tai, Stefan

    2011-01-01

    Cloud computing is a buzz-word in today's information technology (IT) that nobody can escape. But what is really behind it? There are many interpretations of this term, but no standardized or even uniform definition. Instead, as a result of the multi-faceted viewpoints and the diverse interests expressed by the various stakeholders, cloud computing is perceived as a rather fuzzy concept. With this book, the authors deliver an overview of cloud computing architecture, services, and applications. Their aim is to bring readers up to date on this technology and thus to provide a common basis for d

  11. Computational Viscoelasticity

    CERN Document Server

    Marques, Severino P C

    2012-01-01

    This text is a guide how to solve problems in which viscoelasticity is present using existing commercial computational codes. The book gives information on codes’ structure and use, data preparation  and output interpretation and verification. The first part of the book introduces the reader to the subject, and to provide the models, equations and notation to be used in the computational applications. The second part shows the most important Computational techniques: Finite elements formulation, Boundary elements formulation, and presents the solutions of Viscoelastic problems with Abaqus.

  12. Optical computing.

    Science.gov (United States)

    Stroke, G. W.

    1972-01-01

    Applications of the optical computer include an approach for increasing the sharpness of images obtained from the most powerful electron microscopes and fingerprint/credit card identification. The information-handling capability of the various optical computing processes is very great. Modern synthetic-aperture radars scan upward of 100,000 resolvable elements per second. Fields which have assumed major importance on the basis of optical computing principles are optical image deblurring, coherent side-looking synthetic-aperture radar, and correlative pattern recognition. Some examples of the most dramatic image deblurring results are shown.

  13. Computational physics

    International Nuclear Information System (INIS)

    Anon.

    1987-01-01

    Computers have for many years played a vital role in the acquisition and treatment of experimental data, but they have more recently taken up a much more extended role in physics research. The numerical and algebraic calculations now performed on modern computers make it possible to explore consequences of basic theories in a way which goes beyond the limits of both analytic insight and experimental investigation. This was brought out clearly at the Conference on Perspectives in Computational Physics, held at the International Centre for Theoretical Physics, Trieste, Italy, from 29-31 October

  14. Phenomenological Computation?

    DEFF Research Database (Denmark)

    Brier, Søren

    2014-01-01

    Open peer commentary on the article “Info-computational Constructivism and Cognition” by Gordana Dodig-Crnkovic. Upshot: The main problems with info-computationalism are: (1) Its basic concept of natural computing has neither been defined theoretically or implemented practically. (2. It cannot...... encompass human concepts of subjective experience and intersubjective meaningful communication, which prevents it from being genuinely transdisciplinary. (3) Philosophically, it does not sufficiently accept the deep ontological differences between various paradigms such as von Foerster’s second- order...

  15. Computational Fluid Dynamics for Nuclear Reactor Safety-5 (CFD4NRS-5). Workshop Proceedings, 9-11 September 2014, Zurich, Switzerland

    International Nuclear Information System (INIS)

    Smith, Brian L.; Andreani, Michele; Badillo, Arnoldo; Dehbi, Abdel; Sato, Yohei; Smith, Brian L.; Dreier, Joerg; Kapulla, Ralf; Niceno, Bojan; Sharabi, Medhat; Bestion, Dominique; Bieder, Ulrich; Coste, Pierre; Martinez, Jean Marc; Zigh, Ghani; Boyd, Chris; Prasser, Horst-Michael; Kerenyi, Nora; Adams, Robert; Bolesch, Christian; D'Aleo, Paolo; Eismann, Ralph; Kickhofel, John; Lafferty, Nathan; Saxena, Abhishek; Kissane, Martin; ); Ulses, Anthony; ); Bartosiewicz, Yann; Seynhaeve, Jean-Marie; Caraghiaur, Diana; Munoz Cobo, Jose Luis; Glaeser, Horst; Buchholz, Sebastian; Scheuerer, Martina; Hassan, Yassin; In, Wang-Kee; Song, Chul-Hwa; Yoon, Han-Young; Kim, J.W.; Koncar, Bostjan; Tiselj, Iztoc; Lakehal, Djamel; Yadigaroglu, George; Lo, Simon; Manera, Annalisa; Petrov, Victor; Mimouni, Stephane; Benhamadouche, Sofiane; Morii, Tadashi; Suikkanen, Heikki; Toppila, Timo; Angele, Kristian; Baglietto, Emilio; Cheng, Xu; Graffard, Estelle; Ko, Jordan; Hoehne, Thomas; Lucas, Dirk; Krepper, Eckhard; Laurien, Eckart; Moretti, Fabio; Piro, Markus; Roelofs, Ferry; Veber, Pascal; Watanabe, Tadashi; Yan, Jin; Yeoh, Guan

    2016-01-01

    This present workshop, the 5. Computational Fluid Dynamics for Nuclear-Reactor Safety (CFD4NRS-5), in the biennial series of such Nuclear Energy Agency (NEA) and International Atomic Energy Agency (IAEA) sponsored events, a tradition which began in Garching in 2006, follows the format and objectives of its predecessors in creating a forum whereby numerical analysts and experimentalists can exchange information in the application of computational fluid dynamics (CFD) to nuclear power plant (NPP) safety and future design issues. The emphasis, as always, was, in a congenial atmosphere, to offer exposure to state-of-the-art (single-phase and multi-phase) CFD applications reflecting topical issues arising in NPP design and safety, but in particular to promote the release of high-resolution experimental data to continue the CFD validation process in this application area. The reason for the increased use of multi-dimensional CFD methods is that a number of important thermal-hydraulic phenomena occurring in NPPs cannot be adequately predicted using traditional one-dimensional system hydraulics codes with the required accuracy and spatial resolution when strong three-dimensional motions prevail. Established CFD codes already contain empirical models for simulating turbulence, heat transfer, multi-phase interaction and chemical reactions. Nonetheless, such models must be validated against test data before they can be used with confidence. The necessary validation procedure is performed by comparing model predictions against trustworthy experimental data. However, reliable model assessment requires CFD simulations to be undertaken with full control over numerical errors and input uncertainties. The writing groups originally set up by the NEA have been consistently promoting the use of best practice guidelines (BPGs) in the application of CFD for just this purpose, and BPGs remain a central pillar of the simulation material accepted at this current workshop, as it was at its

  16. Essentials of cloud computing

    CERN Document Server

    Chandrasekaran, K

    2014-01-01

    ForewordPrefaceComputing ParadigmsLearning ObjectivesPreambleHigh-Performance ComputingParallel ComputingDistributed ComputingCluster ComputingGrid ComputingCloud ComputingBiocomputingMobile ComputingQuantum ComputingOptical ComputingNanocomputingNetwork ComputingSummaryReview PointsReview QuestionsFurther ReadingCloud Computing FundamentalsLearning ObjectivesPreambleMotivation for Cloud ComputingThe Need for Cloud ComputingDefining Cloud ComputingNIST Definition of Cloud ComputingCloud Computing Is a ServiceCloud Computing Is a Platform5-4-3 Principles of Cloud computingFive Essential Charact

  17. Personal Computers.

    Science.gov (United States)

    Toong, Hoo-min D.; Gupta, Amar

    1982-01-01

    Describes the hardware, software, applications, and current proliferation of personal computers (microcomputers). Includes discussions of microprocessors, memory, output (including printers), application programs, the microcomputer industry, and major microcomputer manufacturers (Apple, Radio Shack, Commodore, and IBM). (JN)

  18. Computational Literacy

    DEFF Research Database (Denmark)

    Chongtay, Rocio; Robering, Klaus

    2016-01-01

    In recent years, there has been a growing interest in and recognition of the importance of Computational Literacy, a skill generally considered to be necessary for success in the 21st century. While much research has concentrated on requirements, tools, and teaching methodologies for the acquisit......In recent years, there has been a growing interest in and recognition of the importance of Computational Literacy, a skill generally considered to be necessary for success in the 21st century. While much research has concentrated on requirements, tools, and teaching methodologies...... for the acquisition of Computational Literacy at basic educational levels, focus on higher levels of education has been much less prominent. The present paper considers the case of courses for higher education programs within the Humanities. A model is proposed which conceives of Computational Literacy as a layered...

  19. Computing Religion

    DEFF Research Database (Denmark)

    Nielbo, Kristoffer Laigaard; Braxton, Donald M.; Upal, Afzal

    2012-01-01

    The computational approach has become an invaluable tool in many fields that are directly relevant to research in religious phenomena. Yet the use of computational tools is almost absent in the study of religion. Given that religion is a cluster of interrelated phenomena and that research...... concerning these phenomena should strive for multilevel analysis, this article argues that the computational approach offers new methodological and theoretical opportunities to the study of religion. We argue that the computational approach offers 1.) an intermediary step between any theoretical construct...... and its targeted empirical space and 2.) a new kind of data which allows the researcher to observe abstract constructs, estimate likely outcomes, and optimize empirical designs. Because sophisticated mulitilevel research is a collaborative project we also seek to introduce to scholars of religion some...

  20. Computational Controversy

    NARCIS (Netherlands)

    Timmermans, Benjamin; Kuhn, Tobias; Beelen, Kaspar; Aroyo, Lora

    2017-01-01

    Climate change, vaccination, abortion, Trump: Many topics are surrounded by fierce controversies. The nature of such heated debates and their elements have been studied extensively in the social science literature. More recently, various computational approaches to controversy analysis have

  1. Grid Computing

    Indian Academy of Sciences (India)

    IAS Admin

    emergence of supercomputers led to the use of computer simula- tion as an .... Scientific and engineering applications (e.g., Tera grid secure gate way). Collaborative ... Encryption, privacy, protection from malicious software. Physical Layer.

  2. Computer tomographs

    International Nuclear Information System (INIS)

    Niedzwiedzki, M.

    1982-01-01

    Physical foundations and the developments in the transmission and emission computer tomography are presented. On the basis of the available literature and private communications a comparison is made of the various transmission tomographs. A new technique of computer emission tomography ECT, unknown in Poland, is described. The evaluation of two methods of ECT, namely those of positron and single photon emission tomography is made. (author)

  3. Computational sustainability

    CERN Document Server

    Kersting, Kristian; Morik, Katharina

    2016-01-01

    The book at hand gives an overview of the state of the art research in Computational Sustainability as well as case studies of different application scenarios. This covers topics such as renewable energy supply, energy storage and e-mobility, efficiency in data centers and networks, sustainable food and water supply, sustainable health, industrial production and quality, etc. The book describes computational methods and possible application scenarios.

  4. Computing farms

    International Nuclear Information System (INIS)

    Yeh, G.P.

    2000-01-01

    High-energy physics, nuclear physics, space sciences, and many other fields have large challenges in computing. In recent years, PCs have achieved performance comparable to the high-end UNIX workstations, at a small fraction of the price. We review the development and broad applications of commodity PCs as the solution to CPU needs, and look forward to the important and exciting future of large-scale PC computing

  5. Computational chemistry

    Science.gov (United States)

    Arnold, J. O.

    1987-01-01

    With the advent of supercomputers, modern computational chemistry algorithms and codes, a powerful tool was created to help fill NASA's continuing need for information on the properties of matter in hostile or unusual environments. Computational resources provided under the National Aerodynamics Simulator (NAS) program were a cornerstone for recent advancements in this field. Properties of gases, materials, and their interactions can be determined from solutions of the governing equations. In the case of gases, for example, radiative transition probabilites per particle, bond-dissociation energies, and rates of simple chemical reactions can be determined computationally as reliably as from experiment. The data are proving to be quite valuable in providing inputs to real-gas flow simulation codes used to compute aerothermodynamic loads on NASA's aeroassist orbital transfer vehicles and a host of problems related to the National Aerospace Plane Program. Although more approximate, similar solutions can be obtained for ensembles of atoms simulating small particles of materials with and without the presence of gases. Computational chemistry has application in studying catalysis, properties of polymers, all of interest to various NASA missions, including those previously mentioned. In addition to discussing these applications of computational chemistry within NASA, the governing equations and the need for supercomputers for their solution is outlined.

  6. Computational creativity

    Directory of Open Access Journals (Sweden)

    López de Mántaras Badia, Ramon

    2013-12-01

    Full Text Available New technologies, and in particular artificial intelligence, are drastically changing the nature of creative processes. Computers are playing very significant roles in creative activities such as music, architecture, fine arts, and science. Indeed, the computer is already a canvas, a brush, a musical instrument, and so on. However, we believe that we must aim at more ambitious relations between computers and creativity. Rather than just seeing the computer as a tool to help human creators, we could see it as a creative entity in its own right. This view has triggered a new subfield of Artificial Intelligence called Computational Creativity. This article addresses the question of the possibility of achieving computational creativity through some examples of computer programs capable of replicating some aspects of creative behavior in the fields of music and science.Las nuevas tecnologías y en particular la Inteligencia Artificial están cambiando de forma importante la naturaleza del proceso creativo. Los ordenadores están jugando un papel muy significativo en actividades artísticas tales como la música, la arquitectura, las bellas artes y la ciencia. Efectivamente, el ordenador ya es el lienzo, el pincel, el instrumento musical, etc. Sin embargo creemos que debemos aspirar a relaciones más ambiciosas entre los ordenadores y la creatividad. En lugar de verlos solamente como herramientas de ayuda a la creación, los ordenadores podrían ser considerados agentes creativos. Este punto de vista ha dado lugar a un nuevo subcampo de la Inteligencia Artificial denominado Creatividad Computacional. En este artículo abordamos la cuestión de la posibilidad de alcanzar dicha creatividad computacional mediante algunos ejemplos de programas de ordenador capaces de replicar algunos aspectos relacionados con el comportamiento creativo en los ámbitos de la música y la ciencia.

  7. Quantum computing

    International Nuclear Information System (INIS)

    Steane, Andrew

    1998-01-01

    The subject of quantum computing brings together ideas from classical information theory, computer science, and quantum physics. This review aims to summarize not just quantum computing, but the whole subject of quantum information theory. Information can be identified as the most general thing which must propagate from a cause to an effect. It therefore has a fundamentally important role in the science of physics. However, the mathematical treatment of information, especially information processing, is quite recent, dating from the mid-20th century. This has meant that the full significance of information as a basic concept in physics is only now being discovered. This is especially true in quantum mechanics. The theory of quantum information and computing puts this significance on a firm footing, and has led to some profound and exciting new insights into the natural world. Among these are the use of quantum states to permit the secure transmission of classical information (quantum cryptography), the use of quantum entanglement to permit reliable transmission of quantum states (teleportation), the possibility of preserving quantum coherence in the presence of irreversible noise processes (quantum error correction), and the use of controlled quantum evolution for efficient computation (quantum computation). The common theme of all these insights is the use of quantum entanglement as a computational resource. It turns out that information theory and quantum mechanics fit together very well. In order to explain their relationship, this review begins with an introduction to classical information theory and computer science, including Shannon's theorem, error correcting codes, Turing machines and computational complexity. The principles of quantum mechanics are then outlined, and the Einstein, Podolsky and Rosen (EPR) experiment described. The EPR-Bell correlations, and quantum entanglement in general, form the essential new ingredient which distinguishes quantum from

  8. Quantum computing

    Energy Technology Data Exchange (ETDEWEB)

    Steane, Andrew [Department of Atomic and Laser Physics, University of Oxford, Clarendon Laboratory, Oxford (United Kingdom)

    1998-02-01

    The subject of quantum computing brings together ideas from classical information theory, computer science, and quantum physics. This review aims to summarize not just quantum computing, but the whole subject of quantum information theory. Information can be identified as the most general thing which must propagate from a cause to an effect. It therefore has a fundamentally important role in the science of physics. However, the mathematical treatment of information, especially information processing, is quite recent, dating from the mid-20th century. This has meant that the full significance of information as a basic concept in physics is only now being discovered. This is especially true in quantum mechanics. The theory of quantum information and computing puts this significance on a firm footing, and has led to some profound and exciting new insights into the natural world. Among these are the use of quantum states to permit the secure transmission of classical information (quantum cryptography), the use of quantum entanglement to permit reliable transmission of quantum states (teleportation), the possibility of preserving quantum coherence in the presence of irreversible noise processes (quantum error correction), and the use of controlled quantum evolution for efficient computation (quantum computation). The common theme of all these insights is the use of quantum entanglement as a computational resource. It turns out that information theory and quantum mechanics fit together very well. In order to explain their relationship, this review begins with an introduction to classical information theory and computer science, including Shannon's theorem, error correcting codes, Turing machines and computational complexity. The principles of quantum mechanics are then outlined, and the Einstein, Podolsky and Rosen (EPR) experiment described. The EPR-Bell correlations, and quantum entanglement in general, form the essential new ingredient which distinguishes quantum from

  9. Multiparty Computations

    DEFF Research Database (Denmark)

    Dziembowski, Stefan

    here and discuss other problems caused by the adaptiveness. All protocols in the thesis are formally specified and the proofs of their security are given. [1]Ronald Cramer, Ivan Damgård, Stefan Dziembowski, Martin Hirt, and Tal Rabin. Efficient multiparty computations with dishonest minority......In this thesis we study a problem of doing Verifiable Secret Sharing (VSS) and Multiparty Computations in a model where private channels between the players and a broadcast channel is available. The adversary is active, adaptive and has an unbounded computing power. The thesis is based on two...... to a polynomial time black-box reduction, the complexity of adaptively secure VSS is the same as that of ordinary secret sharing (SS), where security is only required against a passive, static adversary. Previously, such a connection was only known for linear secret sharing and VSS schemes. We then show...

  10. Scientific computing

    CERN Document Server

    Trangenstein, John A

    2017-01-01

    This is the third of three volumes providing a comprehensive presentation of the fundamentals of scientific computing. This volume discusses topics that depend more on calculus than linear algebra, in order to prepare the reader for solving differential equations. This book and its companions show how to determine the quality of computational results, and how to measure the relative efficiency of competing methods. Readers learn how to determine the maximum attainable accuracy of algorithms, and how to select the best method for computing problems. This book also discusses programming in several languages, including C++, Fortran and MATLAB. There are 90 examples, 200 exercises, 36 algorithms, 40 interactive JavaScript programs, 91 references to software programs and 1 case study. Topics are introduced with goals, literature references and links to public software. There are descriptions of the current algorithms in GSLIB and MATLAB. This book could be used for a second course in numerical methods, for either ...

  11. Computational Psychiatry

    Science.gov (United States)

    Wang, Xiao-Jing; Krystal, John H.

    2014-01-01

    Psychiatric disorders such as autism and schizophrenia arise from abnormalities in brain systems that underlie cognitive, emotional and social functions. The brain is enormously complex and its abundant feedback loops on multiple scales preclude intuitive explication of circuit functions. In close interplay with experiments, theory and computational modeling are essential for understanding how, precisely, neural circuits generate flexible behaviors and their impairments give rise to psychiatric symptoms. This Perspective highlights recent progress in applying computational neuroscience to the study of mental disorders. We outline basic approaches, including identification of core deficits that cut across disease categories, biologically-realistic modeling bridging cellular and synaptic mechanisms with behavior, model-aided diagnosis. The need for new research strategies in psychiatry is urgent. Computational psychiatry potentially provides powerful tools for elucidating pathophysiology that may inform both diagnosis and treatment. To achieve this promise will require investment in cross-disciplinary training and research in this nascent field. PMID:25442941

  12. Computational artifacts

    DEFF Research Database (Denmark)

    Schmidt, Kjeld; Bansler, Jørgen P.

    2016-01-01

    The key concern of CSCW research is that of understanding computing technologies in the social context of their use, that is, as integral features of our practices and our lives, and to think of their design and implementation under that perspective. However, the question of the nature...... of that which is actually integrated in our practices is often discussed in confusing ways, if at all. The article aims to try to clarify the issue and in doing so revisits and reconsiders the notion of ‘computational artifact’....

  13. Computer security

    CERN Document Server

    Gollmann, Dieter

    2011-01-01

    A completely up-to-date resource on computer security Assuming no previous experience in the field of computer security, this must-have book walks you through the many essential aspects of this vast topic, from the newest advances in software and technology to the most recent information on Web applications security. This new edition includes sections on Windows NT, CORBA, and Java and discusses cross-site scripting and JavaScript hacking as well as SQL injection. Serving as a helpful introduction, this self-study guide is a wonderful starting point for examining the variety of competing sec

  14. Cloud Computing

    CERN Document Server

    Antonopoulos, Nick

    2010-01-01

    Cloud computing has recently emerged as a subject of substantial industrial and academic interest, though its meaning and scope is hotly debated. For some researchers, clouds are a natural evolution towards the full commercialisation of grid systems, while others dismiss the term as a mere re-branding of existing pay-per-use technologies. From either perspective, 'cloud' is now the label of choice for accountable pay-per-use access to third party applications and computational resources on a massive scale. Clouds support patterns of less predictable resource use for applications and services a

  15. Computational Logistics

    DEFF Research Database (Denmark)

    Pacino, Dario; Voss, Stefan; Jensen, Rune Møller

    2013-01-01

    This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized in to...... in topical sections named: maritime shipping, road transport, vehicle routing problems, aviation applications, and logistics and supply chain management.......This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized...

  16. Computational Logistics

    DEFF Research Database (Denmark)

    This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized in to...... in topical sections named: maritime shipping, road transport, vehicle routing problems, aviation applications, and logistics and supply chain management.......This book constitutes the refereed proceedings of the 4th International Conference on Computational Logistics, ICCL 2013, held in Copenhagen, Denmark, in September 2013. The 19 papers presented in this volume were carefully reviewed and selected for inclusion in the book. They are organized...

  17. Computational engineering

    CERN Document Server

    2014-01-01

    The book presents state-of-the-art works in computational engineering. Focus is on mathematical modeling, numerical simulation, experimental validation and visualization in engineering sciences. In particular, the following topics are presented: constitutive models and their implementation into finite element codes, numerical models in nonlinear elasto-dynamics including seismic excitations, multiphase models in structural engineering and multiscale models of materials systems, sensitivity and reliability analysis of engineering structures, the application of scientific computing in urban water management and hydraulic engineering, and the application of genetic algorithms for the registration of laser scanner point clouds.

  18. Computer busses

    CERN Document Server

    Buchanan, William

    2000-01-01

    As more and more equipment is interface or'bus' driven, either by the use of controllers or directly from PCs, the question of which bus to use is becoming increasingly important both in industry and in the office. 'Computer Busses' has been designed to help choose the best type of bus for the particular application.There are several books which cover individual busses, but none which provide a complete guide to computer busses. The author provides a basic theory of busses and draws examples and applications from real bus case studies. Busses are analysed using from a top-down approach, helpin

  19. Reconfigurable Computing

    CERN Document Server

    Cardoso, Joao MP

    2011-01-01

    As the complexity of modern embedded systems increases, it becomes less practical to design monolithic processing platforms. As a result, reconfigurable computing is being adopted widely for more flexible design. Reconfigurable Computers offer the spatial parallelism and fine-grained customizability of application-specific circuits with the postfabrication programmability of software. To make the most of this unique combination of performance and flexibility, designers need to be aware of both hardware and software issues. FPGA users must think not only about the gates needed to perform a comp

  20. Riemannian computing in computer vision

    CERN Document Server

    Srivastava, Anuj

    2016-01-01

    This book presents a comprehensive treatise on Riemannian geometric computations and related statistical inferences in several computer vision problems. This edited volume includes chapter contributions from leading figures in the field of computer vision who are applying Riemannian geometric approaches in problems such as face recognition, activity recognition, object detection, biomedical image analysis, and structure-from-motion. Some of the mathematical entities that necessitate a geometric analysis include rotation matrices (e.g. in modeling camera motion), stick figures (e.g. for activity recognition), subspace comparisons (e.g. in face recognition), symmetric positive-definite matrices (e.g. in diffusion tensor imaging), and function-spaces (e.g. in studying shapes of closed contours).   ·         Illustrates Riemannian computing theory on applications in computer vision, machine learning, and robotics ·         Emphasis on algorithmic advances that will allow re-application in other...

  1. Statistical Computing

    Indian Academy of Sciences (India)

    inference and finite population sampling. Sudhakar Kunte. Elements of statistical computing are discussed in this series. ... which captain gets an option to decide whether to field first or bat first ... may of course not be fair, in the sense that the team which wins ... describe two methods of drawing a random number between 0.

  2. Computational biology

    DEFF Research Database (Denmark)

    Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue

    2011-01-01

    Computation via biological devices has been the subject of close scrutiny since von Neumann’s early work some 60 years ago. In spite of the many relevant works in this field, the notion of programming biological devices seems to be, at best, ill-defined. While many devices are claimed or proved t...

  3. Computing News

    CERN Multimedia

    McCubbin, N

    2001-01-01

    We are still five years from the first LHC data, so we have plenty of time to get the computing into shape, don't we? Well, yes and no: there is time, but there's an awful lot to do! The recently-completed CERN Review of LHC Computing gives the flavour of the LHC computing challenge. The hardware scale for each of the LHC experiments is millions of 'SpecInt95' (SI95) units of cpu power and tens of PetaBytes of data storage. PCs today are about 20-30SI95, and expected to be about 100 SI95 by 2005, so it's a lot of PCs. This hardware will be distributed across several 'Regional Centres' of various sizes, connected by high-speed networks. How to realise this in an orderly and timely fashion is now being discussed in earnest by CERN, Funding Agencies, and the LHC experiments. Mixed in with this is, of course, the GRID concept...but that's a topic for another day! Of course hardware, networks and the GRID constitute just one part of the computing. Most of the ATLAS effort is spent on software development. What we ...

  4. Quantum Computation

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 16; Issue 9. Quantum Computation - Particle and Wave Aspects of Algorithms. Apoorva Patel. General Article Volume 16 Issue 9 September 2011 pp 821-835. Fulltext. Click here to view fulltext PDF. Permanent link:

  5. Cloud computing.

    Science.gov (United States)

    Wink, Diane M

    2012-01-01

    In this bimonthly series, the author examines how nurse educators can use Internet and Web-based technologies such as search, communication, and collaborative writing tools; social networking and social bookmarking sites; virtual worlds; and Web-based teaching and learning programs. This article describes how cloud computing can be used in nursing education.

  6. Computer Recreations.

    Science.gov (United States)

    Dewdney, A. K.

    1988-01-01

    Describes the creation of the computer program "BOUNCE," designed to simulate a weighted piston coming into equilibrium with a cloud of bouncing balls. The model follows the ideal gas law. Utilizes the critical event technique to create the model. Discusses another program, "BOOM," which simulates a chain reaction. (CW)

  7. [Grid computing

    CERN Multimedia

    Wolinsky, H

    2003-01-01

    "Turn on a water spigot, and it's like tapping a bottomless barrel of water. Ditto for electricity: Flip the switch, and the supply is endless. But computing is another matter. Even with the Internet revolution enabling us to connect in new ways, we are still limited to self-contained systems running locally stored software, limited by corporate, institutional and geographic boundaries" (1 page).

  8. Computational Finance

    DEFF Research Database (Denmark)

    Rasmussen, Lykke

    One of the major challenges in todays post-crisis finance environment is calculating the sensitivities of complex products for hedging and risk management. Historically, these derivatives have been determined using bump-and-revalue, but due to the increasing magnitude of these computations does...

  9. Optical Computing

    Indian Academy of Sciences (India)

    Optical computing technology is, in general, developing in two directions. One approach is ... current support in many places, with private companies as well as governments in several countries encouraging such research work. For example, much ... which enables more information to be carried and data to be processed.

  10. Computable Frames in Computable Banach Spaces

    Directory of Open Access Journals (Sweden)

    S.K. Kaushik

    2016-06-01

    Full Text Available We develop some parts of the frame theory in Banach spaces from the point of view of Computable Analysis. We define computable M-basis and use it to construct a computable Banach space of scalar valued sequences. Computable Xd frames and computable Banach frames are also defined and computable versions of sufficient conditions for their existence are obtained.

  11. Algebraic computing

    International Nuclear Information System (INIS)

    MacCallum, M.A.H.

    1990-01-01

    The implementation of a new computer algebra system is time consuming: designers of general purpose algebra systems usually say it takes about 50 man-years to create a mature and fully functional system. Hence the range of available systems and their capabilities changes little between one general relativity meeting and the next, despite which there have been significant changes in the period since the last report. The introductory remarks aim to give a brief survey of capabilities of the principal available systems and highlight one or two trends. The reference to the most recent full survey of computer algebra in relativity and brief descriptions of the Maple, REDUCE and SHEEP and other applications are given. (author)

  12. Computational Controversy

    OpenAIRE

    Timmermans, Benjamin; Kuhn, Tobias; Beelen, Kaspar; Aroyo, Lora

    2017-01-01

    Climate change, vaccination, abortion, Trump: Many topics are surrounded by fierce controversies. The nature of such heated debates and their elements have been studied extensively in the social science literature. More recently, various computational approaches to controversy analysis have appeared, using new data sources such as Wikipedia, which help us now better understand these phenomena. However, compared to what social sciences have discovered about such debates, the existing computati...

  13. Computed tomography

    International Nuclear Information System (INIS)

    Andre, M.; Resnick, D.

    1988-01-01

    Computed tomography (CT) has matured into a reliable and prominent tool for study of the muscoloskeletal system. When it was introduced in 1973, it was unique in many ways and posed a challenge to interpretation. It is in these unique features, however, that its advantages lie in comparison with conventional techniques. These advantages will be described in a spectrum of important applications in orthopedics and rheumatology

  14. Computed radiography

    International Nuclear Information System (INIS)

    Pupchek, G.

    2004-01-01

    Computed radiography (CR) is an image acquisition process that is used to create digital, 2-dimensional radiographs. CR employs a photostimulable phosphor-based imaging plate, replacing the standard x-ray film and intensifying screen combination. Conventional radiographic exposure equipment is used with no modification required to the existing system. CR can transform an analog x-ray department into a digital one and eliminates the need for chemicals, water, darkrooms and film processor headaches. (author)

  15. Computational universes

    International Nuclear Information System (INIS)

    Svozil, Karl

    2005-01-01

    Suspicions that the world might be some sort of a machine or algorithm existing 'in the mind' of some symbolic number cruncher have lingered from antiquity. Although popular at times, the most radical forms of this idea never reached mainstream. Modern developments in physics and computer science have lent support to the thesis, but empirical evidence is needed before it can begin to replace our contemporary world view

  16. Mathematical modelling of ultrasound propagation in multi-phase flow

    DEFF Research Database (Denmark)

    Simurda, Matej

    violates the repeatability of the measurements and thus impairs the device accuracy. Development of new flow meter designs for these conditions based on a purely experimental approach is expensive both in terms of time and economy. An attractive alternative is the employment of a mathematical model...

  17. Hydrodynamics of multi-phase packed bed micro-reactors

    NARCIS (Netherlands)

    Márquez Luzardo, N.M.

    2010-01-01

    Why to use packed bed micro-reactors for catalyst testing? Miniaturized packed bed reactors have a large surface-to-volume ratio at the reactor and particle level that favors the heat- and mass-transfer processes at all scales (intra-particle, inter-phase and inter-particle or reactor level). If the

  18. Cryogenic homogenization and sampling of heterogeneous multi-phase feedstock

    Science.gov (United States)

    Doyle, Glenn Michael; Ideker, Virgene Linda; Siegwarth, James David

    2002-01-01

    An apparatus and process for producing a homogeneous analytical sample from a heterogenous feedstock by: providing the mixed feedstock, reducing the temperature of the feedstock to a temperature below a critical temperature, reducing the size of the feedstock components, blending the reduced size feedstock to form a homogeneous mixture; and obtaining a representative sample of the homogeneous mixture. The size reduction and blending steps are performed at temperatures below the critical temperature in order to retain organic compounds in the form of solvents, oils, or liquids that may be adsorbed onto or absorbed into the solid components of the mixture, while also improving the efficiency of the size reduction. Preferably, the critical temperature is less than 77 K (-196.degree. C.). Further, with the process of this invention the representative sample may be maintained below the critical temperature until being analyzed.

  19. Microgravity Multi-Phase Flow Experiment for Suborbital Testing (MFEST)

    Data.gov (United States)

    National Aeronautics and Space Administration — The primary objective is to conduct a pathfinder, suborbital flight experiment for two-phase fluid flow and separator operations.The primary purpose of this test...

  20. Fundamentals of Turbulent and Multi-Phase Combustion

    CERN Document Server

    Kuo, Kenneth Kuan-yun

    2012-01-01

    Detailed coverage of advanced combustion topics from the author of Principles of Combustion, Second Edition Turbulence, turbulent combustion, and multiphase reacting flows have become major research topics in recent decades due to their application across diverse fields, including energy, environment, propulsion, transportation, industrial safety, and nanotechnology. Most of the knowledge accumulated from this research has never been published in book form-until now. Fundamentals of Turbulent and Multiphase Combustion presents up-to-date, integrated coverage of the fundamentals of turbulence

  1. Multi-phase Volume Segmentation with Tetrahedral Mesh

    DEFF Research Database (Denmark)

    Nguyen Trung, Tuan; Dahl, Vedrana Andersen; Bærentzen, Jakob Andreas

    Volume segmentation is efficient for reconstructing material structure, which is important for several analyses, e.g. simulation with finite element method, measurement of quantitative information like surface area, surface curvature, volume, etc. We are concerned about the representations of the 3......D volumes, which can be categorized into two groups: fixed voxel grids [1] and unstructured meshes [2]. Among these two representations, the voxel grids are more popular since manipulating a fixed grid is easier than an unstructured mesh, but they are less efficient for quantitative measurements....... In many cases, the voxel grids are converted to explicit meshes, however the conversion may reduce the accuracy of the segmentations, and the effort for meshing is also not trivial. On the other side, methods using unstructured meshes have difficulty in handling topology changes. To reduce the complexity...

  2. Development of the Multi-Phase/Multi-Dimensional Code BUBBLEX

    International Nuclear Information System (INIS)

    Lee, Sang Yong; Kim, Shin Whan; Kim, Eun Kee

    2005-01-01

    A test version of the two-fluid program has been developed by extending the PISO algorithm. Unlike the conventional industry two-fluid codes, such as, RELAP5 and TRAC, this scheme does not need to develop a pressure matrix. Instead, it adopts the iterative procedure to implement the implicitness of the pressure. In this paper, a brief introduction to the numerical scheme will be presented. Then, its application to bubble column simulation will be described. Some concluding remarks will be followed

  3. Stochastic Rotation Dynamics simulations of wetting multi-phase flows

    Science.gov (United States)

    Hiller, Thomas; Sanchez de La Lama, Marta; Brinkmann, Martin

    2016-06-01

    Multi-color Stochastic Rotation Dynamics (SRDmc) has been introduced by Inoue et al. [1,2] as a particle based simulation method to study the flow of emulsion droplets in non-wetting microchannels. In this work, we extend the multi-color method to also account for different wetting conditions. This is achieved by assigning the color information not only to fluid particles but also to virtual wall particles that are required to enforce proper no-slip boundary conditions. To extend the scope of the original SRDmc algorithm to e.g. immiscible two-phase flow with viscosity contrast we implement an angular momentum conserving scheme (SRD+mc). We perform extensive benchmark simulations to show that a mono-phase SRDmc fluid exhibits bulk properties identical to a standard SRD fluid and that SRDmc fluids are applicable to a wide range of immiscible two-phase flows. To quantify the adhesion of a SRD+mc fluid in contact to the walls we measure the apparent contact angle from sessile droplets in mechanical equilibrium. For a further verification of our wettability implementation we compare the dewetting of a liquid film from a wetting stripe to experimental and numerical studies of interfacial morphologies on chemically structured surfaces.

  4. Crustal Viscosity Structure Estimated from Multi-Phase Mixing Theory

    Science.gov (United States)

    Shinevar, W. J.; Behn, M. D.; Hirth, G.

    2014-12-01

    Estimates of lower crustal viscosity are typically constrained by analyses of isostatic rebound, post seismic creep, and laboratory-derived flow laws for crustal rocks and minerals. Here we follow a new approach for calculating the viscosity structure of the lower continental crust. We use Perple_X to calculate mineral assemblages for different crustal compositions. Effective viscosity is then calculated using the rheologic mixing model of Huet et al. (2014) incorporating flow laws for each mineral phase. Calculations are performed along geotherms appropriate for the Basin and Range, Tibetan Plateau, Colorado Plateau, and the San Andreas Fault. To assess the role of crustal composition on viscosity, we examined two compositional gradients extending from an upper crust with ~67 wt% SiO2 to a lower crust that is either: (i) basaltic with ~53 wt% SiO2 (Rudnick and Gao, 2003), or (ii) andesitic with ~64% SiO2 (Hacker et al., 2011). In all cases, the middle continental crust has a viscosity that is 2-3 orders of magnitude greater than that inferred for wet quartz, a common proxy for mid-crustal viscosities. An andesitic lower crust results in viscosities of 1020-1021 Pa-s and 1021-1022 Pa-s for hotter and colder crustal geotherms, respectively. A mafic lower crust predicts viscosities that are an order of magnitude higher for the same geotherm. In all cases, the viscosity calculated from the mixing model decreases less with depth compared to single-phase estimates. Lastly, for anhydrous conditions in which alpha quartz is stable, we find that there is a strong correlation between Vp/Vs and bulk viscosity; in contrast, little to no correlation exists for hydrous conditions.

  5. Multi-Phase Modeling of Rainbird Water Injection

    Science.gov (United States)

    Vu, Bruce T.; Moss, Nicholas; Sampson, Zoe

    2014-01-01

    This paper describes the use of a Volume of Fluid (VOF) multiphase model to simulate the water injected from a rainbird nozzle used in the sound suppression system during launch. The simulations help determine the projectile motion for different water flow rates employed at the pad, as it is critical to know if water will splash on the first-stage rocket engine during liftoff.

  6. On the potential of computational methods and numerical simulation in ice mechanics

    International Nuclear Information System (INIS)

    Bergan, Paal G; Cammaert, Gus; Skeie, Geir; Tharigopula, Venkatapathi

    2010-01-01

    This paper deals with the challenge of developing better methods and tools for analysing interaction between sea ice and structures and, in particular, to be able to calculate ice loads on these structures. Ice loads have traditionally been estimated using empirical data and 'engineering judgment'. However, it is believed that computational mechanics and advanced computer simulations of ice-structure interaction can play an important role in developing safer and more efficient structures, especially for irregular structural configurations. The paper explains the complexity of ice as a material in computational mechanics terms. Some key words here are large displacements and deformations, multi-body contact mechanics, instabilities, multi-phase materials, inelasticity, time dependency and creep, thermal effects, fracture and crushing, and multi-scale effects. The paper points towards the use of advanced methods like ALE formulations, mesh-less methods, particle methods, XFEM, and multi-domain formulations in order to deal with these challenges. Some examples involving numerical simulation of interaction and loads between level sea ice and offshore structures are presented. It is concluded that computational mechanics may prove to become a very useful tool for analysing structures in ice; however, much research is still needed to achieve satisfactory reliability and versatility of these methods.

  7. Computational Fluid Dynamics (CFD-Based Droplet Size Estimates in Emulsification Equipment

    Directory of Open Access Journals (Sweden)

    Jo Janssen

    2016-12-01

    Full Text Available While academic literature shows steady progress in combining multi-phase computational fluid dynamics (CFD and population balance modelling (PBM of emulsification processes, the computational burden of this approach is still too large for routine use in industry. The challenge, thus, is to link a sufficiently detailed flow analysis to the droplet behavior in a way that is both physically relevant and computationally manageable. In this research article we propose the use of single-phase CFD to map out the local maximum stable droplet diameter within a given device, based on well-known academic droplet break-up studies in quasi-steady 2D linear flows. The results of the latter are represented by analytical correlations for the critical capillary number, which are valid across a wide viscosity ratio range. Additionally, we suggest a parameter to assess how good the assumption of quasi-steady 2D flow is locally. The approach is demonstrated for a common lab-scale rotor-stator device (Ultra-Turrax, IKA-Werke GmbH, Staufen, Germany. It is found to provide useful insights with minimal additional user coding and little increase in computational effort compared to the single-phase CFD simulations of the flow field, as such. Some suggestions for further development are briefly discussed.

  8. Customizable computing

    CERN Document Server

    Chen, Yu-Ting; Gill, Michael; Reinman, Glenn; Xiao, Bingjun

    2015-01-01

    Since the end of Dennard scaling in the early 2000s, improving the energy efficiency of computation has been the main concern of the research community and industry. The large energy efficiency gap between general-purpose processors and application-specific integrated circuits (ASICs) motivates the exploration of customizable architectures, where one can adapt the architecture to the workload. In this Synthesis lecture, we present an overview and introduction of the recent developments on energy-efficient customizable architectures, including customizable cores and accelerators, on-chip memory

  9. Computed tomography

    International Nuclear Information System (INIS)

    Wells, P.; Davis, J.; Morgan, M.

    1994-01-01

    X-ray or gamma-ray transmission computed tomography (CT) is a powerful non-destructive evaluation (NDE) technique that produces two-dimensional cross-sectional images of an object without the need to physically section it. CT is also known by the acronym CAT, for computerised axial tomography. This review article presents a brief historical perspective on CT, its current status and the underlying physics. The mathematical fundamentals of computed tomography are developed for the simplest transmission CT modality. A description of CT scanner instrumentation is provided with an emphasis on radiation sources and systems. Examples of CT images are shown indicating the range of materials that can be scanned and the spatial and contrast resolutions that may be achieved. Attention is also given to the occurrence, interpretation and minimisation of various image artefacts that may arise. A final brief section is devoted to the principles and potential of a range of more recently developed tomographic modalities including diffraction CT, positron emission CT and seismic tomography. 57 refs., 2 tabs., 14 figs

  10. Computing Services and Assured Computing

    Science.gov (United States)

    2006-05-01

    fighters’ ability to execute the mission.” Computing Services 4 We run IT Systems that: provide medical care pay the warfighters manage maintenance...users • 1,400 applications • 18 facilities • 180 software vendors • 18,000+ copies of executive software products • Virtually every type of mainframe and... chocs electriques, de branchez les deux cordons d’al imentation avant de faire le depannage P R IM A R Y SD A S B 1 2 PowerHub 7000 RST U L 00- 00

  11. Computational neuroscience

    CERN Document Server

    Blackwell, Kim L

    2014-01-01

    Progress in Molecular Biology and Translational Science provides a forum for discussion of new discoveries, approaches, and ideas in molecular biology. It contains contributions from leaders in their fields and abundant references. This volume brings together different aspects of, and approaches to, molecular and multi-scale modeling, with applications to a diverse range of neurological diseases. Mathematical and computational modeling offers a powerful approach for examining the interaction between molecular pathways and ionic channels in producing neuron electrical activity. It is well accepted that non-linear interactions among diverse ionic channels can produce unexpected neuron behavior and hinder a deep understanding of how ion channel mutations bring about abnormal behavior and disease. Interactions with the diverse signaling pathways activated by G protein coupled receptors or calcium influx adds an additional level of complexity. Modeling is an approach to integrate myriad data sources into a cohesiv...

  12. Social Computing

    CERN Multimedia

    CERN. Geneva

    2011-01-01

    The past decade has witnessed a momentous transformation in the way people interact with each other. Content is now co-produced, shared, classified, and rated by millions of people, while attention has become the ephemeral and valuable resource that everyone seeks to acquire. This talk will describe how social attention determines the production and consumption of content within both the scientific community and social media, how its dynamics can be used to predict the future and the role that social media plays in setting the public agenda. About the speaker Bernardo Huberman is a Senior HP Fellow and Director of the Social Computing Lab at Hewlett Packard Laboratories. He received his Ph.D. in Physics from the University of Pennsylvania, and is currently a Consulting Professor in the Department of Applied Physics at Stanford University. He originally worked in condensed matter physics, ranging from superionic conductors to two-dimensional superfluids, and made contributions to the theory of critical p...

  13. computer networks

    Directory of Open Access Journals (Sweden)

    N. U. Ahmed

    2002-01-01

    Full Text Available In this paper, we construct a new dynamic model for the Token Bucket (TB algorithm used in computer networks and use systems approach for its analysis. This model is then augmented by adding a dynamic model for a multiplexor at an access node where the TB exercises a policing function. In the model, traffic policing, multiplexing and network utilization are formally defined. Based on the model, we study such issues as (quality of service QoS, traffic sizing and network dimensioning. Also we propose an algorithm using feedback control to improve QoS and network utilization. Applying MPEG video traces as the input traffic to the model, we verify the usefulness and effectiveness of our model.

  14. Computer Tree

    Directory of Open Access Journals (Sweden)

    Onur AĞAOĞLU

    2014-12-01

    Full Text Available It is crucial that gifted and talented students should be supported by different educational methods for their interests and skills. The science and arts centres (gifted centres provide the Supportive Education Program for these students with an interdisciplinary perspective. In line with the program, an ICT lesson entitled “Computer Tree” serves for identifying learner readiness levels, and defining the basic conceptual framework. A language teacher also contributes to the process, since it caters for the creative function of the basic linguistic skills. The teaching technique is applied for 9-11 aged student level. The lesson introduces an evaluation process including basic information, skills, and interests of the target group. Furthermore, it includes an observation process by way of peer assessment. The lesson is considered to be a good sample of planning for any subject, for the unpredicted convergence of visual and technical abilities with linguistic abilities.

  15. Computed tomography

    International Nuclear Information System (INIS)

    Boyd, D.P.

    1989-01-01

    This paper reports on computed tomographic (CT) scanning which has improved computer-assisted imaging modalities for radiologic diagnosis. The advantage of this modality is its ability to image thin cross-sectional planes of the body, thus uncovering density information in three dimensions without tissue superposition problems. Because this enables vastly superior imaging of soft tissues in the brain and body, CT scanning was immediately successful and continues to grow in importance as improvements are made in speed, resolution, and cost efficiency. CT scanners are used for general purposes, and the more advanced machines are generally preferred in large hospitals, where volume and variety of usage justifies the cost. For imaging in the abdomen, a scanner with a rapid speed is preferred because peristalsis, involuntary motion of the diaphram, and even cardiac motion are present and can significantly degrade image quality. When contrast media is used in imaging to demonstrate scanner, immediate review of images, and multiformat hardcopy production. A second console is reserved for the radiologist to read images and perform the several types of image analysis that are available. Since CT images contain quantitative information in terms of density values and contours of organs, quantitation of volumes, areas, and masses is possible. This is accomplished with region-of- interest methods, which involve the electronic outlining of the selected region of the television display monitor with a trackball-controlled cursor. In addition, various image- processing options, such as edge enhancement (for viewing fine details of edges) or smoothing filters (for enhancing the detectability of low-contrast lesions) are useful tools

  16. Cloud Computing: The Future of Computing

    OpenAIRE

    Aggarwal, Kanika

    2013-01-01

    Cloud computing has recently emerged as a new paradigm for hosting and delivering services over the Internet. Cloud computing is attractive to business owners as it eliminates the requirement for users to plan ahead for provisioning, and allows enterprises to start from the small and increase resources only when there is a rise in service demand. The basic principles of cloud computing is to make the computing be assigned in a great number of distributed computers, rather then local computer ...

  17. Computer Refurbishment

    International Nuclear Information System (INIS)

    Ichiyen, Norman; Chan, Dominic; Thompson, Paul

    2004-01-01

    The major activity for the 18-month refurbishment outage at the Point Lepreau Generating Station is the replacement of all 380 fuel channel and calandria tube assemblies and the lower portion of connecting feeder pipes. New Brunswick Power would also take advantage of this outage to conduct a number of repairs, replacements, inspections and upgrades (such as rewinding or replacing the generator, replacement of shutdown system trip computers, replacement of certain valves and expansion joints, inspection of systems not normally accessible, etc.). This would allow for an additional 25 to 30 years. Among the systems to be replaced are the PDC's for both shutdown systems. Assessments have been completed for both the SDS1 and SDS2 PDC's, and it has been decided to replace the SDS2 PDCs with the same hardware and software approach that has been used successfully for the Wolsong 2, 3, and 4 and the Qinshan 1 and 2 SDS2 PDCs. For SDS1, it has been decided to use the same software development methodology that was used successfully for the Wolsong and Qinshan called the I A and to use a new hardware platform in order to ensure successful operation for the 25-30 year station operating life. The selected supplier is Triconex, which uses a triple modular redundant architecture that will enhance the robustness/fault tolerance of the design with respect to equipment failures

  18. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... Physician Resources Professions Site Index A-Z Computed Tomography (CT) - Sinuses Computed tomography (CT) of the sinuses ... CT of the Sinuses? What is CT (Computed Tomography) of the Sinuses? Computed tomography, more commonly known ...

  19. Illustrated computer tomography

    International Nuclear Information System (INIS)

    Takahashi, S.

    1983-01-01

    This book provides the following information: basic aspects of computed tomography; atlas of computed tomography of the normal adult; clinical application of computed tomography; and radiotherapy planning and computed tomography

  20. Analog and hybrid computing

    CERN Document Server

    Hyndman, D E

    2013-01-01

    Analog and Hybrid Computing focuses on the operations of analog and hybrid computers. The book first outlines the history of computing devices that influenced the creation of analog and digital computers. The types of problems to be solved on computers, computing systems, and digital computers are discussed. The text looks at the theory and operation of electronic analog computers, including linear and non-linear computing units and use of analog computers as operational amplifiers. The monograph examines the preparation of problems to be deciphered on computers. Flow diagrams, methods of ampl

  1. Cloud Computing Fundamentals

    Science.gov (United States)

    Furht, Borko

    In the introductory chapter we define the concept of cloud computing and cloud services, and we introduce layers and types of cloud computing. We discuss the differences between cloud computing and cloud services. New technologies that enabled cloud computing are presented next. We also discuss cloud computing features, standards, and security issues. We introduce the key cloud computing platforms, their vendors, and their offerings. We discuss cloud computing challenges and the future of cloud computing.

  2. Unconventional Quantum Computing Devices

    OpenAIRE

    Lloyd, Seth

    2000-01-01

    This paper investigates a variety of unconventional quantum computation devices, including fermionic quantum computers and computers that exploit nonlinear quantum mechanics. It is shown that unconventional quantum computing devices can in principle compute some quantities more rapidly than `conventional' quantum computers.

  3. Computing handbook computer science and software engineering

    CERN Document Server

    Gonzalez, Teofilo; Tucker, Allen

    2014-01-01

    Overview of Computer Science Structure and Organization of Computing Peter J. DenningComputational Thinking Valerie BarrAlgorithms and Complexity Data Structures Mark WeissBasic Techniques for Design and Analysis of Algorithms Edward ReingoldGraph and Network Algorithms Samir Khuller and Balaji RaghavachariComputational Geometry Marc van KreveldComplexity Theory Eric Allender, Michael Loui, and Kenneth ReganFormal Models and Computability Tao Jiang, Ming Li, and Bala

  4. Specialized computer architectures for computational aerodynamics

    Science.gov (United States)

    Stevenson, D. K.

    1978-01-01

    In recent years, computational fluid dynamics has made significant progress in modelling aerodynamic phenomena. Currently, one of the major barriers to future development lies in the compute-intensive nature of the numerical formulations and the relative high cost of performing these computations on commercially available general purpose computers, a cost high with respect to dollar expenditure and/or elapsed time. Today's computing technology will support a program designed to create specialized computing facilities to be dedicated to the important problems of computational aerodynamics. One of the still unresolved questions is the organization of the computing components in such a facility. The characteristics of fluid dynamic problems which will have significant impact on the choice of computer architecture for a specialized facility are reviewed.

  5. Ten Years toward Equity: Preliminary Results from a Follow-Up Case Study of Academic Computing Culture

    Directory of Open Access Journals (Sweden)

    Tanya L. Crenshaw

    2017-05-01

    Full Text Available Just over 10 years ago, we conducted a culture study of the Computer Science Department at the flagship University of Illinois at Urbana-Champaign, one of the top five computing departments in the country. The study found that while the department placed an emphasis on research, it did so in a way that, in conjunction with a lack of communication and transparency, devalued teaching and mentoring, and negatively impacted the professional development, education, and sense of belonging of the students. As one part of a multi-phase case study spanning over a decade, this manuscript presents preliminary findings from our latest work at the university. We detail early comparisons between data gathered at the Department of Computer Science at the University of Illinois at Urbana-Champaign in 2005 and our most recent pilot case study, a follow-up research project completed in 2016. Though we have not yet completed the full data collection, we find it worthwhile to reflect on the pilot case study data we have collected thus far. Our data reveals improvements in the perceptions of undergraduate teaching quality and undergraduate peer mentoring networks. However, we also found evidence of continuing feelings of isolation, incidents of bias, policy opacity, and uneven policy implementation that are areas of concern, particularly with respect to historically underrepresented groups. We discuss these preliminary follow-up findings, offer research and methodological reflections, and share next steps for applied research that aims to create positive cultural change in computing.

  6. Applied Parallel Computing Industrial Computation and Optimization

    DEFF Research Database (Denmark)

    Madsen, Kaj; NA NA NA Olesen, Dorte

    Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)......Proceedings and the Third International Workshop on Applied Parallel Computing in Industrial Problems and Optimization (PARA96)...

  7. Further computer appreciation

    CERN Document Server

    Fry, T F

    2014-01-01

    Further Computer Appreciation is a comprehensive cover of the principles and aspects in computer appreciation. The book starts by describing the development of computers from the first to the third computer generations, to the development of processors and storage systems, up to the present position of computers and future trends. The text tackles the basic elements, concepts and functions of digital computers, computer arithmetic, input media and devices, and computer output. The basic central processor functions, data storage and the organization of data by classification of computer files,

  8. BONFIRE: benchmarking computers and computer networks

    OpenAIRE

    Bouckaert, Stefan; Vanhie-Van Gerwen, Jono; Moerman, Ingrid; Phillips, Stephen; Wilander, Jerker

    2011-01-01

    The benchmarking concept is not new in the field of computing or computer networking. With “benchmarking tools”, one usually refers to a program or set of programs, used to evaluate the performance of a solution under certain reference conditions, relative to the performance of another solution. Since the 1970s, benchmarking techniques have been used to measure the performance of computers and computer networks. Benchmarking of applications and virtual machines in an Infrastructure-as-a-Servi...

  9. Democratizing Computer Science

    Science.gov (United States)

    Margolis, Jane; Goode, Joanna; Ryoo, Jean J.

    2015-01-01

    Computer science programs are too often identified with a narrow stratum of the student population, often white or Asian boys who have access to computers at home. But because computers play such a huge role in our world today, all students can benefit from the study of computer science and the opportunity to build skills related to computing. The…

  10. Computing at Stanford.

    Science.gov (United States)

    Feigenbaum, Edward A.; Nielsen, Norman R.

    1969-01-01

    This article provides a current status report on the computing and computer science activities at Stanford University, focusing on the Computer Science Department, the Stanford Computation Center, the recently established regional computing network, and the Institute for Mathematical Studies in the Social Sciences. Also considered are such topics…

  11. Experimental and computational analysis of pressure response in a multiphase flow loop

    Science.gov (United States)

    Morshed, Munzarin; Amin, Al; Rahman, Mohammad Azizur; Imtiaz, Syed

    2016-07-01

    The characteristics of multiphase fluid flow in pipes are useful to understand fluid mechanics encountered in the oil and gas industries. In the present day oil and gas exploration is successively inducing subsea operation in the deep sea and arctic condition. During the transport of petroleum products, understanding the fluid dynamics inside the pipe network is important for flow assurance. In this case the information regarding static and dynamic pressure response, pressure loss, optimum flow rate, pipe diameter etc. are the important parameter for flow assurance. The principal aim of this research is to represents computational analysis and experimental analysis of multi-phase (L/G) in a pipe network. This computational study considers a two-phase fluid flow through a horizontal flow loop with at different Reynolds number in order to determine the pressure distribution, frictional pressure loss profiles by volume of fluid (VOF) method. However, numerical simulations are validated with the experimental data. The experiment is conducted in 76.20 mm ID transparent circular pipe using water and air in the flow loop. Static pressure transducers are used to measure local pressure response in multiphase pipeline.

  12. Soft computing in computer and information science

    CERN Document Server

    Fray, Imed; Pejaś, Jerzy

    2015-01-01

    This book presents a carefully selected and reviewed collection of papers presented during the 19th Advanced Computer Systems conference ACS-2014. The Advanced Computer Systems conference concentrated from its beginning on methods and algorithms of artificial intelligence. Further future brought new areas of interest concerning technical informatics related to soft computing and some more technological aspects of computer science such as multimedia and computer graphics, software engineering, web systems, information security and safety or project management. These topics are represented in the present book under the categories Artificial Intelligence, Design of Information and Multimedia Systems, Information Technology Security and Software Technologies.

  13. Computational Intelligence, Cyber Security and Computational Models

    CERN Document Server

    Anitha, R; Lekshmi, R; Kumar, M; Bonato, Anthony; Graña, Manuel

    2014-01-01

    This book contains cutting-edge research material presented by researchers, engineers, developers, and practitioners from academia and industry at the International Conference on Computational Intelligence, Cyber Security and Computational Models (ICC3) organized by PSG College of Technology, Coimbatore, India during December 19–21, 2013. The materials in the book include theory and applications for design, analysis, and modeling of computational intelligence and security. The book will be useful material for students, researchers, professionals, and academicians. It will help in understanding current research trends and findings and future scope of research in computational intelligence, cyber security, and computational models.

  14. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... When the image slices are reassembled by computer software, the result is a very detailed multidimensional view ... Safety Images related to Computed Tomography (CT) - Head Videos related to Computed Tomography (CT) - Head Sponsored by ...

  15. Computers: Instruments of Change.

    Science.gov (United States)

    Barkume, Megan

    1993-01-01

    Discusses the impact of computers in the home, the school, and the workplace. Looks at changes in computer use by occupations and by industry. Provides information on new job titles in computer occupations. (JOW)

  16. DNA computing models

    CERN Document Server

    Ignatova, Zoya; Zimmermann, Karl-Heinz

    2008-01-01

    In this excellent text, the reader is given a comprehensive introduction to the field of DNA computing. The book emphasizes computational methods to tackle central problems of DNA computing, such as controlling living cells, building patterns, and generating nanomachines.

  17. Distributed multiscale computing

    NARCIS (Netherlands)

    Borgdorff, J.

    2014-01-01

    Multiscale models combine knowledge, data, and hypotheses from different scales. Simulating a multiscale model often requires extensive computation. This thesis evaluates distributing these computations, an approach termed distributed multiscale computing (DMC). First, the process of multiscale

  18. Computational Modeling | Bioenergy | NREL

    Science.gov (United States)

    cell walls and are the source of biofuels and biomaterials. Our modeling investigates their properties . Quantum Mechanical Models NREL studies chemical and electronic properties and processes to reduce barriers Computational Modeling Computational Modeling NREL uses computational modeling to increase the

  19. Computer Viruses: An Overview.

    Science.gov (United States)

    Marmion, Dan

    1990-01-01

    Discusses the early history and current proliferation of computer viruses that occur on Macintosh and DOS personal computers, mentions virus detection programs, and offers suggestions for how libraries can protect themselves and their users from damage by computer viruses. (LRW)

  20. Efficacy of the biomaterials 3 wt%-nanostrontium-hydroxyapatite-enhanced calcium phosphate cement (nanoSr-CPC) and nanoSr-CPC-incorporated simvastatin-loaded poly(lactic-co-glycolic-acid) microspheres in osteogenesis improvement: An explorative multi-phase experimental in vitro/vivo study

    International Nuclear Information System (INIS)

    Masaeli, Reza; Jafarzadeh Kashi, Tahereh Sadat; Dinarvand, Rassoul; Rakhshan, Vahid; Shahoon, Hossein; Hooshmand, Behzad; Mashhadi Abbas, Fatemeh; Raz, Majid; Rajabnejad, Alireza; Eslami, Hossein; Khoshroo, Kimia

    2016-01-01

    Aims: The purpose of this multi-phase explorative in vivo animal/surgical and in vitro multi-test experimental study was to (1) create a 3 wt%-nanostrontium hydroxyapatite-enhanced calcium phosphate cement (Sr-HA/CPC) for increasing bone formation and (2) creating a simvastatin-loaded poly(lactic-co-glycolic acid) (SIM-loaded PLGA) microspheres plus CPC composite (SIM-loaded PLGA + nanostrontium-CPC). The third goal was the extensive assessment of multiple in vitro and in vivo characteristics of the above experimental explorative products in vitro and in vivo (animal and surgical studies). Methods and results pertaining to Sr-HA/CPC: Physical and chemical properties of the prepared Sr-HA/CPC were evaluated. MTT assay and alkaline phosphatase activities, and radiological and histological examinations of Sr-HA/CPC, CPC and negative control were compared. X-ray diffraction (XRD) indicated that crystallinity of the prepared cement increased by increasing the powder-to-liquid ratio. Incorporation of Sr-HA into CPC increased MTT assay (biocompatibility) and ALP activity (P < 0.05). Histomorphometry showed greater bone formation after 4 weeks, after implantation of Sr-HA/CPC in 10 rats compared to implantations of CPC or empty defects in the same rats (n = 30, ANOVA P < 0.05). Methods and results pertaining to SIM-loaded PLGA microspheres + nanostrontium-CPC composite: After SEM assessment, the produced composite of microspheres and enhanced CPC were implanted for 8 weeks in 10 rabbits, along with positive and negative controls, enhanced CPC, and enhanced CPC plus SIM (n = 50). In the control group, only a small amount of bone had been regenerated (localized at the boundary of the defect); whereas, other groups showed new bone formation within and around the materials. A significant difference was found in the osteogenesis induced by the groups sham control (16.96 ± 1.01), bone materials (32.28 ± 4.03), nanostrontium-CPC (24.84 ± 2.6), nanostrontium-CPC-simvastatin (40

  1. Efficacy of the biomaterials 3 wt%-nanostrontium-hydroxyapatite-enhanced calcium phosphate cement (nanoSr-CPC) and nanoSr-CPC-incorporated simvastatin-loaded poly(lactic-co-glycolic-acid) microspheres in osteogenesis improvement: An explorative multi-phase experimental in vitro/vivo study

    Energy Technology Data Exchange (ETDEWEB)

    Masaeli, Reza [Dental Biomaterials Department, School of Dentistry, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Jafarzadeh Kashi, Tahereh Sadat, E-mail: jafarzat@sina.tums.ac.ir [Dental Biomaterials Department, School of Dentistry, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Iranian Tissue Bank and Research Center, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Dinarvand, Rassoul [Department of Pharmaceutics, Faculty of Pharmacy, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Rakhshan, Vahid [Iranian Tissue Bank and Research Center, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Shahoon, Hossein [Department of Oral and Maxillofacial Surgery, School of Dentistry, Shahed University, Tehran (Iran, Islamic Republic of); Hooshmand, Behzad [Department of Periodontology, School of Dentistry, Shahid Beheshti University of Medical Sciences, Tehran (Iran, Islamic Republic of); Mashhadi Abbas, Fatemeh [Department of Oral and Maxillofacial Pathology, School of Dentistry, Shahid Beheshti Medical Science University, Tehran (Iran, Islamic Republic of); Raz, Majid; Rajabnejad, Alireza; Eslami, Hossein [Biomaterials Group, Faculty of Biomedical Engineering, Amirkabir University of Technology, Tehran (Iran, Islamic Republic of); Khoshroo, Kimia [Dental Biomaterials Department, School of Dentistry, Tehran University of Medical Sciences, Tehran (Iran, Islamic Republic of); Department of Developmental Sciences, School of Dentistry, Marquette University, Milwaukee, WI (United States); and others

    2016-12-01

    Aims: The purpose of this multi-phase explorative in vivo animal/surgical and in vitro multi-test experimental study was to (1) create a 3 wt%-nanostrontium hydroxyapatite-enhanced calcium phosphate cement (Sr-HA/CPC) for increasing bone formation and (2) creating a simvastatin-loaded poly(lactic-co-glycolic acid) (SIM-loaded PLGA) microspheres plus CPC composite (SIM-loaded PLGA + nanostrontium-CPC). The third goal was the extensive assessment of multiple in vitro and in vivo characteristics of the above experimental explorative products in vitro and in vivo (animal and surgical studies). Methods and results pertaining to Sr-HA/CPC: Physical and chemical properties of the prepared Sr-HA/CPC were evaluated. MTT assay and alkaline phosphatase activities, and radiological and histological examinations of Sr-HA/CPC, CPC and negative control were compared. X-ray diffraction (XRD) indicated that crystallinity of the prepared cement increased by increasing the powder-to-liquid ratio. Incorporation of Sr-HA into CPC increased MTT assay (biocompatibility) and ALP activity (P < 0.05). Histomorphometry showed greater bone formation after 4 weeks, after implantation of Sr-HA/CPC in 10 rats compared to implantations of CPC or empty defects in the same rats (n = 30, ANOVA P < 0.05). Methods and results pertaining to SIM-loaded PLGA microspheres + nanostrontium-CPC composite: After SEM assessment, the produced composite of microspheres and enhanced CPC were implanted for 8 weeks in 10 rabbits, along with positive and negative controls, enhanced CPC, and enhanced CPC plus SIM (n = 50). In the control group, only a small amount of bone had been regenerated (localized at the boundary of the defect); whereas, other groups showed new bone formation within and around the materials. A significant difference was found in the osteogenesis induced by the groups sham control (16.96 ± 1.01), bone materials (32.28 ± 4.03), nanostrontium-CPC (24.84 ± 2.6), nanostrontium-CPC-simvastatin (40

  2. Computer Virus and Trends

    OpenAIRE

    Tutut Handayani; Soenarto Usna,Drs.MMSI

    2004-01-01

    Since its appearance the first time in the mid-1980s, computer virus has invited various controversies that still lasts to this day. Along with the development of computer systems technology, viruses komputerpun find new ways to spread itself through a variety of existing communications media. This paper discusses about some things related to computer viruses, namely: the definition and history of computer viruses; the basics of computer viruses; state of computer viruses at this time; and ...

  3. Plasticity: modeling & computation

    National Research Council Canada - National Science Library

    Borja, Ronaldo Israel

    2013-01-01

    .... "Plasticity Modeling & Computation" is a textbook written specifically for students who want to learn the theoretical, mathematical, and computational aspects of inelastic deformation in solids...

  4. Cloud Computing Quality

    Directory of Open Access Journals (Sweden)

    Anamaria Şiclovan

    2013-02-01

    Full Text Available Cloud computing was and it will be a new way of providing Internet services and computers. This calculation approach is based on many existing services, such as the Internet, grid computing, Web services. Cloud computing as a system aims to provide on demand services more acceptable as price and infrastructure. It is exactly the transition from computer to a service offered to the consumers as a product delivered online. This paper is meant to describe the quality of cloud computing services, analyzing the advantages and characteristics offered by it. It is a theoretical paper.Keywords: Cloud computing, QoS, quality of cloud computing

  5. Computer hardware fault administration

    Science.gov (United States)

    Archer, Charles J.; Megerian, Mark G.; Ratterman, Joseph D.; Smith, Brian E.

    2010-09-14

    Computer hardware fault administration carried out in a parallel computer, where the parallel computer includes a plurality of compute nodes. The compute nodes are coupled for data communications by at least two independent data communications networks, where each data communications network includes data communications links connected to the compute nodes. Typical embodiments carry out hardware fault administration by identifying a location of a defective link in the first data communications network of the parallel computer and routing communications data around the defective link through the second data communications network of the parallel computer.

  6. Computer jargon explained

    CERN Document Server

    Enticknap, Nicholas

    2014-01-01

    Computer Jargon Explained is a feature in Computer Weekly publications that discusses 68 of the most commonly used technical computing terms. The book explains what the terms mean and why the terms are important to computer professionals. The text also discusses how the terms relate to the trends and developments that are driving the information technology industry. Computer jargon irritates non-computer people and in turn causes problems for computer people. The technology and the industry are changing so rapidly; it is very hard even for professionals to keep updated. Computer people do not

  7. Computers and data processing

    CERN Document Server

    Deitel, Harvey M

    1985-01-01

    Computers and Data Processing provides information pertinent to the advances in the computer field. This book covers a variety of topics, including the computer hardware, computer programs or software, and computer applications systems.Organized into five parts encompassing 19 chapters, this book begins with an overview of some of the fundamental computing concepts. This text then explores the evolution of modern computing systems from the earliest mechanical calculating devices to microchips. Other chapters consider how computers present their results and explain the storage and retrieval of

  8. Computers in nuclear medicine

    International Nuclear Information System (INIS)

    Giannone, Carlos A.

    1999-01-01

    This chapter determines: capture and observation of images in computers; hardware and software used, personal computers, networks and workstations. The use of special filters determine the quality image

  9. Advances in unconventional computing

    CERN Document Server

    2017-01-01

    The unconventional computing is a niche for interdisciplinary science, cross-bred of computer science, physics, mathematics, chemistry, electronic engineering, biology, material science and nanotechnology. The aims of this book are to uncover and exploit principles and mechanisms of information processing in and functional properties of physical, chemical and living systems to develop efficient algorithms, design optimal architectures and manufacture working prototypes of future and emergent computing devices. This first volume presents theoretical foundations of the future and emergent computing paradigms and architectures. The topics covered are computability, (non-)universality and complexity of computation; physics of computation, analog and quantum computing; reversible and asynchronous devices; cellular automata and other mathematical machines; P-systems and cellular computing; infinity and spatial computation; chemical and reservoir computing. The book is the encyclopedia, the first ever complete autho...

  10. Computability and unsolvability

    CERN Document Server

    Davis, Martin

    1985-01-01

    ""A clearly written, well-presented survey of an intriguing subject."" - Scientific American. Classic text considers general theory of computability, computable functions, operations on computable functions, Turing machines self-applied, unsolvable decision problems, applications of general theory, mathematical logic, Kleene hierarchy, computable functionals, classification of unsolvable decision problems and more.

  11. Mathematics for computer graphics

    CERN Document Server

    Vince, John

    2006-01-01

    Helps you understand the mathematical ideas used in computer animation, virtual reality, CAD, and other areas of computer graphics. This work also helps you to rediscover the mathematical techniques required to solve problems and design computer programs for computer graphic applications

  12. Computations and interaction

    NARCIS (Netherlands)

    Baeten, J.C.M.; Luttik, S.P.; Tilburg, van P.J.A.; Natarajan, R.; Ojo, A.

    2011-01-01

    We enhance the notion of a computation of the classical theory of computing with the notion of interaction. In this way, we enhance a Turing machine as a model of computation to a Reactive Turing Machine that is an abstract model of a computer as it is used nowadays, always interacting with the user

  13. Symbiotic Cognitive Computing

    OpenAIRE

    Farrell, Robert G.; Lenchner, Jonathan; Kephjart, Jeffrey O.; Webb, Alan M.; Muller, MIchael J.; Erikson, Thomas D.; Melville, David O.; Bellamy, Rachel K.E.; Gruen, Daniel M.; Connell, Jonathan H.; Soroker, Danny; Aaron, Andy; Trewin, Shari M.; Ashoori, Maryam; Ellis, Jason B.

    2016-01-01

    IBM Research is engaged in a research program in symbiotic cognitive computing to investigate how to embed cognitive computing in physical spaces. This article proposes 5 key principles of symbiotic cognitive computing.  We describe how these principles are applied in a particular symbiotic cognitive computing environment and in an illustrative application.  

  14. Computer scientist looks at reliability computations

    International Nuclear Information System (INIS)

    Rosenthal, A.

    1975-01-01

    Results from the theory of computational complexity are applied to reliability computations on fault trees and networks. A well known class of problems which almost certainly have no fast solution algorithms is presented. It is shown that even approximately computing the reliability of many systems is difficult enough to be in this class. In the face of this result, which indicates that for general systems the computation time will be exponential in the size of the system, decomposition techniques which can greatly reduce the effective size of a wide variety of realistic systems are explored

  15. Roadmap to greener computing

    CERN Document Server

    Nguemaleu, Raoul-Abelin Choumin

    2014-01-01

    A concise and accessible introduction to green computing and green IT, this book addresses how computer science and the computer infrastructure affect the environment and presents the main challenges in making computing more environmentally friendly. The authors review the methodologies, designs, frameworks, and software development tools that can be used in computer science to reduce energy consumption and still compute efficiently. They also focus on Computer Aided Design (CAD) and describe what design engineers and CAD software applications can do to support new streamlined business directi

  16. Brief: Managing computing technology

    International Nuclear Information System (INIS)

    Startzman, R.A.

    1994-01-01

    While computing is applied widely in the production segment of the petroleum industry, its effective application is the primary goal of computing management. Computing technology has changed significantly since the 1950's, when computers first began to influence petroleum technology. The ability to accomplish traditional tasks faster and more economically probably is the most important effect that computing has had on the industry. While speed and lower cost are important, are they enough? Can computing change the basic functions of the industry? When new computing technology is introduced improperly, it can clash with traditional petroleum technology. This paper examines the role of management in merging these technologies

  17. Computer mathematics for programmers

    CERN Document Server

    Abney, Darrell H; Sibrel, Donald W

    1985-01-01

    Computer Mathematics for Programmers presents the Mathematics that is essential to the computer programmer.The book is comprised of 10 chapters. The first chapter introduces several computer number systems. Chapter 2 shows how to perform arithmetic operations using the number systems introduced in Chapter 1. The third chapter covers the way numbers are stored in computers, how the computer performs arithmetic on real numbers and integers, and how round-off errors are generated in computer programs. Chapter 4 details the use of algorithms and flowcharting as problem-solving tools for computer p

  18. Parallel computing works

    Energy Technology Data Exchange (ETDEWEB)

    1991-10-23

    An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

  19. The digital computer

    CERN Document Server

    Parton, K C

    2014-01-01

    The Digital Computer focuses on the principles, methodologies, and applications of the digital computer. The publication takes a look at the basic concepts involved in using a digital computer, simple autocode examples, and examples of working advanced design programs. Discussions focus on transformer design synthesis program, machine design analysis program, solution of standard quadratic equations, harmonic analysis, elementary wage calculation, and scientific calculations. The manuscript then examines commercial and automatic programming, how computers work, and the components of a computer

  20. Cloud computing for radiologists

    OpenAIRE

    Amit T Kharat; Amjad Safvi; S S Thind; Amarjit Singh

    2012-01-01

    Cloud computing is a concept wherein a computer grid is created using the Internet with the sole purpose of utilizing shared resources such as computer software, hardware, on a pay-per-use model. Using Cloud computing, radiology users can efficiently manage multimodality imaging units by using the latest software and hardware without paying huge upfront costs. Cloud computing systems usually work on public, private, hybrid, or community models. Using the various components of a Cloud, such as...

  1. Toward Cloud Computing Evolution

    OpenAIRE

    Susanto, Heru; Almunawar, Mohammad Nabil; Kang, Chen Chin

    2012-01-01

    -Information Technology (IT) shaped the success of organizations, giving them a solid foundation that increases both their level of efficiency as well as productivity. The computing industry is witnessing a paradigm shift in the way computing is performed worldwide. There is a growing awareness among consumers and enterprises to access their IT resources extensively through a "utility" model known as "cloud computing." Cloud computing was initially rooted in distributed grid-based computing. ...

  2. Algorithmically specialized parallel computers

    CERN Document Server

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  3. Synthetic Computation: Chaos Computing, Logical Stochastic Resonance, and Adaptive Computing

    Science.gov (United States)

    Kia, Behnam; Murali, K.; Jahed Motlagh, Mohammad-Reza; Sinha, Sudeshna; Ditto, William L.

    Nonlinearity and chaos can illustrate numerous behaviors and patterns, and one can select different patterns from this rich library of patterns. In this paper we focus on synthetic computing, a field that engineers and synthesizes nonlinear systems to obtain computation. We explain the importance of nonlinearity, and describe how nonlinear systems can be engineered to perform computation. More specifically, we provide an overview of chaos computing, a field that manually programs chaotic systems to build different types of digital functions. Also we briefly describe logical stochastic resonance (LSR), and then extend the approach of LSR to realize combinational digital logic systems via suitable concatenation of existing logical stochastic resonance blocks. Finally we demonstrate how a chaotic system can be engineered and mated with different machine learning techniques, such as artificial neural networks, random searching, and genetic algorithm, to design different autonomous systems that can adapt and respond to environmental conditions.

  4. Future Computer Requirements for Computational Aerodynamics

    Science.gov (United States)

    1978-01-01

    Recent advances in computational aerodynamics are discussed as well as motivations for and potential benefits of a National Aerodynamic Simulation Facility having the capability to solve fluid dynamic equations at speeds two to three orders of magnitude faster than presently possible with general computers. Two contracted efforts to define processor architectures for such a facility are summarized.

  5. Computers and Computation. Readings from Scientific American.

    Science.gov (United States)

    Fenichel, Robert R.; Weizenbaum, Joseph

    A collection of articles from "Scientific American" magazine has been put together at this time because the current period in computer science is one of consolidation rather than innovation. A few years ago, computer science was moving so swiftly that even the professional journals were more archival than informative; but today it is…

  6. Know Your Personal Computer Introduction to Computers

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 1; Issue 1. Know Your Personal Computer Introduction to Computers. Siddhartha Kumar Ghoshal. Series Article Volume 1 Issue 1 January 1996 pp 48-55. Fulltext. Click here to view fulltext PDF. Permanent link:

  7. Heterotic computing: exploiting hybrid computational devices.

    Science.gov (United States)

    Kendon, Viv; Sebald, Angelika; Stepney, Susan

    2015-07-28

    Current computational theory deals almost exclusively with single models: classical, neural, analogue, quantum, etc. In practice, researchers use ad hoc combinations, realizing only recently that they can be fundamentally more powerful than the individual parts. A Theo Murphy meeting brought together theorists and practitioners of various types of computing, to engage in combining the individual strengths to produce powerful new heterotic devices. 'Heterotic computing' is defined as a combination of two or more computational systems such that they provide an advantage over either substrate used separately. This post-meeting collection of articles provides a wide-ranging survey of the state of the art in diverse computational paradigms, together with reflections on their future combination into powerful and practical applications. © 2015 The Author(s) Published by the Royal Society. All rights reserved.

  8. Cloud Computing for radiologists.

    Science.gov (United States)

    Kharat, Amit T; Safvi, Amjad; Thind, Ss; Singh, Amarjit

    2012-07-01

    Cloud computing is a concept wherein a computer grid is created using the Internet with the sole purpose of utilizing shared resources such as computer software, hardware, on a pay-per-use model. Using Cloud computing, radiology users can efficiently manage multimodality imaging units by using the latest software and hardware without paying huge upfront costs. Cloud computing systems usually work on public, private, hybrid, or community models. Using the various components of a Cloud, such as applications, client, infrastructure, storage, services, and processing power, Cloud computing can help imaging units rapidly scale and descale operations and avoid huge spending on maintenance of costly applications and storage. Cloud computing allows flexibility in imaging. It sets free radiology from the confines of a hospital and creates a virtual mobile office. The downsides to Cloud computing involve security and privacy issues which need to be addressed to ensure the success of Cloud computing in the future.

  9. Cloud Computing for radiologists

    International Nuclear Information System (INIS)

    Kharat, Amit T; Safvi, Amjad; Thind, SS; Singh, Amarjit

    2012-01-01

    Cloud computing is a concept wherein a computer grid is created using the Internet with the sole purpose of utilizing shared resources such as computer software, hardware, on a pay-per-use model. Using Cloud computing, radiology users can efficiently manage multimodality imaging units by using the latest software and hardware without paying huge upfront costs. Cloud computing systems usually work on public, private, hybrid, or community models. Using the various components of a Cloud, such as applications, client, infrastructure, storage, services, and processing power, Cloud computing can help imaging units rapidly scale and descale operations and avoid huge spending on maintenance of costly applications and storage. Cloud computing allows flexibility in imaging. It sets free radiology from the confines of a hospital and creates a virtual mobile office. The downsides to Cloud computing involve security and privacy issues which need to be addressed to ensure the success of Cloud computing in the future

  10. Cloud computing for radiologists

    Directory of Open Access Journals (Sweden)

    Amit T Kharat

    2012-01-01

    Full Text Available Cloud computing is a concept wherein a computer grid is created using the Internet with the sole purpose of utilizing shared resources such as computer software, hardware, on a pay-per-use model. Using Cloud computing, radiology users can efficiently manage multimodality imaging units by using the latest software and hardware without paying huge upfront costs. Cloud computing systems usually work on public, private, hybrid, or community models. Using the various components of a Cloud, such as applications, client, infrastructure, storage, services, and processing power, Cloud computing can help imaging units rapidly scale and descale operations and avoid huge spending on maintenance of costly applications and storage. Cloud computing allows flexibility in imaging. It sets free radiology from the confines of a hospital and creates a virtual mobile office. The downsides to Cloud computing involve security and privacy issues which need to be addressed to ensure the success of Cloud computing in the future.

  11. Review of quantum computation

    International Nuclear Information System (INIS)

    Lloyd, S.

    1992-01-01

    Digital computers are machines that can be programmed to perform logical and arithmetical operations. Contemporary digital computers are ''universal,'' in the sense that a program that runs on one computer can, if properly compiled, run on any other computer that has access to enough memory space and time. Any one universal computer can simulate the operation of any other; and the set of tasks that any such machine can perform is common to all universal machines. Since Bennett's discovery that computation can be carried out in a non-dissipative fashion, a number of Hamiltonian quantum-mechanical systems have been proposed whose time-evolutions over discrete intervals are equivalent to those of specific universal computers. The first quantum-mechanical treatment of computers was given by Benioff, who exhibited a Hamiltonian system with a basis whose members corresponded to the logical states of a Turing machine. In order to make the Hamiltonian local, in the sense that its structure depended only on the part of the computation being performed at that time, Benioff found it necessary to make the Hamiltonian time-dependent. Feynman discovered a way to make the computational Hamiltonian both local and time-independent by incorporating the direction of computation in the initial condition. In Feynman's quantum computer, the program is a carefully prepared wave packet that propagates through different computational states. Deutsch presented a quantum computer that exploits the possibility of existing in a superposition of computational states to perform tasks that a classical computer cannot, such as generating purely random numbers, and carrying out superpositions of computations as a method of parallel processing. In this paper, we show that such computers, by virtue of their common function, possess a common form for their quantum dynamics

  12. Computers for imagemaking

    CERN Document Server

    Clark, D

    1981-01-01

    Computers for Image-Making tells the computer non-expert all he needs to know about Computer Animation. In the hands of expert computer engineers, computer picture-drawing systems have, since the earliest days of computing, produced interesting and useful images. As a result of major technological developments since then, it no longer requires the expert's skill to draw pictures; anyone can do it, provided they know how to use the appropriate machinery. This collection of specially commissioned articles reflects the diversity of user applications in this expanding field

  13. Computer Lexis and Terminology

    Directory of Open Access Journals (Sweden)

    Gintautas Grigas

    2011-04-01

    Full Text Available Computer becomes a widely used tool in everyday work and at home. Every computer user sees texts on its screen containing a lot of words naming new concepts. Those words come from the terminology used by specialists. The common vocabury between computer terminology and lexis of everyday language comes into existence. The article deals with the part of computer terminology which goes to everyday usage and the influence of ordinary language to computer terminology. The relation between English and Lithuanian computer terminology, the construction and pronouncing of acronyms are discussed as well.

  14. Computations in plasma physics

    International Nuclear Information System (INIS)

    Cohen, B.I.; Killeen, J.

    1984-01-01

    A review of computer application in plasma physics is presented. Computer contribution to the investigation of magnetic and inertial confinement of a plasma and charged particle beam propagation is described. Typical utilization of computer for simulation and control of laboratory and cosmic experiments with a plasma and for data accumulation in these experiments is considered. Basic computational methods applied in plasma physics are discussed. Future trends of computer utilization in plasma reseaches are considered in terms of an increasing role of microprocessors and high-speed data plotters and the necessity of more powerful computer application

  15. Quantum computer science

    CERN Document Server

    Lanzagorta, Marco

    2009-01-01

    In this text we present a technical overview of the emerging field of quantum computation along with new research results by the authors. What distinguishes our presentation from that of others is our focus on the relationship between quantum computation and computer science. Specifically, our emphasis is on the computational model of quantum computing rather than on the engineering issues associated with its physical implementation. We adopt this approach for the same reason that a book on computer programming doesn't cover the theory and physical realization of semiconductors. Another distin

  16. Explorations in quantum computing

    CERN Document Server

    Williams, Colin P

    2011-01-01

    By the year 2020, the basic memory components of a computer will be the size of individual atoms. At such scales, the current theory of computation will become invalid. ""Quantum computing"" is reinventing the foundations of computer science and information theory in a way that is consistent with quantum physics - the most accurate model of reality currently known. Remarkably, this theory predicts that quantum computers can perform certain tasks breathtakingly faster than classical computers -- and, better yet, can accomplish mind-boggling feats such as teleporting information, breaking suppos

  17. Physics vs. computer science

    International Nuclear Information System (INIS)

    Pike, R.

    1982-01-01

    With computers becoming more frequently used in theoretical and experimental physics, physicists can no longer afford to be ignorant of the basic techniques and results of computer science. Computing principles belong in a physicist's tool box, along with experimental methods and applied mathematics, and the easiest way to educate physicists in computing is to provide, as part of the undergraduate curriculum, a computing course designed specifically for physicists. As well, the working physicist should interact with computer scientists, giving them challenging problems in return for their expertise. (orig.)

  18. Polymorphous computing fabric

    Science.gov (United States)

    Wolinski, Christophe Czeslaw [Los Alamos, NM; Gokhale, Maya B [Los Alamos, NM; McCabe, Kevin Peter [Los Alamos, NM

    2011-01-18

    Fabric-based computing systems and methods are disclosed. A fabric-based computing system can include a polymorphous computing fabric that can be customized on a per application basis and a host processor in communication with said polymorphous computing fabric. The polymorphous computing fabric includes a cellular architecture that can be highly parameterized to enable a customized synthesis of fabric instances for a variety of enhanced application performances thereof. A global memory concept can also be included that provides the host processor random access to all variables and instructions associated with the polymorphous computing fabric.

  19. Computer ray tracing speeds.

    Science.gov (United States)

    Robb, P; Pawlowski, B

    1990-05-01

    The results of measuring the ray trace speed and compilation speed of thirty-nine computers in fifty-seven configurations, ranging from personal computers to super computers, are described. A correlation of ray trace speed has been made with the LINPACK benchmark which allows the ray trace speed to be estimated using LINPACK performance data. The results indicate that the latest generation of workstations, using CPUs based on RISC (Reduced Instruction Set Computer) technology, are as fast or faster than mainframe computers in compute-bound situations.

  20. BIGHORN Computational Fluid Dynamics Theory, Methodology, and Code Verification & Validation Benchmark Problems

    Energy Technology Data Exchange (ETDEWEB)

    Xia, Yidong [Idaho National Lab. (INL), Idaho Falls, ID (United States); Andrs, David [Idaho National Lab. (INL), Idaho Falls, ID (United States); Martineau, Richard Charles [Idaho National Lab. (INL), Idaho Falls, ID (United States)

    2016-08-01

    This document presents the theoretical background for a hybrid finite-element / finite-volume fluid flow solver, namely BIGHORN, based on the Multiphysics Object Oriented Simulation Environment (MOOSE) computational framework developed at the Idaho National Laboratory (INL). An overview of the numerical methods used in BIGHORN are discussed and followed by a presentation of the formulation details. The document begins with the governing equations for the compressible fluid flow, with an outline of the requisite constitutive relations. A second-order finite volume method used for solving the compressible fluid flow problems is presented next. A Pressure-Corrected Implicit Continuous-fluid Eulerian (PCICE) formulation for time integration is also presented. The multi-fluid formulation is being developed. Although multi-fluid is not fully-developed, BIGHORN has been designed to handle multi-fluid problems. Due to the flexibility in the underlying MOOSE framework, BIGHORN is quite extensible, and can accommodate both multi-species and multi-phase formulations. This document also presents a suite of verification & validation benchmark test problems for BIGHORN. The intent for this suite of problems is to provide baseline comparison data that demonstrates the performance of the BIGHORN solution methods on problems that vary in complexity from laminar to turbulent flows. Wherever possible, some form of solution verification has been attempted to identify sensitivities in the solution methods, and suggest best practices when using BIGHORN.

  1. Computing networks from cluster to cloud computing

    CERN Document Server

    Vicat-Blanc, Pascale; Guillier, Romaric; Soudan, Sebastien

    2013-01-01

    "Computing Networks" explores the core of the new distributed computing infrastructures we are using today:  the networking systems of clusters, grids and clouds. It helps network designers and distributed-application developers and users to better understand the technologies, specificities, constraints and benefits of these different infrastructures' communication systems. Cloud Computing will give the possibility for millions of users to process data anytime, anywhere, while being eco-friendly. In order to deliver this emerging traffic in a timely, cost-efficient, energy-efficient, and

  2. Computing Nash equilibria through computational intelligence methods

    Science.gov (United States)

    Pavlidis, N. G.; Parsopoulos, K. E.; Vrahatis, M. N.

    2005-03-01

    Nash equilibrium constitutes a central solution concept in game theory. The task of detecting the Nash equilibria of a finite strategic game remains a challenging problem up-to-date. This paper investigates the effectiveness of three computational intelligence techniques, namely, covariance matrix adaptation evolution strategies, particle swarm optimization, as well as, differential evolution, to compute Nash equilibria of finite strategic games, as global minima of a real-valued, nonnegative function. An issue of particular interest is to detect more than one Nash equilibria of a game. The performance of the considered computational intelligence methods on this problem is investigated using multistart and deflection.

  3. Reversible computing fundamentals, quantum computing, and applications

    CERN Document Server

    De Vos, Alexis

    2010-01-01

    Written by one of the few top internationally recognized experts in the field, this book concentrates on those topics that will remain fundamental, such as low power computing, reversible programming languages, and applications in thermodynamics. It describes reversible computing from various points of view: Boolean algebra, group theory, logic circuits, low-power electronics, communication, software, quantum computing. It is this multidisciplinary approach that makes it unique.Backed by numerous examples, this is useful for all levels of the scientific and academic community, from undergr

  4. Computing in high energy physics

    Energy Technology Data Exchange (ETDEWEB)

    Watase, Yoshiyuki

    1991-09-15

    The increasingly important role played by computing and computers in high energy physics is displayed in the 'Computing in High Energy Physics' series of conferences, bringing together experts in different aspects of computing - physicists, computer scientists, and vendors.

  5. Searching with Quantum Computers

    OpenAIRE

    Grover, Lov K.

    2000-01-01

    This article introduces quantum computation by analogy with probabilistic computation. A basic description of the quantum search algorithm is given by representing the algorithm as a C program in a novel way.

  6. Book Review: Computational Topology

    DEFF Research Database (Denmark)

    Raussen, Martin

    2011-01-01

    Computational Topology by Herbert Edelsbrunner and John L. Harer. American Matheamtical Society, 2010 - ISBN 978-0-8218-4925-5......Computational Topology by Herbert Edelsbrunner and John L. Harer. American Matheamtical Society, 2010 - ISBN 978-0-8218-4925-5...

  7. Essential numerical computer methods

    CERN Document Server

    Johnson, Michael L

    2010-01-01

    The use of computers and computational methods has become ubiquitous in biological and biomedical research. During the last 2 decades most basic algorithms have not changed, but what has is the huge increase in computer speed and ease of use, along with the corresponding orders of magnitude decrease in cost. A general perception exists that the only applications of computers and computer methods in biological and biomedical research are either basic statistical analysis or the searching of DNA sequence data bases. While these are important applications they only scratch the surface of the current and potential applications of computers and computer methods in biomedical research. The various chapters within this volume include a wide variety of applications that extend far beyond this limited perception. As part of the Reliable Lab Solutions series, Essential Numerical Computer Methods brings together chapters from volumes 210, 240, 321, 383, 384, 454, and 467 of Methods in Enzymology. These chapters provide ...

  8. Know Your Personal Computer

    Indian Academy of Sciences (India)

    computer with IBM PC .... read by a human and not translated by a compiler are called .... by different stages of education becomes a computer scientist. ... ancestors knew and carried out the semantic actions without question or comment.

  9. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... images. These images can be viewed on a computer monitor, printed on film or transferred to a ... other in a ring, called a gantry. The computer workstation that processes the imaging information is located ...

  10. SSCL computer planning

    International Nuclear Information System (INIS)

    Price, L.E.

    1990-01-01

    The SSC Laboratory is in the process of planning the acquisition of a substantial computing system to support the design of detectors. Advice has been sought from users and computer experts in several stages. This paper discuss this process

  11. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... images. These images can be viewed on a computer monitor, printed on film or transferred to a ... other in a ring, called a gantry. The computer workstation that processes the imaging information is located ...

  12. Computational Science Facility (CSF)

    Data.gov (United States)

    Federal Laboratory Consortium — PNNL Institutional Computing (PIC) is focused on meeting DOE's mission needs and is part of PNNL's overarching research computing strategy. PIC supports large-scale...

  13. Quantum Computer Science

    Science.gov (United States)

    Mermin, N. David

    2007-08-01

    Preface; 1. Cbits and Qbits; 2. General features and some simple examples; 3. Breaking RSA encryption with a quantum computer; 4. Searching with a quantum computer; 5. Quantum error correction; 6. Protocols that use just a few Qbits; Appendices; Index.

  14. Computer Vision Syndrome.

    Science.gov (United States)

    Randolph, Susan A

    2017-07-01

    With the increased use of electronic devices with visual displays, computer vision syndrome is becoming a major public health issue. Improving the visual status of workers using computers results in greater productivity in the workplace and improved visual comfort.

  15. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... are the limitations of CT of the Sinuses? What is CT (Computed Tomography) of the Sinuses? Computed ... nasal cavity by small openings. top of page What are some common uses of the procedure? CT ...

  16. Computer Technology Directory.

    Science.gov (United States)

    Exceptional Parent, 1990

    1990-01-01

    This directory lists approximately 300 commercial vendors that offer computer hardware, software, and communication aids for children with disabilities. The company listings indicate computer compatibility and specific disabilities served by their products. (JDD)

  17. My Computer Is Learning.

    Science.gov (United States)

    Good, Ron

    1986-01-01

    Describes instructional uses of computer programs found in David Heiserman's book "Projects in Machine Intelligence for Your Home Computer." The programs feature "creatures" of various colors that move around within a rectangular white border. (JN)

  18. What is Computed Tomography?

    Science.gov (United States)

    ... Imaging Medical X-ray Imaging What is Computed Tomography? Share Tweet Linkedin Pin it More sharing options ... Chest X ray Image back to top Computed Tomography (CT) Although also based on the variable absorption ...

  19. Joint Computing Facility

    Data.gov (United States)

    Federal Laboratory Consortium — Raised Floor Computer Space for High Performance ComputingThe ERDC Information Technology Laboratory (ITL) provides a robust system of IT facilities to develop and...

  20. Computing for Belle

    CERN Multimedia

    CERN. Geneva

    2004-01-01

    2s-1, 10 times as much as we obtain now. This presentation describes Belle's efficient computing operations, struggles to manage large amount of raw and physics data, and plans for Belle computing for Super KEKB/Belle.

  1. Computational Continuum Mechanics

    CERN Document Server

    Shabana, Ahmed A

    2011-01-01

    This text presents the theory of continuum mechanics using computational methods. Ideal for students and researchers, the second edition features a new chapter on computational geometry and finite element analysis.

  2. Applications of computer algebra

    CERN Document Server

    1985-01-01

    Today, certain computer software systems exist which surpass the computational ability of researchers when their mathematical techniques are applied to many areas of science and engineering. These computer systems can perform a large portion of the calculations seen in mathematical analysis. Despite this massive power, thousands of people use these systems as a routine resource for everyday calculations. These software programs are commonly called "Computer Algebra" systems. They have names such as MACSYMA, MAPLE, muMATH, REDUCE and SMP. They are receiving credit as a computational aid with in­ creasing regularity in articles in the scientific and engineering literature. When most people think about computers and scientific research these days, they imagine a machine grinding away, processing numbers arithmetically. It is not generally realized that, for a number of years, computers have been performing non-numeric computations. This means, for example, that one inputs an equa­ tion and obtains a closed for...

  3. ICASE Computer Science Program

    Science.gov (United States)

    1985-01-01

    The Institute for Computer Applications in Science and Engineering computer science program is discussed in outline form. Information is given on such topics as problem decomposition, algorithm development, programming languages, and parallel architectures.

  4. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... ring, called a gantry. The computer workstation that processes the imaging information is located in a separate ... follows a spiral path. A special computer program processes this large volume of data to create two- ...

  5. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... ring, called a gantry. The computer workstation that processes the imaging information is located in a separate ... follows a spiral path. A special computer program processes this large volume of data to create two- ...

  6. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... Stroke Brain Tumors Computer Tomography (CT) Safety During Pregnancy Head and Neck Cancer X-ray, Interventional Radiology and Nuclear Medicine Radiation Safety Images related to Computed Tomography (CT) - ...

  7. Intimacy and Computer Communication.

    Science.gov (United States)

    Robson, Dave; Robson, Maggie

    1998-01-01

    Addresses the relationship between intimacy and communication that is based on computer technology. Discusses definitions of intimacy and the nature of intimate conversations that use computers as a communications medium. Explores implications for counseling. (MKA)

  8. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... other in a ring, called a gantry. The computer workstation that processes the imaging information is located ... ray beam follows a spiral path. A special computer program processes this large volume of data to ...

  9. Cognitive Computing for Security.

    Energy Technology Data Exchange (ETDEWEB)

    Debenedictis, Erik [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Rothganger, Fredrick [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Aimone, James Bradley [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Marinella, Matthew [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Evans, Brian Robert [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Warrender, Christina E. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Mickel, Patrick [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-12-01

    Final report for Cognitive Computing for Security LDRD 165613. It reports on the development of hybrid of general purpose/ne uromorphic computer architecture, with an emphasis on potential implementation with memristors.

  10. Computation: A New Open Access Journal of Computational Chemistry, Computational Biology and Computational Engineering

    OpenAIRE

    Karlheinz Schwarz; Rainer Breitling; Christian Allen

    2013-01-01

    Computation (ISSN 2079-3197; http://www.mdpi.com/journal/computation) is an international scientific open access journal focusing on fundamental work in the field of computational science and engineering. Computational science has become essential in many research areas by contributing to solving complex problems in fundamental science all the way to engineering. The very broad range of application domains suggests structuring this journal into three sections, which are briefly characterized ...

  11. Nanoelectronics: Metrology and Computation

    International Nuclear Information System (INIS)

    Lundstrom, Mark; Clark, Jason V.; Klimeck, Gerhard; Raman, Arvind

    2007-01-01

    Research in nanoelectronics poses new challenges for metrology, but advances in theory, simulation and computing and networking technology provide new opportunities to couple simulation and metrology. This paper begins with a brief overview of current work in computational nanoelectronics. Three examples of how computation can assist metrology will then be discussed. The paper concludes with a discussion of how cyberinfrastructure can help connect computing and metrology using the nanoHUB (www.nanoHUB.org) as a specific example

  12. Foundations of Neuromorphic Computing

    Science.gov (United States)

    2013-05-01

    paradigms: few sensors/complex computations and many sensors/simple computation. Challenges with Nano-enabled Neuromorphic Chips A wide variety of...FOUNDATIONS OF NEUROMORPHIC COMPUTING MAY 2013 FINAL TECHNICAL REPORT APPROVED FOR PUBLIC RELEASE; DISTRIBUTION...2009 – SEP 2012 4. TITLE AND SUBTITLE FOUNDATIONS OF NEUROMORPHIC COMPUTING 5a. CONTRACT NUMBER IN-HOUSE 5b. GRANT NUMBER N/A 5c. PROGRAM

  13. Approximation and Computation

    CERN Document Server

    Gautschi, Walter; Rassias, Themistocles M

    2011-01-01

    Approximation theory and numerical analysis are central to the creation of accurate computer simulations and mathematical models. Research in these areas can influence the computational techniques used in a variety of mathematical and computational sciences. This collection of contributed chapters, dedicated to renowned mathematician Gradimir V. Milovanovia, represent the recent work of experts in the fields of approximation theory and numerical analysis. These invited contributions describe new trends in these important areas of research including theoretic developments, new computational alg

  14. Computed tomography for radiographers

    International Nuclear Information System (INIS)

    Brooker, M.

    1986-01-01

    Computed tomography is regarded by many as a complicated union of sophisticated x-ray equipment and computer technology. This book overcomes these complexities. The rigid technicalities of the machinery and the clinical aspects of computed tomography are discussed including the preparation of patients, both physically and mentally, for scanning. Furthermore, the author also explains how to set up and run a computed tomography department, including advice on how the room should be designed

  15. Quantum computing and probability.

    Science.gov (United States)

    Ferry, David K

    2009-11-25

    Over the past two decades, quantum computing has become a popular and promising approach to trying to solve computationally difficult problems. Missing in many descriptions of quantum computing is just how probability enters into the process. Here, we discuss some simple examples of how uncertainty and probability enter, and how this and the ideas of quantum computing challenge our interpretations of quantum mechanics. It is found that this uncertainty can lead to intrinsic decoherence, and this raises challenges for error correction.

  16. Quantum computing and probability

    International Nuclear Information System (INIS)

    Ferry, David K

    2009-01-01

    Over the past two decades, quantum computing has become a popular and promising approach to trying to solve computationally difficult problems. Missing in many descriptions of quantum computing is just how probability enters into the process. Here, we discuss some simple examples of how uncertainty and probability enter, and how this and the ideas of quantum computing challenge our interpretations of quantum mechanics. It is found that this uncertainty can lead to intrinsic decoherence, and this raises challenges for error correction. (viewpoint)

  17. Quantum mechanics and computation

    International Nuclear Information System (INIS)

    Cirac Sasturain, J. I.

    2000-01-01

    We review how some of the basic principles of Quantum Mechanics can be used in the field of computation. In particular, we explain why a quantum computer can perform certain tasks in a much more efficient way than the computers we have available nowadays. We give the requirements for a quantum system to be able to implement a quantum computer and illustrate these requirements in some particular physical situations. (Author) 16 refs

  18. COMPUTATIONAL SCIENCE CENTER

    Energy Technology Data Exchange (ETDEWEB)

    DAVENPORT,J.

    2004-11-01

    The Brookhaven Computational Science Center brings together researchers in biology, chemistry, physics, and medicine with applied mathematicians and computer scientists to exploit the remarkable opportunities for scientific discovery which have been enabled by modern computers. These opportunities are especially great in computational biology and nanoscience, but extend throughout science and technology and include for example, nuclear and high energy physics, astrophysics, materials and chemical science, sustainable energy, environment, and homeland security.

  19. COMPUTER GAMES AND EDUCATION

    OpenAIRE

    Sukhov, Anton

    2018-01-01

    This paper devoted to the research of educational resources and possibilities of modern computer games. The “internal” educational aspects of computer games include educational mechanism (a separate or integrated “tutorial”) and representation of a real or even fantastic educational process within virtual worlds. The “external” dimension represents educational opportunities of computer games for personal and professional development in different genres of computer games (various transport, so...

  20. Man and computer

    International Nuclear Information System (INIS)

    Fischbach, K.F.

    1981-01-01

    The discussion of cultural and sociological consequences of computer evolution is hindered by human prejudice. For example the sentence 'a computer is at best as intelligent as its programmer' veils actual developments. Theoretical limits of computer intelligence are the limits of intelligence in general. Modern computer systems replace not only human labour, but also human decision making and thereby human responsibility. The historical situation is unique. Human head-work is being automated and man is loosing function. (orig.) [de

  1. Computational physics an introduction

    CERN Document Server

    Vesely, Franz J

    1994-01-01

    Author Franz J. Vesely offers students an introductory text on computational physics, providing them with the important basic numerical/computational techniques. His unique text sets itself apart from others by focusing on specific problems of computational physics. The author also provides a selection of modern fields of research. Students will benefit from the appendixes which offer a short description of some properties of computing and machines and outline the technique of 'Fast Fourier Transformation.'

  2. Computing environment logbook

    Science.gov (United States)

    Osbourn, Gordon C; Bouchard, Ann M

    2012-09-18

    A computing environment logbook logs events occurring within a computing environment. The events are displayed as a history of past events within the logbook of the computing environment. The logbook provides search functionality to search through the history of past events to find one or more selected past events, and further, enables an undo of the one or more selected past events.

  3. The Computer Revolution.

    Science.gov (United States)

    Berkeley, Edmund C.

    "The Computer Revolution", a part of the "Second Industrial Revolution", is examined with reference to the social consequences of computers. The subject is introduced in an opening section which discusses the revolution in the handling of information and the history, powers, uses, and working s of computers. A second section examines in detail the…

  4. Advances in physiological computing

    CERN Document Server

    Fairclough, Stephen H

    2014-01-01

    This edited collection will provide an overview of the field of physiological computing, i.e. the use of physiological signals as input for computer control. It will cover a breadth of current research, from brain-computer interfaces to telemedicine.

  5. Physics of quantum computation

    International Nuclear Information System (INIS)

    Belokurov, V.V.; Khrustalev, O.A.; Sadovnichij, V.A.; Timofeevskaya, O.D.

    2003-01-01

    In the paper, the modern status of the theory of quantum computation is considered. The fundamental principles of quantum computers and their basic notions such as quantum processors and computational basis states of the quantum Turing machine as well as the quantum Fourier transform are discussed. Some possible experimental realizations on the basis of NMR methods are given

  6. Quantum walk computation

    International Nuclear Information System (INIS)

    Kendon, Viv

    2014-01-01

    Quantum versions of random walks have diverse applications that are motivating experimental implementations as well as theoretical studies. Recent results showing quantum walks are “universal for quantum computation” relate to algorithms, to be run on quantum computers. We consider whether an experimental implementation of a quantum walk could provide useful computation before we have a universal quantum computer

  7. The Challenge of Computers.

    Science.gov (United States)

    Leger, Guy

    Computers may change teachers' lifestyles, teaching styles, and perhaps even their personal values. A brief survey of the history of computers demonstrates the incredible pace at which computer technology is moving ahead. The cost and size of microchips will continue to decline dramatically over the next 20 years, while the capability and variety…

  8. Visitor's Computer Guidelines | CTIO

    Science.gov (United States)

    Visitor's Computer Guidelines Network Connection Request Instruments Instruments by Telescope IR Instruments Guidelines Library Facilities Outreach NOAO-S EPO Program team Art of Darkness Image Gallery EPO/CADIAS ‹› You are here CTIO Home » Astronomers » Visitor's Computer Guidelines Visitor's Computer

  9. Medical Computational Thinking

    DEFF Research Database (Denmark)

    Musaeus, Peter; Tatar, Deborah Gail; Rosen, Michael A.

    2017-01-01

    Computational thinking (CT) in medicine means deliberating when to pursue computer-mediated solutions to medical problems and evaluating when such solutions are worth pursuing in order to assist in medical decision making. Teaching computational thinking (CT) at medical school should be aligned...

  10. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... Physician Resources Professions Site Index A-Z Computed Tomography (CT) - Head Computed tomography (CT) of the head uses special x-ray ... What is CT Scanning of the Head? Computed tomography, more commonly known as a CT or CAT ...

  11. Emission computed tomography

    International Nuclear Information System (INIS)

    Ott, R.J.

    1986-01-01

    Emission Computed Tomography is a technique used for producing single or multiple cross-sectional images of the distribution of radionuclide labelled agents in vivo. The techniques of Single Photon Emission Computed Tomography (SPECT) and Positron Emission Tomography (PET) are described with particular regard to the function of the detectors used to produce images and the computer techniques used to build up images. (UK)

  12. Computed Tomography (CT) -- Sinuses

    Medline Plus

    Full Text Available ... Physician Resources Professions Site Index A-Z Computed Tomography (CT) - Sinuses Computed tomography (CT) of the sinuses uses special x-ray equipment ... story here Images × Image Gallery Patient undergoing computed tomography (CT) scan. View full size with caption Pediatric Content ...

  13. Computed Tomography (CT) -- Head

    Medline Plus

    Full Text Available ... Physician Resources Professions Site Index A-Z Computed Tomography (CT) - Head Computed tomography (CT) of the head uses special x-ray equipment ... story here Images × Image Gallery Patient undergoing computed tomography (CT) scan. View full size with caption Pediatric Content ...

  14. Beyond the Computer Literacy.

    Science.gov (United States)

    Streibel, Michael J.; Garhart, Casey

    1985-01-01

    Describes the approach taken in an education computing course for pre- and in-service teachers. Outlines the basic operational, analytical, and evaluation skills that are emphasized in the course, suggesting that these skills go beyond the attainment of computer literacy and can assist in the effective use of computers. (ML)

  15. Computer algebra applications

    International Nuclear Information System (INIS)

    Calmet, J.

    1982-01-01

    A survey of applications based either on fundamental algorithms in computer algebra or on the use of a computer algebra system is presented. Recent work in biology, chemistry, physics, mathematics and computer science is discussed. In particular, applications in high energy physics (quantum electrodynamics), celestial mechanics and general relativity are reviewed. (Auth.)

  16. Computer-assisted instruction

    NARCIS (Netherlands)

    Voogt, J.; Fisser, P.; Wright, J.D.

    2015-01-01

    Since the early days of computer technology in education in the 1960s, it was claimed that computers can assist instructional practice and hence improve student learning. Since then computer technology has developed, and its potential for education has increased. In this article, we first discuss

  17. Designing with computational intelligence

    CERN Document Server

    Lopes, Heitor; Mourelle, Luiza

    2017-01-01

    This book discusses a number of real-world applications of computational intelligence approaches. Using various examples, it demonstrates that computational intelligence has become a consolidated methodology for automatically creating new competitive solutions to complex real-world problems. It also presents a concise and efficient synthesis of different systems using computationally intelligent techniques.

  18. A new computing principle

    International Nuclear Information System (INIS)

    Fatmi, H.A.; Resconi, G.

    1988-01-01

    In 1954 while reviewing the theory of communication and cybernetics the late Professor Dennis Gabor presented a new mathematical principle for the design of advanced computers. During our work on these computers it was found that the Gabor formulation can be further advanced to include more recent developments in Lie algebras and geometric probability, giving rise to a new computing principle

  19. Computers and Information Flow.

    Science.gov (United States)

    Patrick, R. L.

    This paper is designed to fill the need for an easily understood introduction to the computing and data processing field for the layman who has, or can expect to have, some contact with it. Information provided includes the unique terminology and jargon of the field, the various types of computers and the scope of computational capabilities, and…

  20. Computer naratology: narrative templates in computer games

    OpenAIRE

    Praks, Vítězslav

    2009-01-01

    Relations and actions between literature and computer games were examined. Study contains theoretical analysis of game as an aesthetic artefact. To play a game means to leave practical world for sake of a fictional world. Artistic communication has more similarities with game communication than with normal, practical communication. Game study can help us understand basic concepts of art communication (game rules - poetic rules, game world - fiction, function in game - meaning in art). Compute...

  1. Neural Computation and the Computational Theory of Cognition

    Science.gov (United States)

    Piccinini, Gualtiero; Bahar, Sonya

    2013-01-01

    We begin by distinguishing computationalism from a number of other theses that are sometimes conflated with it. We also distinguish between several important kinds of computation: computation in a generic sense, digital computation, and analog computation. Then, we defend a weak version of computationalism--neural processes are computations in the…

  2. Quantum computing and spintronics

    International Nuclear Information System (INIS)

    Kantser, V.

    2007-01-01

    Tentative to build a computer, which can operate according to the quantum laws, has leaded to concept of quantum computing algorithms and hardware. In this review we highlight recent developments which point the way to quantum computing on the basis solid state nanostructures after some general considerations concerning quantum information science and introducing a set of basic requirements for any quantum computer proposal. One of the major direction of research on the way to quantum computing is to exploit the spin (in addition to the orbital) degree of freedom of the electron, giving birth to the field of spintronics. We address some semiconductor approach based on spin orbit coupling in semiconductor nanostructures. (authors)

  3. Theory of computation

    CERN Document Server

    Tourlakis, George

    2012-01-01

    Learn the skills and acquire the intuition to assess the theoretical limitations of computer programming Offering an accessible approach to the topic, Theory of Computation focuses on the metatheory of computing and the theoretical boundaries between what various computational models can do and not do—from the most general model, the URM (Unbounded Register Machines), to the finite automaton. A wealth of programming-like examples and easy-to-follow explanations build the general theory gradually, which guides readers through the modeling and mathematical analysis of computational pheno

  4. Computer Security Handbook

    CERN Document Server

    Bosworth, Seymour; Whyne, Eric

    2012-01-01

    The classic and authoritative reference in the field of computer security, now completely updated and revised With the continued presence of large-scale computers; the proliferation of desktop, laptop, and handheld computers; and the vast international networks that interconnect them, the nature and extent of threats to computer security have grown enormously. Now in its fifth edition, Computer Security Handbook continues to provide authoritative guidance to identify and to eliminate these threats where possible, as well as to lessen any losses attributable to them. With seventy-seven chapter

  5. Secure cloud computing

    CERN Document Server

    Jajodia, Sushil; Samarati, Pierangela; Singhal, Anoop; Swarup, Vipin; Wang, Cliff

    2014-01-01

    This book presents a range of cloud computing security challenges and promising solution paths. The first two chapters focus on practical considerations of cloud computing. In Chapter 1, Chandramouli, Iorga, and Chokani describe the evolution of cloud computing and the current state of practice, followed by the challenges of cryptographic key management in the cloud. In Chapter 2, Chen and Sion present a dollar cost model of cloud computing and explore the economic viability of cloud computing with and without security mechanisms involving cryptographic mechanisms. The next two chapters addres

  6. Scalable optical quantum computer

    Energy Technology Data Exchange (ETDEWEB)

    Manykin, E A; Mel' nichenko, E V [Institute for Superconductivity and Solid-State Physics, Russian Research Centre ' Kurchatov Institute' , Moscow (Russian Federation)

    2014-12-31

    A way of designing a scalable optical quantum computer based on the photon echo effect is proposed. Individual rare earth ions Pr{sup 3+}, regularly located in the lattice of the orthosilicate (Y{sub 2}SiO{sub 5}) crystal, are suggested to be used as optical qubits. Operations with qubits are performed using coherent and incoherent laser pulses. The operation protocol includes both the method of measurement-based quantum computations and the technique of optical computations. Modern hybrid photon echo protocols, which provide a sufficient quantum efficiency when reading recorded states, are considered as most promising for quantum computations and communications. (quantum computer)

  7. Computing meaning v.4

    CERN Document Server

    Bunt, Harry; Pulman, Stephen

    2013-01-01

    This book is a collection of papers by leading researchers in computational semantics. It presents a state-of-the-art overview of recent and current research in computational semantics, including descriptions of new methods for constructing and improving resources for semantic computation, such as WordNet, VerbNet, and semantically annotated corpora. It also presents new statistical methods in semantic computation, such as the application of distributional semantics in the compositional calculation of sentence meanings. Computing the meaning of sentences, texts, and spoken or texted dialogue i

  8. Scalable optical quantum computer

    International Nuclear Information System (INIS)

    Manykin, E A; Mel'nichenko, E V

    2014-01-01

    A way of designing a scalable optical quantum computer based on the photon echo effect is proposed. Individual rare earth ions Pr 3+ , regularly located in the lattice of the orthosilicate (Y 2 SiO 5 ) crystal, are suggested to be used as optical qubits. Operations with qubits are performed using coherent and incoherent laser pulses. The operation protocol includes both the method of measurement-based quantum computations and the technique of optical computations. Modern hybrid photon echo protocols, which provide a sufficient quantum efficiency when reading recorded states, are considered as most promising for quantum computations and communications. (quantum computer)

  9. COMPUTATIONAL SCIENCE CENTER

    Energy Technology Data Exchange (ETDEWEB)

    DAVENPORT, J.

    2005-11-01

    The Brookhaven Computational Science Center brings together researchers in biology, chemistry, physics, and medicine with applied mathematicians and computer scientists to exploit the remarkable opportunities for scientific discovery which have been enabled by modern computers. These opportunities are especially great in computational biology and nanoscience, but extend throughout science and technology and include, for example, nuclear and high energy physics, astrophysics, materials and chemical science, sustainable energy, environment, and homeland security. To achieve our goals we have established a close alliance with applied mathematicians and computer scientists at Stony Brook and Columbia Universities.

  10. Computer algebra and operators

    Science.gov (United States)

    Fateman, Richard; Grossman, Robert

    1989-01-01

    The symbolic computation of operator expansions is discussed. Some of the capabilities that prove useful when performing computer algebra computations involving operators are considered. These capabilities may be broadly divided into three areas: the algebraic manipulation of expressions from the algebra generated by operators; the algebraic manipulation of the actions of the operators upon other mathematical objects; and the development of appropriate normal forms and simplification algorithms for operators and their actions. Brief descriptions are given of the computer algebra computations that arise when working with various operators and their actions.

  11. Cloud Computing Bible

    CERN Document Server

    Sosinsky, Barrie

    2010-01-01

    The complete reference guide to the hot technology of cloud computingIts potential for lowering IT costs makes cloud computing a major force for both IT vendors and users; it is expected to gain momentum rapidly with the launch of Office Web Apps later this year. Because cloud computing involves various technologies, protocols, platforms, and infrastructure elements, this comprehensive reference is just what you need if you'll be using or implementing cloud computing.Cloud computing offers significant cost savings by eliminating upfront expenses for hardware and software; its growing popularit

  12. Design of Computer Experiments

    DEFF Research Database (Denmark)

    Dehlendorff, Christian

    The main topic of this thesis is design and analysis of computer and simulation experiments and is dealt with in six papers and a summary report. Simulation and computer models have in recent years received increasingly more attention due to their increasing complexity and usability. Software...... packages make the development of rather complicated computer models using predefined building blocks possible. This implies that the range of phenomenas that are analyzed by means of a computer model has expanded significantly. As the complexity grows so does the need for efficient experimental designs...... and analysis methods, since the complex computer models often are expensive to use in terms of computer time. The choice of performance parameter is an important part of the analysis of computer and simulation models and Paper A introduces a new statistic for waiting times in health care units. The statistic...

  13. Computer in radiology

    International Nuclear Information System (INIS)

    Kuesters, H.

    1985-01-01

    With this publication, the author presents the requirements that a user specific software should fulfill to reach an effective practice rationalisation through computer usage and the hardware configuration necessary as basic equipment. This should make it more difficult in the future for sales representatives to sell radiologists unusable computer systems. Furthermore, questions shall be answered that were asked by computer interested radiologists during the system presentation. On the one hand there still exists a prejudice against programmes of standard texts and on the other side undefined fears, that handling a computer is to difficult and that one has to learn a computer language first to be able to work with computers. Finally, it i pointed out, the real competitive advantages can be obtained through computer usage. (orig.) [de

  14. Programming in biomolecular computation

    DEFF Research Database (Denmark)

    Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue

    2011-01-01

    Our goal is to provide a top-down approach to biomolecular computation. In spite of widespread discussion about connections between biology and computation, one question seems notable by its absence: Where are the programs? We identify a number of common features in programming that seem...... conspicuously absent from the literature on biomolecular computing; to partially redress this absence, we introduce a model of computation that is evidently programmable, by programs reminiscent of low-level computer machine code; and at the same time biologically plausible: its functioning is defined...... by a single and relatively small set of chemical-like reaction rules. Further properties: the model is stored-program: programs are the same as data, so programs are not only executable, but are also compilable and interpretable. It is universal: all computable functions can be computed (in natural ways...

  15. Computer assisted radiology

    International Nuclear Information System (INIS)

    Lemke, H.U.; Jaffe, C.C.; Felix, R.

    1993-01-01

    The proceedings of the CAR'93 symposium present the 126 oral papers and the 58 posters contributed to the four Technical Sessions entitled: (1) Image Management, (2) Medical Workstations, (3) Digital Image Generation - DIG, and (4) Application Systems - AS. Topics discussed in Session (1) are: picture archiving and communication systems, teleradiology, hospital information systems and radiological information systems, technology assessment and implications, standards, and data bases. Session (2) deals with computer vision, computer graphics, design and application, man computer interaction. Session (3) goes into the details of the diagnostic examination methods such as digital radiography, MRI, CT, nuclear medicine, ultrasound, digital angiography, and multimodality imaging. Session (4) is devoted to computer-assisted techniques, as there are: computer assisted radiological diagnosis, knowledge based systems, computer assisted radiation therapy and computer assisted surgical planning. (UWA). 266 figs [de

  16. Micro-computed tomography pore-scale study of flow in porous media: Effect of voxel resolution

    Science.gov (United States)

    Shah, S. M.; Gray, F.; Crawshaw, J. P.; Boek, E. S.

    2016-09-01

    A fundamental understanding of flow in porous media at the pore-scale is necessary to be able to upscale average displacement processes from core to reservoir scale. The study of fluid flow in porous media at the pore-scale consists of two key procedures: Imaging - reconstruction of three-dimensional (3D) pore space images; and modelling such as with single and two-phase flow simulations with Lattice-Boltzmann (LB) or Pore-Network (PN) Modelling. Here we analyse pore-scale results to predict petrophysical properties such as porosity, single-phase permeability and multi-phase properties at different length scales. The fundamental issue is to understand the image resolution dependency of transport properties, in order to up-scale the flow physics from pore to core scale. In this work, we use a high resolution micro-computed tomography (micro-CT) scanner to image and reconstruct three dimensional pore-scale images of five sandstones (Bentheimer, Berea, Clashach, Doddington and Stainton) and five complex carbonates (Ketton, Estaillades, Middle Eastern sample 3, Middle Eastern sample 5 and Indiana Limestone 1) at four different voxel resolutions (4.4 μm, 6.2 μm, 8.3 μm and 10.2 μm), scanning the same physical field of view. Implementing three phase segmentation (macro-pore phase, intermediate phase and grain phase) on pore-scale images helps to understand the importance of connected macro-porosity in the fluid flow for the samples studied. We then compute the petrophysical properties for all the samples using PN and LB simulations in order to study the influence of voxel resolution on petrophysical properties. We then introduce a numerical coarsening scheme which is used to coarsen a high voxel resolution image (4.4 μm) to lower resolutions (6.2 μm, 8.3 μm and 10.2 μm) and study the impact of coarsening data on macroscopic and multi-phase properties. Numerical coarsening of high resolution data is found to be superior to using a lower resolution scan because it

  17. DCE. Future IHEP's computing environment

    International Nuclear Information System (INIS)

    Zheng Guorui; Liu Xiaoling

    1995-01-01

    IHEP'S computing environment consists of several different computing environments established on IHEP computer networks. In which, the BES environment supported HEP computing is the main part of IHEP computing environment. Combining with the procedure of improvement and extension of BES environment, the authors describe development of computing environments in outline as viewed from high energy physics (HEP) environment establishment. The direction of developing to distributed computing of the IHEP computing environment based on the developing trend of present distributed computing is presented

  18. Natural Computing in Computational Finance Volume 4

    CERN Document Server

    O’Neill, Michael; Maringer, Dietmar

    2012-01-01

    This book follows on from Natural Computing in Computational Finance  Volumes I, II and III.   As in the previous volumes of this series, the  book consists of a series of  chapters each of  which was selected following a rigorous, peer-reviewed, selection process.  The chapters illustrate the application of a range of cutting-edge natural  computing and agent-based methodologies in computational finance and economics.  The applications explored include  option model calibration, financial trend reversal detection, enhanced indexation, algorithmic trading,  corporate payout determination and agent-based modeling of liquidity costs, and trade strategy adaptation.  While describing cutting edge applications, the chapters are  written so that they are accessible to a wide audience. Hence, they should be of interest  to academics, students and practitioners in the fields of computational finance and  economics.  

  19. Computational Biology and High Performance Computing 2000

    Energy Technology Data Exchange (ETDEWEB)

    Simon, Horst D.; Zorn, Manfred D.; Spengler, Sylvia J.; Shoichet, Brian K.; Stewart, Craig; Dubchak, Inna L.; Arkin, Adam P.

    2000-10-19

    The pace of extraordinary advances in molecular biology has accelerated in the past decade due in large part to discoveries coming from genome projects on human and model organisms. The advances in the genome project so far, happening well ahead of schedule and under budget, have exceeded any dreams by its protagonists, let alone formal expectations. Biologists expect the next phase of the genome project to be even more startling in terms of dramatic breakthroughs in our understanding of human biology, the biology of health and of disease. Only today can biologists begin to envision the necessary experimental, computational and theoretical steps necessary to exploit genome sequence information for its medical impact, its contribution to biotechnology and economic competitiveness, and its ultimate contribution to environmental quality. High performance computing has become one of the critical enabling technologies, which will help to translate this vision of future advances in biology into reality. Biologists are increasingly becoming aware of the potential of high performance computing. The goal of this tutorial is to introduce the exciting new developments in computational biology and genomics to the high performance computing community.

  20. COMPUTER-ASSISTED ACCOUNTING

    Directory of Open Access Journals (Sweden)

    SORIN-CIPRIAN TEIUŞAN

    2009-01-01

    Full Text Available What is computer-assisted accounting? Where is the place and what is the role of the computer in the financial-accounting activity? What is the position and importance of the computer in the accountant’s activity? All these are questions that require scientific research in order to find the answers. The paper approaches the issue of the support granted to the accountant to organize and manage the accounting activity by the computer. Starting from the notions of accounting and computer, the concept of computer-assisted accounting is introduced, it has a general character and it refers to the accounting performed with the help of the computer or using the computer to automate the procedures performed by the person who is doing the accounting activity; this is a concept used to define the computer applications of the accounting activity. The arguments regarding the use of the computer to assist accounting targets the accounting informatization, the automating of the financial-accounting activities and the endowment with modern technology of the contemporary accounting.

  1. Quantum analogue computing.

    Science.gov (United States)

    Kendon, Vivien M; Nemoto, Kae; Munro, William J

    2010-08-13

    We briefly review what a quantum computer is, what it promises to do for us and why it is so hard to build one. Among the first applications anticipated to bear fruit is the quantum simulation of quantum systems. While most quantum computation is an extension of classical digital computation, quantum simulation differs fundamentally in how the data are encoded in the quantum computer. To perform a quantum simulation, the Hilbert space of the system to be simulated is mapped directly onto the Hilbert space of the (logical) qubits in the quantum computer. This type of direct correspondence is how data are encoded in a classical analogue computer. There is no binary encoding, and increasing precision becomes exponentially costly: an extra bit of precision doubles the size of the computer. This has important consequences for both the precision and error-correction requirements of quantum simulation, and significant open questions remain about its practicality. It also means that the quantum version of analogue computers, continuous-variable quantum computers, becomes an equally efficient architecture for quantum simulation. Lessons from past use of classical analogue computers can help us to build better quantum simulators in future.

  2. COMPUTATIONAL SCIENCE CENTER

    International Nuclear Information System (INIS)

    DAVENPORT, J.

    2006-01-01

    Computational Science is an integral component of Brookhaven's multi science mission, and is a reflection of the increased role of computation across all of science. Brookhaven currently has major efforts in data storage and analysis for the Relativistic Heavy Ion Collider (RHIC) and the ATLAS detector at CERN, and in quantum chromodynamics. The Laboratory is host for the QCDOC machines (quantum chromodynamics on a chip), 10 teraflop/s computers which boast 12,288 processors each. There are two here, one for the Riken/BNL Research Center and the other supported by DOE for the US Lattice Gauge Community and other scientific users. A 100 teraflop/s supercomputer will be installed at Brookhaven in the coming year, managed jointly by Brookhaven and Stony Brook, and funded by a grant from New York State. This machine will be used for computational science across Brookhaven's entire research program, and also by researchers at Stony Brook and across New York State. With Stony Brook, Brookhaven has formed the New York Center for Computational Science (NYCCS) as a focal point for interdisciplinary computational science, which is closely linked to Brookhaven's Computational Science Center (CSC). The CSC has established a strong program in computational science, with an emphasis on nanoscale electronic structure and molecular dynamics, accelerator design, computational fluid dynamics, medical imaging, parallel computing and numerical algorithms. We have been an active participant in DOES SciDAC program (Scientific Discovery through Advanced Computing). We are also planning a major expansion in computational biology in keeping with Laboratory initiatives. Additional laboratory initiatives with a dependence on a high level of computation include the development of hydrodynamics models for the interpretation of RHIC data, computational models for the atmospheric transport of aerosols, and models for combustion and for energy utilization. The CSC was formed to bring together

  3. COMPUTATIONAL SCIENCE CENTER

    Energy Technology Data Exchange (ETDEWEB)

    DAVENPORT, J.

    2006-11-01

    Computational Science is an integral component of Brookhaven's multi science mission, and is a reflection of the increased role of computation across all of science. Brookhaven currently has major efforts in data storage and analysis for the Relativistic Heavy Ion Collider (RHIC) and the ATLAS detector at CERN, and in quantum chromodynamics. The Laboratory is host for the QCDOC machines (quantum chromodynamics on a chip), 10 teraflop/s computers which boast 12,288 processors each. There are two here, one for the Riken/BNL Research Center and the other supported by DOE for the US Lattice Gauge Community and other scientific users. A 100 teraflop/s supercomputer will be installed at Brookhaven in the coming year, managed jointly by Brookhaven and Stony Brook, and funded by a grant from New York State. This machine will be used for computational science across Brookhaven's entire research program, and also by researchers at Stony Brook and across New York State. With Stony Brook, Brookhaven has formed the New York Center for Computational Science (NYCCS) as a focal point for interdisciplinary computational science, which is closely linked to Brookhaven's Computational Science Center (CSC). The CSC has established a strong program in computational science, with an emphasis on nanoscale electronic structure and molecular dynamics, accelerator design, computational fluid dynamics, medical imaging, parallel computing and numerical algorithms. We have been an active participant in DOES SciDAC program (Scientific Discovery through Advanced Computing). We are also planning a major expansion in computational biology in keeping with Laboratory initiatives. Additional laboratory initiatives with a dependence on a high level of computation include the development of hydrodynamics models for the interpretation of RHIC data, computational models for the atmospheric transport of aerosols, and models for combustion and for energy utilization. The CSC was formed to

  4. Computation as Medium

    DEFF Research Database (Denmark)

    Jochum, Elizabeth Ann; Putnam, Lance

    2017-01-01

    Artists increasingly utilize computational tools to generate art works. Computational approaches to art making open up new ways of thinking about agency in interactive art because they invite participation and allow for unpredictable outcomes. Computational art is closely linked...... to the participatory turn in visual art, wherein spectators physically participate in visual art works. Unlike purely physical methods of interaction, computer assisted interactivity affords artists and spectators more nuanced control of artistic outcomes. Interactive art brings together human bodies, computer code......, and nonliving objects to create emergent art works. Computation is more than just a tool for artists, it is a medium for investigating new aesthetic possibilities for choreography and composition. We illustrate this potential through two artistic projects: an improvisational dance performance between a human...

  5. Introduction to morphogenetic computing

    CERN Document Server

    Resconi, Germano; Xu, Guanglin

    2017-01-01

    This book offers a concise introduction to morphogenetic computing, showing that its use makes global and local relations, defects in crystal non-Euclidean geometry databases with source and sink, genetic algorithms, and neural networks more stable and efficient. It also presents applications to database, language, nanotechnology with defects, biological genetic structure, electrical circuit, and big data structure. In Turing machines, input and output states form a system – when the system is in one state, the input is transformed into output. This computation is always deterministic and without any possible contradiction or defects. In natural computation there are defects and contradictions that have to be solved to give a coherent and effective computation. The new computation generates the morphology of the system that assumes different forms in time. Genetic process is the prototype of the morphogenetic computing. At the Boolean logic truth value, we substitute a set of truth (active sets) values with...

  6. The CMS Computing Model

    International Nuclear Information System (INIS)

    Bonacorsi, D.

    2007-01-01

    The CMS experiment at LHC has developed a baseline Computing Model addressing the needs of a computing system capable to operate in the first years of LHC running. It is focused on a data model with heavy streaming at the raw data level based on trigger, and on the achievement of the maximum flexibility in the use of distributed computing resources. The CMS distributed Computing Model includes a Tier-0 centre at CERN, a CMS Analysis Facility at CERN, several Tier-1 centres located at large regional computing centres, and many Tier-2 centres worldwide. The workflows have been identified, along with a baseline architecture for the data management infrastructure. This model is also being tested in Grid Service Challenges of increasing complexity, coordinated with the Worldwide LHC Computing Grid community

  7. Introduction to reversible computing

    CERN Document Server

    Perumalla, Kalyan S

    2013-01-01

    Few books comprehensively cover the software and programming aspects of reversible computing. Filling this gap, Introduction to Reversible Computing offers an expanded view of the field that includes the traditional energy-motivated hardware viewpoint as well as the emerging application-motivated software approach. Collecting scattered knowledge into one coherent account, the book provides a compendium of both classical and recently developed results on reversible computing. It explores up-and-coming theories, techniques, and tools for the application of rever

  8. Tracking and computing

    International Nuclear Information System (INIS)

    Niederer, J.

    1983-01-01

    This note outlines several ways in which large scale simulation computing and programming support may be provided to the SSC design community. One aspect of the problem is getting supercomputer power without the high cost and long lead times of large scale institutional computing. Another aspect is the blending of modern programming practices with more conventional accelerator design programs in ways that do not also swamp designers with the details of complicated computer technology

  9. Computing at Belle II

    International Nuclear Information System (INIS)

    Kuhr, Thomas

    2012-01-01

    Belle II, a next-generation B-factory experiment, will search for new physics effects in a data sample about 50 times larger than the one collected by its predecessor, the Belle experiment. To match the advances in accelerator and detector technology, the computing system and the software have to be upgraded as well. The Belle II computing model is presented and an overview of the distributed computing system and the offline software framework is given.

  10. Computing Conference at Bologna

    International Nuclear Information System (INIS)

    Anon.

    1980-01-01

    From 9-12 September a Europhysics Conference on Computing in High Energy and Nuclear Physics, organized by the Computational Physics Group of the European Physical Society, was held in Bologna, attracting some 150 participants. Its purpose was contact and exchange of information between experimental physicists (from both fields of research) and computer experts (on whom the successful outcome of the research has become increasingly dependent)

  11. Review on Computational Electromagnetics

    Directory of Open Access Journals (Sweden)

    P. Sumithra

    2017-03-01

    Full Text Available Computational electromagnetics (CEM is applied to model the interaction of electromagnetic fields with the objects like antenna, waveguides, aircraft and their environment using Maxwell equations.  In this paper the strength and weakness of various computational electromagnetic techniques are discussed. Performance of various techniques in terms accuracy, memory and computational time for application specific tasks such as modeling RCS (Radar cross section, space applications, thin wires, antenna arrays are presented in this paper.

  12. CAD on personal computers

    International Nuclear Information System (INIS)

    Lee, Seong U; Cho, Cheol Ho; Ko, Il Du

    1990-02-01

    This book contains four studies of CAD on personal computers. The first thing is computer graphics in computer-aided design by Seong U Lee. The second thing is graphics primer and programming with Fortran by Seong U Lee. The third thing is application of Auto cad by Il Do Ko. The last thing is application of CAD in building construction design by Cheol Ho Cho.

  13. Computational movement analysis

    CERN Document Server

    Laube, Patrick

    2014-01-01

    This SpringerBrief discusses the characteristics of spatiotemporal movement data, including uncertainty and scale. It investigates three core aspects of Computational Movement Analysis: Conceptual modeling of movement and movement spaces, spatiotemporal analysis methods aiming at a better understanding of movement processes (with a focus on data mining for movement patterns), and using decentralized spatial computing methods in movement analysis. The author presents Computational Movement Analysis as an interdisciplinary umbrella for analyzing movement processes with methods from a range of fi

  14. Computational neurogenetic modeling

    CERN Document Server

    Benuskova, Lubica

    2010-01-01

    Computational Neurogenetic Modeling is a student text, introducing the scope and problems of a new scientific discipline - Computational Neurogenetic Modeling (CNGM). CNGM is concerned with the study and development of dynamic neuronal models for modeling brain functions with respect to genes and dynamic interactions between genes. These include neural network models and their integration with gene network models. This new area brings together knowledge from various scientific disciplines, such as computer and information science, neuroscience and cognitive science, genetics and molecular biol

  15. Research in computer forensics

    OpenAIRE

    Wai, Hor Cheong

    2002-01-01

    Approved for public release; distribution is unlimited Computer Forensics involves the preservation, identification, extraction and documentation of computer evidence stored in the form of magnetically encoded information. With the proliferation of E-commerce initiatives and the increasing criminal activities on the web, this area of study is catching on in the IT industry and among the law enforcement agencies. The objective of the study is to explore the techniques of computer forensics ...

  16. Research in computer science

    Science.gov (United States)

    Ortega, J. M.

    1986-01-01

    Various graduate research activities in the field of computer science are reported. Among the topics discussed are: (1) failure probabilities in multi-version software; (2) Gaussian Elimination on parallel computers; (3) three dimensional Poisson solvers on parallel/vector computers; (4) automated task decomposition for multiple robot arms; (5) multi-color incomplete cholesky conjugate gradient methods on the Cyber 205; and (6) parallel implementation of iterative methods for solving linear equations.

  17. Computer information systems framework

    International Nuclear Information System (INIS)

    Shahabuddin, S.

    1989-01-01

    Management information systems (MIS) is a commonly used term in computer profession. The new information technology has caused management to expect more from computer. The process of supplying information follows a well defined procedure. MIS should be capable for providing usable information to the various areas and levels of organization. MIS is different from data processing. MIS and business hierarchy provides a good framework for many organization which are using computers. (A.B.)

  18. Human Computer Music Performance

    OpenAIRE

    Dannenberg, Roger B.

    2012-01-01

    Human Computer Music Performance (HCMP) is the study of music performance by live human performers and real-time computer-based performers. One goal of HCMP is to create a highly autonomous artificial performer that can fill the role of a human, especially in a popular music setting. This will require advances in automated music listening and understanding, new representations for music, techniques for music synchronization, real-time human-computer communication, music generation, sound synt...

  19. Intelligent distributed computing

    CERN Document Server

    Thampi, Sabu

    2015-01-01

    This book contains a selection of refereed and revised papers of the Intelligent Distributed Computing Track originally presented at the third International Symposium on Intelligent Informatics (ISI-2014), September 24-27, 2014, Delhi, India.  The papers selected for this Track cover several Distributed Computing and related topics including Peer-to-Peer Networks, Cloud Computing, Mobile Clouds, Wireless Sensor Networks, and their applications.

  20. Genomics With Cloud Computing

    OpenAIRE

    Sukhamrit Kaur; Sandeep Kaur

    2015-01-01

    Abstract Genomics is study of genome which provides large amount of data for which large storage and computation power is needed. These issues are solved by cloud computing that provides various cloud platforms for genomics. These platforms provides many services to user like easy access to data easy sharing and transfer providing storage in hundreds of terabytes more computational power. Some cloud platforms are Google genomics DNAnexus and Globus genomics. Various features of cloud computin...

  1. Parallel computing works!

    CERN Document Server

    Fox, Geoffrey C; Messina, Guiseppe C

    2014-01-01

    A clear illustration of how parallel computers can be successfully appliedto large-scale scientific computations. This book demonstrates how avariety of applications in physics, biology, mathematics and other scienceswere implemented on real parallel computers to produce new scientificresults. It investigates issues of fine-grained parallelism relevant forfuture supercomputers with particular emphasis on hypercube architecture. The authors describe how they used an experimental approach to configuredifferent massively parallel machines, design and implement basic systemsoftware, and develop

  2. Computer science I essentials

    CERN Document Server

    Raus, Randall

    2012-01-01

    REA's Essentials provide quick and easy access to critical information in a variety of different fields, ranging from the most basic to the most advanced. As its name implies, these concise, comprehensive study guides summarize the essentials of the field covered. Essentials are helpful when preparing for exams, doing homework and will remain a lasting reference source for students, teachers, and professionals. Computer Science I includes fundamental computer concepts, number representations, Boolean algebra, switching circuits, and computer architecture.

  3. Discrete computational structures

    CERN Document Server

    Korfhage, Robert R

    1974-01-01

    Discrete Computational Structures describes discrete mathematical concepts that are important to computing, covering necessary mathematical fundamentals, computer representation of sets, graph theory, storage minimization, and bandwidth. The book also explains conceptual framework (Gorn trees, searching, subroutines) and directed graphs (flowcharts, critical paths, information network). The text discusses algebra particularly as it applies to concentrates on semigroups, groups, lattices, propositional calculus, including a new tabular method of Boolean function minimization. The text emphasize

  4. Cloud Computing: An Overview

    Science.gov (United States)

    Qian, Ling; Luo, Zhiguo; Du, Yujian; Guo, Leitao

    In order to support the maximum number of user and elastic service with the minimum resource, the Internet service provider invented the cloud computing. within a few years, emerging cloud computing has became the hottest technology. From the publication of core papers by Google since 2003 to the commercialization of Amazon EC2 in 2006, and to the service offering of AT&T Synaptic Hosting, the cloud computing has been evolved from internal IT system to public service, from cost-saving tools to revenue generator, and from ISP to telecom. This paper introduces the concept, history, pros and cons of cloud computing as well as the value chain and standardization effort.

  5. Computational mathematics in China

    CERN Document Server

    Shi, Zhong-Ci

    1994-01-01

    This volume describes the most significant contributions made by Chinese mathematicians over the past decades in various areas of computational mathematics. Some of the results are quite important and complement Western developments in the field. The contributors to the volume range from noted senior mathematicians to promising young researchers. The topics include finite element methods, computational fluid mechanics, numerical solutions of differential equations, computational methods in dynamical systems, numerical algebra, approximation, and optimization. Containing a number of survey articles, the book provides an excellent way for Western readers to gain an understanding of the status and trends of computational mathematics in China.

  6. Multidisciplinary Computational Research

    National Research Council Canada - National Science Library

    Visbal, Miguel R

    2006-01-01

    The purpose of this work is to develop advanced multidisciplinary numerical simulation capabilities for aerospace vehicles with emphasis on highly accurate, massively parallel computational methods...

  7. Frontiers in Computer Education

    CERN Document Server

    Zhu, Egui; 2011 International Conference on Frontiers in Computer Education (ICFCE 2011)

    2012-01-01

    This book is the proceedings of the 2011 International Conference on Frontiers in Computer Education (ICFCE 2011) in Sanya, China, December 1-2, 2011. The contributions can be useful for researchers, software engineers, and programmers, all interested in promoting the computer and education development. Topics covered are computing and communication technology, network management, wireless networks, telecommunication, Signal and Image Processing, Machine Learning, educational management, educational psychology, educational system, education engineering, education technology and training.  The emphasis is on methods and calculi for computer science and education technology development, verification and verification tools support, experiences from doing developments, and the associated theoretical problems.

  8. Computers appreciated by marketers

    International Nuclear Information System (INIS)

    Mantho, M.

    1993-01-01

    The computer has been worth its weight in gold to the fueloil man. In fact, with falling prices on both software and machines, the worth is greater than gold. Every so often, about every three years, we ask some questions about the utilization of computers. This time, we looked into the future, to find out the acceptance of other marvels such as the cellular phone and hand held computer. At the moment, there isn't much penetration. Contact by two-way radio as well as computing meters on trucks still reign supreme

  9. Genomics With Cloud Computing

    Directory of Open Access Journals (Sweden)

    Sukhamrit Kaur

    2015-04-01

    Full Text Available Abstract Genomics is study of genome which provides large amount of data for which large storage and computation power is needed. These issues are solved by cloud computing that provides various cloud platforms for genomics. These platforms provides many services to user like easy access to data easy sharing and transfer providing storage in hundreds of terabytes more computational power. Some cloud platforms are Google genomics DNAnexus and Globus genomics. Various features of cloud computing to genomics are like easy access and sharing of data security of data less cost to pay for resources but still there are some demerits like large time needed to transfer data less network bandwidth.

  10. Computer Games and Art

    Directory of Open Access Journals (Sweden)

    Anton Sukhov

    2015-10-01

    Full Text Available This article devoted to the search of relevant sources (primary and secondary and characteristics of computer games that allow to include them in the field of art (such as the creation of artistic games, computer graphics, active interaction with other forms of art, signs of spiritual aesthetic act, own temporality of computer games, “aesthetic illusion”, interactivity. In general, modern computer games can be attributed to commercial art and popular culture (blockbuster games and to elite forms of contemporary media art (author’s games, visionary games.

  11. Computers in engineering. 1988

    International Nuclear Information System (INIS)

    Tipnis, V.A.; Patton, E.M.

    1988-01-01

    These proceedings discuss the following subjects: Knowledge base systems; Computers in designing; uses of artificial intelligence; engineering optimization and expert systems of accelerators; and parallel processing in designing

  12. Numbers and computers

    CERN Document Server

    Kneusel, Ronald T

    2015-01-01

    This is a book about numbers and how those numbers are represented in and operated on by computers. It is crucial that developers understand this area because the numerical operations allowed by computers, and the limitations of those operations, especially in the area of floating point math, affect virtually everything people try to do with computers. This book aims to fill this gap by exploring, in sufficient but not overwhelming detail, just what it is that computers do with numbers. Divided into two parts, the first deals with standard representations of integers and floating point numb

  13. Octopus: LLL's computing utility

    International Nuclear Information System (INIS)

    Anon.

    1978-01-01

    The Laboratory's Octopus network constitutes one of the greatest concentrations of computing power in the world. This power derives from the network's organization as well as from the size and capability of its computers, storage media, input/output devices, and communication channels. Being in a network enables these facilities to work together to form a unified computing utility that is accessible on demand directly from the users' offices. This computing utility has made a major contribution to the pace of research and development at the Laboratory; an adequate rate of progress in research could not be achieved without it. 4 figures

  14. Theory and Computation

    Data.gov (United States)

    Federal Laboratory Consortium — Flexible computational infrastructure, software tools and theoretical consultation are provided to support modeling and understanding of the structure and properties...

  15. Educational Computer Utilization and Computer Communications.

    Science.gov (United States)

    Singh, Jai P.; Morgan, Robert P.

    As part of an analysis of educational needs and telecommunications requirements for future educational satellite systems, three studies were carried out. 1) The role of the computer in education was examined and both current status and future requirements were analyzed. Trade-offs between remote time sharing and remote batch process were explored…

  16. A study to establish international diagnostic reference levels for paediatric computed tomography

    International Nuclear Information System (INIS)

    Vassileva, J.; Rehani, M.; Kostova-Lefterova, D.; Al-Naemi, H.M.; Al Suwaidi, J.S.; Arandjic, D.; Bashier, E.H.O.; Kodlulovich Renha, S.; El-Nachef, L.; Aguilar, J.G.; Gershan, V.; Gershkevitsh, E.; Gruppetta, E.; Hustuc, A.; Jauhari, A.; Hassan Kharita, Mohammad; Khelassi-Toutaoui, N.; Khosravi, H.R.; Khoury, H.; Kralik, I.; Mahere, S.; Mazuoliene, J.; Mora, P.; Muhogora, W.; Muthuvelu, P.; Nikodemova, D.; Novak, L.; Pallewatte, A.; Pekarovic, D.; Shaaban, M.; Shelly, E.; Stepanyan, K.; Thelsy, N.; Visrutaratna, P.; Zaman, A.

    2015-01-01

    The article reports results from the largest international dose survey in paediatric computed tomography (CT) in 32 countries and proposes international diagnostic reference levels (DRLs) in terms of computed tomography dose index (CTDI vol ) and dose length product (DLP). It also assesses whether mean or median values of individual facilities should be used. A total of 6115 individual patient data were recorded among four age groups: <1 y, >1-5 y, >5-10 y and >10-15 y. CTDI w , CTDI vol and DLP from the CT console were recorded in dedicated forms together with patient data and technical parameters. Statistical analysis was performed, and international DRLs were established at rounded 75. percentile values of distribution of median values from all CT facilities. The study presents evidence in favour of using median rather than mean of patient dose indices as the representative of typical local dose in a facility, and for establishing DRLs as third quartile of median values. International DRLs were established for paediatric CT examinations for routine head, chest and abdomen in the four age groups. DRLs for CTDI vol are similar to the reference values from other published reports, with some differences for chest and abdomen CT. Higher variations were observed between DLP values, based on a survey of whole multi-phase exams. It may be noted that other studies in literature were based on single phase only. DRLs reported in this article can be used in countries without sufficient medical physics support to identify non-optimised practice. Recommendations to improve the accuracy and importance of future surveys are provided. (authors)

  17. Computer Aided Mathematics

    DEFF Research Database (Denmark)

    Sinclair, Robert

    1998-01-01

    Course notes of a PhD course held in 1998. The central idea is to introduce students to computational mathematics using object oriented programming in C++.......Course notes of a PhD course held in 1998. The central idea is to introduce students to computational mathematics using object oriented programming in C++....

  18. Computer Network Operations Methodology

    Science.gov (United States)

    2004-03-01

    means of their computer information systems. Disrupt - This type of attack focuses on disrupting as “attackers might surreptitiously reprogram enemy...by reprogramming the computers that control distribution within the power grid. A disruption attack introduces disorder and inhibits the effective...between commanders. The use of methodologies is widespread and done subconsciously to assist individuals in decision making. The processes that

  19. Classroom Computer Network.

    Science.gov (United States)

    Lent, John

    1984-01-01

    This article describes a computer network system that connects several microcomputers to a single disk drive and one copy of software. Many schools are switching to networks as a cheaper and more efficient means of computer instruction. Teachers may be faced with copywriting problems when reproducing programs. (DF)

  20. Hypercard Another Computer Tool.

    Science.gov (United States)

    Geske, Joel

    1991-01-01

    Describes "Hypercard," a computer application package usable in all three modes of instructional computing: tutor, tool, and tutee. Suggests using Hypercard in scholastic journalism programs to teach such topics as news, headlines, design, photography, and advertising. Argues that the ability to access, organize, manipulate, and comprehend…

  1. Can Computers See?

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 4; Issue 6. Can Computers See? - Can Computers Understand Visual Data? Neelima Shrikhande. General Article Volume 4 Issue 6 June 1999 pp 45-56. Fulltext. Click here to view fulltext PDF. Permanent link:

  2. Computational genomics of hyperthermophiles

    NARCIS (Netherlands)

    Werken, van de H.J.G.

    2008-01-01

    With the ever increasing number of completely sequenced prokaryotic genomes and the subsequent use of functional genomics tools, e.g. DNA microarray and proteomics, computational data analysis and the integration of microbial and molecular data is inevitable. This thesis describes the computational

  3. Computer Technology for Industry

    Science.gov (United States)

    1979-01-01

    In this age of the computer, more and more business firms are automating their operations for increased efficiency in a great variety of jobs, from simple accounting to managing inventories, from precise machining to analyzing complex structures. In the interest of national productivity, NASA is providing assistance both to longtime computer users and newcomers to automated operations. Through a special technology utilization service, NASA saves industry time and money by making available already developed computer programs which have secondary utility. A computer program is essentially a set of instructions which tells the computer how to produce desired information or effect by drawing upon its stored input. Developing a new program from scratch can be costly and time-consuming. Very often, however, a program developed for one purpose can readily be adapted to a totally different application. To help industry take advantage of existing computer technology, NASA operates the Computer Software Management and Information Center (COSMIC)(registered TradeMark),located at the University of Georgia. COSMIC maintains a large library of computer programs developed for NASA, the Department of Defense, the Department of Energy and other technology-generating agencies of the government. The Center gets a continual flow of software packages, screens them for adaptability to private sector usage, stores them and informs potential customers of their availability.

  4. Computers in construction

    DEFF Research Database (Denmark)

    Howard, Rob

    The evolution of technology, particularly computing in building, learning from the past in order to anticipate what may happen in the future......The evolution of technology, particularly computing in building, learning from the past in order to anticipate what may happen in the future...

  5. Computer Use Exposed

    NARCIS (Netherlands)

    J.M. Richter (Janneke)

    2009-01-01

    textabstractEver since the introduction of the personal computer, our daily lives are infl uenced more and more by computers. A day in the life of a PhD-student illustrates this: “At the breakfast table, I check my e-mail to see if the meeting later that day has been confi rmed, and I check the time

  6. Can Computers Create?

    Science.gov (United States)

    Hausman, Carl R.

    1985-01-01

    To be creative, an act must have as its outcome something new in the way it is intelligible and valuable. Computers have restricted contexts of information and have no ability to weigh bits of information. Computer optimists presuppose either determinism or indeterminism, either of which abandons creativity. (MT)

  7. Personalized Empathic Computing (PEC)

    NARCIS (Netherlands)

    van Beusekom, W.; van den Broek, Egon; van der Heijden, M.; Janssen, J.H.; Spaak, E.

    2006-01-01

    Until a decade ago, computers were only used by experts, for professional purposes solely. Nowadays, the personal computer (PC) is standard equipment in most western housekeepings and is used to gather information, play games, communicate, etc. In parallel, users' expectations increase and,

  8. Computers and Creativity.

    Science.gov (United States)

    Ten Dyke, Richard P.

    1982-01-01

    A traditional question is whether or not computers shall ever think like humans. This question is redirected to a discussion of whether computers shall ever be truly creative. Creativity is defined and a program is described that is designed to complete creatively a series problem in mathematics. (MP)

  9. Petascale Computational Systems

    OpenAIRE

    Bell, Gordon; Gray, Jim; Szalay, Alex

    2007-01-01

    Computational science is changing to be data intensive. Super-Computers must be balanced systems; not just CPU farms but also petascale IO and networking arrays. Anyone building CyberInfrastructure should allocate resources to support a balanced Tier-1 through Tier-3 design.

  10. Computer Software Reviews.

    Science.gov (United States)

    Hawaii State Dept. of Education, Honolulu. Office of Instructional Services.

    Intended to provide guidance in the selection of the best computer software available to support instruction and to make optimal use of schools' financial resources, this publication provides a listing of computer software programs that have been evaluated according to their currency, relevance, and value to Hawaii's educational programs. The…

  11. Emission computed tomography

    International Nuclear Information System (INIS)

    Budinger, T.F.; Gullberg, G.T.; Huesman, R.H.

    1979-01-01

    This chapter is devoted to the methods of computer assisted tomography for determination of the three-dimensional distribution of gamma-emitting radionuclides in the human body. The major applications of emission computed tomography are in biological research and medical diagnostic procedures. The objectives of these procedures are to make quantitative measurements of in vivo biochemical and hemodynamic functions

  12. Computers in writing instruction

    NARCIS (Netherlands)

    Schwartz, Helen J.; van der Geest, Thea; Smit-Kreuzen, Marlies

    1992-01-01

    For computers to be useful in writing instruction, innovations should be valuable for students and feasible for teachers to implement. Research findings yield contradictory results in measuring the effects of different uses of computers in writing, in part because of the methodological complexity of

  13. Nature, computation and complexity

    International Nuclear Information System (INIS)

    Binder, P-M; Ellis, G F R

    2016-01-01

    The issue of whether the unfolding of events in the world can be considered a computation is explored in this paper. We come to different conclusions for inert and for living systems (‘no’ and ‘qualified yes’, respectively). We suggest that physical computation as we know it exists only as a tool of complex biological systems: us. (paper)

  14. Computational Sociolinguistics: A Survey

    NARCIS (Netherlands)

    Nguyen, Dong-Phuong; Doğruöz, A. Seza; Rosé, Carolyn P.; de Jong, Franciska M.G.

    Language is a social phenomenon and variation is inherent to its social nature. Recently, there has been a surge of interest within the computational linguistics (CL) community in the social dimension of language. In this article we present a survey of the emerging field of “computational

  15. Fault tolerant computing systems

    International Nuclear Information System (INIS)

    Randell, B.

    1981-01-01

    Fault tolerance involves the provision of strategies for error detection damage assessment, fault treatment and error recovery. A survey is given of the different sorts of strategies used in highly reliable computing systems, together with an outline of recent research on the problems of providing fault tolerance in parallel and distributed computing systems. (orig.)

  16. Theory and computational science

    International Nuclear Information System (INIS)

    Durham, P.

    1985-01-01

    The theoretical and computational science carried out at the Daresbury Laboratory in 1984/5 is detailed in the Appendix to the Daresbury Annual Report. The Theory, Computational Science and Applications Groups, provide support work for the experimental projects conducted at Daresbury. Use of the FPS-164 processor is also described. (U.K.)

  17. Selecting Personal Computers.

    Science.gov (United States)

    Djang, Philipp A.

    1993-01-01

    Describes a Multiple Criteria Decision Analysis Approach for the selection of personal computers that combines the capabilities of Analytic Hierarchy Process and Integer Goal Programing. An example of how decision makers can use this approach to determine what kind of personal computers and how many of each type to purchase is given. (nine…

  18. Physicist or computer specialist?

    Energy Technology Data Exchange (ETDEWEB)

    Clifton, J S [University College Hospital, London (United Kingdom)

    1966-06-15

    Since to most clinicians physical and computer science are two of the great mysteries of the world, the physicist in a hospital is expected by clinicians to be fully conversant with, and competent to make profound pronouncements on, all methods of computing. specific computing problems, and the suitability of computing machinery ranging from desk calculators to Atlas. This is not surprising since the proportion of the syllabus devoted to physics and mathematics in an M. B. degree is indeed meagre, and the word 'computer' has been surrounded with an aura of mysticism which suggests that it is some fantastic piece of electronic gadgetry comprehensible only to a veritable genius. The clinician consequently turns to the only scientific colleague with whom he has direct contact - the medical physicist - and expects him to be an authority. The physicist is thus thrust, however unwillingly, into the forefront of the advance of computer assistance to scientific medicine. It is therefore essential for him to acquire sufficient knowledge of computing science to enable him to provide satisfactory answers for the clinicianst queries, to proffer more detailed advice as to programming convince clinicians that the computer is really a 'simpleton' which can only add and subtract and even that only under instruction.

  19. Theory of computational complexity

    CERN Document Server

    Du, Ding-Zhu

    2011-01-01

    DING-ZHU DU, PhD, is a professor in the Department of Computer Science at the University of Minnesota. KER-I KO, PhD, is a professor in the Department of Computer Science at the State University of New York at Stony Brook.

  20. Computer vision for sports

    DEFF Research Database (Denmark)

    Thomas, Graham; Gade, Rikke; Moeslund, Thomas B.

    2017-01-01

    fixed to players or equipment is generally not possible. This provides a rich set of opportunities for the application of computer vision techniques to help the competitors, coaches and audience. This paper discusses a selection of current commercial applications that use computer vision for sports...

  1. Basic principles of computers

    International Nuclear Information System (INIS)

    Royal, H.D.; Parker, J.A.; Holmen, B.L.

    1988-01-01

    This chapter presents preliminary concepts of computer operations. It describes the hardware used in a nuclear medicine computer system. It discusses the software necessary for acquisition and analysis of nuclear medicine studies. The chapter outlines the integrated package of hardware and software that is necessary to perform specific functions in nuclear medicine

  2. Teaching Using Computer Games

    Science.gov (United States)

    Miller, Lee Dee; Shell, Duane; Khandaker, Nobel; Soh, Leen-Kiat

    2011-01-01

    Computer games have long been used for teaching. Current reviews lack categorization and analysis using learning models which would help instructors assess the usefulness of computer games. We divide the use of games into two classes: game playing and game development. We discuss the Input-Process-Outcome (IPO) model for the learning process when…

  3. Text understanding for computers

    NARCIS (Netherlands)

    Kenter, T.M.

    2017-01-01

    A long-standing challenge for computers communicating with humans is to pass the Turing test, i.e., to communicate in such a way that it is impossible for humans to determine whether they are talking to a computer or another human being. The field of natural language understanding — which studies

  4. Advances in Computer Entertainment.

    NARCIS (Netherlands)

    Nijholt, Antinus; Romão, T.; Reidsma, Dennis; Unknown, [Unknown

    2012-01-01

    These are the proceedings of the 9th International Conference on Advances in Computer Entertainment ACE 2012). ACE has become the leading scientific forum for dissemination of cutting-edge research results in the area of entertainment computing. Interactive entertainment is one of the most vibrant

  5. Computers and Classroom Culture.

    Science.gov (United States)

    Schofield, Janet Ward

    This book explores the meaning of computer technology in schools. The book is based on data gathered from a two-year observation of more than 30 different classrooms in an urban high school: geometry classes in which students used artificially intelligent tutors; business classes in which students learned word processing; and computer science…

  6. Computer Literacy Education

    Science.gov (United States)

    1989-01-01

    Cognitive Aspect ," AEDS Journal, 18, 3 (Spring 1985) 150. "°Geoffrey Akst, "Computer Literacy: An Interview with Dr. Michael Hoban." Journal of Develop- m...1984. Cheng, Tina T.; Plake, Barbara; and Stevens, Dorothy Jo. "A Validation Study of the Computer Literacy Examination: Cognitive Aspect ." AEDS

  7. Ubiquitous human computing.

    Science.gov (United States)

    Zittrain, Jonathan

    2008-10-28

    Ubiquitous computing means network connectivity everywhere, linking devices and systems as small as a drawing pin and as large as a worldwide product distribution chain. What could happen when people are so readily networked? This paper explores issues arising from two possible emerging models of ubiquitous human computing: fungible networked brainpower and collective personal vital sign monitoring.

  8. Learning with Ubiquitous Computing

    Science.gov (United States)

    Rosenheck, Louisa

    2008-01-01

    If ubiquitous computing becomes a reality and is widely adopted, it will inevitably have an impact on education. This article reviews the background of ubiquitous computing and current research projects done involving educational "ubicomp." Finally it explores how ubicomp may and may not change education in both formal and informal settings and…

  9. Quantum Analog Computing

    Science.gov (United States)

    Zak, M.

    1998-01-01

    Quantum analog computing is based upon similarity between mathematical formalism of quantum mechanics and phenomena to be computed. It exploits a dynamical convergence of several competing phenomena to an attractor which can represent an externum of a function, an image, a solution to a system of ODE, or a stochastic process.

  10. Computing in Research.

    Science.gov (United States)

    Ashenhurst, Robert L.

    The introduction and diffusion of automatic computing facilities during the 1960's is reviewed; it is described as a time when research strategies in a broad variety of disciplines changed to take advantage of the newfound power provided by the computer. Several types of typical problems encountered by researchers who adopted the new technologies,…

  11. Computational Cognitive Color Perception

    NARCIS (Netherlands)

    Ciftcioglu, O.; Bittermann, M.S.

    2016-01-01

    Comprehension of aesthetical color characteristics based on a computational model of visual perception and color cognition are presented. The computational comprehension is manifested by the machine’s capability of instantly assigning appropriate colors to the objects perceived. They form a scene

  12. Thinking about computational thinking

    NARCIS (Netherlands)

    Lu, J.J.; Fletcher, G.H.L.; Fitzgerald, S.; Guzdial, M.; Lewandowski, G.; Wolfman, S.A.

    2009-01-01

    Jeannette Wing's call for teaching Computational Thinking (CT) as a formative skill on par with reading, writing, and arithmetic places computer science in the category of basic knowledge. Just as proficiency in basic language arts helps us to effectively communicate and in basic math helps us to

  13. Computer Operating System Maintenance.

    Science.gov (United States)

    1982-06-01

    FACILITY The Computer Management Information Facility ( CMIF ) system was developed by Rapp Systems to fulfill the need at the CRF to record and report on...computer center resource usage and utilization. The foundation of the CMIF system is a System 2000 data base (CRFMGMT) which stores and permits access

  14. Computational Sociolinguistics: A Survey.

    NARCIS (Netherlands)

    de Jong, F.M.G.; Nguyen, Dong

    2016-01-01

    Language is a social phenomenon and variation is inherent to its social nature. Recently, there has been a surge of interest within the computational linguistics (CL) community in the social dimension of language. In this article we present a survey of the emerging field of “computational

  15. Simulation of quantum computers

    NARCIS (Netherlands)

    De Raedt, H; Michielsen, K; Hams, AH; Miyashita, S; Saito, K; Landau, DP; Lewis, SP; Schuttler, HB

    2001-01-01

    We describe a simulation approach to study the functioning of Quantum Computer hardware. The latter is modeled by a collection of interacting spin-1/2 objects. The time evolution of this spin system maps one-to-one to a quantum program carried out by the Quantum Computer. Our simulation software

  16. Simulation of quantum computers

    NARCIS (Netherlands)

    Raedt, H. De; Michielsen, K.; Hams, A.H.; Miyashita, S.; Saito, K.

    2000-01-01

    We describe a simulation approach to study the functioning of Quantum Computer hardware. The latter is modeled by a collection of interacting spin-1/2 objects. The time evolution of this spin system maps one-to-one to a quantum program carried out by the Quantum Computer. Our simulation software

  17. Exercises in Computational Chemistry

    DEFF Research Database (Denmark)

    Spanget-Larsen, Jens

    2016-01-01

    A selection of HyperChem© PC-exercises in computational chemistry. Answers to most questions are appended (Roskilde University 2014-16).......A selection of HyperChem© PC-exercises in computational chemistry. Answers to most questions are appended (Roskilde University 2014-16)....

  18. The Computational Materials Repository

    DEFF Research Database (Denmark)

    Landis, David D.; Hummelshøj, Jens S.; Nestorov, Svetlozar

    2012-01-01

    The possibilities for designing new materials based on quantum physics calculations are rapidly growing, but these design efforts lead to a significant increase in the amount of computational data created. The Computational Materials Repository (CMR) addresses this data challenge and provides...

  19. Programming in biomolecular computation

    DEFF Research Database (Denmark)

    Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue

    2010-01-01

    in a strong sense: a universal algorithm exists, that is able to execute any program, and is not asymptotically inefficient. A prototype model has been implemented (for now in silico on a conventional computer). This work opens new perspectives on just how computation may be specified at the biological level......., by programs reminiscent of low-level computer machine code; and at the same time biologically plausible: its functioning is defined by a single and relatively small set of chemical-like reaction rules. Further properties: the model is stored-program: programs are the same as data, so programs are not only...... executable, but are also compilable and interpretable. It is universal: all computable functions can be computed (in natural ways and without arcane encodings of data and algorithm); it is also uniform: new “hardware” is not needed to solve new problems; and (last but not least) it is Turing complete...

  20. Place-Specific Computing

    DEFF Research Database (Denmark)

    Messeter, Jörn; Johansson, Michael

    project place- specific computing is explored through design oriented research. This article reports six pilot studies where design students have designed concepts for place-specific computing in Berlin (Germany), Cape Town (South Africa), Rome (Italy) and Malmö (Sweden). Background and arguments...... for place-specific computing as a genre of interaction design are described. A total number of 36 design concepts designed for 16 designated zones in the four cities are presented. An analysis of the design concepts is presented indicating potentials, possibilities and problems as directions for future......An increased interest in the notion of place has evolved in interaction design. Proliferation of wireless infrastructure, developments in digital media, and a ‘spatial turn’ in computing provides the base for place-specific computing as a suggested new genre of interaction design. In the REcult...