Final Report for ?Queuing Network Models of Performance of High End Computing Systems?
Energy Technology Data Exchange (ETDEWEB)
Buckwalter, J
2005-09-28
The primary objective of this project is to perform general research into queuing network models of performance of high end computing systems. A related objective is to investigate and predict how an increase in the number of nodes of a supercomputer will decrease the running time of a user's software package, which is often referred to as the strong scaling problem. We investigate the large, MPI-based Linux cluster MCR at LLNL, running the well-known NAS Parallel Benchmark (NPB) applications. Data is collected directly from NPB and also from the low-overhead LLNL profiling tool mpiP. For a run, we break the wall clock execution time of the benchmark into four components: switch delay, MPI contention time, MPI service time, and non-MPI computation time. Switch delay is estimated from message statistics. MPI service time and non-MPI computation time are calculated directly from measurement data. MPI contention is estimated by means of a queuing network model (QNM), based in part on MPI service time. This model of execution time validates reasonably well against the measured execution time, usually within 10%. Since the number of nodes used to run the application is a major input to the model, we can use the model to predict application execution times for various numbers of nodes. We also investigate how the four components of execution time scale individually as the number of nodes increases. Switch delay and MPI service time scale regularly. MPI contention is estimated by the QNM submodel and also has a fairly regular pattern. However, non-MPI compute time has a somewhat irregular pattern, possibly due to caching effects in the memory hierarchy. In contrast to some other performance modeling methods, this method is relatively fast to set up, fast to calculate, simple for data collection, and yet accurate enough to be quite useful.
Time-Predictable Computer Architecture
Directory of Open Access Journals (Sweden)
Schoeberl Martin
2009-01-01
Full Text Available Today's general-purpose processors are optimized for maximum throughput. Real-time systems need a processor with both a reasonable and a known worst-case execution time (WCET. Features such as pipelines with instruction dependencies, caches, branch prediction, and out-of-order execution complicate WCET analysis and lead to very conservative estimates. In this paper, we evaluate the issues of current architectures with respect to WCET analysis. Then, we propose solutions for a time-predictable computer architecture. The proposed architecture is evaluated with implementation of some features in a Java processor. The resulting processor is a good target for WCET analysis and still performs well in the average case.
Directory of Open Access Journals (Sweden)
L. S. Schulman
2005-10-01
Full Text Available Abstract: Some researchers believe that the psychological or consciousness arrow of time is a consequence of the thermodynamic arrow. Some don't. As for many issues in this area, the disagreement revolves about fundamental and undebatable assumptions. As a contribution to this standoff I consider the extent to which a computer---presumably governed by nothing more than the thermodynamic arrow---can be said to possess a psychological arrow. My contention is that the parallels are sufficiently strong as to leave little room for an independent psychological arrow. Reservations are nevertheless expressed on the complete objectivity of the thermodynamic arrow.
12 CFR 1102.27 - Computing time.
2010-01-01
... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 1102.27 Section 1102.27 Banks... for Proceedings § 1102.27 Computing time. (a) General rule. In computing any period of time prescribed... time begins to run is not included. The last day so computed is included, unless it is a...
12 CFR 622.21 - Computing time.
2010-01-01
... 12 Banks and Banking 6 2010-01-01 2010-01-01 false Computing time. 622.21 Section 622.21 Banks and... Formal Hearings § 622.21 Computing time. (a) General rule. In computing any period of time prescribed or allowed by this subpart, the date of the act or event from which the designated period of time begins...
12 CFR 908.27 - Computing time.
2010-01-01
... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 908.27 Section 908.27 Banks and... PRACTICE AND PROCEDURE IN HEARINGS ON THE RECORD General Rules § 908.27 Computing time. (a) General rule. In computing any period of time prescribed or allowed by this subpart, the date of the act or...
Time Traveling: Historians and Computers.
Semonche, John E.
1995-01-01
Presents a brief overview of the recent history and current state of computers as instructional tools for history. Maintains that, although computer use in social science research has increased, most classroom instruction remains static. Includes several descriptions of programs and applications for classrooms. (MJP)
Applied time series analysis and innovative computing
Ao, Sio-Iong
2010-01-01
This text is a systematic, state-of-the-art introduction to the use of innovative computing paradigms as an investigative tool for applications in time series analysis. It includes frontier case studies based on recent research.
Computer network time synchronization the network time protocol
Mills, David L
2006-01-01
What started with the sundial has, thus far, been refined to a level of precision based on atomic resonance: Time. Our obsession with time is evident in this continued scaling down to nanosecond resolution and beyond. But this obsession is not without warrant. Precision and time synchronization are critical in many applications, such as air traffic control and stock trading, and pose complex and important challenges in modern information networks.Penned by David L. Mills, the original developer of the Network Time Protocol (NTP), Computer Network Time Synchronization: The Network Time Protocol
12 CFR 1780.11 - Computing time.
2010-01-01
... the prescribed period for the responsive filing. (3) If service was made by electronic media... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 1780.11 Section 1780.11 Banks and Banking OFFICE OF FEDERAL HOUSING ENTERPRISE OVERSIGHT, DEPARTMENT OF HOUSING AND URBAN...
Time reversibility, computer simulation, and chaos
Hoover, William Graham
1999-01-01
A small army of physicists, chemists, mathematicians, and engineers has joined forces to attack a classic problem, the "reversibility paradox", with modern tools. This book describes their work from the perspective of computer simulation, emphasizing the author's approach to the problem of understanding the compatibility, and even inevitability, of the irreversible second law of thermodynamics with an underlying time-reversible mechanics. Computer simulation has made it possible to probe reversibility from a variety of directions and "chaos theory" or "nonlinear dynamics" has supplied a useful
Time series modeling, computation, and inference
Prado, Raquel
2010-01-01
The authors systematically develop a state-of-the-art analysis and modeling of time series. … this book is well organized and well written. The authors present various statistical models for engineers to solve problems in time series analysis. Readers no doubt will learn state-of-the-art techniques from this book.-Hsun-Hsien Chang, Computing Reviews, March 2012My favorite chapters were on dynamic linear models and vector AR and vector ARMA models.-William Seaver, Technometrics, August 2011… a very modern entry to the field of time-series modelling, with a rich reference list of the current lit
Perspective: Computer simulations of long time dynamics
Energy Technology Data Exchange (ETDEWEB)
Elber, Ron [Department of Chemistry, The Institute for Computational Engineering and Sciences, University of Texas at Austin, Austin, Texas 78712 (United States)
2016-02-14
Atomically detailed computer simulations of complex molecular events attracted the imagination of many researchers in the field as providing comprehensive information on chemical, biological, and physical processes. However, one of the greatest limitations of these simulations is of time scales. The physical time scales accessible to straightforward simulations are too short to address many interesting and important molecular events. In the last decade significant advances were made in different directions (theory, software, and hardware) that significantly expand the capabilities and accuracies of these techniques. This perspective describes and critically examines some of these advances.
Time reversibility, computer simulation, algorithms, chaos
Hoover, William Graham
2012-01-01
A small army of physicists, chemists, mathematicians, and engineers has joined forces to attack a classic problem, the "reversibility paradox", with modern tools. This book describes their work from the perspective of computer simulation, emphasizing the author's approach to the problem of understanding the compatibility, and even inevitability, of the irreversible second law of thermodynamics with an underlying time-reversible mechanics. Computer simulation has made it possible to probe reversibility from a variety of directions and "chaos theory" or "nonlinear dynamics" has supplied a useful vocabulary and a set of concepts, which allow a fuller explanation of irreversibility than that available to Boltzmann or to Green, Kubo and Onsager. Clear illustration of concepts is emphasized throughout, and reinforced with a glossary of technical terms from the specialized fields which have been combined here to focus on a common theme. The book begins with a discussion, contrasting the idealized reversibility of ba...
A Linear Time Algorithm for Seeds Computation
Kociumaka, Tomasz; Radoszewski, Jakub; Rytter, Wojciech; Walen, Tomasz
2011-01-01
Periodicity in words is one of the most fundamental areas of text algorithms and combinatorics. Two classical and natural variations of periodicity are seeds and covers (also called quasiperiods). Linear-time algorithms are known for finding all the covers of a word, however in case of seeds, for the past 15 years only an $O(n\\log{n})$ time algorithm was known (Iliopoulos, Moore and Park, 1996). Finding an $o(n\\log{n})$ time algorithm for the all-seeds problem was mentioned as one of the most important open problems related to repetitions in words in a survey by Smyth (2000). We show a linear-time algorithm computing all the seeds of a word, in particular, the shortest seed. Our approach is based on the use of a version of LZ-factorization and non-trivial combinatorial relations between the LZ-factorization and seeds. It is used here for the first time in context of seeds. It saves the work done for factors processed earlier, similarly as in Crochemore's square-free testing.
Variational optical flow computation in real time.
Bruhn, Andrés; Weickert, Joachim; Feddern, Christian; Kohlberger, Timo; Schnörr, Christoph
2005-05-01
This paper investigates the usefulness of bidirectional multigrid methods for variational optical flow computations. Although these numerical schemes are among the fastest methods for solving equation systems, they are rarely applied in the field of computer vision. We demonstrate how to employ those numerical methods for the treatment of variational optical flow formulations and show that the efficiency of this approach even allows for real-time performance on standard PCs. As a representative for variational optic flow methods, we consider the recently introduced combined local-global method. It can be considered as a noise-robust generalization of the Horn and Schunck technique. We present a decoupled, as well as a coupled, version of the classical Gauss-Seidel solver, and we develop several multgrid implementations based on a discretization coarse grid approximation. In contrast, with standard bidirectional multigrid algorithms, we take advantage of intergrid transfer operators that allow for nondyadic grid hierarchies. As a consequence, no restrictions concerning the image size or the number of traversed levels have to be imposed. In the experimental section, we juxtapose the developed multigrid schemes and demonstrate their superior performance when compared to unidirectional multgrid methods and nonhierachical solvers. For the well-known 316 x 252 Yosemite sequence, we succeeded in computing the complete set of dense flow fields in three quarters of a second on a 3.06-GHz Pentium4 PC. This corresponds to a frame rate of 18 flow fields per second which outperforms the widely-used Gauss-Seidel method by almost three orders of magnitude.
Real-Time Accumulative Computation Motion Detectors
Directory of Open Access Journals (Sweden)
Saturnino Maldonado-Bascón
2009-12-01
Full Text Available The neurally inspired accumulative computation (AC method and its application to motion detection have been introduced in the past years. This paper revisits the fact that many researchers have explored the relationship between neural networks and finite state machines. Indeed, finite state machines constitute the best characterized computational model, whereas artificial neural networks have become a very successful tool for modeling and problem solving. The article shows how to reach real-time performance after using a model described as a finite state machine. This paper introduces two steps towards that direction: (a A simplification of the general AC method is performed by formally transforming it into a finite state machine. (b A hardware implementation in FPGA of such a designed AC module, as well as an 8-AC motion detector, providing promising performance results. We also offer two case studies of the use of AC motion detectors in surveillance applications, namely infrared-based people segmentation and color-based people tracking, respectively.
Response time accuracy in Apple Macintosh computers.
Neath, Ian; Earle, Avery; Hallett, Darcy; Surprenant, Aimée M
2011-06-01
The accuracy and variability of response times (RTs) collected on stock Apple Macintosh computers using USB keyboards was assessed. A photodiode detected a change in the screen's luminosity and triggered a solenoid that pressed a key on the keyboard. The RTs collected in this way were reliable, but could be as much as 100 ms too long. The standard deviation of the measured RTs varied between 2.5 and 10 ms, and the distributions approximated a normal distribution. Surprisingly, two recent Apple-branded USB keyboards differed in their accuracy by as much as 20 ms. The most accurate RTs were collected when an external CRT was used to display the stimuli and Psychtoolbox was able to synchronize presentation with the screen refresh. We conclude that RTs collected on stock iMacs can detect a difference as small as 5-10 ms under realistic conditions, and this dictates which types of research should or should not use these systems.
Cluster Computing for Embedded/Real-Time Systems
Katz, D.; Kepner, J.
1999-01-01
Embedded and real-time systems, like other computing systems, seek to maximize computing power for a given price, and thus can significantly benefit from the advancing capabilities of cluster computing.
Software Accelerates Computing Time for Complex Math
2014-01-01
Ames Research Center awarded Newark, Delaware-based EM Photonics Inc. SBIR funding to utilize graphic processing unit (GPU) technology- traditionally used for computer video games-to develop high-computing software called CULA. The software gives users the ability to run complex algorithms on personal computers with greater speed. As a result of the NASA collaboration, the number of employees at the company has increased 10 percent.
Reversibility and Adiabatic Computation Trading Time and Space for Energy
Li, Maozhen; Li, Ming; Vitanyi, Paul
1996-01-01
Future miniaturization and mobilization of computing devices requires energy parsimonious `adiabatic' computation. This is contingent on logical reversibility of computation. An example is the idea of quantum computations which are reversible except for the irreversible observation steps. We propose to study quantitatively the exchange of computational resources like time and space for irreversibility in computations. Reversible simulations of irreversible computations are memory intensive. Such (polynomial time) simulations are analysed here in terms of `reversible' pebble games. We show that Bennett's pebbling strategy uses least additional space for the greatest number of simulated steps. We derive a trade-off for storage space versus irreversible erasure. Next we consider reversible computation itself. An alternative proof is provided for the precise expression of the ultimate irreversibility cost of an otherwise reversible computation without restrictions on time and space use. A time-irreversibility tra...
Hwang, Yaw-Huei; Chen, Yen-Ting; Yeh, Jao-Yu; Liang, Huey-Wen
2010-10-01
This study aimed to examine the effects of passive and non-computer work time on the estimation of computer use times by electronic activity monitoring. A total of 20 subjects with computers were monitored for 3 h. Average relative error for total computer use time estimation was about 4%, given that non-computer work time was 20% of the 3-h monitored period. No significant impact of passive computer use time was found in this study. Non-computer work time of 40% or less is suggested as criteria for the application of electronic activity monitoring to ensure reliability in the physical work loading assessment. Statement of Relevance: This research studied the criteria of non-computer work time for the appropriate use of electronic activity monitoring to ensure reliability in the assessment of physical work loading. It is suggested that it should be set to 40% or less of the 3-h monitoring period.
Infinite Time Cellular Automata: A Real Computation Model
Givors, Fabien; Ollinger, Nicolas
2010-01-01
We define a new transfinite time model of computation, infinite time cellular automata. The model is shown to be as powerful than infinite time Turing machines, both on finite and infinite inputs; thus inheriting many of its properties. We then show how to simulate the canonical real computation model, BSS machines, with infinite time cellular automata in exactly \\omega steps.
Real-time data-intensive computing
Energy Technology Data Exchange (ETDEWEB)
Parkinson, Dilworth Y., E-mail: dyparkinson@lbl.gov; Chen, Xian; Hexemer, Alexander; MacDowell, Alastair A.; Padmore, Howard A.; Shapiro, David; Tamura, Nobumichi [Advanced Light Source, Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Beattie, Keith; Krishnan, Harinarayan; Patton, Simon J.; Perciano, Talita; Stromsness, Rune; Tull, Craig E.; Ushizima, Daniela [Computational Research Division, Lawrence Berkeley National Laboratory Berkeley CA 94720 (United States); Correa, Joaquin; Deslippe, Jack R. [National Energy Research Scientific Computing Center, Berkeley, CA 94720 (United States); Dart, Eli; Tierney, Brian L. [Energy Sciences Network, Berkeley, CA 94720 (United States); Daurer, Benedikt J.; Maia, Filipe R. N. C. [Uppsala University, Uppsala (Sweden); and others
2016-07-27
Today users visit synchrotrons as sources of understanding and discovery—not as sources of just light, and not as sources of data. To achieve this, the synchrotron facilities frequently provide not just light but often the entire end station and increasingly, advanced computational facilities that can reduce terabytes of data into a form that can reveal a new key insight. The Advanced Light Source (ALS) has partnered with high performance computing, fast networking, and applied mathematics groups to create a “super-facility”, giving users simultaneous access to the experimental, computational, and algorithmic resources to make this possible. This combination forms an efficient closed loop, where data—despite its high rate and volume—is transferred and processed immediately and automatically on appropriate computing resources, and results are extracted, visualized, and presented to users or to the experimental control system, both to provide immediate insight and to guide decisions about subsequent experiments during beamtime. We will describe our work at the ALS ptychography, scattering, micro-diffraction, and micro-tomography beamlines.
Computer Aided Continuous Time Stochastic Process Modelling
DEFF Research Database (Denmark)
Kristensen, N.R.; Madsen, Henrik; Jørgensen, Sten Bay
2001-01-01
A grey-box approach to process modelling that combines deterministic and stochastic modelling is advocated for identification of models for model-based control of batch and semi-batch processes. A computer-aided tool designed for supporting decision-making within the corresponding modelling cycle...
Decoherence and Zeno time in quantum computations
Antoniou, I; Pronko, G; Yarevsky, E
2003-01-01
We analyze the short-time behaviour of the survival probability in the frame of the Friedrichs model for different form factors. We have shown that the probability may not be quadratic for the short times while the quantum Zeno effect (QZE) still exists in this case. We have found that the time when the QZE could be observed is much smaller than usually assumed. We have studied the anti-Zeno era and have estimated its duration. Related decoherence processes are also discussed.
Influence of computer work under time pressure on cardiac activity.
Shi, Ping; Hu, Sijung; Yu, Hongliu
2015-03-01
Computer users are often under stress when required to complete computer work within a required time. Work stress has repeatedly been associated with an increased risk for cardiovascular disease. The present study examined the effects of time pressure workload during computer tasks on cardiac activity in 20 healthy subjects. Heart rate, time domain and frequency domain indices of heart rate variability (HRV) and Poincaré plot parameters were compared among five computer tasks and two rest periods. Faster heart rate and decreased standard deviation of R-R interval were noted in response to computer tasks under time pressure. The Poincaré plot parameters showed significant differences between different levels of time pressure workload during computer tasks, and between computer tasks and the rest periods. In contrast, no significant differences were identified for the frequency domain indices of HRV. The results suggest that the quantitative Poincaré plot analysis used in this study was able to reveal the intrinsic nonlinear nature of the autonomically regulated cardiac rhythm. Specifically, heightened vagal tone occurred during the relaxation computer tasks without time pressure. In contrast, the stressful computer tasks with added time pressure stimulated cardiac sympathetic activity. Copyright © 2015 Elsevier Ltd. All rights reserved.
12 CFR 308.526 - Computation of time.
2010-01-01
... 12 Banks and Banking 4 2010-01-01 2010-01-01 false Computation of time. 308.526 Section 308.526 Banks and Banking FEDERAL DEPOSIT INSURANCE CORPORATION PROCEDURE AND RULES OF PRACTICE RULES OF PRACTICE AND PROCEDURE Program Fraud Civil Remedies and Procedures § 308.526 Computation of time. (a) In...
Time computations in anuran auditory systems
Directory of Open Access Journals (Sweden)
Gary J Rose
2014-05-01
Full Text Available Temporal computations are important in the acoustic communication of anurans. In many cases, calls between closely related species are nearly identical spectrally but differ markedly in temporal structure. Depending on the species, calls can differ in pulse duration, shape and/or rate (i.e., amplitude modulation, direction and rate of frequency modulation, and overall call duration. Also, behavioral studies have shown that anurans are able to discriminate between calls that differ in temporal structure. In the peripheral auditory system, temporal information is coded primarily in the spatiotemporal patterns of activity of auditory-nerve fibers. However, major transformations in the representation of temporal information occur in the central auditory system. In this review I summarize recent advances in understanding how temporal information is represented in the anuran midbrain, with particular emphasis on mechanisms that underlie selectivity for pulse duration and pulse rate (i.e., intervals between onsets of successive pulses. Two types of neurons have been identified that show selectivity for pulse rate: long-interval cells respond well to slow pulse rates but fail to spike or respond phasically to fast pulse rates; conversely, interval-counting neurons respond to intermediate or fast pulse rates, but only after a threshold number of pulses, presented at optimal intervals, have occurred. Duration-selectivity is manifest as short-pass, band-pass or long-pass tuning. Whole-cell patch recordings, in vivo, suggest that excitation and inhibition are integrated in diverse ways to generate temporal selectivity. In many cases, activity-related enhancement or depression of excitatory or inhibitory processes appear to contribute to selective responses.
Computing the Matched Filter in Linear Time
Fish, Alexander; Hadani, Ronny; Sayeed, Akbar; Schwartz, Oded
2011-01-01
A fundamental problem in wireless communication is the time-frequency shift (TFS) problem: Find the time-frequency shift of a signal in a noisy environment. The shift is the result of time asynchronization of a sender with a receiver, and of non-zero speed of a sender with respect to a receiver. A classical solution of a discrete analog of the TFS problem is called the matched filter algorithm. It uses a pseudo-random waveform S(t) of the length p, and its arithmetic complexity is O(p^{2} \\cdot log (p)), using fast Fourier transform. In these notes we introduce a novel approach of designing new waveforms that allow faster matched filter algorithm. We use techniques from group representation theory to design waveforms S(t), which enable us to introduce two fast matched filter (FMF) algorithms, called the flag algorithm, and the cross algorithm. These methods solve the TFS problem in O(p\\cdot log (p)) operations. We discuss applications of the algorithms to mobile communication, GPS, and radar.
Recent achievements in real-time computational seismology in Taiwan
Lee, S.; Liang, W.; Huang, B.
2012-12-01
Real-time computational seismology is currently possible to be achieved which needs highly connection between seismic database and high performance computing. We have developed a real-time moment tensor monitoring system (RMT) by using continuous BATS records and moment tensor inversion (CMT) technique. The real-time online earthquake simulation service is also ready to open for researchers and public earthquake science education (ROS). Combine RMT with ROS, the earthquake report based on computational seismology can provide within 5 minutes after an earthquake occurred (RMT obtains point source information < 120 sec; ROS completes a 3D simulation < 3 minutes). All of these computational results are posted on the internet in real-time now. For more information, welcome to visit real-time computational seismology earthquake report webpage (RCS).
Time-Domain Computation Of Electromagnetic Fields In MMICs
Lansing, Faiza S.; Rascoe, Daniel L.
1995-01-01
Maxwell's equations solved on three-dimensional, conformed orthogonal grids by finite-difference techniques. Method of computing frequency-dependent electrical parameters of monolithic microwave integrated circuit (MMIC) involves time-domain computation of propagation of electromagnetic field in response to excitation by single pulse at input terminal, followed by computation of Fourier transforms to obtain frequency-domain response from time-domain response. Parameters computed include electric and magnetic fields, voltages, currents, impedances, scattering parameters, and effective dielectric constants. Powerful and efficient means for analyzing performance of even complicated MMIC.
Noise-constrained switching times for heteroclinic computing
Neves, Fabio Schittler; Voit, Maximilian; Timme, Marc
2017-03-01
Heteroclinic computing offers a novel paradigm for universal computation by collective system dynamics. In such a paradigm, input signals are encoded as complex periodic orbits approaching specific sequences of saddle states. Without inputs, the relevant states together with the heteroclinic connections between them form a network of states—the heteroclinic network. Systems of pulse-coupled oscillators or spiking neurons naturally exhibit such heteroclinic networks of saddles, thereby providing a substrate for general analog computations. Several challenges need to be resolved before it becomes possible to effectively realize heteroclinic computing in hardware. The time scales on which computations are performed crucially depend on the switching times between saddles, which in turn are jointly controlled by the system's intrinsic dynamics and the level of external and measurement noise. The nonlinear dynamics of pulse-coupled systems often strongly deviate from that of time-continuously coupled (e.g., phase-coupled) systems. The factors impacting switching times in pulse-coupled systems are still not well understood. Here we systematically investigate switching times in dependence of the levels of noise and intrinsic dissipation in the system. We specifically reveal how local responses to pulses coact with external noise. Our findings confirm that, like in time-continuous phase-coupled systems, piecewise-continuous pulse-coupled systems exhibit switching times that transiently increase exponentially with the number of switches up to some order of magnitude set by the noise level. Complementarily, we show that switching times may constitute a good predictor for the computation reliability, indicating how often an input signal must be reiterated. By characterizing switching times between two saddles in conjunction with the reliability of a computation, our results provide a first step beyond the coding of input signal identities toward a complementary coding for
Time Domain Terahertz Axial Computed Tomography Non Destructive Evaluation Project
National Aeronautics and Space Administration — We propose to demonstrate key elements of feasibility for a high speed automated time domain terahertz computed axial tomography (TD-THz CT) non destructive...
Time Domain Terahertz Axial Computed Tomography Non Destructive Evaluation Project
National Aeronautics and Space Administration — In this Phase 2 project, we propose to develop, construct, and deliver to NASA a computed axial tomography time-domain terahertz (CT TD-THz) non destructive...
29 CFR 1603.106 - Computation of time.
2010-07-01
... EXEMPT STATE AND LOCAL GOVERNMENT EMPLOYEE COMPLAINTS OF EMPLOYMENT DISCRIMINATION UNDER SECTION 304 OF THE GOVERNMENT EMPLOYEE RIGHTS ACT OF 1991 Administrative Process § 1603.106 Computation of time....
GPU and APU computations of Finite Time Lyapunov Exponent fields
Conti, Christian; Rossinelli, Diego; Koumoutsakos, Petros
2012-03-01
We present GPU and APU accelerated computations of Finite-Time Lyapunov Exponent (FTLE) fields. The calculation of FTLEs is a computationally intensive process, as in order to obtain the sharp ridges associated with the Lagrangian Coherent Structures an extensive resampling of the flow field is required. The computational performance of this resampling is limited by the memory bandwidth of the underlying computer architecture. The present technique harnesses data-parallel execution of many-core architectures and relies on fast and accurate evaluations of moment conserving functions for the mesh to particle interpolations. We demonstrate how the computation of FTLEs can be efficiently performed on a GPU and on an APU through OpenCL and we report over one order of magnitude improvements over multi-threaded executions in FTLE computations of bluff body flows.
CPU timing routines for a CONVEX C220 computer system
Bynum, Mary Ann
1989-01-01
The timing routines available on the CONVEX C220 computer system in the Structural Mechanics Division (SMD) at NASA Langley Research Center are examined. The function of the timing routines, the use of the timing routines in sequential, parallel, and vector code, and the interpretation of the results from the timing routines with respect to the CONVEX model of computing are described. The timing routines available on the SMD CONVEX fall into two groups. The first group includes standard timing routines generally available with UNIX 4.3 BSD operating systems, while the second group includes routines unique to the SMD CONVEX. The standard timing routines described in this report are /bin/csh time,/bin/time, etime, and ctime. The routines unique to the SMD CONVEX are getinfo, second, cputime, toc, and a parallel profiling package made up of palprof, palinit, and palsum.
An approximate 3D computational method for real-time computation of induction logging responses
Bensdorp, S.; Petersen, S.A.; Van den Berg, P.M.; Fokkema, J.T.
2014-01-01
Over many years, induction logging systems have been used to create well formation logs. The major drawback for the utilization of these tools is the long simulation time for a single forward computation. We proposed an efficient computational method based on a contrast-type of integral-equation for
39 CFR 3001.15 - Computation of time.
2010-07-01
... 39 Postal Service 1 2010-07-01 2010-07-01 false Computation of time. 3001.15 Section 3001.15 Postal Service POSTAL REGULATORY COMMISSION PERSONNEL RULES OF PRACTICE AND PROCEDURE Rules of General... time prescribed or allowed by this part, or by any notice, order, rule or regulation of the Commission...
The computational power of time dilation in special relativity
Biamonte, Jacob
2014-03-01
The Lorentzian length of a timelike curve connecting both endpoints of a classical computation is a function of the path taken through Minkowski spacetime. The associated runtime difference is due to time-dilation: the phenomenon whereby an observer finds that another's physically identical ideal clock has ticked at a different rate than their own clock. Using ideas appearing in the framework of computational complexity theory, time-dilation is quantified as an algorithmic resource by relating relativistic energy to an nth order polynomial time reduction at the completion of an observer's journey. These results enable a comparison between the optimal quadratic Grover speedup from quantum computing and an n=2 speedup using classical computers and relativistic effects. The goal is not to propose a practical model of computation, but to probe the ultimate limits physics places on computation. Parts of this talk are based on [J.Phys.Conf.Ser. 229:012020 (2010), arXiv:0907.1579]. Support is acknowledged from the Foundational Questions Institute (FQXi) and the Compagnia di San Paolo Foundation.
The design of a scalable, fixed-time computer benchmark
Energy Technology Data Exchange (ETDEWEB)
Gustafson, J.; Rover, D.; Elbert, S.; Carter, M.
1990-10-01
By using the principle of fixed time benchmarking, it is possible to compare a very wide range of computers, from a small personal computer to the most powerful parallel supercomputer, an a single scale. Fixed-time benchmarks promise far greater longevity than those based on a particular problem size, and are more appropriate for grand challenge'' capability comparison. We present the design of a benchmark, SLALOM{trademark}, that scales automatically to the computing power available, and corrects several deficiencies in various existing benchmarks: it is highly scalable, it solves a real problem, it includes input and output times, and it can be run on parallel machines of all kinds, using any convenient language. The benchmark provides a reasonable estimate of the size of problem solvable on scientific computers. Results are presented that span six orders of magnitude for contemporary computers of various architectures. The benchmarks also can be used to demonstrate a new source of superlinear speedup in parallel computers. 15 refs., 14 figs., 3 tabs.
Lightweight distributed computing for intraoperative real-time image guidance
Suwelack, Stefan; Katic, Darko; Wagner, Simon; Spengler, Patrick; Bodenstedt, Sebastian; Röhl, Sebastian; Dillmann, Rüdiger; Speidel, Stefanie
2012-02-01
In order to provide real-time intraoperative guidance, computer assisted surgery (CAS) systems often rely on computationally expensive algorithms. The real-time constraint is especially challenging if several components such as intraoperative image processing, soft tissue registration or context aware visualization are combined in a single system. In this paper, we present a lightweight approach to distribute the workload over several workstations based on the OpenIGTLink protocol. We use XML-based message passing for remote procedure calls and native types for transferring data such as images, meshes or point coordinates. Two different, but typical scenarios are considered in order to evaluate the performance of the new system. First, we analyze a real-time soft tissue registration algorithm based on a finite element (FE) model. Here, we use the proposed approach to distribute the computational workload between a primary workstation that handles sensor data processing and visualization and a dedicated workstation that runs the real-time FE algorithm. We show that the additional overhead that is introduced by the technique is small compared to the total execution time. Furthermore, the approach is used to speed up a context aware augmented reality based navigation system for dental implant surgery. In this scenario, the additional delay for running the computationally expensive reasoning server on a separate workstation is less than a millisecond. The results show that the presented approach is a promising strategy to speed up real-time CAS systems.
A new computational structure for real-time dynamics
Energy Technology Data Exchange (ETDEWEB)
Izaguirre, A. (New Jersey Inst. of Tech., Newark (United States)); Hashimoto, Minoru (Univ. of Electrocommunications, Tokyo (Japan))
1992-08-01
The authors present an efficient structure for the computation of robot dynamics in real time. The fundamental characteristic of this structure is the division of the computation into a high-priority synchronous task and low-priority background tasks, possibly sharing the resources of a conventional computing unit based on commercial microprocessors. The background tasks compute the inertial and gravitational coefficients as well as the forces due to the velocities of the joints. In each control sample period, the high-priority synchronous task computes the product of the inertial coefficients by the accelerations of the joints and performs the summation of the torques due to the velocities and gravitational forces. Kircanski et al. (1986) have shown that the bandwidth of the variation of joint angles and of their velocities is an order of magnitude less than the variation of joint accelerations. This result agrees with the experiments the authors have carried out using a PUMA 260 robot. Two main strategies contribute to reduce the computational burden associated with the evaluation of the dynamic equations. The first involves the use of efficient algorithms for the evaluation of the equations. The second is aimed at reducing the number of dynamic parameters by identifying beforehand the linear dependencies among these parameters, as well as carrying out a significance analysis of the parameters' contribution to the final joint torques. The actual code used to evaluate this dynamic model is entirely computer generated from experimental data, requiring no other manual intervention than performing a campaign of measurements.
Pattern recognition computation using action potential timing for stimulus representation
Hopfield, J. J.
1995-07-01
A computational model is described in which the sizes of variables are represented by the explicit times at which action potentials occur, rather than by the more usual 'firing rate' of neurons. The comparison of patterns over sets of analogue variables is done by a network using different delays for different information paths. This mode of computation explains how one scheme of neuroarchitecture can be used for very different sensory modalities and seemingly different computations. The oscillations and anatomy of the mammalian olfactory systems have a simple interpretation in terms of this representation, and relate to processing in the auditory system. Single-electrode recording would not detect such neural computing. Recognition 'units' in this style respond more like radial basis function units than elementary sigmoid units.
Fast computation of recurrences in long time series
Rawald, Tobias; Sips, Mike; Marwan, Norbert; Dransch, Doris
2014-05-01
The quadratic time complexity of calculating basic RQA measures, doubling the size of the input time series leads to a quadrupling in operations, impairs the fast computation of RQA in many application scenarios. As an example, we analyze the Potsdamer Reihe, an ongoing non-interrupted hourly temperature profile since 1893, consisting of 1,043,112 data points. Using an optimized single-threaded CPU implementation this analysis requires about six hours. Our approach conducts RQA for the Potsdamer Reihe in five minutes. We automatically split a long time series into smaller chunks (Divide) and distribute the computation of RQA measures across multiple GPU devices. To guarantee valid RQA results, we employ carryover buffers that allow sharing information between pairs of chunks (Recombine). We demonstrate the capabilities of our Divide and Recombine approach to process long time series by comparing the runtime of our implementation to existing RQA tools. We support a variety of platforms by employing the computing framework OpenCL. Our current implementation supports the computation of standard RQA measures (recurrence rate, determinism, laminarity, ratio, average diagonal line length, trapping time, longest diagonal line, longest vertical line, divergence, entropy, trend) and also calculates recurrence times. To utilize the potential of our approach for a number of applications, we plan to release our implementation under an Open Source software license. It will be available at http://www.gfz-potsdam.de/fast-rqa/. Since our approach allows to compute RQA measures for a long time series fast, we plan to extend our implementation to support multi-scale RQA.
GPS computer navigators to shorten EMS response and transport times.
Ota, F S; Muramatsu, R S; Yoshida, B H; Yamamoto, L G
2001-05-01
GPS (global positioning satellite system to determine one's position on earth) units have become inexpensive and compact. The purpose of this study is to assess the effectiveness of a GPS enhanced computer street map navigator to improve the ability of EMS drivers in an urban setting to locate their destination and shorten response times. For part I, residential addresses in the city were randomly selected from a telephone directory. Two driver/navigator teams were assigned to drive to the address adhering to speed limits. One team used a standard street map, whereas the other team used a GPS computer navigator. The travel time and distance of the runs were compared. For part II, the computer GPS navigator was placed on an ambulance to supplement their normal methods of navigation to find the address requesting EMS. After the run was completed, EMS providers were interviewed to determine their opinion of whether the GPS navigator was helpful. For part I the results showed that in the 29 initial test runs, comparing the GPS team versus the standard map team, the mean distances traveled were 8.7 versus 9.0 kilometers (not significant) and the mean travel times were 13.5 versus 14.6 minutes (P=.02), respectively. The GPS team arrived faster in 72% runs. For part II the results showed that most EMS providers surveyed noted that the GPS computer navigator enhanced their ability to find the destination and all EMS providers acknowledged that it would enhance their ability to find a destination in an area in which they were unfamiliar. These results suggest that a portable GPS computer navigator system is helpful and can enhance the ability of prehospital care providers to locate their destination. Because these units are accurate and inexpensive, GPS computer navigators may be a valuable tool in reducing pre-hospital transport times.
Instructional Advice, Time Advice and Learning Questions in Computer Simulations
Rey, Gunter Daniel
2010-01-01
Undergraduate students (N = 97) used an introductory text and a computer simulation to learn fundamental concepts about statistical analyses (e.g., analysis of variance, regression analysis and General Linear Model). Each learner was randomly assigned to one cell of a 2 (with or without instructional advice) x 2 (with or without time advice) x 2…
Time Advice and Learning Questions in Computer Simulations
Rey, Gunter Daniel
2011-01-01
Students (N = 101) used an introductory text and a computer simulation to learn fundamental concepts about statistical analyses (e.g., analysis of variance, regression analysis and General Linear Model). Each learner was randomly assigned to one cell of a 2 (with or without time advice) x 3 (with learning questions and corrective feedback, with…
41 CFR 105-70.027 - Computation of time.
2010-07-01
... 41 Public Contracts and Property Management 3 2010-07-01 2010-07-01 false Computation of time. 105-70.027 Section 105-70.027 Public Contracts and Property Management Federal Property Management Regulations System (Continued) GENERAL SERVICES ADMINISTRATION Regional Offices-General...
Computing Finite-Time Lyapunov Exponents with Optimally Time Dependent Reduction
Babaee, Hessam; Farazmand, Mohammad; Sapsis, Themis; Haller, George
2016-11-01
We present a method to compute Finite-Time Lyapunov Exponents (FTLE) of a dynamical system using Optimally Time-Dependent (OTD) reduction recently introduced by H. Babaee and T. P. Sapsis. The OTD modes are a set of finite-dimensional, time-dependent, orthonormal basis {ui (x , t) } |i=1N that capture the directions associated with transient instabilities. The evolution equation of the OTD modes is derived from a minimization principle that optimally approximates the most unstable directions over finite times. To compute the FTLE, we evolve a single OTD mode along with the nonlinear dynamics. We approximate the FTLE from the reduced system obtained from projecting the instantaneous linearized dynamics onto the OTD mode. This results in a significant reduction in the computational cost compared to conventional methods for computing FTLE. We demonstrate the efficiency of our method for double Gyre and ABC flows. ARO project 66710-EG-YIP.
Computational Complexity of Decoding Orthogonal Space-Time Block Codes
Ayanoglu, Ender; Karipidis, Eleftherios
2009-01-01
The computational complexity of optimum decoding for an orthogonal space-time block code G satisfying the orthogonality property that the Hermitian transpose of G multiplied by G is equal to a constant c times the sum of the squared symbols of the code times an identity matrix, where c is a positive integer is quantified. Four equivalent techniques of optimum decoding which have the same computational complexity are specified. Modifications to the basic formulation in special cases are calculated and illustrated by means of examples. This paper corrects and extends [1],[2], and unifies them with the results from the literature. In addition, a number of results from the literature are extended to the case c > 1.
Efficient quantum algorithm for computing n-time correlation functions.
Pedernales, J S; Di Candia, R; Egusquiza, I L; Casanova, J; Solano, E
2014-07-11
We propose a method for computing n-time correlation functions of arbitrary spinorial, fermionic, and bosonic operators, consisting of an efficient quantum algorithm that encodes these correlations in an initially added ancillary qubit for probe and control tasks. For spinorial and fermionic systems, the reconstruction of arbitrary n-time correlation functions requires the measurement of two ancilla observables, while for bosonic variables time derivatives of the same observables are needed. Finally, we provide examples applicable to different quantum platforms in the frame of the linear response theory.
Soft Real-Time PID Control on a VME Computer
Karayan, Vahag; Sander, Stanley; Cageao, Richard
2007-01-01
microPID (uPID) is a computer program for real-time proportional + integral + derivative (PID) control of a translation stage in a Fourier-transform ultraviolet spectrometer. microPID implements a PID control loop over a position profile at sampling rate of 8 kHz (sampling period 125microseconds). The software runs in a strippeddown Linux operating system on a VersaModule Eurocard (VME) computer operating in real-time priority queue using an embedded controller, a 16-bit digital-to-analog converter (D/A) board, and a laser-positioning board (LPB). microPID consists of three main parts: (1) VME device-driver routines, (2) software that administers a custom protocol for serial communication with a control computer, and (3) a loop section that obtains the current position from an LPB-driver routine, calculates the ideal position from the profile, and calculates a new voltage command by use of an embedded PID routine all within each sampling period. The voltage command is sent to the D/A board to control the stage. microPID uses special kernel headers to obtain microsecond timing resolution. Inasmuch as microPID implements a single-threaded process and all other processes are disabled, the Linux operating system acts as a soft real-time system.
Computational intelligence in time series forecasting theory and engineering applications
Palit, Ajoy K
2005-01-01
Foresight in an engineering enterprise can make the difference between success and failure, and can be vital to the effective control of industrial systems. Applying time series analysis in the on-line milieu of most industrial plants has been problematic owing to the time and computational effort required. The advent of soft computing tools offers a solution. The authors harness the power of intelligent technologies individually and in combination. Examples of the particular systems and processes susceptible to each technique are investigated, cultivating a comprehensive exposition of the improvements on offer in quality, model building and predictive control and the selection of appropriate tools from the plethora available. Application-oriented engineers in process control, manufacturing, production industry and research centres will find much to interest them in this book. It is suitable for industrial training purposes, as well as serving as valuable reference material for experimental researchers.
Lily Pad: Towards Real-time Interactive Computational Fluid Dynamics
Weymouth, Gabriel D
2015-01-01
Despite the fact that computational fluid dynamics (CFD) software is now (relatively) fast and freely available, it is still amazingly difficult to use. Inaccessible software imposes a significant entry barrier on students and junior engineers, and even senior researchers spend less time developing insights and more time on software issues. Lily Pad was developed as an initial attempt to address some of these problems. The goal of Lily Pad is to lower the barrier to CFD by adopting simple high-speed methods, utilising modern programming features and environments, and giving immediate visual feed-back to the user. The resulting software focuses on the fluid dynamics instead of the computation, making it useful for both education and research. LilyPad is open source and available online at https://github.com/weymouth/lily-pad for all use under the MIT license.
Real-Time Visualization System for Computational Offloading
2015-01-01
dependencies are hard- coded into the visualization system. The remainder of this report is organized as follows. In Section 2, we discuss the...timer. Such a driver has access to all the visualization functionality present in the visualization pane. In Fig. 8, we show a code snippet required...Real-Time Visualization System for Computational Offloading by Bryan Dawson and David L Doria ARL-TN-0655 January 2015
Spike-timing-based computation in sound localization.
Directory of Open Access Journals (Sweden)
Dan F M Goodman
2010-11-01
Full Text Available Spike timing is precise in the auditory system and it has been argued that it conveys information about auditory stimuli, in particular about the location of a sound source. However, beyond simple time differences, the way in which neurons might extract this information is unclear and the potential computational advantages are unknown. The computational difficulty of this task for an animal is to locate the source of an unexpected sound from two monaural signals that are highly dependent on the unknown source signal. In neuron models consisting of spectro-temporal filtering and spiking nonlinearity, we found that the binaural structure induced by spatialized sounds is mapped to synchrony patterns that depend on source location rather than on source signal. Location-specific synchrony patterns would then result in the activation of location-specific assemblies of postsynaptic neurons. We designed a spiking neuron model which exploited this principle to locate a variety of sound sources in a virtual acoustic environment using measured human head-related transfer functions. The model was able to accurately estimate the location of previously unknown sounds in both azimuth and elevation (including front/back discrimination in a known acoustic environment. We found that multiple representations of different acoustic environments could coexist as sets of overlapping neural assemblies which could be associated with spatial locations by Hebbian learning. The model demonstrates the computational relevance of relative spike timing to extract spatial information about sources independently of the source signal.
Modern EMC analysis I time-domain computational schemes
Kantartzis, Nikolaos V
2008-01-01
The objective of this two-volume book is the systematic and comprehensive description of the most competitive time-domain computational methods for the efficient modeling and accurate solution of contemporary real-world EMC problems. Intended to be self-contained, it performs a detailed presentation of all well-known algorithms, elucidating on their merits or weaknesses, and accompanies the theoretical content with a variety of applications. Outlining the present volume, the analysis covers the theory of the finite-difference time-domain, the transmission-line matrix/modeling, and the finite i
A heterogeneous hierarchical architecture for real-time computing
Energy Technology Data Exchange (ETDEWEB)
Skroch, D.A.; Fornaro, R.J.
1988-12-01
The need for high-speed data acquisition and control algorithms has prompted continued research in the area of multiprocessor systems and related programming techniques. The result presented here is a unique hardware and software architecture for high-speed real-time computer systems. The implementation of a prototype of this architecture has required the integration of architecture, operating systems and programming languages into a cohesive unit. This report describes a Heterogeneous Hierarchial Architecture for Real-Time (H{sup 2} ART) and system software for program loading and interprocessor communication.
Grigoryeva, Lyudmila; Henriques, Julie; Larger, Laurent; Ortega, Juan-Pablo
2014-07-01
Reservoir computing is a recently introduced machine learning paradigm that has already shown excellent performances in the processing of empirical data. We study a particular kind of reservoir computers called time-delay reservoirs that are constructed out of the sampling of the solution of a time-delay differential equation and show their good performance in the forecasting of the conditional covariances associated to multivariate discrete-time nonlinear stochastic processes of VEC-GARCH type as well as in the prediction of factual daily market realized volatilities computed with intraday quotes, using as training input daily log-return series of moderate size. We tackle some problems associated to the lack of task-universality for individually operating reservoirs and propose a solution based on the use of parallel arrays of time-delay reservoirs. Copyright © 2014 Elsevier Ltd. All rights reserved.
Baier, Christel; Hermanns, H.; Katoen, Joost P.; Haverkort, Boudewijn R.H.M.
2005-01-01
A continuous-time Markov decision process (CTMDP) is a generalization of a continuous-time Markov chain in which both probabilistic and nondeterministic choices co-exist. This paper presents an efficient algorithm to compute the maximum (or minimum) probability to reach a set of goal states within a
Heterogeneous computing for a real-time pig monitoring system
Choi, Younchang; Kim, Jinseong; Kim, Jaehak; Chung, Yeonwoo; Chung, Yongwha; Park, Daihee; Kim, Hakjae
2017-06-01
Video sensor data has been widely used in automatic surveillance applications. In this study, we present a method that automatically detects pigs in a pig room by using depth information obtained from a Kinect sensor. For a real-time implementation, we propose a means of reducing the execution time by applying parallel processing techniques. In general, most parallel processing techniques have been used to parallelize a specific task. In this study, we consider parallelization of an entire system that consists of several tasks. By applying a scheduling strategy to identify a computing device for each task and implementing it with OpenCL, we can reduce the total execution time efficiently. Experimental results reveal that the proposed method can automatically detect pigs using a CPU-GPU hybrid system in real time, regardless of the relative performance between the CPU and GPU.
A note on computing average state occupation times
Directory of Open Access Journals (Sweden)
Jan Beyersmann
2014-05-01
Full Text Available Objective: This review discusses how biometricians would probably compute or estimate expected waiting times, if they had the data. Methods: Our framework is a time-inhomogeneous Markov multistate model, where all transition hazards are allowed to be time-varying. We assume that the cumulative transition hazards are given. That is, they are either known, as in a simulation, determined by expert guesses, or obtained via some method of statistical estimation. Our basic tool is product integration, which transforms the transition hazards into the matrix of transition probabilities. Product integration enjoys a rich mathematical theory, which has successfully been used to study probabilistic and statistical aspects of multistate models. Our emphasis will be on practical implementation of product integration, which allows us to numerically approximate the transition probabilities. Average state occupation times and other quantities of interest may then be derived from the transition probabilities.
Computer network time synchronization the network time protocol on earth and in space
Mills, David L
2010-01-01
Carefully coordinated, reliable, and accurate time synchronization is vital to a wide spectrum of fields-from air and ground traffic control, to buying and selling goods and services, to TV network programming. Ill-gotten time could even lead to the unimaginable and cause DNS caches to expire, leaving the entire Internet to implode on the root servers.Written by the original developer of the Network Time Protocol (NTP), Computer Network Time Synchronization: The Network Time Protocol on Earth and in Space, Second Edition addresses the technological infrastructure of time dissemination, distrib
Maass, Wolfgang; Natschläger, Thomas; Markram, Henry
2002-11-01
A key challenge for neural modeling is to explain how a continuous stream of multimodal input from a rapidly changing environment can be processed by stereotypical recurrent circuits of integrate-and-fire neurons in real time. We propose a new computational model for real-time computing on time-varying input that provides an alternative to paradigms based on Turing machines or attractor neural networks. It does not require a task-dependent construction of neural circuits. Instead, it is based on principles of high-dimensional dynamical systems in combination with statistical learning theory and can be implemented on generic evolved or found recurrent circuitry. It is shown that the inherent transient dynamics of the high-dimensional dynamical system formed by a sufficiently large and heterogeneous neural circuit may serve as universal analog fading memory. Readout neurons can learn to extract in real time from the current state of such recurrent neural circuit information about current and past inputs that may be needed for diverse tasks. Stable internal states are not required for giving a stable output, since transient internal states can be transformed by readout neurons into stable target outputs due to the high dimensionality of the dynamical system. Our approach is based on a rigorous computational model, the liquid state machine, that, unlike Turing machines, does not require sequential transitions between well-defined discrete internal states. It is supported, as the Turing machine is, by rigorous mathematical results that predict universal computational power under idealized conditions, but for the biologically more realistic scenario of real-time processing of time-varying inputs. Our approach provides new perspectives for the interpretation of neural coding, the design of experiments and data analysis in neurophysiology, and the solution of problems in robotics and neurotechnology.
Computational electrodynamics the finite-difference time-domain method
Taflove, Allen
2005-01-01
This extensively revised and expanded third edition of the Artech House bestseller, Computational Electrodynamics: The Finite-Difference Time-Domain Method, offers engineers the most up-to-date and definitive resource on this critical method for solving Maxwell's equations. The method helps practitioners design antennas, wireless communications devices, high-speed digital and microwave circuits, and integrated optical devices with unsurpassed efficiency. There has been considerable advancement in FDTD computational technology over the past few years, and the third edition brings professionals the very latest details with entirely new chapters on important techniques, major updates on key topics, and new discussions on emerging areas such as nanophotonics. What's more, to supplement the third edition, the authors have created a Web site with solutions to problems, downloadable graphics and videos, and updates, making this new edition the ideal textbook on the subject as well.
In this issue: Time to replace doctors’ judgement with computers
Directory of Open Access Journals (Sweden)
Simon de Lusignan
2015-11-01
Full Text Available Informaticians continue to rise to the challenge, set by the English Health Minister, of trying to replace doctors’ judgement with computers. This issue describes successes and where there are barriers. However, whilst there is progress this tends to be incremental and there are grand challenges to be overcome before computers can replace clinician. These grand challenges include: (1 improving usability so it is possible to more readily incorporate technology into clinical workflow; (2 rigorous new analytic methods that make use of the mass of available data, ‘Big data’, to create real-world evidence; (3 faster ways of meeting regulatory and legal requirements including ensuring privacy; (4 provision of reimbursement models to fund innovative technology that can substitute for clinical time and (5 recognition that innovations that improve quality also often increase cost. Informatics more is likely to support and augment clinical decision making rather than replace clinicians.
Computation Offloading for Frame-Based Real-Time Tasks under Given Server Response Time Guarantees
Directory of Open Access Journals (Sweden)
Anas S. M. Toma
2014-11-01
Full Text Available Computation offloading has been adopted to improve the performance of embedded systems by offloading the computation of some tasks, especially computation-intensive tasks, to servers or clouds. This paper explores computation offloading for real-time tasks in embedded systems, provided given response time guarantees from the servers, to decide which tasks should be offloaded to get the results in time. We consider frame-based real-time tasks with the same period and relative deadline. When the execution order of the tasks is given, the problem can be solved in linear time. However, when the execution order is not specified, we prove that the problem is NP-complete. We develop a pseudo-polynomial-time algorithm for deriving feasible schedules, if they exist. An approximation scheme is also developed to trade the error made from the algorithm and the complexity. Our algorithms are extended to minimize the period/relative deadline of the tasks for performance maximization. The algorithms are evaluated with a case study for a surveillance system and synthesized benchmarks.
ANIMAC: a multiprocessor architecture for real time computer animation
Energy Technology Data Exchange (ETDEWEB)
Whelan, D.S.
1985-01-01
Advances in integrated circuit technology have been largely responsible for the growth of the computer graphic industry. This technology promises additional growth through the remainder of the century. This dissertation addresses how this future technology can be harnessed and used to construct very high performance real-time computer graphics systems. A new architecture is proposed for real-time animation engines. The ANIMAC architecture achieves high performance by utilizing a two-dimensional array of processors that determine visible surfaces in parallel. An array of sixteen processors with only nearest neighbor interprocessor communications can produce real-time shadowed images of scenes containing 100,000 triangles. The ANIMAC architecture is based upon analysis and simulations of various parallelization techniques. These simulations suggest that the viewing space be spatially subdivided and that each processor produce a visible surface image for several viewing space subvolumes. Simple assignments of viewing space subvolumes to processors are shown to offer high parallel efficiencies. Simulations of parallel algorithms were driven with data derived from real scenes since analysis of scene composition suggested that using simplistic models of scene composition might lead to incorrect results.
Energy and time determine scaling in biological and computer designs.
Moses, Melanie; Bezerra, George; Edwards, Benjamin; Brown, James; Forrest, Stephanie
2016-08-19
Metabolic rate in animals and power consumption in computers are analogous quantities that scale similarly with size. We analyse vascular systems of mammals and on-chip networks of microprocessors, where natural selection and human engineering, respectively, have produced systems that minimize both energy dissipation and delivery times. Using a simple network model that simultaneously minimizes energy and time, our analysis explains empirically observed trends in the scaling of metabolic rate in mammals and power consumption and performance in microprocessors across several orders of magnitude in size. Just as the evolutionary transitions from unicellular to multicellular animals in biology are associated with shifts in metabolic scaling, our model suggests that the scaling of power and performance will change as computer designs transition to decentralized multi-core and distributed cyber-physical systems. More generally, a single energy-time minimization principle may govern the design of many complex systems that process energy, materials and information.This article is part of the themed issue 'The major synthetic evolutionary transitions'.
A computer program for the estimation of time of death
DEFF Research Database (Denmark)
Lynnerup, N
1993-01-01
and that the temperature at death is known. Also, Marshall and Hoare's formula expresses the temperature as a function of time, and not vice versa, the latter being the problem most often encountered by forensic scientists. A simple BASIC program that enables solving of Marshall and Hoare's equation for the postmortem...... cooling of bodies is presented. It is proposed that by having a computer program that solves the equation, giving the length of the cooling period in response to a certain rectal temperature, and which allows easy comparison of multiple solutions, the uncertainties related to ambience temperature...
Computing Resolution-Path Dependencies in Linear Time
Slivovsky, Friedrich
2012-01-01
The alternation of existential and universal quantifiers in a quantified boolean formula (QBF) generates dependencies among variables that must be respected when evaluating the formula. Dependency schemes provide a general framework for representing such dependencies. Since it is generally intractable to determine dependencies exactly, a set of potential dependencies is computed instead, which may include false positives. Among the schemes proposed so far, resolution-path dependencies introduce the fewest spurious dependencies. In this work, we describe an algorithm that detects resolution-path dependencies in linear time, resolving a problem posed by Van Gelder (CP 2011)
RighTime: A real time clock correcting program for MS-DOS-based computer systems
Becker, G. Thomas
1993-01-01
A computer program is described which effectively eliminates the misgivings of the DOS system clock in PC/AT-class computers. RighTime is a small, sophisticated memory-resident program that automatically corrects both the DOS system clock and the hardware 'CMOS' real time clock (RTC) in real time. RighTime learns what corrections are required without operator interaction beyond the occasional accurate time set. Both warm (power on) and cool (power off) errors are corrected, usually yielding better than one part per million accuracy in the typical desktop computer with no additional hardware, and RighTime increases the system clock resolution from approximately 0.0549 second to 0.01 second. Program tools are also available which allow visualization of RighTime's actions, verification of its performance, display of its history log, and which provide data for graphing of the system clock behavior. The program has found application in a wide variety of industries, including astronomy, satellite tracking, communications, broadcasting, transportation, public utilities, manufacturing, medicine, and the military.
Time Series Analysis, Modeling and Applications A Computational Intelligence Perspective
Chen, Shyi-Ming
2013-01-01
Temporal and spatiotemporal data form an inherent fabric of the society as we are faced with streams of data coming from numerous sensors, data feeds, recordings associated with numerous areas of application embracing physical and human-generated phenomena (environmental data, financial markets, Internet activities, etc.). A quest for a thorough analysis, interpretation, modeling and prediction of time series comes with an ongoing challenge for developing models that are both accurate and user-friendly (interpretable). The volume is aimed to exploit the conceptual and algorithmic framework of Computational Intelligence (CI) to form a cohesive and comprehensive environment for building models of time series. The contributions covered in the volume are fully reflective of the wealth of the CI technologies by bringing together ideas, algorithms, and numeric studies, which convincingly demonstrate their relevance, maturity and visible usefulness. It reflects upon the truly remarkable diversity of methodological a...
Computing the Tutte polynomial in vertex-exponential time
Björklund, Andreas; Kaski, Petteri; Koivisto, Mikko
2007-01-01
The deletion--contraction algorithm is perhaps the most popular method for computing a host of fundamental graph invariants such as the chromatic, flow, and reliability polynomials in graph theory, the Jones polynomial of an alternating link in knot theory, and the partition functions of the models of Ising, Potts, and Furtuin--Kasteleyn in statistical physics. Prior to this work, deletion--contraction was also the fastest known general-purpose algorithm for these invariants, running in time roughly proportional to the number of spanning trees in the input graph. Here, we provide a substantially faster algorithm that computes the Tutte polynomial--and hence, all the aforementioned invariants and more--of an arbitrary graph in time within a polynomial factor of the number of connected induced subgraphs. The algorithm is based on a new recurrence formula that alternates between partitioning an induced subgraph into components and a subtraction step to solve the connected case. For bounded-degree graphs on $n$ v...
Bielski, Conrad; Lemoine, Guido; Syryczynski, Jacek
2009-09-01
High Performance Computing (HPC) hardware solutions such as grid computing and General Processing on a Graphics Processing Unit (GPGPU) are now accessible to users with general computing needs. Grid computing infrastructures in the form of computing clusters or blades are becoming common place and GPGPU solutions that leverage the processing power of the video card are quickly being integrated into personal workstations. Our interest in these HPC technologies stems from the need to produce near real-time maps from a combination of pre- and post-event satellite imagery in support of post-disaster management. Faster processing provides a twofold gain in this situation: 1. critical information can be provided faster and 2. more elaborate automated processing can be performed prior to providing the critical information. In our particular case, we test the use of the PANTEX index which is based on analysis of image textural measures extracted using anisotropic, rotation-invariant GLCM statistics. The use of this index, applied in a moving window, has been shown to successfully identify built-up areas in remotely sensed imagery. Built-up index image masks are important input to the structuring of damage assessment interpretation because they help optimise the workload. The performance of computing the PANTEX workflow is compared on two different HPC hardware architectures: (1) a blade server with 4 blades, each having dual quad-core CPUs and (2) a CUDA enabled GPU workstation. The reference platform is a dual CPU-quad core workstation and the PANTEX workflow total computing time is measured. Furthermore, as part of a qualitative evaluation, the differences in setting up and configuring various hardware solutions and the related software coding effort is presented.
Multiscale Methods, Parallel Computation, and Neural Networks for Real-Time Computer Vision.
Battiti, Roberto
1990-01-01
This thesis presents new algorithms for low and intermediate level computer vision. The guiding ideas in the presented approach are those of hierarchical and adaptive processing, concurrent computation, and supervised learning. Processing of the visual data at different resolutions is used not only to reduce the amount of computation necessary to reach the fixed point, but also to produce a more accurate estimation of the desired parameters. The presented adaptive multiple scale technique is applied to the problem of motion field estimation. Different parts of the image are analyzed at a resolution that is chosen in order to minimize the error in the coefficients of the differential equations to be solved. Tests with video-acquired images show that velocity estimation is more accurate over a wide range of motion with respect to the homogeneous scheme. In some cases introduction of explicit discontinuities coupled to the continuous variables can be used to avoid propagation of visual information from areas corresponding to objects with different physical and/or kinematic properties. The human visual system uses concurrent computation in order to process the vast amount of visual data in "real -time." Although with different technological constraints, parallel computation can be used efficiently for computer vision. All the presented algorithms have been implemented on medium grain distributed memory multicomputers with a speed-up approximately proportional to the number of processors used. A simple two-dimensional domain decomposition assigns regions of the multiresolution pyramid to the different processors. The inter-processor communication needed during the solution process is proportional to the linear dimension of the assigned domain, so that efficiency is close to 100% if a large region is assigned to each processor. Finally, learning algorithms are shown to be a viable technique to engineer computer vision systems for different applications starting from
A shortcut through time The path to the quantum computer
Johnson, George
2003-01-01
The newest Pentium chip powering PCs and laptops contains 40 million electronic switches packed onto a piece of silicon about the size of a thumbnail. Several years from now, if this incredible shrinking continues, a single chip will hold a billion switches, then a trillion. The logical culmination is a computer in which the switches are so tiny that each consists of an individual atom. At that point something miraculous happens: Quantum mechanics kick in. Anyone who follows the science news or watches 'Star Trek' has at least a notion of what that means: particles that can be in two or more places at once, that can seem one moment like hard little specks of matter and the next like waves. Atoms obey a peculiar logic of their own - and if it can be harnessed society will be transformed. Problems that would now take for ever even on a supercomputer would be solved almost instantly. Quantum computing promises nothing less than a shortcut through time. In this brief and totally absorbing book, we are brought to ...
Krishnanathan, Kirubhakaran; Anderson, Sean R.; Billings, Stephen A.; Kadirkamanathan, Visakan
2016-11-01
In this paper, we derive a system identification framework for continuous-time nonlinear systems, for the first time using a simulation-focused computational Bayesian approach. Simulation approaches to nonlinear system identification have been shown to outperform regression methods under certain conditions, such as non-persistently exciting inputs and fast-sampling. We use the approximate Bayesian computation (ABC) algorithm to perform simulation-based inference of model parameters. The framework has the following main advantages: (1) parameter distributions are intrinsically generated, giving the user a clear description of uncertainty, (2) the simulation approach avoids the difficult problem of estimating signal derivatives as is common with other continuous-time methods, and (3) as noted above, the simulation approach improves identification under conditions of non-persistently exciting inputs and fast-sampling. Term selection is performed by judging parameter significance using parameter distributions that are intrinsically generated as part of the ABC procedure. The results from a numerical example demonstrate that the method performs well in noisy scenarios, especially in comparison to competing techniques that rely on signal derivative estimation.
3D Vectorial Time Domain Computational Integrated Photonics
Energy Technology Data Exchange (ETDEWEB)
Kallman, J S; Bond, T C; Koning, J M; Stowell, M L
2007-02-16
The design of integrated photonic structures poses considerable challenges. 3D-Time-Domain design tools are fundamental in enabling technologies such as all-optical logic, photonic bandgap sensors, THz imaging, and fast radiation diagnostics. Such technologies are essential to LLNL and WFO sponsors for a broad range of applications: encryption for communications and surveillance sensors (NSA, NAI and IDIV/PAT); high density optical interconnects for high-performance computing (ASCI); high-bandwidth instrumentation for NIF diagnostics; micro-sensor development for weapon miniaturization within the Stockpile Stewardship and DNT programs; and applications within HSO for CBNP detection devices. While there exist a number of photonics simulation tools on the market, they primarily model devices of interest to the communications industry. We saw the need to extend our previous software to match the Laboratory's unique emerging needs. These include modeling novel material effects (such as those of radiation induced carrier concentrations on refractive index) and device configurations (RadTracker bulk optics with radiation induced details, Optical Logic edge emitting lasers with lateral optical inputs). In addition we foresaw significant advantages to expanding our own internal simulation codes: parallel supercomputing could be incorporated from the start, and the simulation source code would be accessible for modification and extension. This work addressed Engineering's Simulation Technology Focus Area, specifically photonics. Problems addressed from the Engineering roadmap of the time included modeling the Auston switch (an important THz source/receiver), modeling Vertical Cavity Surface Emitting Lasers (VCSELs, which had been envisioned as part of fast radiation sensors), and multi-scale modeling of optical systems (for a variety of applications). We proposed to develop novel techniques to numerically solve the 3D multi-scale propagation problem for both the
29 CFR 459.1 - Computation of time for filing papers.
2010-07-01
... 29 Labor 2 2010-07-01 2010-07-01 false Computation of time for filing papers. 459.1 Section 459.1... OF CONDUCT MISCELLANEOUS § 459.1 Computation of time for filing papers. In computing any period of... computations. When these regulations require the filing of any paper, such document must be received by...
Time-Domain Terahertz Computed Axial Tomography NDE System
Zimdars, David
2012-01-01
NASA has identified the need for advanced non-destructive evaluation (NDE) methods to characterize aging and durability in aircraft materials to improve the safety of the nation's airline fleet. 3D THz tomography can play a major role in detection and characterization of flaws and degradation in aircraft materials, including Kevlar-based composites and Kevlar and Zylon fabric covers for soft-shell fan containment where aging and durability issues are critical. A prototype computed tomography (CT) time-domain (TD) THz imaging system has been used to generate 3D images of several test objects including a TUFI tile (a thermal protection system tile used on the Space Shuttle and possibly the Orion or similar capsules). This TUFI tile had simulated impact damage that was located and the depth of damage determined. The CT motion control gan try was designed and constructed, and then integrated with a T-Ray 4000 control unit and motion controller to create a complete CT TD-THz imaging system prototype. A data collection software script was developed that takes multiple z-axis slices in sequence and saves the data for batch processing. The data collection software was integrated with the ability to batch process the slice data with the CT TD-THz image reconstruction software. The time required to take a single CT slice was decreased from six minutes to approximately one minute by replacing the 320 ps, 100-Hz waveform acquisition system with an 80 ps, 1,000-Hz waveform acquisition system. The TD-THZ computed tomography system was built from pre-existing commercial off-the-shelf subsystems. A CT motion control gantry was constructed from COTS components that can handle larger samples. The motion control gantry allows inspection of sample sizes of up to approximately one cubic foot (.0.03 cubic meters). The system reduced to practice a CT-TDTHz system incorporating a COTS 80- ps/l-kHz waveform scanner. The incorporation of this scanner in the system allows acquisition of 3D
Continuous-time cross-phase modulation and quantum computation
Shapiro, J H; Razavi, Mohsen; Shapiro, Jeffrey H.
2006-01-01
The weak nonlinear Kerr interaction between single photons and intense laser fields has been recently proposed as a basis for distributed optics-based solutions to few-qubit applications in quantum communication and computation. Here, we analyze the above Kerr interaction by employing a continuous-time multi-mode model for the input/output fields to/from the nonlinear medium. In contrast to previous single-mode treatments of this problem, our analysis takes into account the full temporal content of the free-field input beams as well as the non-instantaneous response of the medium. The main implication of this model, in which the cross-Kerr phase shift on one input is proportional to the photon flux of the other input, is the existence of phase noise terms at the output. We show that these phase noise terms will degrade the performance of the parity gate proposed by Munro, Nemoto, and Spiller [New J. Phys. 7, 137 (2005)].
Delivery of Educational Materials in Real Time via the Computer.
Himmelblau, D. M.
1985-01-01
Discusses preparation for and problems of electronic publishing applied to text modules for individual/classroom use in undergraduate chemical engineering. Also describes how the modules are authored and evaluated and how text, equations, and figures are entered into the computer for use with a wide variety of computers and terminals. (Author/JN)
Cluster Computing For Real Time Seismic Array Analysis.
Martini, M.; Giudicepietro, F.
A seismic array is an instrument composed by a dense distribution of seismic sen- sors that allow to measure the directional properties of the wavefield (slowness or wavenumber vector) radiated by a seismic source. Over the last years arrays have been widely used in different fields of seismological researches. In particular they are applied in the investigation of seismic sources on volcanoes where they can be suc- cessfully used for studying the volcanic microtremor and long period events which are critical for getting information on the volcanic systems evolution. For this reason arrays could be usefully employed for the volcanoes monitoring, however the huge amount of data produced by this type of instruments and the processing techniques which are quite time consuming limited their potentiality for this application. In order to favor a direct application of arrays techniques to continuous volcano monitoring we designed and built a small PC cluster able to near real time computing the kinematics properties of the wavefield (slowness or wavenumber vector) produced by local seis- mic source. The cluster is composed of 8 Intel Pentium-III bi-processors PC working at 550 MHz, and has 4 Gigabytes of RAM memory. It runs under Linux operating system. The developed analysis software package is based on the Multiple SIgnal Classification (MUSIC) algorithm and is written in Fortran. The message-passing part is based upon the LAM programming environment package, an open-source imple- mentation of the Message Passing Interface (MPI). The developed software system includes modules devote to receiving date by internet and graphical applications for the continuous displaying of the processing results. The system has been tested with a data set collected during a seismic experiment conducted on Etna in 1999 when two dense seismic arrays have been deployed on the northeast and the southeast flanks of this volcano. A real time continuous acquisition system has been simulated by
Time independent universal computing with spin chains: quantum plinko machine
Thompson, K. F.; Gokler, C.; Lloyd, S.; Shor, P. W.
2016-07-01
We present a scheme for universal quantum computing using XY Heisenberg spin chains. Information is encoded into packets propagating down these chains, and they interact with each other to perform universal quantum computation. A circuit using g gate blocks on m qubits can be encoded into chains of length O({g}3+δ {m}3+δ ) for all δ \\gt 0 with vanishingly small error.
5 CFR 2429.21 - Computation of time for filing papers.
2010-01-01
... 5 Administrative Personnel 3 2010-01-01 2010-01-01 false Computation of time for filing papers... REQUIREMENTS General Requirements § 2429.21 Computation of time for filing papers. (a) In computing any period... § 2429.23(a) of this part, when this subchapter requires the filing of any paper with the Authority,...
A general algorithm for computing distance transforms in linear time
Meijster, A.; Roerdink, J.B.T.M.; Hesselink, W.H.; Goutsias, J; Vincent, L; Bloomberg, DS
2000-01-01
A new general algorithm fur computing distance transforms of digital images is presented. The algorithm consists of two phases. Both phases consist of two scans, a forward and a backward scan. The first phase scans the image column-wise, while the second phase scans the image row-wise. Since the com
A general algorithm for computing distance transforms in linear time
Meijster, A.; Roerdink, J.B.T.M.; Hesselink, W.H.; Goutsias, J; Vincent, L; Bloomberg, DS
2000-01-01
A new general algorithm fur computing distance transforms of digital images is presented. The algorithm consists of two phases. Both phases consist of two scans, a forward and a backward scan. The first phase scans the image column-wise, while the second phase scans the image row-wise. Since the
Computing the Tutte Polynomial in Vertex-Exponential Time
DEFF Research Database (Denmark)
Björklund, Andreas; Husfeldt, Thore; Kaski, Petteri
2008-01-01
The deletion–contraction algorithm is perhapsthe most popular method for computing a host of fundamental graph invariants such as the chromatic, flow, and reliability polynomials in graph theory, the Jones polynomial of an alternating link in knot theory, and the partition functions of the models...
Fault recovery for real-time, multi-tasking computer system
Hess, Richard (Inventor); Kelly, Gerald B. (Inventor); Rogers, Randy (Inventor); Stange, Kent A. (Inventor)
2011-01-01
System and methods for providing a recoverable real time multi-tasking computer system are disclosed. In one embodiment, a system comprises a real time computing environment, wherein the real time computing environment is adapted to execute one or more applications and wherein each application is time and space partitioned. The system further comprises a fault detection system adapted to detect one or more faults affecting the real time computing environment and a fault recovery system, wherein upon the detection of a fault the fault recovery system is adapted to restore a backup set of state variables.
Tempel, David G; Aspuru-Guzik, Alán
2012-01-01
We prove that the theorems of TDDFT can be extended to a class of qubit Hamiltonians that are universal for quantum computation. The theorems of TDDFT applied to universal Hamiltonians imply that single-qubit expectation values can be used as the basic variables in quantum computation and information theory, rather than wavefunctions. From a practical standpoint this opens the possibility of approximating observables of interest in quantum computations directly in terms of single-qubit quantities (i.e. as density functionals). Additionally, we also demonstrate that TDDFT provides an exact prescription for simulating universal Hamiltonians with other universal Hamiltonians that have different, and possibly easier-to-realize two-qubit interactions. This establishes the foundations of TDDFT for quantum computation and opens the possibility of developing density functionals for use in quantum algorithms.
A real-time transaction service for grid computing
Institute of Scientific and Technical Information of China (English)
Tang Feilong; Li Minglu; Joshua Huang Zhexue; Frank Tong; Cao Jian
2005-01-01
A real-time transaction service aims at ensuring the consistency of real-time and transactional Grid applications. This paper proposes an efficient real-time transaction service (RTTS). It provides abilities to discover qualified services, I.e., participants, to execute specified sub-transactions, coordinate real-time Grid transactions, calculate deadline and assign priority for scheduling concurrent transactions. The service discovery adopts the two-level registry mechanismto adapt to transient Grid services. The coordination algorithm invokes multiplefunctional alternative services for each sub-transaction so that it remarkablyimproves the successful ratio of real-time transactions. The experiment result shows that the RTTS can well coordinate real-time transactions for Grid service environment and shield users from the complex process.
Real-time exposure fusion on a mobile computer
CSIR Research Space (South Africa)
Bachoo, AK
2009-12-01
Full Text Available Romeny, J. Zimmer- man, and K. Zuiderveld, “Adaptive histogram equalization and its variations,” Computer Vision, Graphics and Image Processing, vol. 39, pp. 355–368, 1987. [8] R. Gonzalez and R. Woods, Digital image processing. Addison-Wesley... Publishing Company, 2002. [9] R. Rost, OpenGL(R) Shading Language (2nd Edition). Addison-Wesley Professional, 2006. [10] D. Shreiner, M. Woo, J. Neider, and T. Davis, OpenGL Programming Guide. Addison-Wesley Professional, 2007. [11] A. Reza...
Computer Modeling of Real-Time Dynamic Lighting
Maida, James C.; Pace, J.; Novak, J.; Russo, Dane M. (Technical Monitor)
2000-01-01
Space Station tasks involve procedures that are very complex and highly dependent on the availability of visual information. In many situations, cameras are used as tools to help overcome the visual and physical restrictions associated with space flight. However, these cameras are effected by the dynamic lighting conditions of space. Training for these is conditions is necessary. The current project builds on the findings of an earlier NRA funded project, which revealed improved performance by humans when trained with computer graphics and lighting effects such as shadows and glare.
Storm blueprints patterns for distributed real-time computation
Goetz, P Taylor
2014-01-01
A blueprints book with 10 different projects built in 10 different chapters which demonstrate the various use cases of storm for both beginner and intermediate users, grounded in real-world example applications.Although the book focuses primarily on Java development with Storm, the patterns are more broadly applicable and the tips, techniques, and approaches described in the book apply to architects, developers, and operations.Additionally, the book should provoke and inspire applications of distributed computing to other industries and domains. Hadoop enthusiasts will also find this book a go
Macroprocessing is the computing design principle for the times
2001-01-01
In a keynote speech, Intel Corporation CEO Craig Barrett emphasized that "macroprocessing" provides innovative and cost effective solutions to companies that they can customize and scale to match their own data needs. Barrett showcased examples of macroprocessing implementations from business, government and the scientific community, which use the power of Intel Architecture and Oracle9i Real Application Clusters to build large complex and scalable database solutions. A testimonial from CERN explained how the need for high performance computing to perform scientific research on sub-atomic particles was accomplished by using clusters of Xeon processor-based servers.
Zimovets, Artem; Matviychuk, Alexander; Ushakov, Vladimir
2016-12-01
The paper presents two different approaches to reduce the time of computer calculation of reachability sets. First of these two approaches use different data structures for storing the reachability sets in the computer memory for calculation in single-threaded mode. Second approach is based on using parallel algorithms with reference to the data structures from the first approach. Within the framework of this paper parallel algorithm of approximate reachability set calculation on computer with SMP-architecture is proposed. The results of numerical modelling are presented in the form of tables which demonstrate high efficiency of parallel computing technology and also show how computing time depends on the used data structure.
7 CFR 900.15 - Filing; extensions of time; effective date of filing; and computation of time.
2010-01-01
...; and computation of time. 900.15 Section 900.15 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements and Orders; Fruits, Vegetables, Nuts), DEPARTMENT OF AGRICULTURE GENERAL REGULATIONS Rules of Practice and Procedure...
A strategy for reducing turnaround time in design optimization using a distributed computer system
Young, Katherine C.; Padula, Sharon L.; Rogers, James L.
1988-01-01
There is a need to explore methods for reducing lengthly computer turnaround or clock time associated with engineering design problems. Different strategies can be employed to reduce this turnaround time. One strategy is to run validated analysis software on a network of existing smaller computers so that portions of the computation can be done in parallel. This paper focuses on the implementation of this method using two types of problems. The first type is a traditional structural design optimization problem, which is characterized by a simple data flow and a complicated analysis. The second type of problem uses an existing computer program designed to study multilevel optimization techniques. This problem is characterized by complicated data flow and a simple analysis. The paper shows that distributed computing can be a viable means for reducing computational turnaround time for engineering design problems that lend themselves to decomposition. Parallel computing can be accomplished with a minimal cost in terms of hardware and software.
Newmark local time stepping on high-performance computing architectures
Rietmann, Max
2016-11-25
In multi-scale complex media, finite element meshes often require areas of local refinement, creating small elements that can dramatically reduce the global time-step for wave-propagation problems due to the CFL condition. Local time stepping (LTS) algorithms allow an explicit time-stepping scheme to adapt the time-step to the element size, allowing near-optimal time-steps everywhere in the mesh. We develop an efficient multilevel LTS-Newmark scheme and implement it in a widely used continuous finite element seismic wave-propagation package. In particular, we extend the standard LTS formulation with adaptations to continuous finite element methods that can be implemented very efficiently with very strong element-size contrasts (more than 100×). Capable of running on large CPU and GPU clusters, we present both synthetic validation examples and large scale, realistic application examples to demonstrate the performance and applicability of the method and implementation on thousands of CPU cores and hundreds of GPUs.
Newmark local time stepping on high-performance computing architectures
Rietmann, Max; Grote, Marcus; Peter, Daniel; Schenk, Olaf
2017-04-01
In multi-scale complex media, finite element meshes often require areas of local refinement, creating small elements that can dramatically reduce the global time-step for wave-propagation problems due to the CFL condition. Local time stepping (LTS) algorithms allow an explicit time-stepping scheme to adapt the time-step to the element size, allowing near-optimal time-steps everywhere in the mesh. We develop an efficient multilevel LTS-Newmark scheme and implement it in a widely used continuous finite element seismic wave-propagation package. In particular, we extend the standard LTS formulation with adaptations to continuous finite element methods that can be implemented very efficiently with very strong element-size contrasts (more than 100x). Capable of running on large CPU and GPU clusters, we present both synthetic validation examples and large scale, realistic application examples to demonstrate the performance and applicability of the method and implementation on thousands of CPU cores and hundreds of GPUs.
Computation and evaluation of scheduled waiting time for railway networks
DEFF Research Database (Denmark)
Landex, Alex
2010-01-01
Timetables are affected by scheduled waiting time (SWT) that prolongs the travel times for trains and thereby passengers. SWT occurs when a train hinders another train to run with the wanted speed. The SWT affects both the trains and the passengers in the trains. The passengers may be further aff...... timetable by analysing different timetables and/or plans of operation. This article presents methods to examine SWT by simulation for both trains and passengers in entire railway networks....... affected due to longer transfer times to other trains. SWT can be estimated analytically for a given timetable or by simulation of timetables and/or plans of operation. The simulation of SWT has the benefit that it is possible to examine the entire network. This makes it possible to improve the future...
Time-of-Flight Sensors in Computer Graphics
DEFF Research Database (Denmark)
Kolb, Andreas; Barth, Erhardt; Koch, Reinhard
2009-01-01
Abstract A growing number of applications depend on accurate and fast 3D scene analysis. Examples are model and lightfield acquisition, collision prevention, mixed reality, and gesture recognition. The estimation of a range map by image analysis or laser scan techniques is still a time-consuming ......-time geometry devices” for gaming, web-conferencing, and numerous other applications. This STAR gives an account of recent developments in ToF-technology and discusses the current state of the integration of this technology into various graphics-related applications.......-consuming and expensive part of such systems. A lower-priced, fast and robust alternative for distance measurements are Time-of-Flight (ToF) cameras. Recently, significant improvements have been made in order to achieve low-cost and compact ToF-devices, that have the potential to revolutionize many fields of research...
Computing the Discrete Fr\\'echet Distance in Subquadratic Time
Agarwal, Pankaj K; Kaplan, Haim; Sharir, Micha
2012-01-01
The Fr\\'echet distance is a similarity measure between two curves $A$ and $B$: Informally, it is the minimum length of a leash required to connect a dog, constrained to be on $A$, and its owner, constrained to be on $B$, as they walk without backtracking along their respective curves from one endpoint to the other. The advantage of this measure on other measures such as the Hausdorff distance is that it takes into account the ordering of the points along the curves. The discrete Fr\\'echet distance replaces the dog and its owner by a pair of frogs that can only reside on $n$ and $m$ specific pebbles on the curves $A$ and $B$, respectively. These frogs hop from a pebble to the next without backtracking. The discrete Fr\\'echet distance can be computed by a rather straightforward quadratic dynamic programming algorithm. However, despite a considerable amount of work on this problem and its variations, there is no subquadratic algorithm known, even for approximation versions of the problem. In this paper we presen...
20 CFR 655.1030 - Service and computation of time.
2010-04-01
... EMPLOYMENT OF FOREIGN WORKERS IN THE UNITED STATES Enforcement of the Attestation Process for Attestations... party is complete upon mailing to the last known address. No additional time for filing or response is authorized where service is by mail. In the interest of expeditious proceedings, the administrative law judge...
Electromagnetic space-time crystals. II. Fractal computational approach
Borzdov, G. N.
2014-01-01
A fractal approach to numerical analysis of electromagnetic space-time crystals, created by three standing plane harmonic waves with mutually orthogonal phase planes and the same frequency, is presented. Finite models of electromagnetic crystals are introduced, which make possible to obtain various approximate solutions of the Dirac equation. A criterion for evaluating accuracy of these approximate solutions is suggested.
Time-of-Flight Sensors in Computer Graphics
DEFF Research Database (Denmark)
Kolb, Andreas; Barth, Erhardt; Koch, Reinhard
2009-01-01
Abstract A growing number of applications depend on accurate and fast 3D scene analysis. Examples are model and lightfield acquisition, collision prevention, mixed reality, and gesture recognition. The estimation of a range map by image analysis or laser scan techniques is still a time...
Electromagnetic space-time crystals. II. Fractal computational approach
2014-01-01
A fractal approach to numerical analysis of electromagnetic space-time crystals, created by three standing plane harmonic waves with mutually orthogonal phase planes and the same frequency, is presented. Finite models of electromagnetic crystals are introduced, which make possible to obtain various approximate solutions of the Dirac equation. A criterion for evaluating accuracy of these approximate solutions is suggested.
12 CFR 269b.720 - Computation of time for filing papers.
2010-01-01
... 12 Banks and Banking 3 2010-01-01 2010-01-01 false Computation of time for filing papers. 269b.720... papers. In computing any period of time prescribed by or allowed by the panel, the day of the act, event... regulations in this subchapter require the filing of any paper, such document must be received by the panel...
I. Fisk
2011-01-01
Introduction CMS distributed computing system performed well during the 2011 start-up. The events in 2011 have more pile-up and are more complex than last year; this results in longer reconstruction times and harder events to simulate. Significant increases in computing capacity were delivered in April for all computing tiers, and the utilisation and load is close to the planning predictions. All computing centre tiers performed their expected functionalities. Heavy-Ion Programme The CMS Heavy-Ion Programme had a very strong showing at the Quark Matter conference. A large number of analyses were shown. The dedicated heavy-ion reconstruction facility at the Vanderbilt Tier-2 is still involved in some commissioning activities, but is available for processing and analysis. Facilities and Infrastructure Operations Facility and Infrastructure operations have been active with operations and several important deployment tasks. Facilities participated in the testing and deployment of WMAgent and WorkQueue+Request...
Power computations in time series analyses for traffic safety interventions.
McLeod, A Ian; Vingilis, E R
2008-05-01
The evaluation of traffic safety interventions or other policies that can affect road safety often requires the collection of administrative time series data, such as monthly motor vehicle collision data that may be difficult and/or expensive to collect. Furthermore, since policy decisions may be based on the results found from the intervention analysis of the policy, it is important to ensure that the statistical tests have enough power, that is, that we have collected enough time series data both before and after the intervention so that a meaningful change in the series will likely be detected. In this short paper, we present a simple methodology for doing this. It is expected that the methodology presented will be useful for sample size determination in a wide variety of traffic safety intervention analysis applications. Our method is illustrated with a proposed traffic safety study that was funded by NIH.
Time-of-Flight Cameras in Computer Graphics
DEFF Research Database (Denmark)
Kolb, Andreas; Barth, Erhardt; Koch, Reinhard
2010-01-01
A growing number of applications depend on accurate and fast 3D scene analysis. Examples are model and lightfield acquisition, collision prevention, mixed reality, and gesture recognition. The estimation of a range map by image analysis or laser scan techniques is still a time-consuming and expen...... devices” for gaming, web-conferencing, and numerous other applications. This STAR gives an account of recent developments in ToF-technology and discusses the current state of the integration of this technology into various graphics-related applications.......-consuming and expensive part of such systems. A lower-priced, fast and robust alternative for distance measurements are Time-of-Flight (ToF) cameras. Recently, significant advances have been made in producing low-cost and compact ToF-devices, which have the potential to revolutionize many fields of research, including...
Wake force computation in the time domain for long structures
Energy Technology Data Exchange (ETDEWEB)
Bane, K.; Weiland, T.
1983-07-01
One is often interested in calculating the wake potentials for short bunches in long structures using TBCI. For ultra-relativistic particles it is sufficient to solve for the fields only over a window containing the bunch and moving along with it. This technique reduces both the memory and the running time required by a factor that equals the ratio of the structure length to the window length. For example, for a bunch with sigma/sub z/ of one picosecond traversing a single SLAC cell this improvement factor is 15. It is thus possible to solve for the wakefields in very long structures: for a given problem, increasing the structure length will not change the memory required while only adding linearly to the CPU time needed.
Time-ordered product expansions for computational stochastic system biology.
Mjolsness, Eric
2013-06-01
The time-ordered product framework of quantum field theory can also be used to understand salient phenomena in stochastic biochemical networks. It is used here to derive Gillespie's stochastic simulation algorithm (SSA) for chemical reaction networks; consequently, the SSA can be interpreted in terms of Feynman diagrams. It is also used here to derive other, more general simulation and parameter-learning algorithms including simulation algorithms for networks of stochastic reaction-like processes operating on parameterized objects, and also hybrid stochastic reaction/differential equation models in which systems of ordinary differential equations evolve the parameters of objects that can also undergo stochastic reactions. Thus, the time-ordered product expansion can be used systematically to derive simulation and parameter-fitting algorithms for stochastic systems.
Real-time brain computer interface using imaginary movements
DEFF Research Database (Denmark)
El-Madani, Ahmad; Sørensen, Helge Bjarup Dissing; Kjær, Troels W.
2015-01-01
, the implementation of a 2-class real-time BCI system based on the event related desynchronization (ERD) of the sensorimotor rhythms (SMR) is described. Methods: Off-line measurements were conducted on 12 healthy test subjects with 3 different feedback systems (cross, basket and bars). From the collected...... classifier (BLC) was developed and used for signal classification. These three subject-specific settings were preserved for the on-line experiments with the same feedback systems. Results: Six of the 12 subjects were qualified for the on-line experiments based on their high off-line classification accuracies...... (CAs > 75 %). The overall mean on-line accuracy was found to be 80%. Conclusions: The subject-specific settings applied on the feedback systems have resulted in the development of a successful real-time BCI system with high accuracies...
Wavelet analysis on paleomagnetic (and computer simulated VGP time series
Directory of Open Access Journals (Sweden)
A. Siniscalchi
2003-06-01
Full Text Available We present Continuous Wavelet Transform (CWT data analysis of Virtual Geomagnetic Pole (VGP latitude time series. The analyzed time series are sedimentary paleomagnetic and geodynamo simulated data. Two mother wavelets (the Morlet function and the first derivative of a Gaussian function are used in order to detect features related to the spectral content as well as polarity excursions and reversals. By means of the Morlet wavelet, we estimate both the global spectrum and the time evolution of the spectral content of the paleomagnetic data series. Some peaks corresponding to the orbital components are revealed by the spectra and the local analysis helped disclose their statistical significance. Even if this feature could be an indication of orbital influence on geodynamo, other interpretations are possible. In particular, we note a correspondence of local spectral peaks with the appearance of the excursions in the series. The comparison among the paleomagnetic and simulated spectra shows a similarity in the high frequency region indicating that their degree of regularity is analogous. By means of Gaussian first derivative wavelet, reversals and excursions of polarity were sought. The analysis was performed first on the simulated data, to have a guide in understanding the features present in the more complex paleomagnetic data. Various excursions and reversals have been identified, despite of the prevalent normality of the series and its inherent noise. The found relative chronology of the paleomagnetic data reversals was compared with a coeval global polarity time scale (Channel et al., 1995. The relative lengths of polarity stability intervals are found similar, but a general shift appears between the two scales, that could be due to the datation uncertainties of the Hauterivian/Barremian boundary.
HOPE: Just-in-time Python compiler for astrophysical computations
Akeret, Joel; Gamper, Lukas; Amara, Adam; Refregier, Alexandre
2014-11-01
HOPE is a specialized Python just-in-time (JIT) compiler designed for numerical astrophysical applications. HOPE focuses on a subset of the language and is able to translate Python code into C++ while performing numerical optimization on mathematical expressions at runtime. To enable the JIT compilation, the user only needs to add a decorator to the function definition. By using HOPE, the user benefits from being able to write common numerical code in Python while getting the performance of compiled implementation.
Reducing Computational Time of Basic Encryption and Authentication Algorithms
Directory of Open Access Journals (Sweden)
Sandeep Kumar,
2011-04-01
Full Text Available Today most of data are sending via the internet for sharing, so the trust of data files is decreased. For the trust more security and authentication is needed, less security increase the liability ofattacks on data. Digital signature of the data is a solution to this security problem which provides the reliability, authenticity and accuracy. Most basic algorithm for security and authentication is RSA, DSA, algorithms which uses the different key of different sizes. This work presents ECDSA algorithm to encrypt the data, use parameterized hash algorithm to authenticate the data and also compare both RSA and ECDSA methods in respect of time parameters.
Imprecise Computation Based Real-time Fault Tolerant Implementation for Model Predictive Control
Institute of Scientific and Technical Information of China (English)
无
2006-01-01
Model predictive control (MPC) could not be deployed in real-time control systems for its computation time is not well defined. A real-time fault tolerant implementation algorithm based on imprecise computation is proposed for MPC,according to the solving process of quadratic programming (QP) problem. In this algorithm, system stability is guaranteed even when computation resource is not enough to finish optimization completely. By this kind of graceful degradation, the behavior of real-time control systems is still predictable and determinate. The algorithm is demonstrated by experiments on servomotor, and the simulation results show its effectiveness.
2001-01-01
The first phase of the LHC Computing Grid project was approved at an extraordinary meeting of the Council on 20 September 2001. CERN is preparing for the unprecedented avalanche of data that will be produced by the Large Hadron Collider experiments. A thousand times more computer power will be needed by 2006! CERN's need for a dramatic advance in computing capacity is urgent. As from 2006, the four giant detectors observing trillions of elementary particle collisions at the LHC will accumulate over ten million Gigabytes of data, equivalent to the contents of about 20 million CD-ROMs, each year of its operation. A thousand times more computing power will be needed than is available to CERN today. The strategy the collabortations have adopted to analyse and store this unprecedented amount of data is the coordinated deployment of Grid technologies at hundreds of institutes which will be able to search out and analyse information from an interconnected worldwide grid of tens of thousands of computers and storag...
Using Brain Waves as New Biometric Feature for Authenticating a Computer User in Real-Time
Kusuma Mohanchandra; Lingaraju G M; Prashanth Kambli; Vinay Krishnamurthy
2013-01-01
In this paper we propose an Electroencephalogram based Brain Computer Interface as a new modality for Person Authentication and develop a screen lock application that will lock and unlock the computer screen at the users will. The brain waves of the person, recorded in real time are used as password to unlock the screen. Data fusion from 14 sensors of the Emotiv headset is done to enhance the signal features. The power spectral density of the intermingle signals is computed. The channel spect...
An Approach to Integrate a Space-Time GIS Data Model with High Performance Computers
Energy Technology Data Exchange (ETDEWEB)
Wang, Dali [ORNL; Zhao, Ziliang [University of Tennessee, Knoxville (UTK); Shaw, Shih-Lung [ORNL
2011-01-01
In this paper, we describe an approach to integrate a Space-Time GIS data model on a high performance computing platform. The Space-Time GIS data model has been developed on a desktop computing environment. We use the Space-Time GIS data model to generate GIS module, which organizes a series of remote sensing data. We are in the process of porting the GIS module into an HPC environment, in which the GIS modules handle large dataset directly via parallel file system. Although it is an ongoing project, authors hope this effort can inspire further discussions on the integration of GIS on high performance computing platforms.
A time lag study of the vertical motion simulator computer system
Cleveland, W. B.
1981-01-01
A study was performed to evaluate an experimental method to determine time lags in real-time computer systems as the one associated with the Vertical Motion Simulator at Ames Research Center. The approach was to use an ordinary frequency analyzer to measure the phase difference between inputs and outputs of the computer system. The various elements of the program and computational architecture were modeled. Various factors, such as computer frame time and input frequency, were varied so that they were representative of the operational use of the simulator facilities. Experimentally determined results were compared with predictions derived from the simulation models. The results indicate that the frequency analyzer can be readily used to evaluate time lags in systems of this type. Differences between predicted and measured phase values indicate that the hardware and software imparts a time lag of about 5 msec to this facility.
Real Time Animation of Trees Based on BBSC in Computer Games
Directory of Open Access Journals (Sweden)
Xuefeng Ao
2009-01-01
Full Text Available That researchers in the field of computer games usually find it is difficult to simulate the motion of actual 3D model trees lies in the fact that the tree model itself has very complicated structure, and many sophisticated factors need to be considered during the simulation. Though there are some works on simulating 3D tree and its motion, few of them are used in computer games due to the high demand for real-time in computer games. In this paper, an approach of animating trees in computer games based on a novel tree model representation—Ball B-Spline Curves (BBSCs are proposed. By taking advantage of the good features of the BBSC-based model, physical simulation of the motion of leafless trees with wind blowing becomes easier and more efficient. The method can generate realistic 3D tree animation in real-time, which meets the high requirement for real time in computer games.
I. Fisk
2013-01-01
Computing activity had ramped down after the completion of the reprocessing of the 2012 data and parked data, but is increasing with new simulation samples for analysis and upgrade studies. Much of the Computing effort is currently involved in activities to improve the computing system in preparation for 2015. Operations Office Since the beginning of 2013, the Computing Operations team successfully re-processed the 2012 data in record time, not only by using opportunistic resources like the San Diego Supercomputer Center which was accessible, to re-process the primary datasets HTMHT and MultiJet in Run2012D much earlier than planned. The Heavy-Ion data-taking period was successfully concluded in February collecting almost 500 T. Figure 3: Number of events per month (data) In LS1, our emphasis is to increase efficiency and flexibility of the infrastructure and operation. Computing Operations is working on separating disk and tape at the Tier-1 sites and the full implementation of the xrootd federation ...
The finite difference time domain method on a massively parallel computer
Ewijk, L.J. van
1996-01-01
At the Physics and Electronics Laboratory TNO much research is done in the field of computational electromagnetics (CEM). One of the tools in this field is the Finite Difference Time Domain method (FDTD), a method that has been implemented in a program in order to be able to compute electromagnetic
Real-World Experimentation Comparing Time-Sharing and Batch Processing in Teaching Computer Science,
effectiveness of time-sharing and batch processing in teaching computer science . The experimental design was centered on direct, ’real world’ comparison...ALGOL). The experimental sample involved all introductory computer science courses with a total population of 415 cadets. The results generally
Online Operation Guidance of Computer System Used in Real-Time Distance Education Environment
He, Aiguo
2011-01-01
Computer system is useful for improving real time and interactive distance education activities. Especially in the case that a large number of students participate in one distance lecture together and every student uses their own computer to share teaching materials or control discussions over the virtual classrooms. The problem is that within…
2013-06-28
... National Institute of Standards and Technology Computer Security Incident Coordination (CSIC): Providing Timely Cyber Incident Response AGENCY: National Institute of Standards and Technology, U.S. Department of... Technology (NIST) is seeking information relating to Computer Security Incident Coordination (CSIC). NIST is...
I. Fisk
2010-01-01
Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion An activity that is still in progress is computing for the heavy-ion program. The heavy-ion events are collected without zero suppression, so the event size is much large at roughly 11 MB per event of RAW. The central collisions are more complex and...
Computational Methods for Stability and Control (COMSAC): The Time Has Come
Hall, Robert M.; Biedron, Robert T.; Ball, Douglas N.; Bogue, David R.; Chung, James; Green, Bradford E.; Grismer, Matthew J.; Brooks, Gregory P.; Chambers, Joseph R.
2005-01-01
Powerful computational fluid dynamics (CFD) tools have emerged that appear to offer significant benefits as an adjunct to the experimental methods used by the stability and control community to predict aerodynamic parameters. The decreasing costs for and increasing availability of computing hours are making these applications increasingly viable as time goes on and the cost of computing continues to drop. This paper summarizes the efforts of four organizations to utilize high-end computational fluid dynamics (CFD) tools to address the challenges of the stability and control arena. General motivation and the backdrop for these efforts will be summarized as well as examples of current applications.
Wan, Junwei; Chen, Hongyan; Zhao, Jing
2017-08-01
According to the requirements of real-time, reliability and safety for aerospace experiment, the single center cloud computing technology application verification platform is constructed. At the IAAS level, the feasibility of the cloud computing technology be applied to the field of aerospace experiment is tested and verified. Based on the analysis of the test results, a preliminary conclusion is obtained: Cloud computing platform can be applied to the aerospace experiment computing intensive business. For I/O intensive business, it is recommended to use the traditional physical machine.
Method reduces computer time for smoothing functions and derivatives through ninth order polynomials
Glauz, R. D.; Wilgus, C. A.
1969-01-01
Analysis presented is an efficient technique to adjust previously calculated orthogonal polynomial coefficients for an odd number of equally spaced data points. The adjusting technique derivation is for a ninth order polynomial. It reduces computer time for smoothing functions.
The New Screen Time: Computers, Tablets, and Smartphones Enter the Equation
Wiles, Bradford B.; Schachtner, Laura; Pentz, Julie L.
2016-01-01
Emerging technologies attract children and push parents' and caregivers' abilities to attend to their families. This article presents recommendations related to the new version of screen time, which includes time with computers, tablets, and smartphones. Recommendations are provided for screen time for very young children and those in middle and…
The New Screen Time: Computers, Tablets, and Smartphones Enter the Equation
Wiles, Bradford B.; Schachtner, Laura; Pentz, Julie L.
2016-01-01
Emerging technologies attract children and push parents' and caregivers' abilities to attend to their families. This article presents recommendations related to the new version of screen time, which includes time with computers, tablets, and smartphones. Recommendations are provided for screen time for very young children and those in middle and…
The Effects of Time Scarcity on Conflict and Compromise in Computer Conferencing.
Reid, Fraser J. M.; Hards, Rachael
1998-01-01
Examines the effects of time scarcity on the way disagreement is managed in synchronous computer conferencing; reports an experiment in which pairs of undergraduates used keyboard-based conferencing software to resolve disputes on two controversial discussion topics under conditions either of time scarcity, or time abundance; and discusses…
Accelerating the discovery of space-time patterns of infectious diseases using parallel computing.
Hohl, Alexander; Delmelle, Eric; Tang, Wenwu; Casas, Irene
2016-11-01
Infectious diseases have complex transmission cycles, and effective public health responses require the ability to monitor outbreaks in a timely manner. Space-time statistics facilitate the discovery of disease dynamics including rate of spread and seasonal cyclic patterns, but are computationally demanding, especially for datasets of increasing size, diversity and availability. High-performance computing reduces the effort required to identify these patterns, however heterogeneity in the data must be accounted for. We develop an adaptive space-time domain decomposition approach for parallel computation of the space-time kernel density. We apply our methodology to individual reported dengue cases from 2010 to 2011 in the city of Cali, Colombia. The parallel implementation reaches significant speedup compared to sequential counterparts. Density values are visualized in an interactive 3D environment, which facilitates the identification and communication of uneven space-time distribution of disease events. Our framework has the potential to enhance the timely monitoring of infectious diseases.
An Algorithm for Optimized Time, Cost, and Reliability in a Distributed Computing System
Directory of Open Access Journals (Sweden)
Pankaj Saxena
2013-03-01
Full Text Available Distributed Computing System (DCS refers to multiple computer systems working on a single problem. A distributed system consists of a collection of autonomous computers, connected through a network which enables computers to coordinate their activities and to share the resources of the system. In distributed computing, a single problem is divided into many parts, and each part is solved by different computers. As long as the computers are networked, they can communicate with each other to solve the problem. DCS consists of multiple software components that are on multiple computers, but run as a single system. The computers that are in a distributed system can be physically close together and connected by a local network, or they can be geographically distant and connected by a wide area network. The ultimate goal of distributed computing is to maximize performance in a time effective, cost-effective, and reliability effective manner. In DCS the whole workload is divided into small and independent units, called tasks and it allocates onto the available processors. It also ensures fault tolerance and enables resource accessibility in the event that one of the components fails. The problem is addressed of assigning a task to a distributed computing system. The assignment of the modules of tasks is done statically. We have to give an algorithm to solve the problem of static task assignment in DCS, i.e. given a set of communicating tasks to be executed on a distributed system on a set of processors, to which processor should each task be assigned to get the more reliable results in lesser time and cost. In this paper an efficient algorithm for task allocation in terms of optimum time or optimum cost or optimum reliability is presented where numbers of tasks are more then the number of processors.
Directory of Open Access Journals (Sweden)
Helena Lindskog
2011-02-01
Traditionally, time-rich people were also money-rich, but the big majority was both time-poor and money-poor. This is completely turned upside down in the Information society. Nowadays people are equally divided between time-rich and money-poor, and time-poor and money-rich. Both time-rich and time-poor need entertainment and are potential consumers of entertainment especially games as e-services but in different ways. This is a conceptual paper and its purpose is to investigate the importance of the division on time-rich and time-poor for the entertainment industry and practical implications with a special focus on computer games as products and e-services.
Reconciling fault-tolerant distributed algorithms and real-time computing.
Moser, Heinrich; Schmid, Ulrich
We present generic transformations, which allow to translate classic fault-tolerant distributed algorithms and their correctness proofs into a real-time distributed computing model (and vice versa). Owing to the non-zero-time, non-preemptible state transitions employed in our real-time model, scheduling and queuing effects (which are inherently abstracted away in classic zero step-time models, sometimes leading to overly optimistic time complexity results) can be accurately modeled. Our results thus make fault-tolerant distributed algorithms amenable to a sound real-time analysis, without sacrificing the wealth of algorithms and correctness proofs established in classic distributed computing research. By means of an example, we demonstrate that real-time algorithms generated by transforming classic algorithms can be competitive even w.r.t. optimal real-time algorithms, despite their comparatively simple real-time analysis.
Real-Time Brain-Computer Interface System Based on Motor Imagery
Institute of Scientific and Technical Information of China (English)
Tie-Jun Liu; Ping Yang; Xu-Yong Peng; Yu Huang; De-Zhong Yao
2009-01-01
A brain-computer interface (BCI) real-time system based on motor imagery translates the user's motor intention into a real-time control signal for peripheral equipments.A key problem to be solved for practical applications is real-time data collection and processing.In this paper,a real-time BCI system is implemented on computer with electroencephalogram amplifier.In our implementation,the on-line voting method is adopted for feedback control strategy,and the voting results are used to control the cursor horizontal movement.Three subjects take part in the experiment.The results indicate that the best accuracy is 90%.
M. Kasemann P. McBride Edited by M-C. Sawley with contributions from: P. Kreuzer D. Bonacorsi S. Belforte F. Wuerthwein L. Bauerdick K. Lassila-Perini M-C. Sawley
Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the comput...
The application of queue theory in cloud computing to reduce the waiting time
Directory of Open Access Journals (Sweden)
N.N. Bharkad
2014-10-01
Full Text Available Cloud computing is a new technology in computer field to provide on line service to the customers. -Cloud computing has got enormous popularity as it offers dynamic, low-cost computing solutions. To get the service of cloud the user has to be in queue until he is served. Each arriving Cloud computing User (CCU requests Cloud computing Service Provider (CCSP to use the resources, if server is available, the arriving user will seize and hold it for a length of time, which leads to queue length and more waiting time. A new arrival leaves the queue with no service. After service completion the server is made immediately available to others. From the user’s point of view he needs to be served immediately and to prevent waiting the CCSP’s can use infinite servers to reduce waiting time & queue length. The arrival pattern is often Poisson in queuing theory. In this article we analyzed the dynamic behavior of the system with infinite servers by finding various effective measures like response time, average time spend in the system, utilization and throughput.
Energy Technology Data Exchange (ETDEWEB)
Cline, M.C.
1981-08-01
VNAP2 is a computer program for calculating turbulent (as well as laminar and inviscid), steady, and unsteady flow. VNAP2 solves the two-dimensional, time-dependent, compressible Navier-Stokes equations. The turbulence is modeled with either an algebraic mixing-length model, a one-equation model, or the Jones-Launder two-equation model. The geometry may be a single- or a dual-flowing stream. The interior grid points are computed using the unsplit MacCormack scheme. Two options to speed up the calculations for high Reynolds number flows are included. The boundary grid points are computed using a reference-plane-characteristic scheme with the viscous terms treated as source functions. An explicit artificial viscosity is included for shock computations. The fluid is assumed to be a perfect gas. The flow boundaries may be arbitrary curved solid walls, inflow/outflow boundaries, or free-jet envelopes. Typical problems that can be solved concern nozzles, inlets, jet-powered afterbodies, airfoils, and free-jet expansions. The accuracy and efficiency of the program are shown by calculations of several inviscid and turbulent flows. The program and its use are described completely, and six sample cases and a code listing are included.
I. Fisk
2011-01-01
Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion The Tier 0 infrastructure was able to repack and promptly reconstruct heavy-ion collision data. Two copies were made of the data at CERN using a large CASTOR disk pool, and the core physics sample was replicated ...
Time expenditure in computer aided time studies implemented for highly mechanized forest equipment
Directory of Open Access Journals (Sweden)
Elena Camelia Mușat
2016-06-01
Full Text Available Time studies represent important tools that are used in forest operations research to produce empirical models or to comparatively assess the performance of two or more operational alternatives with the general aim to predict the performance of operational behavior, choose the most adequate equipment or eliminate the useless time. There is a long tradition in collecting the needed data in a traditional fashion, but this approach has its limitations, and it is likely that in the future the use of professional software would be extended is such preoccupations as this kind of tools have been already implemented. However, little to no information is available in what concerns the performance of data analyzing tasks when using purpose-built professional time studying software in such research preoccupations, while the resources needed to conduct time studies, including here the time may be quite intensive. Our study aimed to model the relations between the variation of time needed to analyze the video-recorded time study data and the variation of some measured independent variables for a complex organization of a work cycle. The results of our study indicate that the number of work elements which were separated within a work cycle as well as the delay-free cycle time and the software functionalities that were used during data analysis, significantly affected the time expenditure needed to analyze the data (α=0.01, p<0.01. Under the conditions of this study, where the average duration of a work cycle was of about 48 seconds and the number of separated work elements was of about 14, the speed that was usedto replay the video files significantly affected the mean time expenditure which averaged about 273 seconds for half of the real speed and about 192 seconds for an analyzing speed that equaled the real speed. We argue that different study designs as well as the parameters used within the software are likely to produce
A computational approach to real-time image processing for serial time-encoded amplified microscopy
Oikawa, Minoru; Hiyama, Daisuke; Hirayama, Ryuji; Hasegawa, Satoki; Endo, Yutaka; Sugie, Takahisa; Tsumura, Norimichi; Kuroshima, Mai; Maki, Masanori; Okada, Genki; Lei, Cheng; Ozeki, Yasuyuki; Goda, Keisuke; Shimobaba, Tomoyoshi
2016-03-01
High-speed imaging is an indispensable technique, particularly for identifying or analyzing fast-moving objects. The serial time-encoded amplified microscopy (STEAM) technique was proposed to enable us to capture images with a frame rate 1,000 times faster than using conventional methods such as CCD (charge-coupled device) cameras. The application of this high-speed STEAM imaging technique to a real-time system, such as flow cytometry for a cell-sorting system, requires successively processing a large number of captured images with high throughput in real time. We are now developing a high-speed flow cytometer system including a STEAM camera. In this paper, we describe our approach to processing these large amounts of image data in real time. We use an analog-to-digital converter that has up to 7.0G samples/s and 8-bit resolution for capturing the output voltage signal that involves grayscale images from the STEAM camera. Therefore the direct data output from the STEAM camera generates 7.0G byte/s continuously. We provided a field-programmable gate array (FPGA) device as a digital signal pre-processor for image reconstruction and finding objects in a microfluidic channel with high data rates in real time. We also utilized graphics processing unit (GPU) devices for accelerating the calculation speed of identification of the reconstructed images. We built our prototype system, which including a STEAM camera, a FPGA device and a GPU device, and evaluated its performance in real-time identification of small particles (beads), as virtual biological cells, owing through a microfluidic channel.
Computer Evaluation Of Real-Time X-Ray And Acoustic Images
Jacoby, M. H.; Loe, R. S.; Dondes, P. A.
1983-03-01
The weakest link in the inspection process is the subjective interpretation of data by inspectors. To overcome this troublesome fact computer based analysis systems have been developed. In the field of nondestructive evaluation (NDE) there is a large class of inspections that can benefit from computer analysis. X-ray images (both film and fluoroscopic) and acoustic images lend themselves to automatic analysis as do the one-dimensional signals associated with ultrasonic, eddy current and acoustic emission testing. Computer analysis can enhance and evaluate subtle details. Flaws can be located and measured, and accept-ance decisions made by computer in a consistent and objective manner. This paper describes the interactive, computer-based analysis of real-time x-ray images and acoustic images of graphite/epoxy adhesively bonded structures.
Continuous-variable quantum computing in optical time-frequency modes using quantum memories.
Humphreys, Peter C; Kolthammer, W Steven; Nunn, Joshua; Barbieri, Marco; Datta, Animesh; Walmsley, Ian A
2014-09-26
We develop a scheme for time-frequency encoded continuous-variable cluster-state quantum computing using quantum memories. In particular, we propose a method to produce, manipulate, and measure two-dimensional cluster states in a single spatial mode by exploiting the intrinsic time-frequency selectivity of Raman quantum memories. Time-frequency encoding enables the scheme to be extremely compact, requiring a number of memories that are a linear function of only the number of different frequencies in which the computational state is encoded, independent of its temporal duration. We therefore show that quantum memories can be a powerful component for scalable photonic quantum information processing architectures.
Reducing the throughput time of the diagnostic track involving CT scanning with computer simulation
Energy Technology Data Exchange (ETDEWEB)
Lent, Wineke A.M. van, E-mail: w.v.lent@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); University of Twente, IGS Institute for Innovation and Governance Studies, Department of Health Technology Services Research (HTSR), Enschede (Netherlands); Deetman, Joost W., E-mail: j.deetman@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Teertstra, H. Jelle, E-mail: h.teertstra@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Muller, Sara H., E-mail: s.muller@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Hans, Erwin W., E-mail: e.w.hans@utwente.nl [University of Twente, School of Management and Governance, Dept. of Industrial Engineering and Business Intelligence Systems, Enschede (Netherlands); Harten, Wim H. van, E-mail: w.v.harten@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); University of Twente, IGS Institute for Innovation and Governance Studies, Department of Health Technology Services Research (HTSR), Enschede (Netherlands)
2012-11-15
Introduction: To examine the use of computer simulation to reduce the time between the CT request and the consult in which the CT report is discussed (diagnostic track) while restricting idle time and overtime. Methods: After a pre implementation analysis in our case study hospital, by computer simulation three scenarios were evaluated on access time, overtime and idle time of the CT; after implementation these same aspects were evaluated again. Effects on throughput time were measured for outpatient short-term and urgent requests only. Conclusion: The pre implementation analysis showed an average CT access time of 9.8 operating days and an average diagnostic track of 14.5 operating days. Based on the outcomes of the simulation, management changed the capacity for the different patient groups to facilitate a diagnostic track of 10 operating days, with a CT access time of 7 days. After the implementation of changes, the average diagnostic track duration was 12.6 days with an average CT access time of 7.3 days. The fraction of patients with a total throughput time within 10 days increased from 29% to 44% while the utilization remained equal with 82%, the idle time increased by 11% and the overtime decreased by 82%. The fraction of patients that completed the diagnostic track within 10 days improved with 52%. Computer simulation proved useful for studying the effects of proposed scenarios in radiology management. Besides the tangible effects, the simulation increased the awareness that optimizing capacity allocation can reduce access times.
Goings, Joshua J; Li, Xiaosong
2016-06-21
One of the challenges of interpreting electronic circular dichroism (ECD) band spectra is that different states may have different rotatory strength signs, determined by their absolute configuration. If the states are closely spaced and opposite in sign, observed transitions may be washed out by nearby states, unlike absorption spectra where transitions are always positive additive. To accurately compute ECD bands, it is necessary to compute a large number of excited states, which may be prohibitively costly if one uses the linear-response time-dependent density functional theory (TDDFT) framework. Here we implement a real-time, atomic-orbital based TDDFT method for computing the entire ECD spectrum simultaneously. The method is advantageous for large systems with a high density of states. In contrast to previous implementations based on real-space grids, the method is variational, independent of nuclear orientation, and does not rely on pseudopotential approximations, making it suitable for computation of chiroptical properties well into the X-ray regime.
Local algorithm for computing complex travel time based on the complex eikonal equation.
Huang, Xingguo; Sun, Jianguo; Sun, Zhangqing
2016-04-01
The traditional algorithm for computing the complex travel time, e.g., dynamic ray tracing method, is based on the paraxial ray approximation, which exploits the second-order Taylor expansion. Consequently, the computed results are strongly dependent on the width of the ray tube and, in regions with dramatic velocity variations, it is difficult for the method to account for the velocity variations. When solving the complex eikonal equation, the paraxial ray approximation can be avoided and no second-order Taylor expansion is required. However, this process is time consuming. In this case, we may replace the global computation of the whole model with local computation by taking both sides of the ray as curved boundaries of the evanescent wave. For a given ray, the imaginary part of the complex travel time should be zero on the central ray. To satisfy this condition, the central ray should be taken as a curved boundary. We propose a nonuniform grid-based finite difference scheme to solve the curved boundary problem. In addition, we apply the limited-memory Broyden-Fletcher-Goldfarb-Shanno technology for obtaining the imaginary slowness used to compute the complex travel time. The numerical experiments show that the proposed method is accurate. We examine the effectiveness of the algorithm for the complex travel time by comparing the results with those from the dynamic ray tracing method and the Gauss-Newton Conjugate Gradient fast marching method.
Real-time data acquisition and feedback control using Linux Intel computers
Energy Technology Data Exchange (ETDEWEB)
Penaflor, B.G. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Ferron, J.R. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Piglowski, D.A. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Johnson, R.D. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States); Walker, M.L. [General Atomics, P.O. Box 85608, San Diego, CA 92186-5608 (United States)
2006-07-15
This paper describes the experiences of the DIII-D programming staff in adapting Linux based Intel computing hardware for use in real-time data acquisition and feedback control systems. Due to the highly dynamic and unstable nature of magnetically confined plasmas in tokamak fusion experiments, real-time data acquisition and feedback control systems are in routine use with all major tokamaks. At DIII-D, plasmas are created and sustained using a real-time application known as the digital plasma control system (PCS). During each experiment, the PCS periodically samples data from hundreds of diagnostic signals and provides these data to control algorithms implemented in software. These algorithms compute the necessary commands to send to various actuators that affect plasma performance. The PCS consists of a group of rack mounted Intel Xeon computer systems running an in-house customized version of the Linux operating system tailored specifically to meet the real-time performance needs of the plasma experiments. This paper provides a more detailed description of the real-time computing hardware and custom developed software, including recent work to utilize dual Intel Xeon equipped computers within the PCS.
Local algorithm for computing complex travel time based on the complex eikonal equation
Huang, Xingguo; Sun, Jianguo; Sun, Zhangqing
2016-04-01
The traditional algorithm for computing the complex travel time, e.g., dynamic ray tracing method, is based on the paraxial ray approximation, which exploits the second-order Taylor expansion. Consequently, the computed results are strongly dependent on the width of the ray tube and, in regions with dramatic velocity variations, it is difficult for the method to account for the velocity variations. When solving the complex eikonal equation, the paraxial ray approximation can be avoided and no second-order Taylor expansion is required. However, this process is time consuming. In this case, we may replace the global computation of the whole model with local computation by taking both sides of the ray as curved boundaries of the evanescent wave. For a given ray, the imaginary part of the complex travel time should be zero on the central ray. To satisfy this condition, the central ray should be taken as a curved boundary. We propose a nonuniform grid-based finite difference scheme to solve the curved boundary problem. In addition, we apply the limited-memory Broyden-Fletcher-Goldfarb-Shanno technology for obtaining the imaginary slowness used to compute the complex travel time. The numerical experiments show that the proposed method is accurate. We examine the effectiveness of the algorithm for the complex travel time by comparing the results with those from the dynamic ray tracing method and the Gauss-Newton Conjugate Gradient fast marching method.
Supporting Real-Time Computer Vision Workloads using OpenVX on Multicore+GPU Platforms
2015-05-01
workloads specified using OpenVX to be supported in a predictable way. I. INTRODUCTION In the automotive industry today, vision-based sensing through cameras...Supporting Real-Time Computer Vision Workloads using OpenVX on Multicore+GPU Platforms Glenn A. Elliott, Kecheng Yang, and James H. Anderson...Department of Computer Science, University of North Carolina at Chapel Hill Abstract—In the automotive industry, there is currently great interest in
Kajian dan Implementasi Real Time Operating System pada Single Board Computer Berbasis ARM
Wiedjaja; Handi Muljoredjo; Jonathan Lukas; Benyamin Christian; Luis Kristofel
2014-01-01
Operating System is an important software in computer system. For personal and office use the operating system is sufficient. However, to critical mission applications such as nuclear power plants and braking system on the car (auto braking system) which need a high level of reliability, it requires operating system which operates in real time. The study aims to assess the implementation of the Linux-based operating system on a Single Board Computer (SBC) ARM-based, namely Pandaboard ES with ...
Energy Technology Data Exchange (ETDEWEB)
Gurzadyan, V.G. [Sapienza University of Rome, SIA, Rome (Italy); Alikhanian National Laboratory and Yerevan State University, Center for Cosmology and Astrophysics, Yerevan (Armenia); Harutyunyan, V.V. [Alikhanian National Laboratory and Yerevan State University, Center for Cosmology and Astrophysics, Yerevan (Armenia); Kocharyan, A.A. [Alikhanian National Laboratory and Yerevan State University, Center for Cosmology and Astrophysics, Yerevan (Armenia); Monash University, School of Mathematical Sciences, Clayton (Australia)
2015-06-15
Extensive N-body simulations are among the key means for the study of numerous astrophysical and cosmological phenomena, so various schemes are developed for possibly higher accuracy computations. We demonstrate the principal possibility for revealing the evolution of a perturbed Hamiltonian system with an accuracy independent on time. The method is based on the Laplace transform and the derivation and analytical solution of an evolution equation in the phase space for the resolvent and using computer algebra. (orig.)
P. McBride
It has been a very active year for the computing project with strong contributions from members of the global community. The project has focused on site preparation and Monte Carlo production. The operations group has begun processing data from P5 as part of the global data commissioning. Improvements in transfer rates and site availability have been seen as computing sites across the globe prepare for large scale production and analysis as part of CSA07. Preparations for the upcoming Computing Software and Analysis Challenge CSA07 are progressing. Ian Fisk and Neil Geddes have been appointed as coordinators for the challenge. CSA07 will include production tests of the Tier-0 production system, reprocessing at the Tier-1 sites and Monte Carlo production at the Tier-2 sites. At the same time there will be a large analysis exercise at the Tier-2 centres. Pre-production simulation of the Monte Carlo events for the challenge is beginning. Scale tests of the Tier-0 will begin in mid-July and the challenge it...
M. Kasemann
Introduction During the past six months, Computing participated in the STEP09 exercise, had a major involvement in the October exercise and has been working with CMS sites on improving open issues relevant for data taking. At the same time operations for MC production, real data reconstruction and re-reconstructions and data transfers at large scales were performed. STEP09 was successfully conducted in June as a joint exercise with ATLAS and the other experiments. It gave good indication about the readiness of the WLCG infrastructure with the two major LHC experiments stressing the reading, writing and processing of physics data. The October Exercise, in contrast, was conducted as an all-CMS exercise, where Physics, Computing and Offline worked on a common plan to exercise all steps to efficiently access and analyze data. As one of the major results, the CMS Tier-2s demonstrated to be fully capable for performing data analysis. In recent weeks, efforts were devoted to CMS Computing readiness. All th...
M. Kasemann
CCRC’08 challenges and CSA08 During the February campaign of the Common Computing readiness challenges (CCRC’08), the CMS computing team had achieved very good results. The link between the detector site and the Tier0 was tested by gradually increasing the number of parallel transfer streams well beyond the target. Tests covered the global robustness at the Tier0, processing a massive number of very large files and with a high writing speed to tapes. Other tests covered the links between the different Tiers of the distributed infrastructure and the pre-staging and reprocessing capacity of the Tier1’s: response time, data transfer rate and success rate for Tape to Buffer staging of files kept exclusively on Tape were measured. In all cases, coordination with the sites was efficient and no serious problem was found. These successful preparations prepared the ground for the second phase of the CCRC’08 campaign, in May. The Computing Software and Analysis challen...
I. Fisk
2012-01-01
Introduction Computing continued with a high level of activity over the winter in preparation for conferences and the start of the 2012 run. 2012 brings new challenges with a new energy, more complex events, and the need to make the best use of the available time before the Long Shutdown. We expect to be resource constrained on all tiers of the computing system in 2012 and are working to ensure the high-priority goals of CMS are not impacted. Heavy ions After a successful 2011 heavy-ion run, the programme is moving to analysis. During the run, the CAF resources were well used for prompt analysis. Since then in 2012 on average 200 job slots have been used continuously at Vanderbilt for analysis workflows. Operations Office As of 2012, the Computing Project emphasis has moved from commissioning to operation of the various systems. This is reflected in the new organisation structure where the Facilities and Data Operations tasks have been merged into a common Operations Office, which now covers everything ...
M. Kasemann
Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the co...
Challenges in reducing the computational time of QSTS simulations for distribution system analysis.
Energy Technology Data Exchange (ETDEWEB)
Deboever, Jeremiah [Georgia Inst. of Technology, Atlanta, GA (United States); Zhang, Xiaochen [Georgia Inst. of Technology, Atlanta, GA (United States); Reno, Matthew J. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Broderick, Robert Joseph [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Grijalva, Santiago [Georgia Inst. of Technology, Atlanta, GA (United States); Therrien, Francis [CME International T& D, St. Bruno, QC (Canada)
2017-06-01
The rapid increase in penetration of distributed energy resources on the electric power distribution system has created a need for more comprehensive interconnection modelling and impact analysis. Unlike conventional scenario - based studies , quasi - static time - series (QSTS) simulation s can realistically model time - dependent voltage controllers and the diversity of potential impacts that can occur at different times of year . However, to accurately model a distribution system with all its controllable devices, a yearlong simulation at 1 - second resolution is often required , which could take conventional computers a computational time of 10 to 120 hours when an actual unbalanced distribution feeder is modeled . This computational burden is a clear l imitation to the adoption of QSTS simulation s in interconnection studies and for determining optimal control solutions for utility operations . Our ongoing research to improve the speed of QSTS simulation has revealed many unique aspects of distribution system modelling and sequential power flow analysis that make fast QSTS a very difficult problem to solve. In this report , the most relevant challenges in reducing the computational time of QSTS simulations are presented: number of power flows to solve, circuit complexity, time dependence between time steps, multiple valid power flow solutions, controllable element interactions, and extensive accurate simulation analysis.
Directory of Open Access Journals (Sweden)
Shunji Sako
2014-08-01
Full Text Available Objectives: This study investigated the association between task-induced stress and fatigue by examining the cardiovascular responses of subjects using different mouse positions while operating a computer under time constraints. Material and Methods: The study was participated by 16 young, healthy men and examined the use of optical mouse devices affixed to laptop computers. Two mouse positions were investigated: (1 the distal position (DP, in which the subjects place their forearms on the desk accompanied by the abduction and flexion of their shoulder joints, and (2 the proximal position (PP, in which the subjects place only their wrists on the desk without using an armrest. The subjects continued each task for 16 min. We assessed differences in several characteristics according to mouse position, including expired gas values, autonomic nerve activities (based on cardiorespiratory responses, operating efficiencies (based on word counts, and fatigue levels (based on the visual analog scale – VAS. Results: Oxygen consumption (VO2, the ratio of inspiration time to respiration time (Ti/Ttotal, respiratory rate (RR, minute ventilation (VE, and the ratio of expiration to inspiration (Te/Ti were significantly lower when the participants were performing the task in the DP than those obtained in the PP. Tidal volume (VT, carbon dioxide output rates (VCO2/VE, and oxygen extraction fractions (VO2/VE were significantly higher for the DP than they were for the PP. No significant difference in VAS was observed between the positions; however, as the task progressed, autonomic nerve activities were lower and operating efficiencies were significantly higher for the DP than they were for the PP. Conclusions: Our results suggest that the DP has fewer effects on cardiorespiratory functions, causes lower levels of sympathetic nerve activity and mental stress, and produces a higher total workload than the PP. This suggests that the DP is preferable to the PP when
Spatiotemporal Domain Decomposition for Massive Parallel Computation of Space-Time Kernel Density
Hohl, A.; Delmelle, E. M.; Tang, W.
2015-07-01
Accelerated processing capabilities are deemed critical when conducting analysis on spatiotemporal datasets of increasing size, diversity and availability. High-performance parallel computing offers the capacity to solve computationally demanding problems in a limited timeframe, but likewise poses the challenge of preventing processing inefficiency due to workload imbalance between computing resources. Therefore, when designing new algorithms capable of implementing parallel strategies, careful spatiotemporal domain decomposition is necessary to account for heterogeneity in the data. In this study, we perform octtree-based adaptive decomposition of the spatiotemporal domain for parallel computation of space-time kernel density. In order to avoid edge effects near subdomain boundaries, we establish spatiotemporal buffers to include adjacent data-points that are within the spatial and temporal kernel bandwidths. Then, we quantify computational intensity of each subdomain to balance workloads among processors. We illustrate the benefits of our methodology using a space-time epidemiological dataset of Dengue fever, an infectious vector-borne disease that poses a severe threat to communities in tropical climates. Our parallel implementation of kernel density reaches substantial speedup compared to sequential processing, and achieves high levels of workload balance among processors due to great accuracy in quantifying computational intensity. Our approach is portable of other space-time analytical tests.
Computational modeling and real-time control of patient-specific laser treatment of cancer.
Fuentes, D; Oden, J T; Diller, K R; Hazle, J D; Elliott, A; Shetty, A; Stafford, R J
2009-04-01
An adaptive feedback control system is presented which employs a computational model of bioheat transfer in living tissue to guide, in real-time, laser treatments of prostate cancer monitored by magnetic resonance thermal imaging. The system is built on what can be referred to as cyberinfrastructure-a complex structure of high-speed network, large-scale parallel computing devices, laser optics, imaging, visualizations, inverse-analysis algorithms, mesh generation, and control systems that guide laser therapy to optimally control the ablation of cancerous tissue. The computational system has been successfully tested on in vivo, canine prostate. Over the course of an 18 min laser-induced thermal therapy performed at M.D. Anderson Cancer Center (MDACC) in Houston, Texas, the computational models were calibrated to intra-operative real-time thermal imaging treatment data and the calibrated models controlled the bioheat transfer to within 5 degrees C of the predetermined treatment plan. The computational arena is in Austin, Texas and managed at the Institute for Computational Engineering and Sciences (ICES). The system is designed to control the bioheat transfer remotely while simultaneously providing real-time remote visualization of the on-going treatment. Post-operative histology of the canine prostate reveal that the damage region was within the targeted 1.2 cm diameter treatment objective.
Frantzeskakis, Emmanuel N.; Baras, John S.; Liu, Kuo Juey R.
1993-11-01
In this paper, we establish an architectural framework for parallel time-recursive computation. We consider a class of linear operators that consists of the discrete time, time invariant, compactly supported, but otherwise arbitrary kernel functions. We specify the properties of the linear operators that can be implemented efficiently in a time-recursive way. Based on these properties, we develop a routine that produces a time-recursive architectural implementation for a given operator. This routine is instructive for the design of a CAD tool that will facilitate the architecture derivation. Using this background, we design an architecture for the Modulated Lapped Transform (commonly called Modified Discrete Cosine Transform), which has linear cost in operator counts.
Stepnicka, M.; Cortez, Paulo; Peralta Donate, Juan; Stepnickova, Lenka
2013-01-01
Accurate time series forecasting is a key issue to support individual and or- ganizational decision making. In this paper, we introduce novel methods for multi-step seasonal time series forecasting. All the presented methods stem from computational intelligence techniques: evolutionary artificial neu- ral networks, support vector machines and genuine linguistic fuzzy rules. Performance of the suggested methods is experimentally justified on sea- sonal time series from distinct domains on thre...
Universal adiabatic quantum computation via the space-time circuit-to-Hamiltonian construction.
Gosset, David; Terhal, Barbara M; Vershynina, Anna
2015-04-10
We show how to perform universal adiabatic quantum computation using a Hamiltonian which describes a set of particles with local interactions on a two-dimensional grid. A single parameter in the Hamiltonian is adiabatically changed as a function of time to simulate the quantum circuit. We bound the eigenvalue gap above the unique ground state by mapping our model onto the ferromagnetic XXZ chain with kink boundary conditions; the gap of this spin chain was computed exactly by Koma and Nachtergaele using its q-deformed version of SU(2) symmetry. We also discuss a related time-independent Hamiltonian which was shown by Janzing to be capable of universal computation. We observe that in the limit of large system size, the time evolution is equivalent to the exactly solvable quantum walk on Young's lattice.
The reliable solution and computation time of variable parameters Logistic model
Pengfei, Wang
2016-01-01
The reliable computation time (RCT, marked as Tc) when applying a double precision computation of a variable parameters logistic map (VPLM) is studied. First, using the method proposed, the reliable solutions for the logistic map are obtained. Second, for a time-dependent non-stationary parameters VPLM, 10000 samples of reliable experiments are constructed, and the mean Tc is then computed. The results indicate that for each different initial value, the Tcs of the VPLM are generally different. However, the mean Tc trends to a constant value once the sample number is large enough. The maximum, minimum and probable distribution function of Tc is also obtained, which can help us to identify the robustness of applying a nonlinear time series theory to forecasting while using the VPLM output. In addition, the Tc of the fixed parameter experiments of the logistic map was obtained, and the results suggested that this Tc matches the theoretical formula predicted value.
Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure
Wang, Henry; Ma, Yunzhi; Pratx, Guillem; Xing, Lei
2011-09-01
Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47× speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed. This work was presented in part at the 2010 Annual Meeting of the American Association of Physicists in Medicine (AAPM), Philadelphia, PA.
Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure
Energy Technology Data Exchange (ETDEWEB)
Wang, Henry [Department of Electrical Engineering, Stanford University, Stanford, CA 94305 (United States); Ma Yunzhi; Pratx, Guillem; Xing Lei, E-mail: hwang41@stanford.edu [Department of Radiation Oncology, Stanford University School of Medicine, Stanford, CA 94305-5847 (United States)
2011-09-07
Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47x speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed. (note)
Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure.
Wang, Henry; Ma, Yunzhi; Pratx, Guillem; Xing, Lei
2011-09-07
Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47× speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed.
Mayorga, René V; Carrera, Jonathan
2007-06-01
This Paper presents an efficient approach for the fast computation of inverse continuous time variant functions with the proper use of Radial Basis Function Networks (RBFNs). The approach is based on implementing RBFNs for computing inverse continuous time variant functions via an overall damped least squares solution that includes a novel null space vector for singularities prevention. The singularities avoidance null space vector is derived from developing a sufficiency condition for singularities prevention that conduces to establish some characterizing matrices and an associated performance index.
Real Time Efficient Scheduling Algorithm for Load Balancing in Fog Computing Environment
Directory of Open Access Journals (Sweden)
Manisha Verma
2016-04-01
Full Text Available Cloud computing is the new era technology, which is entirely dependent on the internet to maintain large applications, where data is shared over one platform to provide better services to clients belonging to a different organization. It ensures maximum utilization of computational resources by making availability of data, software and infrastructure with lower cost in a secure, reliable and flexible manner. Though cloud computing offers many advantages, but it suffers from certain limitation too, that during load balancing of data in cloud data centers the internet faces problems of network congestion, less bandwidth utilization, fault tolerance and security etc. To get rid out of this issue new computing model called Fog Computing is introduced which easily transfer sensitive data without delaying to distributed devices. Fog is similar to the cloud only difference lies in the fact that it is located more close to end users to process and give response to the client in less time. Secondly, it is beneficial to the real time streaming applications, sensor networks, Internet of things which need high speed and reliable internet connectivity. Our proposed architecture introduced a new scheduling policy for load balancing in Fog Computing environment, which complete real tasks within deadline, increase throughput and network utilization, maintaining data consistency with less complexity to meet the present day demand of end users
Real Time Multiple Hand Gesture Recognition System for Human Computer Interaction
Directory of Open Access Journals (Sweden)
Siddharth S. Rautaray
2012-05-01
Full Text Available With the increasing use of computing devices in day to day life, the need of user friendly interfaces has lead towards the evolution of different types of interfaces for human computer interaction. Real time vision based hand gesture recognition affords users the ability to interact with computers in more natural and intuitive ways. Direct use of hands as an input device is an attractive method which can communicate much more information by itself in comparison to mice, joysticks etc allowing a greater number of recognition system that can be used in a variety of human computer interaction applications. The gesture recognition system consist of three main modules like hand segmentation, hand tracking and gesture recognition from hand features. The designed system further integrated with different applications like image browser, virtual game etc. possibilities for human computer interaction. Computer Vision based systems has the potential to provide more natural, non-contact solutions. The present research work focuses on to design and develops a practical framework for real time hand gesture.
Fijany, Amir (Inventor); Bejczy, Antal K. (Inventor)
1993-01-01
This is a real-time robotic controller and simulator which is a MIMD-SIMD parallel architecture for interfacing with an external host computer and providing a high degree of parallelism in computations for robotic control and simulation. It includes a host processor for receiving instructions from the external host computer and for transmitting answers to the external host computer. There are a plurality of SIMD microprocessors, each SIMD processor being a SIMD parallel processor capable of exploiting fine grain parallelism and further being able to operate asynchronously to form a MIMD architecture. Each SIMD processor comprises a SIMD architecture capable of performing two matrix-vector operations in parallel while fully exploiting parallelism in each operation. There is a system bus connecting the host processor to the plurality of SIMD microprocessors and a common clock providing a continuous sequence of clock pulses. There is also a ring structure interconnecting the plurality of SIMD microprocessors and connected to the clock for providing the clock pulses to the SIMD microprocessors and for providing a path for the flow of data and instructions between the SIMD microprocessors. The host processor includes logic for controlling the RRCS by interpreting instructions sent by the external host computer, decomposing the instructions into a series of computations to be performed by the SIMD microprocessors, using the system bus to distribute associated data among the SIMD microprocessors, and initiating activity of the SIMD microprocessors to perform the computations on the data by procedure call.
Babaee, Hessam; Farazmand, Mohamad; Haller, George; Sapsis, Themistoklis P.
2017-06-01
High-dimensional chaotic dynamical systems can exhibit strongly transient features. These are often associated with instabilities that have a finite-time duration. Because of the finite-time character of these transient events, their detection through infinite-time methods, e.g., long term averages, Lyapunov exponents or information about the statistical steady-state, is not possible. Here, we utilize a recently developed framework, the Optimally Time-Dependent (OTD) modes, to extract a time-dependent subspace that spans the modes associated with transient features associated with finite-time instabilities. As the main result, we prove that the OTD modes, under appropriate conditions, converge exponentially fast to the eigendirections of the Cauchy-Green tensor associated with the most intense finite-time instabilities. Based on this observation, we develop a reduced-order method for the computation of finite-time Lyapunov exponents (FTLE) and vectors. In high-dimensional systems, the computational cost of the reduced-order method is orders of magnitude lower than the full FTLE computation. We demonstrate the validity of the theoretical findings on two numerical examples.
Babaee, Hessam; Farazmand, Mohamad; Haller, George; Sapsis, Themistoklis P
2017-06-01
High-dimensional chaotic dynamical systems can exhibit strongly transient features. These are often associated with instabilities that have a finite-time duration. Because of the finite-time character of these transient events, their detection through infinite-time methods, e.g., long term averages, Lyapunov exponents or information about the statistical steady-state, is not possible. Here, we utilize a recently developed framework, the Optimally Time-Dependent (OTD) modes, to extract a time-dependent subspace that spans the modes associated with transient features associated with finite-time instabilities. As the main result, we prove that the OTD modes, under appropriate conditions, converge exponentially fast to the eigendirections of the Cauchy-Green tensor associated with the most intense finite-time instabilities. Based on this observation, we develop a reduced-order method for the computation of finite-time Lyapunov exponents (FTLE) and vectors. In high-dimensional systems, the computational cost of the reduced-order method is orders of magnitude lower than the full FTLE computation. We demonstrate the validity of the theoretical findings on two numerical examples.
Computer-games for gravitational wave science outreach: Black Hole Pong and Space Time Quest
Carbone, L.; Bond, C.; Brown, D.; Brückner, F.; Grover, K.; Lodhia, D.; Mingarelli, C. M. F.; Fulda, P.; Smith, R. J. E.; Unwin, R.; Vecchio, A.; Wang, M.; Whalley, L.; Freise, A.
2012-06-01
We have established a program aimed at developing computer applications and web applets to be used for educational purposes as well as gravitational wave outreach activities. These applications and applets teach gravitational wave physics and technology. The computer programs are generated in collaboration with undergraduates and summer students as part of our teaching activities, and are freely distributed on a dedicated website. As part of this program, we have developed two computer-games related to gravitational wave science: 'Black Hole Pong' and 'Space Time Quest'. In this article we present an overview of our computer related outreach activities and discuss the games and their educational aspects, and report on some positive feedback received.
Computational Procedures for a Class of GI/D/k Systems in Discrete Time
Directory of Open Access Journals (Sweden)
Md. Mostafizur Rahman
2009-01-01
Full Text Available A class of discrete time GI/D/k systems is considered for which the interarrival times have finite support and customers are served in first-in first-out (FIFO order. The system is formulated as a single server queue with new general independent interarrival times and constant service duration by assuming cyclic assignment of customers to the identical servers. Then the queue length is set up as a quasi-birth-death (QBD type Markov chain. It is shown that this transformed GI/D/1 system has special structures which make the computation of the matrix R simple and efficient, thereby reducing the number of multiplications in each iteration significantly. As a result we were able to keep the computation time very low. Moreover, use of the resulting structural properties makes the computation of the distribution of queue length of the transformed system efficient. The computation of the distribution of waiting time is also shown to be simple by exploiting the special structures.
Directory of Open Access Journals (Sweden)
Dimitris Kugiumtzis
2010-02-01
Full Text Available In many applications, such as physiology and finance, large time series data bases are to be analyzed requiring the computation of linear, nonlinear and other measures. Such measures have been developed and implemented in commercial and freeware softwares rather selectively and independently. The Measures of Analysis of Time Series (MATS MATLAB toolkit is designed to handle an arbitrary large set of scalar time series and compute a large variety of measures on them, allowing for the specification of varying measure parameters as well. The variety of options with added facilities for visualization of the results support different settings of time series analysis, such as the detection of dynamics changes in long data records, resampling (surrogate or bootstrap tests for independence and linearity with various test statistics, and discrimination power of different measures and for different combinations of their parameters. The basic features of MATS are presented and the implemented measures are briefly described. The usefulness of MATS is illustrated on some empirical examples along with screenshots.
Reservoir computing with a single time-delay autonomous Boolean node
Haynes, Nicholas D.; Soriano, Miguel C.; Rosin, David P.; Fischer, Ingo; Gauthier, Daniel J.
2015-02-01
We demonstrate reservoir computing with a physical system using a single autonomous Boolean logic element with time-delay feedback. The system generates a chaotic transient with a window of consistency lasting between 30 and 300 ns, which we show is sufficient for reservoir computing. We then characterize the dependence of computational performance on system parameters to find the best operating point of the reservoir. When the best parameters are chosen, the reservoir is able to classify short input patterns with performance that decreases over time. In particular, we show that four distinct input patterns can be classified for 70 ns, even though the inputs are only provided to the reservoir for 7.5 ns.
Real-time dynamics of lattice gauge theories with a few-qubit quantum computer
Martinez, E A; Schindler, P; Nigg, D; Erhard, A; Heyl, M; Hauke, P; Dalmonte, M; Monz, T; Zoller, P; Blatt, R
2016-01-01
Gauge theories are fundamental to our understanding of interactions between the elementary constituents of matter as mediated by gauge bosons. However, computing the real-time dynamics in gauge theories is a notorious challenge for classical computational methods. In the spirit of Feynman's vision of a quantum simulator, this has recently stimulated theoretical effort to devise schemes for simulating such theories on engineered quantum-mechanical devices, with the difficulty that gauge invariance and the associated local conservation laws (Gauss laws) need to be implemented. Here we report the first experimental demonstration of a digital quantum simulation of a lattice gauge theory, by realising 1+1-dimensional quantum electrodynamics (Schwinger model) on a few-qubit trapped-ion quantum computer. We are interested in the real-time evolution of the Schwinger mechanism, describing the instability of the bare vacuum due to quantum fluctuations, which manifests itself in the spontaneous creation of electron-posi...
Tuning Time-Domain Pseudospectral Computations of the Self-Force on a Charged Scalar Particle
Canizares, Priscilla
2011-01-01
The computation of the self-force constitutes one of the main challenges for the construction of precise theoretical waveform templates in order to detect and analyze extreme-mass-ratio inspirals with the future space-based gravitational-wave observatory LISA. Since the number of templates required is quite high, it is important to develop fast algorithms both for the computation of the self-force and the production of waveforms. In this article we show how to tune a recent time-domain technique for the computation of the self-force, what we call the Particle without Particle scheme, in order to make it very precise and at the same time very efficient. We also extend this technique in order to allow for highly eccentric orbits.
Tuning time-domain pseudospectral computations of the self-force on a charged scalar particle
Energy Technology Data Exchange (ETDEWEB)
Canizares, Priscilla; Sopuerta, Carlos F, E-mail: pcm@ieec.uab.es, E-mail: sopuerta@ieec.uab.es [Facultat de Ciencies, Institut de Ciencies de I' Espai (CSIC-IEEC), Campus UAB, Torre C5 parells, Bellaterra, 08193 Barcelona (Spain)
2011-07-07
The computation of the self-force constitutes one of the main challenges for the construction of precise theoretical waveform templates in order to detect and analyze extreme-mass-ratio inspirals with the future space-based gravitational-wave observatory LISA. Since the number of templates required is quite high, it is important to develop fast algorithms both for the computation of the self-force and the production of waveforms. In this paper, we show how to tune a recent time-domain technique for the computation of the self-force, what we call the particle without particle scheme, in order to make it very precise and at the same time very efficient. We also extend this technique in order to allow for highly eccentric orbits.
I. Fisk
2012-01-01
Introduction Computing activity has been running at a sustained, high rate as we collect data at high luminosity, process simulation, and begin to process the parked data. The system is functional, though a number of improvements are planned during LS1. Many of the changes will impact users, we hope only in positive ways. We are trying to improve the distributed analysis tools as well as the ability to access more data samples more transparently. Operations Office Figure 2: Number of events per month, for 2012 Since the June CMS Week, Computing Operations teams successfully completed data re-reconstruction passes and finished the CMSSW_53X MC campaign with over three billion events available in AOD format. Recorded data was successfully processed in parallel, exceeding 1.2 billion raw physics events per month for the first time in October 2012 due to the increase in data-parking rate. In parallel, large efforts were dedicated to WMAgent development and integrati...
2010-01-01
Introduction Just two months after the “LHC First Physics” event of 30th March, the analysis of the O(200) million 7 TeV collision events in CMS accumulated during the first 60 days is well under way. The consistency of the CMS computing model has been confirmed during these first weeks of data taking. This model is based on a hierarchy of use-cases deployed between the different tiers and, in particular, the distribution of RECO data to T1s, who then serve data on request to T2s, along a topology known as “fat tree”. Indeed, during this period this model was further extended by almost full “mesh” commissioning, meaning that RECO data were shipped to T2s whenever possible, enabling additional physics analyses compared with the “fat tree” model. Computing activities at the CMS Analysis Facility (CAF) have been marked by a good time response for a load almost evenly shared between ALCA (Alignment and Calibration tasks - highest p...
Efficient Geo-Computational Algorithms for Constructing Space-Time Prisms in Road Networks
Directory of Open Access Journals (Sweden)
Hui-Ping Chen
2016-11-01
Full Text Available The Space-time prism (STP is a key concept in time geography for analyzing human activity-travel behavior under various Space-time constraints. Most existing time-geographic studies use a straightforward algorithm to construct STPs in road networks by using two one-to-all shortest path searches. However, this straightforward algorithm can introduce considerable computational overhead, given the fact that accessible links in a STP are generally a small portion of the whole network. To address this issue, an efficient geo-computational algorithm, called NTP-A*, is proposed. The proposed NTP-A* algorithm employs the A* and branch-and-bound techniques to discard inaccessible links during two shortest path searches, and thereby improves the STP construction performance. Comprehensive computational experiments are carried out to demonstrate the computational advantage of the proposed algorithm. Several implementation techniques, including the label-correcting technique and the hybrid link-node labeling technique, are discussed and analyzed. Experimental results show that the proposed NTP-A* algorithm can significantly improve STP construction performance in large-scale road networks by a factor of 100, compared with existing algorithms.
A computationally simple and robust method to detect determinism in a time series
DEFF Research Database (Denmark)
Lu, Sheng; Ju, Ki Hwan; Kanters, Jørgen K.;
2006-01-01
We present a new, simple, and fast computational technique, termed the incremental slope (IS), that can accurately distinguish between deterministic from stochastic systems even when the variance of noise is as large or greater than the signal, and remains robust for time-varying signals. The IS ...
Computation of non-monotonic Lyapunov functions for continuous-time systems
Li, Huijuan; Liu, AnPing
2017-09-01
In this paper, we propose two methods to compute non-monotonic Lyapunov functions for continuous-time systems which are asymptotically stable. The first method is to solve a linear optimization problem on a compact and bounded set. The proposed linear programming based algorithm delivers a CPA1
A sub-cubic time algorithm for computing the quartet distance between two general trees
DEFF Research Database (Denmark)
Nielsen, Jesper; Kristensen, Anders Kabell; Mailund, Thomas;
2011-01-01
derived a new algorithm for computing the quartet distance between a pair of general trees, i.e. trees where inner nodes can have any degree ≥ 3. The time and space complexity of our algorithm is sub-cubic in the number of leaves and does not depend on the degree of the inner nodes. This makes...
48 CFR 6302.6 - Computation and extension of time limits (Rule 6).
2010-10-01
... Saturday, Sunday, or a legal holiday, in which case the period runs to the end of the next business day. (b... 48 Federal Acquisition Regulations System 7 2010-10-01 2010-10-01 false Computation and extension of time limits (Rule 6). 6302.6 Section 6302.6 Federal Acquisition Regulations System DEPARTMENT OF...
Bekooij, Marco; Wiggers, Maarten; Meerbergen, van Jef; Falk, H.; Marwedel, P.
2007-01-01
Soft real-time applications that process data streams can often be intuitively described as dataflow process networks. In this paper we present a novel analysis technique to compute conservative estimates of the required buffer capacities in such process networks. With the same analysis technique sc
Fukuda, Masahiro; Ichikawa, Kazuhide; Tachibana, Akitomo
2016-01-01
We discuss the method to compute the integrals which appear in the retarded potential term for a real-time simulation based on QED (Quantum Electrodynamics). We show that the oscillatory integrals over the infinite interval involved in them can be efficiently performed by the method developed by Ooura and Mori based on the double exponential (DE) formula.
Directory of Open Access Journals (Sweden)
Abeer Hamdy
2010-10-01
Full Text Available The paper presents a heuristic algorithm to schedule real time indivisible loads represented as directed sequential task graph on a cluster computing. One of the cluster nodes has some special resources (denoted by special node that may be needed by one of the indivisible loads
Eckhardt, D. E., Jr.
1979-01-01
A model of a central processor (CPU) which services background applications in the presence of time critical activity is presented. The CPU is viewed as an M/M/1 queueing system subject to periodic interrupts by deterministic, time critical process. The Laplace transform of the distribution of service times for the background applications is developed. The use of state of the art queueing models for studying the background processing capability of time critical computer systems is discussed and the results of a model validation study which support this application of queueing models are presented.
Comparison of computer codes for estimates of the symmetric coupled bunch instabilities growth times
Angal-Kalinin, Deepa
2002-01-01
The standard computer codes used for estimating the growth times of the symmetric coupled bunch instabilities are ZAP and BBI.The code Vlasov was earlier used for the LHC for the estimates of the coupled bunch instabilities growth time[1]. The results obtained by these three codes have been compared and the options under which their results can be compared are discussed. The differences in the input and the output for these three codes are given for a typical case.
Computing the Quartet Distance Between Evolutionary Trees in Time O(n log n)
DEFF Research Database (Denmark)
Brodal, Gerth Sølfting; Fagerberg, Rolf; Pedersen, Christian Nørgaard Storm
2003-01-01
Morris, and Meacham. The quartet distance between two unrooted evolutionary trees is the number of quartet topology differences between the two trees, where a quartet topology is the topological subtree induced by four species. In this paper we present an algorithm for computing the quartet distance between two...... unrooted evolutionary trees of n species, where all internal nodes have degree three, in time O(n log n. The previous best algorithm for the problem uses time O(n 2)....
Dynamic acoustics for the STAR-100. [computer algorithms for time dependent sound waves in jet
Bayliss, A.; Turkel, E.
1979-01-01
An algorithm is described to compute time dependent acoustic waves in a jet. The method differs from previous methods in that no harmonic time dependence is assumed, thus permitting the study of nonharmonic acoustical behavior. Large grids are required to resolve the acoustic waves. Since the problem is nonstiff, explicit high order schemes can be used. These have been adapted to the STAR-100 with great efficiencies and permitted the efficient solution of problems which would not be feasible on a scalar machine.
I. Fisk
2011-01-01
Introduction The Computing Team successfully completed the storage, initial processing, and distribution for analysis of proton-proton data in 2011. There are still a variety of activities ongoing to support winter conference activities and preparations for 2012. Heavy ions The heavy-ion run for 2011 started in early November and has already demonstrated good machine performance and success of some of the more advanced workflows planned for 2011. Data collection will continue until early December. Facilities and Infrastructure Operations Operational and deployment support for WMAgent and WorkQueue+Request Manager components, routinely used in production by Data Operations, are provided. The GlideInWMS and components installation are now deployed at CERN, which is added to the GlideInWMS factory placed in the US. There has been new operational collaboration between the CERN team and the UCSD GlideIn factory operators, covering each others time zones by monitoring/debugging pilot jobs sent from the facto...
Reducing the throughput time of the diagnostic track involving CT scanning with computer simulation.
van Lent, Wineke A M; Deetman, Joost W; Teertstra, H Jelle; Muller, Sara H; Hans, Erwin W; van Harten, Wim H
2012-11-01
To examine the use of computer simulation to reduce the time between the CT request and the consult in which the CT report is discussed (diagnostic track) while restricting idle time and overtime. After a pre implementation analysis in our case study hospital, by computer simulation three scenarios were evaluated on access time, overtime and idle time of the CT; after implementation these same aspects were evaluated again. Effects on throughput time were measured for outpatient short-term and urgent requests only. The pre implementation analysis showed an average CT access time of 9.8 operating days and an average diagnostic track of 14.5 operating days. Based on the outcomes of the simulation, management changed the capacity for the different patient groups to facilitate a diagnostic track of 10 operating days, with a CT access time of 7 days. After the implementation of changes, the average diagnostic track duration was 12.6 days with an average CT access time of 7.3 days. The fraction of patients with a total throughput time within 10 days increased from 29% to 44% while the utilization remained equal with 82%, the idle time increased by 11% and the overtime decreased by 82%. The fraction of patients that completed the diagnostic track within 10 days improved with 52%. Computer simulation proved useful for studying the effects of proposed scenarios in radiology management. Besides the tangible effects, the simulation increased the awareness that optimizing capacity allocation can reduce access times. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Real-time computer treatment of THz passive device images with the high image quality
Trofimov, Vyacheslav A.; Trofimov, Vladislav V.
2012-06-01
We demonstrate real-time computer code improving significantly the quality of images captured by the passive THz imaging system. The code is not only designed for a THz passive device: it can be applied to any kind of such devices and active THz imaging systems as well. We applied our code for computer processing of images captured by four passive THz imaging devices manufactured by different companies. It should be stressed that computer processing of images produced by different companies requires using the different spatial filters usually. The performance of current version of the computer code is greater than one image per second for a THz image having more than 5000 pixels and 24 bit number representation. Processing of THz single image produces about 20 images simultaneously corresponding to various spatial filters. The computer code allows increasing the number of pixels for processed images without noticeable reduction of image quality. The performance of the computer code can be increased many times using parallel algorithms for processing the image. We develop original spatial filters which allow one to see objects with sizes less than 2 cm. The imagery is produced by passive THz imaging devices which captured the images of objects hidden under opaque clothes. For images with high noise we develop an approach which results in suppression of the noise after using the computer processing and we obtain the good quality image. With the aim of illustrating the efficiency of the developed approach we demonstrate the detection of the liquid explosive, ordinary explosive, knife, pistol, metal plate, CD, ceramics, chocolate and other objects hidden under opaque clothes. The results demonstrate the high efficiency of our approach for the detection of hidden objects and they are a very promising solution for the security problem.
Time-Dependent Computed Tomographic Perfusion Thresholds for Patients With Acute Ischemic Stroke.
d'Esterre, Christopher D; Boesen, Mari E; Ahn, Seong Hwan; Pordeli, Pooneh; Najm, Mohamed; Minhas, Priyanka; Davari, Paniz; Fainardi, Enrico; Rubiera, Marta; Khaw, Alexander V; Zini, Andrea; Frayne, Richard; Hill, Michael D; Demchuk, Andrew M; Sajobi, Tolulope T; Forkert, Nils D; Goyal, Mayank; Lee, Ting Y; Menon, Bijoy K
2015-12-01
Among patients with acute ischemic stroke, we determine computed tomographic perfusion (CTP) thresholds associated with follow-up infarction at different stroke onset-to-CTP and CTP-to-reperfusion times. Acute ischemic stroke patients with occlusion on computed tomographic angiography were acutely imaged with CTP. Noncontrast computed tomography and magnectic resonance diffusion-weighted imaging between 24 and 48 hours were used to delineate follow-up infarction. Reperfusion was assessed on conventional angiogram or 4-hour repeat computed tomographic angiography. Tmax, cerebral blood flow, and cerebral blood volume derived from delay-insensitive CTP postprocessing were analyzed using receiver-operator characteristic curves to derive optimal thresholds for combined patient data (pooled analysis) and individual patients (patient-level analysis) based on time from stroke onset-to-CTP and CTP-to-reperfusion. One-way ANOVA and locally weighted scatterplot smoothing regression was used to test whether the derived optimal CTP thresholds were different by time. One hundred and thirty-two patients were included. Tmax thresholds of >16.2 and >15.8 s and absolute cerebral blood flow thresholds of stroke onset-to-CTP time and the optimal CTP thresholds for all parameters based on discrete or continuous time analysis (P>0.05). A statistically significant relationship existed between CTP-to-reperfusion time and the optimal thresholds for cerebral blood flow (P<0.001; r=0.59 and 0.77 for gray and white matter, respectively) and Tmax (P<0.001; r=-0.68 and -0.60 for gray and white matter, respectively) parameters. Optimal CTP thresholds associated with follow-up infarction depend on time from imaging to reperfusion. © 2015 American Heart Association, Inc.
Renormalization and Computation II: Time Cut-off and the Halting Problem
Manin, Yuri I
2009-01-01
This is the second installment to the project initiated in [Ma3]. In the first Part, I argued that both philosophy and technique of the perturbative renormalization in quantum field theory could be meaningfully transplanted to the theory of computation, and sketched several contexts supporting this view. In this second part, I address some of the issues raised in [Ma3] and provide their development in three contexts: a categorification of the algorithmic computations; time cut--off and Anytime Algorithms; and finally, a Hopf algebra renormalization of the Halting Problem.
Solution of 3-dimensional time-dependent viscous flows. Part 2: Development of the computer code
Weinberg, B. C.; Mcdonald, H.
1980-01-01
There is considerable interest in developing a numerical scheme for solving the time dependent viscous compressible three dimensional flow equations to aid in the design of helicopter rotors. The development of a computer code to solve a three dimensional unsteady approximate form of the Navier-Stokes equations employing a linearized block emplicit technique in conjunction with a QR operator scheme is described. Results of calculations of several Cartesian test cases are presented. The computer code can be applied to more complex flow fields such as these encountered on rotating airfoils.
An assessment of the real-time application capabilities of the SIFT computer system
Butler, R. W.
1982-01-01
The real-time capabilities of the SIFT computer system, a highly reliable multicomputer architecture developed to support the flight controls of a relaxed static stability aircraft, are discussed. The SIFT computer system was designed to meet extremely high reliability requirements and to facilitate a formal proof of its correctness. Although SIFT represents a significant achievement in fault-tolerant system research it presents an unusual and restrictive interface to its users. The characteristics of the user interface and its impact on application system design are assessed.
Theory and computation of disturbance invariant sets for discrete-time linear systems
Directory of Open Access Journals (Sweden)
Kolmanovsky Ilya
1998-01-01
Full Text Available This paper considers the characterization and computation of invariant sets for discrete-time, time-invariant, linear systems with disturbance inputs whose values are confined to a specified compact set but are otherwise unknown. The emphasis is on determining maximal disturbance-invariant sets X that belong to a specified subset Γ of the state space. Such d-invariant sets have important applications in control problems where there are pointwise-in-time state constraints of the form χ ( t ∈ Γ . One purpose of the paper is to unite and extend in a rigorous way disparate results from the prior literature. In addition there are entirely new results. Specific contributions include: exploitation of the Pontryagin set difference to clarify conceptual matters and simplify mathematical developments, special properties of maximal invariant sets and conditions for their finite determination, algorithms for generating concrete representations of maximal invariant sets, practical computational questions, extension of the main results to general Lyapunov stable systems, applications of the computational techniques to the bounding of state and output response. Results on Lyapunov stable systems are applied to the implementation of a logic-based, nonlinear multimode regulator. For plants with disturbance inputs and state-control constraints it enlarges the constraint-admissible domain of attraction. Numerical examples illustrate the various theoretical and computational results.
Joint Time-Frequency-Space Classification of EEG in a Brain-Computer Interface Application
Directory of Open Access Journals (Sweden)
Molina Gary N Garcia
2003-01-01
Full Text Available Brain-computer interface is a growing field of interest in human-computer interaction with diverse applications ranging from medicine to entertainment. In this paper, we present a system which allows for classification of mental tasks based on a joint time-frequency-space decorrelation, in which mental tasks are measured via electroencephalogram (EEG signals. The efficiency of this approach was evaluated by means of real-time experimentations on two subjects performing three different mental tasks. To do so, a number of protocols for visualization, as well as training with and without feedback, were also developed. Obtained results show that it is possible to obtain good classification of simple mental tasks, in view of command and control, after a relatively small amount of training, with accuracies around 80%, and in real time.
Time domain measurement representation in computer system diagnostics and performance analysis
Directory of Open Access Journals (Sweden)
Stanisław Wideł
2013-06-01
Full Text Available Time analysis is a common approach for testing and detecting methods for the performance analysis of computer systems. In the article it is shown, that measuring and identifying performances based on a benchmark is not sufficient for the proper analysis of the computer systems behavior. The response time of the process is often composed of the execution of many subprocesses or many paths of execution. Under this assumption, it is presented, that both convolution and deconvolution methods can be helpful in obtaining time distributions and modeling of complex processes. In such a modeling the analysis of measurement errors is very important and was taken into consideration. The example of using the methods in buffering process is also discussed.
MINIMIZING COMPUTATIONAL ERRORS OF TSUNAMI WAVE-RAY AND TRAVEL TIME
Directory of Open Access Journals (Sweden)
Andrei G. Marchuk
2008-01-01
Full Text Available There are many methods for computing tsunami kinematics directly and inversely. The direct detection of waves in the deep ocean makes it possible to establish tsunami source characteristics and origin. Thus, accuracy of computational methods is very important in obtaining reliable results. In a non-homogeneous medium where tsunami wave propagation velocity varies, it is not very easy to determine a wave-ray that connects two given points along a path. The present study proposes modification in the methodology of determining tsunami travel-times and of wave-ray paths. An approximate ray trace path can be developed from a source origin point to any other point on a computational grid by solving directly the problem - and thus obtain the tsunami travel- times. The initial ray approximation can be optimized with the use of an algorithm that calculates all potential variations and applies corrections to travel-time values. Such an algorithm was tested in an area with model bathymetry and compared with a non-optimized method. The latter exceeded the optimized method by one minute of travel-time for every hour of tsunami propagation time.
Generalized computer-aided discrete time domain modeling and analysis of dc-dc converters
Lee, F. C.; Iwens, R. P.; Yu, Y.; Triner, J. E.
1977-01-01
A generalized discrete time domain modeling and analysis technique is presented for all types of switching regulators using any type of duty-cycle controller, and operating in both continuous and discontinuous inductor current. State space techniques are employed to derive an equivalent nonlinear discrete time model that describes the converter exactly. The system is linearized about its equilibrium state to obtain a linear discrete time model for small signal performance evaluations, such as stability, audiosusceptibility and transient response. The analysis makes extensive use of the digital computer as an analytical tool. It is universal, exact and easy to use.
Computation of the acoustic radiation force using the finite-difference time-domain method.
Cai, Feiyan; Meng, Long; Jiang, Chunxiang; Pan, Yu; Zheng, Hairong
2010-10-01
The computational details related to calculating the acoustic radiation force on an object using a 2-D grid finite-difference time-domain method (FDTD) are presented. The method is based on propagating the stress and velocity fields through the grid and determining the energy flow with and without the object. The axial and radial acoustic radiation forces predicted by FDTD method are in excellent agreement with the results obtained by analytical evaluation of the scattering method. In particular, the results indicate that it is possible to trap the steel cylinder in the radial direction by optimizing the width of Gaussian source and the operation frequency. As the sizes of the relating objects are smaller than or comparable to wavelength, the algorithm presented here can be easily extended to 3-D and include torque computation algorithms, thus providing a highly flexible and universally usable computation engine.
Real-Time Nonlinear Finite Element Computations on GPU - Application to Neurosurgical Simulation.
Joldes, Grand Roman; Wittek, Adam; Miller, Karol
2010-12-15
Application of biomechanical modeling techniques in the area of medical image analysis and surgical simulation implies two conflicting requirements: accurate results and high solution speeds. Accurate results can be obtained only by using appropriate models and solution algorithms. In our previous papers we have presented algorithms and solution methods for performing accurate nonlinear finite element analysis of brain shift (which includes mixed mesh, different non-linear material models, finite deformations and brain-skull contacts) in less than a minute on a personal computer for models having up to 50.000 degrees of freedom. In this paper we present an implementation of our algorithms on a Graphics Processing Unit (GPU) using the new NVIDIA Compute Unified Device Architecture (CUDA) which leads to more than 20 times increase in the computation speed. This makes possible the use of meshes with more elements, which better represent the geometry, are easier to generate, and provide more accurate results.
A novel class of highly efficient and accurate time-integrators in nonlinear computational mechanics
Wang, Xuechuan; Atluri, Satya N.
2017-05-01
A new class of time-integrators is presented for strongly nonlinear dynamical systems. These algorithms are far superior to the currently common time integrators in computational efficiency and accuracy. These three algorithms are based on a local variational iteration method applied over a finite interval of time. By using Chebyshev polynomials as trial functions and Dirac-Delta functions as the test functions over the finite time interval, the three algorithms are developed into three different discrete time-integrators through the collocation method. These time integrators are labeled as Chebyshev local iterative collocation methods. Through examples of the forced Duffing oscillator, the Lorenz system, and the multiple coupled Duffing equations (which arise as semi-discrete equations for beams, plates and shells undergoing large deformations), it is shown that the new algorithms are far superior to the 4th order Runge-Kutta and ODE45 of MATLAB, in predicting the chaotic responses of strongly nonlinear dynamical systems.
A novel class of highly efficient and accurate time-integrators in nonlinear computational mechanics
Wang, Xuechuan; Atluri, Satya N.
2017-01-01
A new class of time-integrators is presented for strongly nonlinear dynamical systems. These algorithms are far superior to the currently common time integrators in computational efficiency and accuracy. These three algorithms are based on a local variational iteration method applied over a finite interval of time. By using Chebyshev polynomials as trial functions and Dirac-Delta functions as the test functions over the finite time interval, the three algorithms are developed into three different discrete time-integrators through the collocation method. These time integrators are labeled as Chebyshev local iterative collocation methods. Through examples of the forced Duffing oscillator, the Lorenz system, and the multiple coupled Duffing equations (which arise as semi-discrete equations for beams, plates and shells undergoing large deformations), it is shown that the new algorithms are far superior to the 4th order Runge-Kutta and ODE45 of MATLAB, in predicting the chaotic responses of strongly nonlinear dynamical systems.
A unified method for evaluating real-time computer controllers: A case study. [aircraft control
Shin, K. G.; Krishna, C. M.; Lee, Y. H.
1982-01-01
A real time control system consists of a synergistic pair, that is, a controlled process and a controller computer. Performance measures for real time controller computers are defined on the basis of the nature of this synergistic pair. A case study of a typical critical controlled process is presented in the context of new performance measures that express the performance of both controlled processes and real time controllers (taken as a unit) on the basis of a single variable: controller response time. Controller response time is a function of current system state, system failure rate, electrical and/or magnetic interference, etc., and is therefore a random variable. Control overhead is expressed as a monotonically nondecreasing function of the response time and the system suffers catastrophic failure, or dynamic failure, if the response time for a control task exceeds the corresponding system hard deadline, if any. A rigorous probabilistic approach is used to estimate the performance measures. The controlled process chosen for study is an aircraft in the final stages of descent, just prior to landing. First, the performance measures for the controller are presented. Secondly, control algorithms for solving the landing problem are discussed and finally the impact of the performance measures on the problem is analyzed.
Computation of stabilizing PI and PID controllers for processes with time delay.
Tan, Nusret
2005-04-01
In this paper, a new method for the computation of all stabilizing PI controllers for processes with time delay is given. The proposed method is based on plotting the stability boundary locus in the (kp, ki) plane and then computing the stabilizing values of the parameters of a PI controller for a given time delay system. The technique presented does not need to use Pade approximation and does not require sweeping over the parameters and also does not use linear programming to solve a set of inequalities. Thus it offers several important advantages over existing results obtained in this direction. Beyond stabilization, the method is used to compute stabilizing PI controllers which achieve user specified gain and phase margins. The proposed method is also used to design PID controllers for control systems with time delay. The limiting values of a PID controller which stabilize a given system with time delay are obtained in the (kp, ki) plane, (kp, kd) plane, and (ki, kd) plane. Examples are given to show the benefits of the method presented.
Analog quantum computing (AQC) and the need for time-symmetric physics
Werbos, Paul J.; Dolmatova, Ludmilla
2016-03-01
This paper discusses what will be necessary to achieve the full potential capabilities of analog quantum computing (AQC), which is defined here as the enrichment of continuous-variable computing to include stochastic, nonunitary circuit elements such as dissipative spin gates and address the wider range of tasks emerging from new trends in engineering, such as approximation of stochastic maps, ghost imaging and new forms of neural networks and intelligent control. This paper focuses especially on what is needed in terms of new experiments to validate remarkable new results in the modeling of triple entanglement, and in creating a pathway which links fundamental theoretical work with hard core experimental work, on a pathway to AQC similar to the pathway to digital quantum computing already blazed by Zeilinger's group. It discusses the most recent experiments and reviews two families of alternative models based on the traditional eigenvector projection model of polarizers and on a new family of local realistic models based on Markov Random Fields across space-time adhering to the rules of time-symmetric physics. For both families, it reviews lumped parameter versions, continuous time extension and possibilities for extension to continuous space and time.
Abuter, Roberto; Dembet, Roderick; Lacour, Sylvestre; di Lieto, Nicola; Woillez, Julien; Eisenhauer, Frank; Fedou, Pierre; Phan Duc, Than
2016-08-01
The new VLTI (Very Large Telescope Interferometer) 1 instrument GRAVITY5, 22, 23 is equipped with a fringe tracker16 able to stabilize the K-band fringes on six baselines at the same time. It has been designed to achieve a performance for average seeing conditions of a residual OPD (Optical Path Difference) lower than 300 nm with objects brighter than K = 10. The control loop implementing the tracking is composed of a four stage real time computer system compromising: a sensor where the detector pixels are read in and the OPD and GD (Group Delay) are calculated; a controller receiving the computed sensor quantities and producing commands for the piezo actuators; a concentrator which combines both the OPD commands with the real time tip/tilt corrections offloading them to the piezo actuator; and finally a Kalman15 parameter estimator. This last stage is used to monitor current measurements over a window of few seconds and estimate new values for the main Kalman15 control loop parameters. The hardware and software implementation of this design runs asynchronously and communicates the four computers for data transfer via the Reflective Memory Network3. With the purpose of improving the performance of the GRAVITY5, 23 fringe tracking16, 22 control loop, a deviation from the standard asynchronous communication mechanism has been proposed and implemented. This new scheme operates the four independent real time computers involved in the tracking loop synchronously using the Reflective Memory Interrupts2 as the coordination signal. This synchronous mechanism had the effect of reducing the total pure delay of the loop from 3.5 [ms] to 2.0 [ms] which then translates on a better stabilization of the fringes as the bandwidth of the system is substantially improved. This paper will explain in detail the real time architecture of the fringe tracker in both is synchronous and synchronous implementation. The achieved improvements on reducing the delay via this mechanism will be
Pande, Vijay S; Baker, Ian; Chapman, Jarrod; Elmer, Sidney P; Khaliq, Siraj; Larson, Stefan M; Rhee, Young Min; Shirts, Michael R; Snow, Christopher D; Sorin, Eric J; Zagrovic, Bojan
2003-01-01
Atomistic simulations of protein folding have the potential to be a great complement to experimental studies, but have been severely limited by the time scales accessible with current computer hardware and algorithms. By employing a worldwide distributed computing network of tens of thousands of PCs and algorithms designed to efficiently utilize this new many-processor, highly heterogeneous, loosely coupled distributed computing paradigm, we have been able to simulate hundreds of microseconds of atomistic molecular dynamics. This has allowed us to directly simulate the folding mechanism and to accurately predict the folding rate of several fast-folding proteins and polymers, including a nonbiological helix, polypeptide alpha-helices, a beta-hairpin, and a three-helix bundle protein from the villin headpiece. Our results demonstrate that one can reach the time scales needed to simulate fast folding using distributed computing, and that potential sets used to describe interatomic interactions are sufficiently accurate to reach the folded state with experimentally validated rates, at least for small proteins.
Prospects for Finite-Difference Time-Domain (FDTD) Computational Electrodynamics
Taflove, Allen
2002-08-01
FDTD is the most powerful numerical solution of Maxwell's equations for structures having internal details. Relative to moment-method and finite-element techniques, FDTD can accurately model such problems with 100-times more field unknowns and with nonlinear and/or time-variable parameters. Hundreds of FDTD theory and applications papers are published each year. Currently, there are at least 18 commercial FDTD software packages for solving problems in: defense (especially vulnerability to electromagnetic pulse and high-power microwaves); design of antennas and microwave devices/circuits; electromagnetic compatibility; bioelectromagnetics (especially assessment of cellphone-generated RF absorption in human tissues); signal integrity in computer interconnects; and design of micro-photonic devices (especially photonic bandgap waveguides, microcavities; and lasers). This paper explores emerging prospects for FDTD computational electromagnetics brought about by continuing advances in computer capabilities and FDTD algorithms. We conclude that advances already in place point toward the usage by 2015 of ultralarge-scale (up to 1E11 field unknowns) FDTD electromagnetic wave models covering the frequency range from about 0.1 Hz to 1E17 Hz. We expect that this will yield significant benefits for our society in areas as diverse as computing, telecommunications, defense, and public health and safety.
Courboulay, Vincent; Perreira Da Silva, Matthieu
2012-06-01
Providing real time analysis of the huge amount of data generated by computer vision algorithms in interactive applications is still an open problem. It promises great advances across a wide variety of fields. When using dynamics scene analysis algorithms for computer vision, a trade-off must be found between the quality of the results expected, and the amount of computer resources allocated for each task. It is usually a design time decision, implemented through the choice of pre-defined algorithms and parameters. However, this way of doing limits the generality of the system. Using an adaptive vision system provides a more flexible solution as its analysis strategy can be changed according to the new information available. As a consequence, such a system requires some kind of guiding mechanism to explore the scene faster and more efficiently. We propose a visual attention system that it adapts its processing according to the interest (or salience) of each element of the dynamic scene. Somewhere in between hierarchical salience based and competitive distributed, we propose a hierarchical yet competitive and non salience based model. Our original approach allows the generation of attentional focus points without the need of neither saliency map nor explicit inhibition of return mechanism. This new realtime computational model is based on a preys / predators system. The use of this kind of dynamical system is justified by an adjustable trade-off between nondeterministic attentional behavior and properties of stability, reproducibility and reactiveness.
1988-05-31
COMPUTER ALGORITHMS FOR AUTOMATIC REAL-TIME DETERMINATION OF SPACE VEHICLE POTENTIALS IN VARIOUS PLASMA ENVIRONMENTS May 31, 1988 Stanley L. Spiegel...crrnaion DiviSiofl 838 12 2 DERIVATION AND TESTING OF COMPUTER ALGORITHMS FOR AUTOMATIC REAL-TIME DETERMINATION OF SPACE VEHICLE POTENTIALS IN VARIOUS...S.L., "Derivation and testing of computer algorithms for automatic real time determination of space vehicle poteuatials in various plasma
An approach to experimental evaluation of real-time fault-tolerant distributed computing schemes
Kim, K. H.
1989-01-01
A testbed-based approach to the evaluation of fault-tolerant distributed computing schemes is discussed. The approach is based on experimental incorporation of system structuring and design techniques into real-time distributed-computing testbeds centered around tightly coupled microcomputer networks. The effectiveness of this approach has been experimentally confirmed. Primary advantages of this approach include the accuracy of the timing and logical-complexity data and the degree of assurance of the practical effectiveness of the scheme evaluated. Various design issues encountered in the course of establishing the network testbed facilities are discussed, along with their augmentation to support some experiments. The shortcomings of the testbeds are also discussed together with the desired extensions of the testbeds.
Ponce, Brent A; Menendez, Mariano E; Oladeji, Lasun O; Fryberger, Charles T; Dantuluri, Phani K
2014-11-01
The authors describe the first surgical case adopting the combination of real-time augmented reality and wearable computing devices such as Google Glass (Google Inc, Mountain View, California). A 66-year-old man presented to their institution for a total shoulder replacement after 5 years of progressive right shoulder pain and decreased range of motion. Throughout the surgical procedure, Google Glass was integrated with the Virtual Interactive Presence and Augmented Reality system (University of Alabama at Birmingham, Birmingham, Alabama), enabling the local surgeon to interact with the remote surgeon within the local surgical field. Surgery was well tolerated by the patient and early surgical results were encouraging, with an improvement of shoulder pain and greater range of motion. The combination of real-time augmented reality and wearable computing devices such as Google Glass holds much promise in the field of surgery.
A sub-cubic time algorithm for computing the quartet distance between two general trees
Directory of Open Access Journals (Sweden)
Mailund Thomas
2011-06-01
Full Text Available Abstract Background When inferring phylogenetic trees different algorithms may give different trees. To study such effects a measure for the distance between two trees is useful. Quartet distance is one such measure, and is the number of quartet topologies that differ between two trees. Results We have derived a new algorithm for computing the quartet distance between a pair of general trees, i.e. trees where inner nodes can have any degree ≥ 3. The time and space complexity of our algorithm is sub-cubic in the number of leaves and does not depend on the degree of the inner nodes. This makes it the fastest algorithm so far for computing the quartet distance between general trees independent of the degree of the inner nodes. Conclusions We have implemented our algorithm and two of the best competitors. Our new algorithm is significantly faster than the competition and seems to run in close to quadratic time in practice.
Gust Acoustics Computation with a Space-Time CE/SE Parallel 3D Solver
Wang, X. Y.; Himansu, A.; Chang, S. C.; Jorgenson, P. C. E.; Reddy, D. R. (Technical Monitor)
2002-01-01
The benchmark Problem 2 in Category 3 of the Third Computational Aero-Acoustics (CAA) Workshop is solved using the space-time conservation element and solution element (CE/SE) method. This problem concerns the unsteady response of an isolated finite-span swept flat-plate airfoil bounded by two parallel walls to an incident gust. The acoustic field generated by the interaction of the gust with the flat-plate airfoil is computed by solving the 3D (three-dimensional) Euler equations in the time domain using a parallel version of a 3D CE/SE solver. The effect of the gust orientation on the far-field directivity is studied. Numerical solutions are presented and compared with analytical solutions, showing a reasonable agreement.
2015-05-28
recognition is simpler and requires less computational resources compared to other inputs such as facial expressions. The Berlin database of Emotional...Network for Real-Time Speech-Emotion Recognition 5a. CONTRACT NUMBER IN-HOUSE 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 62788F 6. AUTHOR(S) Q...domains such as image and video analysis, anomaly detection, and speech recognition . In this research, a hardware architecture was explored for
Environment-sensitive manipulator control. [real time, decision making computer aided control
Bejczy, A. K.
1974-01-01
Environment-sensitive manipulator control (control systems capable of controlling manipulator motion based on real-time response to sensor data obtained during the attempt to perform a requested task) is described, and experiments on (1) proximity control in manipulation and (2) application of an articulated and adaptively controlled hand-to-environment-sensitive manipulator control are reported. The efficiency of such systems is determined by the separation of control and data processing functions between operator and computer.
M. Kasemann
CMS relies on a well functioning, distributed computing infrastructure. The Site Availability Monitoring (SAM) and the Job Robot submission have been very instrumental for site commissioning in order to increase availability of more sites such that they are available to participate in CSA07 and are ready to be used for analysis. The commissioning process has been further developed, including "lessons learned" documentation via the CMS twiki. Recently the visualization, presentation and summarizing of SAM tests for sites has been redesigned, it is now developed by the central ARDA project of WLCG. Work to test the new gLite Workload Management System was performed; a 4 times increase in throughput with respect to LCG Resource Broker is observed. CMS has designed and launched a new-generation traffic load generator called "LoadTest" to commission and to keep exercised all data transfer routes in the CMS PhE-DEx topology. Since mid-February, a transfer volume of about 12 P...
Li, Xiangrui; Lu, Zhong-Lin
2012-02-29
Display systems based on conventional computer graphics cards are capable of generating images with 8-bit gray level resolution. However, most experiments in vision research require displays with more than 12 bits of luminance resolution. Several solutions are available. Bit++ (1) and DataPixx (2) use the Digital Visual Interface (DVI) output from graphics cards and high resolution (14 or 16-bit) digital-to-analog converters to drive analog display devices. The VideoSwitcher (3) described here combines analog video signals from the red and blue channels of graphics cards with different weights using a passive resister network (4) and an active circuit to deliver identical video signals to the three channels of color monitors. The method provides an inexpensive way to enable high-resolution monochromatic displays using conventional graphics cards and analog monitors. It can also provide trigger signals that can be used to mark stimulus onsets, making it easy to synchronize visual displays with physiological recordings or response time measurements. Although computer keyboards and mice are frequently used in measuring response times (RT), the accuracy of these measurements is quite low. The RTbox is a specialized hardware and software solution for accurate RT measurements. Connected to the host computer through a USB connection, the driver of the RTbox is compatible with all conventional operating systems. It uses a microprocessor and high-resolution clock to record the identities and timing of button events, which are buffered until the host computer retrieves them. The recorded button events are not affected by potential timing uncertainties or biases associated with data transmission and processing in the host computer. The asynchronous storage greatly simplifies the design of user programs. Several methods are available to synchronize the clocks of the RTbox and the host computer. The RTbox can also receive external triggers and be used to measure RT with respect
A Unified Computational Architecture for Preprocessing Visual Information in Space and Time.
Skrzypek, Josef
1986-06-01
The success of autonomous mobile robots depends on the ability to understand continuously changing scenery. Present techniques for analysis of images are not always suitable because in sequential paradigm, computation of visual functions based on absolute values of stimuli is inefficient. Important aspects of visual information are encoded in discontinuities of intensity, hence a representation in terms of relative values seems advantageous. We present the computing architecture of a massively parallel vision module which optimizes the detection of relative intensity changes in space and time. Visual information must remain constant despite variation in ambient light level or velocity of target and robot. Constancy can be achieved by normalizing motion and lightness scales. In both cases, basic computation involves a comparison of the center pixels with the context of surrounding values. Therefore, a similar computing architecture, composed of three functionally-different and hierarchically-arranged layers of overlapping operators, can be used for two integrated parts of the module. The first part maintains high sensitivity to spatial changes by reducing noise and normalizing the lightness scale. The result is used by the second part to maintain high sensitivity to temporal discontinuities and to compute relative motion information. Simulation results show that response of the module is proportional to contrast of the stimulus and remains constant over the whole domain of intensity. It is also proportional to velocity of motion limited to any small portion of the visual field. Uniform motion throughout the visual field results in constant response, independent of velocity. Spatial and temporal intensity changes are enhanced because computationally, the module resembles the behavior of a DOG function.
Reaction time for processing visual stimulus in a computer-assisted rehabilitation environment.
Sanchez, Yerly; Pinzon, David; Zheng, Bin
2017-10-01
To examine the reaction time when human subjects process information presented in the visual channel under both a direct vision and a virtual rehabilitation environment when walking was performed. Visual stimulus included eight math problems displayed on the peripheral vision to seven healthy human subjects in a virtual rehabilitation training (computer-assisted rehabilitation environment (CAREN)) and a direct vision environment. Subjects were required to verbally report the results of these math calculations in a short period of time. Reaction time measured by Tobii Eye tracker and calculation accuracy were recorded and compared between the direct vision and virtual rehabilitation environment. Performance outcomes measured for both groups included reaction time, reading time, answering time and the verbal answer score. A significant difference between the groups was only found for the reaction time (p = .004). Participants had more difficulty recognizing the first equation of the virtual environment. Participants reaction time was faster in the direct vision environment. This reaction time delay should be kept in mind when designing skill training scenarios in virtual environments. This was a pilot project to a series of studies assessing cognition ability of stroke patients who are undertaking a rehabilitation program with a virtual training environment. Implications for rehabilitation Eye tracking is a reliable tool that can be employed in rehabilitation virtual environments. Reaction time changes between direct vision and virtual environment.
Approximating response time distributions in closed queueing network models of computer performance
Energy Technology Data Exchange (ETDEWEB)
Salza, S.; Lavenberg, S.S.
1981-01-01
Hierarchical decomposition methods for approximating response time distributions in certain closed queueing network models of computer performance are investigated. The methods investigated apply whenever part of a customer's response time consists of a geometrically distributed number of successive cycles within a subnetwork. The key step involves replacing the subnetwork with parallel exponential servers having queue-size dependent service rates. Results on thinning stochastic point processes are used to justify this replacement when the mean number of cycles is large. Preliminary numerical comparisons of the approximations with simulation results indicate that the approximations are quite accurate even when the mean number of cycles is small. 17 references.
Martín Furones, Angel; Anquela Julián, Ana Belén; Dimas-Pages, Alejandro; Cos-Gayón, Fernando
2017-08-01
Precise point positioning (PPP) is a well established Global Navigation Satellite System (GNSS) technique that only requires information from the receiver (or rover) to obtain high-precision position coordinates. This is a very interesting and promising technique because eliminates the need for a reference station near the rover receiver or a network of reference stations, thus reducing the cost of a GNSS survey. From a computational perspective, there are two ways to solve the system of observation equations produced by static PPP either in a single step (so-called batch adjustment) or with a sequential adjustment/filter. The results of each should be the same if they are both well implemented. However, if a sequential solution (that is, not only the final coordinates, but also those observed in previous GNSS epochs), is needed, as for convergence studies, finding a batch solution becomes a very time consuming task owing to the need for matrix inversion that accumulates with each consecutive epoch. This is not a problem for the filter solution, which uses information computed in the previous epoch for the solution of the current epoch. Thus filter implementations need extra considerations of user dynamics and parameter state variations between observation epochs with appropriate stochastic update parameter variances from epoch to epoch. These filtering considerations are not needed in batch adjustment, which makes it attractive. The main objective of this research is to significantly reduce the computation time required to obtain sequential results using batch adjustment. The new method we implemented in the adjustment process led to a mean reduction in computational time by 45%.
Space-Time Fluid-Structure Interaction Computation of Flapping-Wing Aerodynamics
Kostov, Nikolay M.
We present a sequentially-coupled space-time (ST) computational fluid-structure interaction (FSI) analysis of flapping-wing aerodynamics of a micro aerial vehicle (MAV). The wing motion and deformation data, whether prescribed fully or partially, is from an actual locust, extracted from high-speed, multi-camera video recordings of the locust in a wind tunnel. The core computational FSI technology is based on the Deforming-Spatial-Domain/Stabilized ST (DSD/SST) formulation. This is supplemented with using NURBS basis functions in temporal representation of the wing and mesh motion, and in remeshing. Here we use the version of the DSD/SST formulation derived in conjunction with the variational multiscale (VMS) method, and this version is called "DSD/SST-VMST." The structural mechanics computations are based on the Kirchhoff-Love shell model. The sequential-coupling technique is applicable to some classes of FSI problems, especially those with temporally-periodic behavior. We show that it performs well in FSI computations of the flapping-wing aerodynamics we consider here. In addition to the straight-flight case, we analyze cases where the MAV body has rolling, pitching, or rolling and pitching motion. We study how all these influence the lift and thrust.
Warner, James E.; Zubair, Mohammad; Ranjan, Desh
2017-01-01
This work investigates novel approaches to probabilistic damage diagnosis that utilize surrogate modeling and high performance computing (HPC) to achieve substantial computational speedup. Motivated by Digital Twin, a structural health management (SHM) paradigm that integrates vehicle-specific characteristics with continual in-situ damage diagnosis and prognosis, the methods studied herein yield near real-time damage assessments that could enable monitoring of a vehicle's health while it is operating (i.e. online SHM). High-fidelity modeling and uncertainty quantification (UQ), both critical to Digital Twin, are incorporated using finite element method simulations and Bayesian inference, respectively. The crux of the proposed Bayesian diagnosis methods, however, is the reformulation of the numerical sampling algorithms (e.g. Markov chain Monte Carlo) used to generate the resulting probabilistic damage estimates. To this end, three distinct methods are demonstrated for rapid sampling that utilize surrogate modeling and exploit various degrees of parallelism for leveraging HPC. The accuracy and computational efficiency of the methods are compared on the problem of strain-based crack identification in thin plates. While each approach has inherent problem-specific strengths and weaknesses, all approaches are shown to provide accurate probabilistic damage diagnoses and several orders of magnitude computational speedup relative to a baseline Bayesian diagnosis implementation.
An Enhanced Tree-Shaped Adachi-Like Chaotic Neural Network Requiring Linear-Time Computations
Qin, Ke; Oommen, B. John
The Adachi Neural Network (AdNN) [1-5], is a fascinating Neural Network (NN) which has been shown to possess chaotic properties, and to also demonstrate Associative Memory (AM) and Pattern Recognition (PR) characteristics. Variants of the AdNN [6,7] have also been used to obtain other PR phenomena, and even blurring. A significant problem associated with the AdNN and its variants, is that all of them require a quadratic number of computations. This is essentially because all their NNs are completely connected graphs. In this paper we consider how the computations can be significantly reduced by merely using a linear number of computations. To do this, we extract from the original complete graph, one of its spanning trees. We then compute the weights for this spanning tree in such a manner that the modified tree-based NN has approximately the same input-output characteristics, and thus the new weights are themselves calculated using a gradient-based algorithm. By a detailed experimental analysis, we show that the new linear-time AdNN-like network possesses chaotic and PR properties for different settings. As far as we know, such a tree-based AdNN has not been reported, and the results given here are novel.
Theoretical Considerations on the Computation of Generalized Time-Periodic Waves
Pauly, Dirk
2011-01-01
We present both, theory and an algorithm for solving time-harmonic wave problems in a general setting. The time-harmonic solutions will be achieved by computing time-periodic solutions of the original wave equations. Thus, an exact controllability technique is proposed to solve the time-dependent wave equations. We discuss a first order Maxwell type system, which will be formulated in the framework of alternating differential forms. This enables us to investigate different kinds of classical wave problems in one fell swoop, such as acoustic, electro-magnetic or elastic wave problems. After a sufficient theory is established, we formulate our exact controllability problem and suggest a least-squares optimization procedure for its solution, which itself is solved in a natural way by a conjugate gradient algorithm operating in the canonical Hilbert space. Therefore, it might be one of the biggest advances of this approach that the proposed conjugate gradient algorithm does not need any preconditioning.
Time sparsification of EEG signals in motor-imagery based brain computer interfaces.
Higashi, Hiroshi; Tanaka, Toshihisa
2012-01-01
We propose a method of sparsifying EEG signals in the time domain for common spatial patterns (CSP) which are often used for feature extraction in brain computer interfaces (BCI). For accurate classification, it is important to analyze the period of time when a BCI user performs a mental task. We address this problem by optimizing the CSP cost with a time sparsification that removes unnecessary samples from the classification. We design a cost function that has CSP spatial weights and time window as optimization parameters. To find these parameters, we use alternating optimization. In an experiment on classification of motor-imagery EEG signals, the proposed method increased classification accuracy by 6% averaged over five subjects.
Institute of Scientific and Technical Information of China (English)
Ji Xu; Jing hai Li; Hua biao Qi; Xiao jian Fang; Li qiang Lu; Wei Ge; Xiao wei Wang; Ming Xu; Fei guo Chen; Xian feng He
2011-01-01
Real-time simulation of industrial equipment is a huge challenge nowadays.The high performance and fine-grained parallel computing provided by graphics processing units (GPUs) bring us closer to our goals.In this article,an industrial-scale rotating drum is simulated using simplified discrete element method (DEM) without consideration of the tangential components of contact force and particle rotation.A single GPU is used first to simulate a small model system with about 8000 particles in real-time,and the simulation is then scaled up to industrial scale using more than 200 GPUs in a 1D domain-decomposition parallelization mode.The overall speed is about 1/11 of the real-time.Optimization of the communication part of the parallel GPU codes can speed up the simulation further,indicating that such real-time simulations have not only methodological but also industrial implications in the near future.
User's manual for THPLOT, A FORTRAN 77 Computer program for time history plotting
Murray, J. E.
1982-01-01
A general purpose FORTRAN 77 computer program (THPLOT) for plotting time histories using Calcomp pen plotters is described. The program is designed to read a time history data file and to generate time history plots for selected time intervals and/or selected data channels. The capabilities of the program are described. The card input required to define the plotting operation is described and examples of card input and the resulting plotted output are given. The examples are followed by a description of the printed output, including both normal output and error messages. Lastly, implementation of the program is described. A complete listing of the program with reference maps produced by the CDC FTN 5.0 compiler is included.
Martins, Goncalo; Moondra, Arul; Dubey, Abhishek; Bhattacharjee, Anirban; Koutsoukos, Xenofon D.
2016-01-01
In modern networked control applications, confidentiality and integrity are important features to address in order to prevent against attacks. Moreover, network control systems are a fundamental part of the communication components of current cyber-physical systems (e.g., automotive communications). Many networked control systems employ Time-Triggered (TT) architectures that provide mechanisms enabling the exchange of precise and synchronous messages. TT systems have computation and communication constraints, and with the aim to enable secure communications in the network, it is important to evaluate the computational and communication overhead of implementing secure communication mechanisms. This paper presents a comprehensive analysis and evaluation of the effects of adding a Hash-based Message Authentication (HMAC) to TT networked control systems. The contributions of the paper include (1) the analysis and experimental validation of the communication overhead, as well as a scalability analysis that utilizes the experimental result for both wired and wireless platforms and (2) an experimental evaluation of the computational overhead of HMAC based on a kernel-level Linux implementation. An automotive application is used as an example, and the results show that it is feasible to implement a secure communication mechanism without interfering with the existing automotive controller execution times. The methods and results of the paper can be used for evaluating the performance impact of security mechanisms and, thus, for the design of secure wired and wireless TT networked control systems. PMID:27463718
Martins, Goncalo; Moondra, Arul; Dubey, Abhishek; Bhattacharjee, Anirban; Koutsoukos, Xenofon D
2016-07-25
In modern networked control applications, confidentiality and integrity are important features to address in order to prevent against attacks. Moreover, network control systems are a fundamental part of the communication components of current cyber-physical systems (e.g., automotive communications). Many networked control systems employ Time-Triggered (TT) architectures that provide mechanisms enabling the exchange of precise and synchronous messages. TT systems have computation and communication constraints, and with the aim to enable secure communications in the network, it is important to evaluate the computational and communication overhead of implementing secure communication mechanisms. This paper presents a comprehensive analysis and evaluation of the effects of adding a Hash-based Message Authentication (HMAC) to TT networked control systems. The contributions of the paper include (1) the analysis and experimental validation of the communication overhead, as well as a scalability analysis that utilizes the experimental result for both wired and wireless platforms and (2) an experimental evaluation of the computational overhead of HMAC based on a kernel-level Linux implementation. An automotive application is used as an example, and the results show that it is feasible to implement a secure communication mechanism without interfering with the existing automotive controller execution times. The methods and results of the paper can be used for evaluating the performance impact of security mechanisms and, thus, for the design of secure wired and wireless TT networked control systems.
M. Kasemann
Overview In autumn the main focus was to process and handle CRAFT data and to perform the Summer08 MC production. The operational aspects were well covered by regular Computing Shifts, experts on duty and Computing Run Coordination. At the Computing Resource Board (CRB) in October a model to account for service work at Tier 2s was approved. The computing resources for 2009 were reviewed for presentation at the C-RRB. The quarterly resource monitoring is continuing. Facilities/Infrastructure operations Operations during CRAFT data taking ran fine. This proved to be a very valuable experience for T0 workflows and operations. The transfers of custodial data to most T1s went smoothly. A first round of reprocessing started at the Tier-1 centers end of November; it will take about two weeks. The Computing Shifts procedure was tested full scale during this period and proved to be very efficient: 30 Computing Shifts Persons (CSP) and 10 Computing Resources Coordinators (CRC). The shift program for the shut down w...
Automated selection of brain regions for real-time fMRI brain-computer interfaces
Lührs, Michael; Sorger, Bettina; Goebel, Rainer; Esposito, Fabrizio
2017-02-01
Objective. Brain-computer interfaces (BCIs) implemented with real-time functional magnetic resonance imaging (rt-fMRI) use fMRI time-courses from predefined regions of interest (ROIs). To reach best performances, localizer experiments and on-site expert supervision are required for ROI definition. To automate this step, we developed two unsupervised computational techniques based on the general linear model (GLM) and independent component analysis (ICA) of rt-fMRI data, and compared their performances on a communication BCI. Approach. 3 T fMRI data of six volunteers were re-analyzed in simulated real-time. During a localizer run, participants performed three mental tasks following visual cues. During two communication runs, a letter-spelling display guided the subjects to freely encode letters by performing one of the mental tasks with a specific timing. GLM- and ICA-based procedures were used to decode each letter, respectively using compact ROIs and whole-brain distributed spatio-temporal patterns of fMRI activity, automatically defined from subject-specific or group-level maps. Main results. Letter-decoding performances were comparable to supervised methods. In combination with a similarity-based criterion, GLM- and ICA-based approaches successfully decoded more than 80% (average) of the letters. Subject-specific maps yielded optimal performances. Significance. Automated solutions for ROI selection may help accelerating the translation of rt-fMRI BCIs from research to clinical applications.
Subramanian, S. V.; Bozzola, R.
1987-01-01
Numerical solutions of the unsteady Euler equations are obtained using the classical fourth order Runge Kutta time marching scheme. This method is fully explicit and is applied to the governing equations in the finite volume, conservation law form. In order to determine the efficiency of this scheme for solving turbomachinery flows, steady blade-to-blade solutions are obtained for compressor and turbine cascades under subsonic and transonic flow conditions. Computed results are compared with other numerical methods and wind tunnel measurements. The study also focuses on other important numerical aspects influencing the performance of the algorithm and the solution accuracy such as grid types, boundary conditions and artificial viscosity. For this purpose, H, O, and C type computational grids as well as characteristic and extrapolation type boundary conditions are included in solution procedures.
Improving Overhead Computation and pre-processing Time for Grid Scheduling System
Bouyer, Asgarali; Abdullah, Abdul Hanan
2010-01-01
Computational Grid is enormous environments with heterogeneous resources and stable infrastructures among other Internet-based computing systems. However, the managing of resources in such systems has its special problems. Scheduler systems need to get last information about participant nodes from information centers for the purpose of firmly job scheduling. In this paper, we focus on online updating resource information centers with processed and provided data based on the assumed hierarchical model. A hybrid knowledge extraction method has been used to classifying grid nodes based on prediction of jobs' features. An affirmative point of this research is that scheduler systems don't waste extra time for getting up-to-date information of grid nodes. The experimental result shows the advantages of our approach compared to other conservative methods, especially due to its ability to predict the behavior of nodes based on comprehensive data tables on each node.
Computer vision system in real-time for color determination on flat surface food
Directory of Open Access Journals (Sweden)
Erick Saldaña
2013-03-01
Full Text Available Artificial vision systems also known as computer vision are potent quality inspection tools, which can be applied in pattern recognition for fruits and vegetables analysis. The aim of this research was to design, implement and calibrate a new computer vision system (CVS in real-time for the color measurement on flat surface food. For this purpose was designed and implemented a device capable of performing this task (software and hardware, which consisted of two phases: a image acquisition and b image processing and analysis. Both the algorithm and the graphical interface (GUI were developed in Matlab. The CVS calibration was performed using a conventional colorimeter (Model CIEL* a* b*, where were estimated the errors of the color parameters: eL* = 5.001%, and ea* = 2.287%, and eb* = 4.314 % which ensure adequate and efficient automation application in industrial processes in the quality control in the food industry sector.
Computer vision system in real-time for color determination on flat surface food
Directory of Open Access Journals (Sweden)
Erick Saldaña
2013-01-01
Full Text Available Artificial vision systems also known as computer vision are potent quality inspection tools, which can be applied in pattern recognition for fruits and vegetables analysis. The aim of this research was to design, implement and calibrate a new computer vision system (CVS in real - time f or the color measurement on flat surface food. For this purpose was designed and implemented a device capable of performing this task (software and hardware, which consisted of two phases: a image acquisition and b image processing and analysis. Both th e algorithm and the graphical interface (GUI were developed in Matlab. The CVS calibration was performed using a conventional colorimeter (Model CIEL* a* b*, where were estimated the errors of the color parameters: e L* = 5.001%, and e a* = 2.287%, and e b* = 4.314 % which ensure adequate and efficient automation application in industrial processes in the quality control in the food industry sector.
Subramanian, S. V.; Bozzola, R.
1985-01-01
Numerical solutions of the unsteady Euler equations are obtained using the classical fourth order Runge Kutta time marching scheme. This method is fully explicit and is applied to the governing equations in the finite volume, conservation law form. In order to determine the efficiency of this scheme for solving turbomachinery flows, steady blade-to-blade solutions are obtained for compressor and turbine cascades under subsonic and transonic flow conditions. Computed results are compared with other numerical methods and wind tunnel measurements. The present study also focuses on other important numerical aspects influencing the performance of the algorithm and the solution accuracy such as grid types, boundary conditions, and artificial viscosity. For this purpose, H, O, and C type computational grids as well as characteristic and extrapolation type boundary conditions are included in the solution procedure.
Computing moment-to-moment BOLD activation for real-time neurofeedback.
Hinds, Oliver; Ghosh, Satrajit; Thompson, Todd W; Yoo, Julie J; Whitfield-Gabrieli, Susan; Triantafyllou, Christina; Gabrieli, John D E
2011-01-01
Estimating moment-to-moment changes in blood oxygenation level dependent (BOLD) activation levels from functional magnetic resonance imaging (fMRI) data has applications for learned regulation of regional activation, brain state monitoring, and brain-machine interfaces. In each of these contexts, accurate estimation of the BOLD signal in as little time as possible is desired. This is a challenging problem due to the low signal-to-noise ratio of fMRI data. Previous methods for real-time fMRI analysis have either sacrificed the ability to compute moment-to-moment activation changes by averaging several acquisitions into a single activation estimate or have sacrificed accuracy by failing to account for prominent sources of noise in the fMRI signal. Here we present a new method for computing the amount of activation present in a single fMRI acquisition that separates moment-to-moment changes in the fMRI signal intensity attributable to neural sources from those due to noise, resulting in a feedback signal more reflective of neural activation. This method computes an incremental general linear model fit to the fMRI time series, which is used to calculate the expected signal intensity at each new acquisition. The difference between the measured intensity and the expected intensity is scaled by the variance of the estimator in order to transform this residual difference into a statistic. Both synthetic and real data were used to validate this method and compare it to the only other published real-time fMRI method. Copyright © 2010 Elsevier Inc. All rights reserved.
Real-time dynamics of lattice gauge theories with a few-qubit quantum computer
Martinez, Esteban A.; Muschik, Christine A.; Schindler, Philipp; Nigg, Daniel; Erhard, Alexander; Heyl, Markus; Hauke, Philipp; Dalmonte, Marcello; Monz, Thomas; Zoller, Peter; Blatt, Rainer
2016-06-01
Gauge theories are fundamental to our understanding of interactions between the elementary constituents of matter as mediated by gauge bosons. However, computing the real-time dynamics in gauge theories is a notorious challenge for classical computational methods. This has recently stimulated theoretical effort, using Feynman’s idea of a quantum simulator, to devise schemes for simulating such theories on engineered quantum-mechanical devices, with the difficulty that gauge invariance and the associated local conservation laws (Gauss laws) need to be implemented. Here we report the experimental demonstration of a digital quantum simulation of a lattice gauge theory, by realizing (1 + 1)-dimensional quantum electrodynamics (the Schwinger model) on a few-qubit trapped-ion quantum computer. We are interested in the real-time evolution of the Schwinger mechanism, describing the instability of the bare vacuum due to quantum fluctuations, which manifests itself in the spontaneous creation of electron-positron pairs. To make efficient use of our quantum resources, we map the original problem to a spin model by eliminating the gauge fields in favour of exotic long-range interactions, which can be directly and efficiently implemented on an ion trap architecture. We explore the Schwinger mechanism of particle-antiparticle generation by monitoring the mass production and the vacuum persistence amplitude. Moreover, we track the real-time evolution of entanglement in the system, which illustrates how particle creation and entanglement generation are directly related. Our work represents a first step towards quantum simulation of high-energy theories using atomic physics experiments—the long-term intention is to extend this approach to real-time quantum simulations of non-Abelian lattice gauge theories.
Real-time dynamics of lattice gauge theories with a few-qubit quantum computer.
Martinez, Esteban A; Muschik, Christine A; Schindler, Philipp; Nigg, Daniel; Erhard, Alexander; Heyl, Markus; Hauke, Philipp; Dalmonte, Marcello; Monz, Thomas; Zoller, Peter; Blatt, Rainer
2016-06-23
Gauge theories are fundamental to our understanding of interactions between the elementary constituents of matter as mediated by gauge bosons. However, computing the real-time dynamics in gauge theories is a notorious challenge for classical computational methods. This has recently stimulated theoretical effort, using Feynman's idea of a quantum simulator, to devise schemes for simulating such theories on engineered quantum-mechanical devices, with the difficulty that gauge invariance and the associated local conservation laws (Gauss laws) need to be implemented. Here we report the experimental demonstration of a digital quantum simulation of a lattice gauge theory, by realizing (1 + 1)-dimensional quantum electrodynamics (the Schwinger model) on a few-qubit trapped-ion quantum computer. We are interested in the real-time evolution of the Schwinger mechanism, describing the instability of the bare vacuum due to quantum fluctuations, which manifests itself in the spontaneous creation of electron-positron pairs. To make efficient use of our quantum resources, we map the original problem to a spin model by eliminating the gauge fields in favour of exotic long-range interactions, which can be directly and efficiently implemented on an ion trap architecture. We explore the Schwinger mechanism of particle-antiparticle generation by monitoring the mass production and the vacuum persistence amplitude. Moreover, we track the real-time evolution of entanglement in the system, which illustrates how particle creation and entanglement generation are directly related. Our work represents a first step towards quantum simulation of high-energy theories using atomic physics experiments-the long-term intention is to extend this approach to real-time quantum simulations of non-Abelian lattice gauge theories.
40 CFR 270.215 - How are time periods in the requirements in this subpart and my RAP computed?
2010-07-01
... requirements in this subpart and my RAP computed? 270.215 Section 270.215 Protection of Environment... HAZARDOUS WASTE PERMIT PROGRAM Remedial Action Plans (RAPs) Operating Under Your Rap § 270.215 How are time periods in the requirements in this subpart and my RAP computed? (a) Any time period scheduled to begin on...
M. Kasemann
Overview During the past three months activities were focused on data operations, testing and re-enforcing shift and operational procedures for data production and transfer, MC production and on user support. Planning of the computing resources in view of the new LHC calendar in ongoing. Two new task forces were created for supporting the integration work: Site Commissioning, which develops tools helping distributed sites to monitor job and data workflows, and Analysis Support, collecting the user experience and feedback during analysis activities and developing tools to increase efficiency. The development plan for DMWM for 2009/2011 was developed at the beginning of the year, based on the requirements from the Physics, Computing and Offline groups (see Offline section). The Computing management meeting at FermiLab on February 19th and 20th was an excellent opportunity discussing the impact and for addressing issues and solutions to the main challenges facing CMS computing. The lack of manpower is particul...
P. McBride
The Computing Project is preparing for a busy year where the primary emphasis of the project moves towards steady operations. Following the very successful completion of Computing Software and Analysis challenge, CSA06, last fall, we have reorganized and established four groups in computing area: Commissioning, User Support, Facility/Infrastructure Operations and Data Operations. These groups work closely together with groups from the Offline Project in planning for data processing and operations. Monte Carlo production has continued since CSA06, with about 30M events produced each month to be used for HLT studies and physics validation. Monte Carlo production will continue throughout the year in the preparation of large samples for physics and detector studies ramping to 50 M events/month for CSA07. Commissioning of the full CMS computing system is a major goal for 2007. Site monitoring is an important commissioning component and work is ongoing to devise CMS specific tests to be included in Service Availa...
Monte Carlo Computation of Spectral Density Function in Real-Time Scalar Field Theory
Abbasi, Navid
2014-01-01
Non-perturbative study of "real-time" field theories is difficult due to the sign problem. We use Bold Schwinger-Dyson (SD) equations to study the real-time $\\phi^4$ theory in $d=4$ beyond the perturbative regime. Combining SD equations in a particular way, we derive a non-linear integral equation for the two-point function. Then we introduce a new method by which one can analytically perform the momentum part of loop integrals in this equation. The price we must pay for such simplification is to numerically solve a non-linear integral equation for the spectral density function. Using Bold diagrammatic Monte Carlo method we find non-perturbative spectral function of theory and compare it with the one obtained from perturbation theory. At the end we utilize our Monte Carlo result to find the full vertex function as the basis for the computation of real-time scattering amplitudes.
On the Computation of $\\pi$-Flat Outputs for Linear Time-Delay Systems
Morio, Vincent; Lévine, Jean
2009-01-01
This paper deals with linear time-varying, delay systems. Extensions of the concept of differential flatness \\cite{Fliess_95} to this context have been first proposed in \\cite{Mounier_95,Fliess_96} (see also \\cite{Rudolph_03,Chyzak_05}), by the introduction of $\\pi$-flat output. Roughly speaking, it means that every system variable may be expressed as a function of a particular output $y$, a finite number of its time derivatives, time delays, and predictions, the latter resulting from the prediction operator $\\pi^{-1}$. We propose a simple and constructive algorithm for the computation of $\\pi$-flat outputs based on concepts of polynomial algebra, in particular Smith-Jacobson decomposition of polynomial matrices. Some examples are provided to illustrate the proposed methodology.
Institute of Scientific and Technical Information of China (English)
刘晨; 季振林; 程垠钟; 刘胜兰
2016-01-01
The multi-dimensional time-domain computational fluid dynamics (CFD) approach is extended to calculate the acoustic attenuation performance of water-filled piping silencers. Transmission loss predictions from the time-domain CFD approach and the frequency-domain finite element method (FEM) agree well with each other for the dual expansion chamber silencer, straight-through and cross-flow perforated tube silencers without flow. Then, the time-domain CFD approach is used to investigate the effect of flow on the acoustic attenuation characteristics of perforated tube silencers. The numerical predictions demonstrate that the mean flow increases the transmission loss, especially at higher frequencies, and shifts the transmission loss curve to lower frequencies.
Computational imaging with multi-camera time-of-flight systems
Shrestha, Shikhar
2016-07-11
Depth cameras are a ubiquitous technology used in a wide range of applications, including robotic and machine vision, human computer interaction, autonomous vehicles as well as augmented and virtual reality. In this paper, we explore the design and applications of phased multi-camera time-of-flight (ToF) systems. We develop a reproducible hardware system that allows for the exposure times and waveforms of up to three cameras to be synchronized. Using this system, we analyze waveform interference between multiple light sources in ToF applications and propose simple solutions to this problem. Building on the concept of orthogonal frequency design, we demonstrate state-of-the-art results for instantaneous radial velocity capture via Doppler time-of-flight imaging and we explore new directions for optically probing global illumination, for example by de-scattering dynamic scenes and by non-line-of-sight motion detection via frequency gating. © 2016 ACM.
Long-range correlations in heart rate variability during computer-mouse work under time pressure
Jiang, Dineng; He, Mulu; Qiu, Yihong; Zhu, Yisheng; Tong, Shanbao
2009-04-01
The aim of this study was to investigate the influences of time pressure on long-range correlations in heart rate variability (HRV), the effects of relaxation on the cardiovascular regulation system and the advantages of detrended fluctuation analysis (DFA) over the conventional power spectral analysis in discriminating states of the cardiovascular systems under different levels of time pressure. Volunteer subjects ( n=10, male/female=5/5) participated in a computer-mouse task consisting of five sessions, i.e. baseline session (BSS) which was free of time pressure, followed by sessions with 80% (SS80), 100% (SS100), 90% (SS90) and 150% (SS150) of the baseline time. Electrocardiogram (ECG) and task performance were recorded throughout the experiments. Two rest sessions before and after the computer-mouse work, i.e. RS1 and RS2, were also recorded as comparison. HRV series were subsequently analyzed by both conventional power spectral analysis and detrended fluctuation analysis (DFA). The long-term scaling exponent α2 by DFA was significantly lower in SS80 than that in other sessions. It was also found that short-term release of time pressure had positive influences on the cardiovascular system, i.e. the α2 in RS2 was significantly higher than that in SS80, SS100 and SS90. No significant differences were found between any two sessions by conventional power spectral analysis. Our results showed that DFA performed better in discriminating the states of cardiovascular autonomic modulation under time pressure than the conventional power spectral analysis.
On the Computational Complexity of Sphere Decoder for Lattice Space-Time Coded MIMO Channel
Abediseid, Walid
2011-01-01
The exact complexity analysis of the basic sphere decoder for general space-time codes applied to multi-input multi-output (MIMO) wireless channel is known to be difficult. In this work, we shed the light on the computational complexity of sphere decoding for the quasi-static, LAttice Space-Time (LAST) coded MIMO channel. Specifically, we derive the asymptotic tail distribution of the decoder's computational complexity in the high signal-to-noise ratio (SNR) regime. For the uncoded $M\\times N$ MIMO channel (e.g., V-BLAST), the analysis in [6] revealed that the tail distribution of such a decoder is of a Pareto-type with tail exponent that is equivalent to $N-M+1$. In our analysis, we show that the tail exponent of the sphere decoder's complexity distribution is equivalent to the diversity-multiplexing tradeoff achieved by LAST coding and lattice decoding schemes. This leads to extend the channel's tradeoff to include the decoding complexity. Moreover, we show analytically how minimum-mean square-error decisio...
Müller, Andy; Osterhage, Hannes; Sowa, Robert; Andrzejak, Ralph G; Mormann, Florian; Lehnertz, Klaus
2006-04-15
We present a client-server application for the distributed multivariate analysis of time series using standard PCs. We here concentrate on analyses of multichannel EEG/MEG data, but our method can easily be adapted to other time series. Due to the rapid development of new analysis techniques, the focus in the design of our application was not only on computational performance, but also on high flexibility and expandability of both the client and the server programs. For this purpose, the communication between the server and the clients as well as the building of the computational tasks has been realized via the Extensible Markup Language (XML). Running our newly developed method in an asynchronous distributed environment with random availability of remote and heterogeneous resources, we tested the system's performance for a number of different univariate and bivariate analysis techniques. Results indicate that for most of the currently available analysis techniques, calculations can be performed in real time, which, in principle, allows on-line analyses at relatively low cost.
The reliable solution and computation time of variable parameters logistic model
Wang, Pengfei; Pan, Xinnong
2017-04-01
The study investigates the reliable computation time (RCT, termed as T c) by applying a double-precision computation of a variable parameters logistic map (VPLM). Firstly, by using the proposed method, we obtain the reliable solutions for the logistic map. Secondly, we construct 10,000 samples of reliable experiments from a time-dependent non-stationary parameters VPLM and then calculate the mean T c. The results indicate that, for each different initial value, the T cs of the VPLM are generally different. However, the mean T c trends to a constant value when the sample number is large enough. The maximum, minimum, and probable distribution functions of T c are also obtained, which can help us to identify the robustness of applying a nonlinear time series theory to forecasting by using the VPLM output. In addition, the T c of the fixed parameter experiments of the logistic map is obtained, and the results suggest that this T c matches the theoretical formula-predicted value.
Saxena, R. K.; Mathai, A. M.; Haubold, H. J.
2015-10-01
This paper deals with the investigation of the computational solutions of an unified fractional reaction-diffusion equation, which is obtained from the standard diffusion equation by replacing the time derivative of first order by the generalized fractional time-derivative defined by Hilfer (2000), the space derivative of second order by the Riesz-Feller fractional derivative and adding the function ϕ (x, t) which is a nonlinear function governing reaction. The solution is derived by the application of the Laplace and Fourier transforms in a compact and closed form in terms of the H-function. The main result obtained in this paper provides an elegant extension of the fundamental solution for the space-time fractional diffusion equation obtained earlier by Mainardi et al. (2001, 2005) and a result very recently given by Tomovski et al. (2011). Computational representation of the fundamental solution is also obtained explicitly. Fractional order moments of the distribution are deduced. At the end, mild extensions of the derived results associated with a finite number of Riesz-Feller space fractional derivatives are also discussed.
Computing network-based features from physiological time series: application to sepsis detection.
Santaniello, Sabato; Granite, Stephen J; Sarma, Sridevi V; Winslow, Raimond L
2014-01-01
Sepsis is a systemic deleterious host response to infection. It is a major healthcare problem that affects millions of patients every year in the intensive care units (ICUs) worldwide. Despite the fact that ICU patients are heavily instrumented with physiological sensors, early sepsis detection remains challenging, perhaps because clinicians identify sepsis by using static scores derived from bed-side measurements individually, i.e., without systematically accounting for potential interactions between these signals and their dynamics. In this study, we apply network-based data analysis to take into account interactions between bed-side physiological time series (PTS) data collected in ICU patients, and we investigate features to distinguish between sepsis and non-sepsis conditions. We treated each PTS source as a node on a graph and we retrieved the graph connectivity matrix over time by tracking the correlation between each pair of sources' signals over consecutive time windows. Then, for each connectivity matrix, we computed the eigenvalue decomposition. We found that, even though raw PTS measurements may have indistinguishable distributions in non-sepsis and early sepsis states, the median /I of the eigenvalues computed from the same data is statistically different (p sepsis detection.
Sundareshan, Malur K.
2002-07-01
Computational complexity is a major impediment to the real- time implementation of image restoration and super- resolution algorithms. Although powerful restoration algorithms have been developed within the last few years utilizing sophisticated mathematical machinery (based on statistical optimization and convex set theory), these algorithms are typically iterative in nature and require enough number of iterations to be executed to achieve desired resolution gains in order to meaningfully perform detection and recognition tasks in practice. Additionally, recent technological breakthroughs have facilitated novel sensor designs (focal plane arrays, for instance) that make it possible to capture mega-pixel imagery data at video frame rates. A major challenge in the processing of these large format images is to complete the execution of the image processing steps within the frame capture times and to keep up with the output rate of the sensor so that all data captured by the sensor can be efficiently utilized. Consequently, development of novel methods that facilitate real-time implementation of image restoration and super- resolution algorithms is of significant practical interest and will be the primary focus of this paper. The key to designing computationally efficient processing schemes lies in strategically introducing appropriate pre-processing and post-processing steps together with the super-resolution iterations in order to tailor optimized overall processing sequences for imagery data of specific formats. Three distinct methods for tailoring a pre-processing filter and integrating it with the super-resolution processing steps will be outlined in this paper. These methods consist of a Region-of-Interest (ROI) extraction scheme, a background- detail separation procedure, and a scene-derived information extraction step for implementing a set-theoretic restoration of the image that is less demanding in computation compared to the super-resolution iterations. A
Real-Time Simulation Computation System. [for digital flight simulation of research aircraft
Fetter, J. L.
1981-01-01
The Real-Time Simulation Computation System, which will provide the flexibility necessary for operation in the research environment at the Ames Research Center is discussed. Designing the system with common subcomponents and using modular construction techniques enhances expandability and maintainability qualities. The 10-MHz series transmission scheme is the basis of the Input/Output Unit System and is the driving force providing the system flexibility. Error checking and detection performed on the transmitted data provide reliability measurements and assurances that accurate data are received at the simulators.
Evolution of perturbed dynamical systems: analytical computation with time independent accuracy
Gurzadyan, A. V.; Kocharyan, A. A.
2016-12-01
An analytical method for investigation of the evolution of dynamical systems with independent on time accuracy is developed for perturbed Hamiltonian systems. The error-free estimation using of computer algebra enables the application of the method to complex multi-dimensional Hamiltonian and dissipative systems. It also opens principal opportunities for the qualitative study of chaotic trajectories. The performance of the method is demonstrated on perturbed two-oscillator systems. It can be applied to various non-linear physical and astrophysical systems, e.g. to long-term planetary dynamics.
Soft Computing Based Procurement Planning of Time-variable Demand in Manufacturing Systems
Institute of Scientific and Technical Information of China (English)
Kai Leung Yung; Wai Hung Ip; Ding-Wei Wang
2007-01-01
Procurement planning with discrete time varying demand is an important problem in Enterprise Resource Planning (ERP). It can be described using the non-analytic mathematical programming model proposed in this paper. To solve the model we propose to use a fuzzy decision embedded genetic algorithm. The algorithm adopts an order strategy selection to simplify the original real optimization problem into binary ones. Then, a fuzzy decision quantification method is used to quantify experience from planning experts. Thus, decision rules can easily be embedded in the computation of genetic operations. This approach is applied to purchase planning problem in a practical machine tool works, where satisfactory results have been achieved.
Evolution of perturbed dynamical systems: analytical computation with time independent accuracy
Energy Technology Data Exchange (ETDEWEB)
Gurzadyan, A.V. [Russian-Armenian (Slavonic) University, Department of Mathematics and Mathematical Modelling, Yerevan (Armenia); Kocharyan, A.A. [Monash University, School of Physics and Astronomy, Clayton (Australia)
2016-12-15
An analytical method for investigation of the evolution of dynamical systems with independent on time accuracy is developed for perturbed Hamiltonian systems. The error-free estimation using of computer algebra enables the application of the method to complex multi-dimensional Hamiltonian and dissipative systems. It also opens principal opportunities for the qualitative study of chaotic trajectories. The performance of the method is demonstrated on perturbed two-oscillator systems. It can be applied to various non-linear physical and astrophysical systems, e.g. to long-term planetary dynamics. (orig.)
Evolution of perturbed dynamical systems: analytical computation with time independent accuracy
Gurzadyan, A V
2016-01-01
An analytical method for investigation of the evolution of dynamical systems {\\it with independent on time accuracy} is developed for perturbed Hamiltonian systems. The error-free estimation using of computer algebra enables the application of the method to complex multi-dimensional Hamiltonian and dissipative systems. It also opens principal opportunities for the qualitative study of chaotic trajectories. The performance of the method is demonstrated on perturbed two-oscillator systems. It can be applied to various non-linear physical and astrophysical systems, e.g. to the long-term planetary dynamics.
High-power graphic computers for visual simulation: a real-time--rendering revolution
Kaiser, M. K.
1996-01-01
Advances in high-end graphics computers in the past decade have made it possible to render visual scenes of incredible complexity and realism in real time. These new capabilities make it possible to manipulate and investigate the interactions of observers with their visual world in ways once only dreamed of. This paper reviews how these developments have affected two preexisting domains of behavioral research (flight simulation and motion perception) and have created a new domain (virtual environment research) which provides tools and challenges for the perceptual psychologist. Finally, the current limitations of these technologies are considered, with an eye toward how perceptual psychologist might shape future developments.
Magnetic Photon Splitting Computations of Proper-time Rates and Spectra
Baring, M G; Baring, Matthew G.; Harding, Alice K.
1997-01-01
The splitting of photons in the presence of an intense magnetic field has recently found astrophysical applications in polar cap models of gamma-ray pulsars and in magnetar scenarios for soft gamma repeaters. Numerical computation of the polarization-dependent rates of this third order QED process for arbitrary field strengths and energies below pair creation threshold is difficult: thus early analyses focused on analytic developments and simpler asymptotic forms. The recent astrophysical interest spurred the use of the S-matrix approach by Mentzel, Berg and Wunner to determine splitting rates. In this paper, we present numerical computations of a full proper-time expression for the rate of splitting that was obtained by Stoneham, and is exact up to the pair creation threshold. While the numerical results derived here are in accord with the earlier asymptotic forms due to Adler, our computed rates still differ by as much as factors of 3 from the S-matrix re-evaluation of Wilke and Wunner, reflecting the extre...
JPL control/structure interaction test bed real-time control computer architecture
Briggs, Hugh C.
1989-01-01
The Control/Structure Interaction Program is a technology development program for spacecraft that exhibit interactions between the control system and structural dynamics. The program objectives include development and verification of new design concepts - such as active structure - and new tools - such as combined structure and control optimization algorithm - and their verification in ground and possibly flight test. A focus mission spacecraft was designed based upon a space interferometer and is the basis for design of the ground test article. The ground test bed objectives include verification of the spacecraft design concepts, the active structure elements and certain design tools such as the new combined structures and controls optimization tool. In anticipation of CSI technology flight experiments, the test bed control electronics must emulate the computation capacity and control architectures of space qualifiable systems as well as the command and control networks that will be used to connect investigators with the flight experiment hardware. The Test Bed facility electronics were functionally partitioned into three units: a laboratory data acquisition system for structural parameter identification and performance verification; an experiment supervisory computer to oversee the experiment, monitor the environmental parameters and perform data logging; and a multilevel real-time control computing system. The design of the Test Bed electronics is presented along with hardware and software component descriptions. The system should break new ground in experimental control electronics and is of interest to anyone working in the verification of control concepts for large structures.
Minsker, B. S.; Zimmer, A. L.; Ostfeld, A.; Schmidt, A.
2014-12-01
Enabling real-time decision support, particularly under conditions of uncertainty, requires computationally efficient algorithms that can rapidly generate recommendations. In this paper, a suite of model predictive control (MPC) genetic algorithms are developed and tested offline to explore their value for reducing CSOs during real-time use in a deep-tunnel sewer system. MPC approaches include the micro-GA, the probability-based compact GA, and domain-specific GA methods that reduce the number of decision variable values analyzed within the sewer hydraulic model, thus reducing algorithm search space. Minimum fitness and constraint values achieved by all GA approaches, as well as computational times required to reach the minimum values, are compared to large population sizes with long convergence times. Optimization results for a subset of the Chicago combined sewer system indicate that genetic algorithm variations with coarse decision variable representation, eventually transitioning to the entire range of decision variable values, are most efficient at addressing the CSO control problem. Although diversity-enhancing micro-GAs evaluate a larger search space and exhibit shorter convergence times, these representations do not reach minimum fitness and constraint values. The domain-specific GAs prove to be the most efficient and are used to test CSO sensitivity to energy costs, CSO penalties, and pressurization constraint values. The results show that CSO volumes are highly dependent on the tunnel pressurization constraint, with reductions of 13% to 77% possible with less conservative operational strategies. Because current management practices may not account for varying costs at CSO locations and electricity rate changes in the summer and winter, the sensitivity of the results is evaluated for variable seasonal and diurnal CSO penalty costs and electricity-related system maintenance costs, as well as different sluice gate constraint levels. These findings indicate
Energy Technology Data Exchange (ETDEWEB)
Zhang, Xue; Xiao, Yang [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China); Zeng, Jie [Department of Medical Ultrasonics, Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou (China); Qiu, Weibao; Qian, Ming; Wang, Congzhi [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China); Zheng, Rongqin, E-mail: zhengronggin@hotmail.com [Department of Medical Ultrasonics, Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou (China); Zheng, Hairong, E-mail: hr.zheng@siat.ac.cn [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China)
2014-01-15
Purpose: To develop and evaluate a computer-assisted method of quantifying five-point elasticity scoring system based on ultrasound real-time elastography (RTE), for classifying benign and malignant breast lesions, with pathologic results as the reference standard. Materials and methods: Conventional ultrasonography (US) and RTE images of 145 breast lesions (67 malignant, 78 benign) were performed in this study. Each lesion was automatically contoured on the B-mode image by the level set method and mapped on the RTE image. The relative elasticity value of each pixel was reconstructed and classified into hard or soft by the fuzzy c-means clustering method. According to the hardness degree inside lesion and its surrounding tissue, the elasticity score of the RTE image was computed in an automatic way. Visual assessments of the radiologists were used for comparing the diagnostic performance. Histopathologic examination was used as the reference standard. The Student's t test and receiver operating characteristic (ROC) curve analysis were performed for statistical analysis. Results: Considering score 4 or higher as test positive for malignancy, the diagnostic accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) were 93.8% (136/145), 92.5% (62/67), 94.9% (74/78), 93.9% (62/66), and 93.7% (74/79) for the computer-assisted scheme, and 89.7% (130/145), 85.1% (57/67), 93.6% (73/78), 92.0% (57/62), and 88.0% (73/83) for manual assessment. Area under ROC curve (A{sub z} value) for the proposed method was higher than the A{sub z} value for visual assessment (0.96 vs. 0.93). Conclusion: Computer-assisted quantification of classical five-point scoring system can significantly eliminate the interobserver variability and thereby improve the diagnostic confidence of classifying the breast lesions to avoid unnecessary biopsy.
Atkinson, Paul
2011-01-01
The pixelated rectangle we spend most of our day staring at in silence is not the television as many long feared, but the computer-the ubiquitous portal of work and personal lives. At this point, the computer is almost so common we don't notice it in our view. It's difficult to envision that not that long ago it was a gigantic, room-sized structure only to be accessed by a few inspiring as much awe and respect as fear and mystery. Now that the machine has decreased in size and increased in popular use, the computer has become a prosaic appliance, little-more noted than a toaster. These dramati
Hoekstra, A.G.; Sloot, P.M.A.; Haan, M.J.; Hertzberger, L.O.; van Leeuwen, J.
1991-01-01
New developments in Computer Science, both hardware and software, offer researchers, such as physicists, unprecedented possibilities to solve their computational intensive problems.However, full exploitation of e.g. new massively parallel computers, parallel languages or runtime environments require
A Computational Method for 3D Anisotropic Travel-time Tomography of Rocks in the Laboratory
Ghofranitabari, Mehdi; Young, R. Paul
2013-04-01
True triaxial loading in the laboratory applies three principal stresses on a cubic rock specimen. Elliptical anisotropy and distributed heterogeneities are introduced in the rock due to closure and opening of the pre-existing cracks and creation and growth of the new aligned cracks. The rock sample is tested in a Geophysical Imaging Cell that is armed with an Acoustic Emission monitoring system which can perform transducer to transducer velocity surveys to image velocity structure of the sample during the experiment. Ultrasonic travel-time tomography as a non-destructive method outfits a map of wave propagation velocity in the sample in order to detect the uniformly distributed or localised heterogeneities and provide the spatial variation and temporal evolution of induced damages in rocks at various stages of loading. The rock sample is partitioned into cubic grid cells as model space. Ray-based tomography method measuring body wave travel time along ray paths between pairs of emitting and receiving transducers is used to calculate isotropic ray-path segment matrix elements (Gij) which contain segment lengths of the ith ray in the jth cell in three dimensions. Synthetic P wave travel times are computed between pairs of transducers in a hypothetical isotropic heterogeneous cubic sample as data space along with an error due to precision of measurement. 3D strain of the squeezed rock and the consequent geometrical deformation is also included in computations for further accuracy. Singular Value Decomposition method is used for the inversion from data space to model space. In the next step, the anisotropic ray-path segment matrix and the corresponded data space are computed for hypothetical anisotropic heterogeneous samples based on the elliptical anisotropic model of velocity which is obtained from the real laboratory experimental data. The method is examined for several different synthetic heterogeneous models. An "Inaccuracy factor" is utilized to inquire the
Zhang, Baohui
Modeling has been promoted by major policy organizations as important for science learning. The purpose of this dissertation is to describe and explore middle school science students' computer-based modeling practices and their changes over time using a scaffolded modeling program. Following a "design-based research" approach, this study was conducted at an independent school. Seventh graders from three classes taught by two experienced teachers participated. Two pairs of target students were chosen from each class for observation. Students created computer-based models after their investigations in a water quality unit and a decomposition unit. The initial modeling cycle for water quality lasted for four days in the fall season, the second cycle for water quality lasted three days in the winter season, and the third cycle for decomposition lasted two days in the spring season. The major data source is video that captured student pairs' computer screen activities and their conversations. Supplementary data include classroom videos of those modeling cycles, replicated students' final models, and models in production. The data were analyzed in terms of the efficiency, meaningfulness, and purposefulness of students' modeling practices. Students' understanding of content, models and modeling, metacognition, and collaboration and their changes were analyzed as secondary learning outcomes. This dissertation shows that with appropriate scaffolding from the modeling program and the teachers, students performed a variety of modeling practices that are valued by science educators, such as planning, analyzing, synthesizing, evaluating, and publicizing. In general, student modeling practices became more efficient, meaningful, and purposeful over time. During their modeling practices, students also made use of and improved content knowledge, understanding of models and modeling, metacognition, and collaboration. Suggestions for improving the modeling program and the learning
A 2D Time Domain DRBEM Computer Model for MagnetoThermoelastic Coupled Wave Propagation Problems
Directory of Open Access Journals (Sweden)
Mohamed Abdelsabour Fahmy
2014-07-01
Full Text Available A numerical computer model based on the dual reciprocity boundary element method (DRBEM is extended to study magneto-thermoelastic coupled wave propagation problems with relaxation times involving anisotropic functionally graded solids. The model formulation is tested through its application to the problem of a solid placed in a constant primary magnetic field acting in the direction of the z-axis and rotating about this axis with a constant angular velocity. In the case of two-dimensional deformation, an implicit-explicit time domain DRBEM was presented and implemented to obtain the solution for the displacement and temperature fields. A comparison of the results is presented graphically in the context of Lord and Shulman (LS and Green and Lindsay (GL theories. Numerical results that demonstrate the validity of the proposed method are also presented graphically.
Rowland, Mark S.; Howard, Douglas E.; Wong, James L.; Jessup, James L.; Bianchini, Greg M.; Miller, Wayne O.
2007-10-23
A real-time method and computer system for identifying radioactive materials which collects gamma count rates from a HPGe gamma-radiation detector to produce a high-resolution gamma-ray energy spectrum. A library of nuclear material definitions ("library definitions") is provided, with each uniquely associated with a nuclide or isotope material and each comprising at least one logic condition associated with a spectral parameter of a gamma-ray energy spectrum. The method determines whether the spectral parameters of said high-resolution gamma-ray energy spectrum satisfy all the logic conditions of any one of the library definitions, and subsequently uniquely identifies the material type as that nuclide or isotope material associated with the satisfied library definition. The method is iteratively repeated to update the spectrum and identification in real time.
Huijse, Pablo; Protopapas, Pavlos; Principe, Jose C; Zegers, Pablo
2015-01-01
Time-domain astronomy (TDA) is facing a paradigm shift caused by the exponential growth of the sample size, data complexity and data generation rates of new astronomical sky surveys. For example, the Large Synoptic Survey Telescope (LSST), which will begin operations in northern Chile in 2022, will generate a nearly 150 Petabyte imaging dataset of the southern hemisphere sky. The LSST will stream data at rates of 2 Terabytes per hour, effectively capturing an unprecedented movie of the sky. The LSST is expected not only to improve our understanding of time-varying astrophysical objects, but also to reveal a plethora of yet unknown faint and fast-varying phenomena. To cope with a change of paradigm to data-driven astronomy, the fields of astroinformatics and astrostatistics have been created recently. The new data-oriented paradigms for astronomy combine statistics, data mining, knowledge discovery, machine learning and computational intelligence, in order to provide the automated and robust methods needed for...
Computational Models of X-Ray Burst Quenching Times and 12C Nucleosynthesis Following a Superburst
Energy Technology Data Exchange (ETDEWEB)
Fisker, J L
2009-03-19
Superbursts are energetic events on neutron stars that are a thousand times more powerful than ordinary type I X-ray bursts. They are believed to be powered by a thermonuclear explosion of accumulated {sup 12}C. However, the source of this {sup 12}C remains elusive to theoretical calculations and its concentration and ignition depth are both unknown. Here we present the first computational simulations of the nucleosynthesis during the thermal decay of a superbust, where X-ray bursts are quenched. Our calculations of the quenching time verify previous analytical calculations and shed new light on the physics of stable burning at low accretion rates. We show that concentrated (X{sub {sup 12}C} {approx}> 0.40), although insufficient, amounts of {sup 12}C are generated during the several weeks following the superburst where the decaying thermal flux of the superburst stabilizes the burning of the accreted material.
Computer-Aided Software Engineering - An approach to real-time software development
Walker, Carrie K.; Turkovich, John J.
1989-01-01
A new software engineering discipline is Computer-Aided Software Engineering (CASE), a technology aimed at automating the software development process. This paper explores the development of CASE technology, particularly in the area of real-time/scientific/engineering software, and a history of CASE is given. The proposed software development environment for the Advanced Launch System (ALS CASE) is described as an example of an advanced software development system for real-time/scientific/engineering (RT/SE) software. The Automated Programming Subsystem of ALS CASE automatically generates executable code and corresponding documentation from a suitably formatted specification of the software requirements. Software requirements are interactively specified in the form of engineering block diagrams. Several demonstrations of the Automated Programming Subsystem are discussed.
Static security based available transfer capability (ATC computation for real-time power markets
Directory of Open Access Journals (Sweden)
Venkaiah Chintham
2010-01-01
Full Text Available In power system deregulation, the Independent System Operator (ISO has the responsibility to control the power transactions and avoid overloading of the transmission lines beyond their thermal limits. To achieve this, the ISO has to update in real-time periodically Available Transfer Capability (ATC index for enabling market participants to reserve the transmission service. In this paper Static Security based ATC has been computed for real-time applications using three artificial intelligent methods viz.: i Back Propagation Algorithm (BPA; ii Radial Basis Function (RBF Neural network; and iii Adaptive Neuro Fuzzy Inference System (ANFIS. These three different intelligent methods are tested on IEEE 24-bus Reliability Test System (RTS and 75-bus practical System for the base case and critical line outage cases for different transactions. The results are compared with the conventional full AC Load Flow method for different transactions.
Energy Technology Data Exchange (ETDEWEB)
Watson, S.B.; Snyder, W.S.; Ford, M.R.
1976-12-01
TIMED is a computer program designed to calculate cumulated radioactivity in the various source organs at various times after radionuclide deposition. TIMED embodies a system of differential equations which describes activity transfer in the lungs, gastrointestinal tract, and other organs of the body. This system accounts for delay of transfer of activity between compartments of the body and radioactive daughters.
Just-in-Time Compilation-Inspired Methodology for Parallelization of Compute Intensive Java Code
Directory of Open Access Journals (Sweden)
GHULAM MUSTAFA
2017-01-01
Full Text Available Compute intensive programs generally consume significant fraction of execution time in a small amount of repetitive code. Such repetitive code is commonly known as hotspot code. We observed that compute intensive hotspots often possess exploitable loop level parallelism. A JIT (Just-in-Time compiler profiles a running program to identify its hotspots. Hotspots are then translated into native code, for efficient execution. Using similar approach, we propose a methodology to identify hotspots and exploit their parallelization potential on multicore systems. Proposed methodology selects and parallelizes each DOALL loop that is either contained in a hotspot method or calls a hotspot method. The methodology could be integrated in front-end of a JIT compiler to parallelize sequential code, just before native translation. However, compilation to native code is out of scope of this work. As a case study, we analyze eighteen JGF (Java Grande Forum benchmarks to determine parallelization potential of hotspots. Eight benchmarks demonstrate a speedup of up to 7.6x on an 8-core system
Applications of soft computing in time series forecasting simulation and modeling techniques
Singh, Pritpal
2016-01-01
This book reports on an in-depth study of fuzzy time series (FTS) modeling. It reviews and summarizes previous research work in FTS modeling and also provides a brief introduction to other soft-computing techniques, such as artificial neural networks (ANNs), rough sets (RS) and evolutionary computing (EC), focusing on how these techniques can be integrated into different phases of the FTS modeling approach. In particular, the book describes novel methods resulting from the hybridization of FTS modeling approaches with neural networks and particle swarm optimization. It also demonstrates how a new ANN-based model can be successfully applied in the context of predicting Indian summer monsoon rainfall. Thanks to its easy-to-read style and the clear explanations of the models, the book can be used as a concise yet comprehensive reference guide to fuzzy time series modeling, and will be valuable not only for graduate students, but also for researchers and professionals working for academic, business and governmen...
Kajian dan Implementasi Real Time Operating System pada Single Board Computer Berbasis ARM
Directory of Open Access Journals (Sweden)
Wiedjaja
2014-06-01
Full Text Available Operating System is an important software in computer system. For personal and office use the operating system is sufficient. However, to critical mission applications such as nuclear power plants and braking system on the car (auto braking system which need a high level of reliability, it requires operating system which operates in real time. The study aims to assess the implementation of the Linux-based operating system on a Single Board Computer (SBC ARM-based, namely Pandaboard ES with the Dual-core ARM Cortex-A9, TI OMAP 4460 type. Research was conducted by the method of implementation of the General Purpose OS Ubuntu 12:04 OMAP4-armhf-RTOS and Linux 3.4.0-rt17 + on PandaBoard ES. Then research compared the latency value of each OS on no-load and with full-load condition. The results obtained show the maximum latency value of RTOS on full load condition is at 45 uS, much smaller than the maximum value of GPOS at full-load at 17.712 uS. The lower value of latency demontrates that the RTOS has ability to run the process in a certain period of time much better than the GPOS.
Real-Time and High-Accuracy Arctangent Computation Using CORDIC and Fast Magnitude Estimation
Directory of Open Access Journals (Sweden)
Luca Pilato
2017-03-01
Full Text Available This paper presents an improved VLSI (Very Large Scale of Integration architecture for real-time and high-accuracy computation of trigonometric functions with fixed-point arithmetic, particularly arctangent using CORDIC (Coordinate Rotation Digital Computer and fast magnitude estimation. The standard CORDIC implementation suffers of a loss of accuracy when the magnitude of the input vector becomes small. Using a fast magnitude estimator before running the standard algorithm, a pre-processing magnification is implemented, shifting the input coordinates by a proper factor. The entire architecture does not use a multiplier, it uses only shift and add primitives as the original CORDIC, and it does not change the data path precision of the CORDIC core. A bit-true case study is presented showing a reduction of the maximum phase error from 414 LSB (angle error of 0.6355 rad to 4 LSB (angle error of 0.0061 rad, with small overheads of complexity and speed. Implementation of the new architecture in 0.18 µm CMOS technology allows for real-time and low-power processing of CORDIC and arctangent, which are key functions in many embedded DSP systems. The proposed macrocell has been verified by integration in a system-on-chip, called SENSASIP (Sensor Application Specific Instruction-set Processor, for position sensor signal processing in automotive measurement applications.
Real-Time Compressive Sensing MRI Reconstruction Using GPU Computing and Split Bregman Methods
Directory of Open Access Journals (Sweden)
David S. Smith
2012-01-01
Full Text Available Compressive sensing (CS has been shown to enable dramatic acceleration of MRI acquisition in some applications. Being an iterative reconstruction technique, CS MRI reconstructions can be more time-consuming than traditional inverse Fourier reconstruction. We have accelerated our CS MRI reconstruction by factors of up to 27 by using a split Bregman solver combined with a graphics processing unit (GPU computing platform. The increases in speed we find are similar to those we measure for matrix multiplication on this platform, suggesting that the split Bregman methods parallelize efficiently. We demonstrate that the combination of the rapid convergence of the split Bregman algorithm and the massively parallel strategy of GPU computing can enable real-time CS reconstruction of even acquisition data matrices of dimension 40962 or more, depending on available GPU VRAM. Reconstruction of two-dimensional data matrices of dimension 10242 and smaller took ~0.3 s or less, showing that this platform also provides very fast iterative reconstruction for small-to-moderate size images.
Time Is Not Space: Core Computations and Domain-Specific Networks for Mental Travels.
Gauthier, Baptiste; van Wassenhove, Virginie
2016-11-23
Humans can consciously project themselves in the future and imagine themselves at different places. Do mental time travel and mental space navigation abilities share common cognitive and neural mechanisms? To test this, we recorded fMRI while participants mentally projected themselves in time or in space (e.g., 9 years ago, in Paris) and ordered historical events from their mental perspective. Behavioral patterns were comparable for mental time and space and shaped by self-projection and by the distance of historical events to the mental position of the self, suggesting the existence of egocentric mapping in both dimensions. Nonetheless, self-projection in space engaged the medial and lateral parietal cortices, whereas self-projection in time engaged a widespread parietofrontal network. Moreover, while a large distributed network was found for spatial distances, temporal distances specifically engaged the right inferior parietal cortex and the anterior insula. Across these networks, a robust overlap was only found in a small region of the inferior parietal lobe, adding evidence for its role in domain-general egocentric mapping. Our findings suggest that mental travel in time or space capitalizes on egocentric remapping and on distance computation, which are implemented in distinct dimension-specific cortical networks converging in inferior parietal lobe.
Mamykina, Lena; Vawdrey, David K; Hripcsak, George
2016-06-01
To understand how much time residents spend using computers compared with other activities, and what residents use computers for. This time and motion study was conducted in June and July 2010 at NewYork-Presbyterian/Columbia University Medical Center with seven residents (first-, second-, and third-year) on the general medicine service. An experienced observer shadowed residents during a single day shift, captured all their activities using an iPad application, and took field notes. The activities were captured using a validated taxonomy of clinical activities, expanded to describe computer-based activities with a greater level of detail. Residents spent 364.5 minutes (50.6%) of their shift time using computers, compared with 67.8 minutes (9.4%) interacting with patients. In addition, they spent 292.3 minutes (40.6%) talking with others in person, 186.0 minutes (25.8%) handling paper notes, 79.7 minutes (11.1%) in rounds, 80.0 minutes (11.1%) walking or waiting, and 54.0 minutes (7.5%) talking on the phone. Residents spent 685 minutes (59.6%) multitasking. Computer-based documentation activities amounted to 189.9 minutes (52.1%) of all computer-based activities time, with 128.7 minutes (35.3%) spent writing notes and 27.3 minutes (7.5%) reading notes composed by others. The study showed that residents spent considerably more time interacting with computers (over 50% of their shift time) than in direct contact with patients (less than 10% of their shift time). Some of this may be due to an increasing reliance on computing systems for access to patient data, further exacerbated by inefficiencies in the design of the electronic health record.
Percutaneous drainage of psoas abscess under real-time computed tomography fluoroscopic guidance
Energy Technology Data Exchange (ETDEWEB)
Yamagami, Takuji; Terayama, Koshi; Yoshimatsu, Rika; Matsumoto, Tomohiro; Miura, Hiroshi; Nishimura, Tsunehiko [Kyoto Prefectural University of Medicine, Department of Radiology, Graduate School of Medical Science, Kyoto (Japan)
2009-03-15
The goal of our study was to determine the usefulness of percutaneous abscess drainage under guidance of computed tomography (CT) fluoroscopy. Our subjects were 21 patients (seven women, 14 men; mean age 64 years; age range 30-87 years) who had undergone percutaneous drainage of 26 psoas abscess lesions under CT fluoroscopic guidance between May 2001 and January 2008. Drainage methods involved a needle puncture, insertion of a guidewire, serial dilations, and the exchange of the needle with a drainage tube. The procedures were guided by use of a helical CT scanner that provided real-time fluoroscopic reconstruction. Percutaneous drainage under real-time CT fluoroscopic guidance was successfully performed in every procedure. Use of real-time CT fluoroscopy allowed rapid assessment of needle, guidewire, and catheter placement. No patient had serious complications related to the drainage procedure. The mean procedure time required to drain one lesion was 35.6 {+-} 13.6 min. Mean period after the drainage procedure until complete disappearance of the abscess as confirmed by CT was 12.4 {+-} 10.0 days (range 3-46 days). CT fluoroscopy is useful in achieving accurate and safe drainage of abnormal psoas fluid collections. (orig.)
Zatarain-Salazar, J.; Reed, P. M.; Quinn, J.; Giuliani, M.; Castelletti, A.
2016-12-01
As we confront the challenges of managing river basin systems with a large number of reservoirs and increasingly uncertain tradeoffs impacting their operations (due to, e.g. climate change, changing energy markets, population pressures, ecosystem services, etc.), evolutionary many-objective direct policy search (EMODPS) solution strategies will need to address the computational demands associated with simulating more uncertainties and therefore optimizing over increasingly noisy objective evaluations. Diagnostic assessments of state-of-the-art many-objective evolutionary algorithms (MOEAs) to support EMODPS have highlighted that search time (or number of function evaluations) and auto-adaptive search are key features for successful optimization. Furthermore, auto-adaptive MOEA search operators are themselves sensitive to having a sufficient number of function evaluations to learn successful strategies for exploring complex spaces and for escaping from local optima when stagnation is detected. Fortunately, recent parallel developments allow coordinated runs that enhance auto-adaptive algorithmic learning and can handle scalable and reliable search with limited wall-clock time, but at the expense of the total number of function evaluations. In this study, we analyze this tradeoff between parallel coordination and depth of search using different parallelization schemes of the Multi-Master Borg on a many-objective stochastic control problem. We also consider the tradeoff between better representing uncertainty in the stochastic optimization, and simplifying this representation to shorten the function evaluation time and allow for greater search. Our analysis focuses on the Lower Susquehanna River Basin (LSRB) system where multiple competing objectives for hydropower production, urban water supply, recreation and environmental flows need to be balanced. Our results provide guidance for balancing exploration, uncertainty, and computational demands when using the EMODPS
I. Fisk
2010-01-01
Introduction The first data taking period of November produced a first scientific paper, and this is a very satisfactory step for Computing. It also gave the invaluable opportunity to learn and debrief from this first, intense period, and make the necessary adaptations. The alarm procedures between different groups (DAQ, Physics, T0 processing, Alignment/calibration, T1 and T2 communications) have been reinforced. A major effort has also been invested into remodeling and optimizing operator tasks in all activities in Computing, in parallel with the recruitment of new Cat A operators. The teams are being completed and by mid year the new tasks will have been assigned. CRB (Computing Resource Board) The Board met twice since last CMS week. In December it reviewed the experience of the November data-taking period and could measure the positive improvements made for the site readiness. It also reviewed the policy under which Tier-2 are associated with Physics Groups. Such associations are decided twice per ye...
Time-Of-Flight Camera, Optical Tracker and Computed Tomography in Pairwise Data Registration.
Directory of Open Access Journals (Sweden)
Bartlomiej Pycinski
Full Text Available A growing number of medical applications, including minimal invasive surgery, depends on multi-modal or multi-sensors data processing. Fast and accurate 3D scene analysis, comprising data registration, seems to be crucial for the development of computer aided diagnosis and therapy. The advancement of surface tracking system based on optical trackers already plays an important role in surgical procedures planning. However, new modalities, like the time-of-flight (ToF sensors, widely explored in non-medical fields are powerful and have the potential to become a part of computer aided surgery set-up. Connection of different acquisition systems promises to provide a valuable support for operating room procedures. Therefore, the detailed analysis of the accuracy of such multi-sensors positioning systems is needed.We present the system combining pre-operative CT series with intra-operative ToF-sensor and optical tracker point clouds. The methodology contains: optical sensor set-up and the ToF-camera calibration procedures, data pre-processing algorithms, and registration technique. The data pre-processing yields a surface, in case of CT, and point clouds for ToF-sensor and marker-driven optical tracker representation of an object of interest. An applied registration technique is based on Iterative Closest Point algorithm.The experiments validate the registration of each pair of modalities/sensors involving phantoms of four various human organs in terms of Hausdorff distance and mean absolute distance metrics. The best surface alignment was obtained for CT and optical tracker combination, whereas the worst for experiments involving ToF-camera.The obtained accuracies encourage to further develop the multi-sensors systems. The presented substantive discussion concerning the system limitations and possible improvements mainly related to the depth information produced by the ToF-sensor is useful for computer aided surgery developers.
A State-of-the-Art Review of the Real-Time Computer-Aided Study of the Writing Process
Abdel Latif, Muhammad M.
2008-01-01
Writing researchers have developed various methods for investigating the writing process since the 1970s. The early 1980s saw the occurrence of the real-time computer-aided study of the writing process that relies on the protocols generated by recording the computer screen activities as writers compose using the word processor. This article…
Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces
Bocquelet, Florent; Hueber, Thomas; Girin, Laurent; Savariaux, Christophe; Yvert, Blaise
2016-01-01
Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI) controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips) into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN) trained on electromagnetic articulography (EMA) data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer. PMID:27880768
Real-Time Control of an Articulatory-Based Speech Synthesizer for Brain Computer Interfaces.
Directory of Open Access Journals (Sweden)
Florent Bocquelet
2016-11-01
Full Text Available Restoring natural speech in paralyzed and aphasic people could be achieved using a Brain-Computer Interface (BCI controlling a speech synthesizer in real-time. To reach this goal, a prerequisite is to develop a speech synthesizer producing intelligible speech in real-time with a reasonable number of control parameters. We present here an articulatory-based speech synthesizer that can be controlled in real-time for future BCI applications. This synthesizer converts movements of the main speech articulators (tongue, jaw, velum, and lips into intelligible speech. The articulatory-to-acoustic mapping is performed using a deep neural network (DNN trained on electromagnetic articulography (EMA data recorded on a reference speaker synchronously with the produced speech signal. This DNN is then used in both offline and online modes to map the position of sensors glued on different speech articulators into acoustic parameters that are further converted into an audio signal using a vocoder. In offline mode, highly intelligible speech could be obtained as assessed by perceptual evaluation performed by 12 listeners. Then, to anticipate future BCI applications, we further assessed the real-time control of the synthesizer by both the reference speaker and new speakers, in a closed-loop paradigm using EMA data recorded in real time. A short calibration period was used to compensate for differences in sensor positions and articulatory differences between new speakers and the reference speaker. We found that real-time synthesis of vowels and consonants was possible with good intelligibility. In conclusion, these results open to future speech BCI applications using such articulatory-based speech synthesizer.
Diffuse globally, compute locally: a cyclist approach to modeling long time robot locomotion
Zhang, Tingnan; Goldman, Daniel; Cvitanović, Predrag
2015-03-01
To advance autonomous robots we are interested to develop a statistical/dynamical description of diffusive self-propulsion on heterogeneous terrain. We consider a minimal model for such diffusion, the 2-dimensional Lorentz gas, which abstracts the motion of a light, point-like particle bouncing within a large number of heavy scatters (e.g. small robots in a boulder field). We present a precise computation (based on exact periodic orbit theory formula for the diffusion constant) for a periodic triangular Lorentz gas with finite horizon. We formulate a new approach to tiling the plane in terms of three elementary tiling generators which, for the first time, enables use of periodic orbits computed in the fundamental domain (that is, 1 / 12 of the hexagonal elementary cell whose translations tile the entire plane). Compared with previous literature, our fundamental domain value of the diffusion constant converges quickly for inter-disk separation/disk radius > 0 . 2 , with the cycle expansion truncated to only a few hundred periodic orbits of up to 5 billiard wall bounces. For small inter-disk separations, with periodic orbits up to 6 bounces, our diffusion constants are close (< 10 %) to direct numerical simulation estimates and the recent literature probabilistic estimates.
Elliott, Kenny B.; Ugoletti, Roberto; Sulla, Jeff
1992-01-01
The evolution and optimization of a real-time digital control system is presented. The control system is part of a testbed used to perform focused technology research on the interactions of spacecraft platform and instrument controllers with the flexible-body dynamics of the platform and platform appendages. The control system consists of Computer Automated Measurement and Control (CAMAC) standard data acquisition equipment interfaced to a workstation computer. The goal of this work is to optimize the control system's performance to support controls research using controllers with up to 50 states and frame rates above 200 Hz. The original system could support a 16-state controller operating at a rate of 150 Hz. By using simple yet effective software improvements, Input/Output (I/O) latencies and contention problems are reduced or eliminated in the control system. The final configuration can support a 16-state controller operating at 475 Hz. Effectively the control system's performance was increased by a factor of 3.
Computing the time-continuous Optimal Mass Transport Problem without Lagrangian techniques
Besson, Olivier; Pousin, Jérôme
2010-01-01
This work originates from a heart's images tracking which is to generate an apparent continuous motion, observable through intensity variation from one starting image to an ending one both supposed segmented. Given two images p0 and p1, we calculate an evolution process p(t, \\cdot) which transports p0 to p1 by using the optimal extended optical flow. In this paper we propose an algorithm based on a fixed point formulation and a time-space least squares formulation of the mass conservation equation for computing the optimal mass transport problem. The strategy is implemented in a 2D case and numerical results are presented with a first order Lagrange finite element, showing the efficiency of the proposed strategy.
The Impact of Mutation Rate on the Computation Time of Evolutionary Dynamic Optimization
Chen, Tianshi; Tang, Ke; Chen, Guoliang; Yao, Xin
2011-01-01
Mutation has traditionally been regarded as an important operator in evolutionary algorithms. In particular, there have been many experimental studies which showed the effectiveness of adapting mutation rates for various static optimization problems. Given the perceived effectiveness of adaptive and self-adaptive mutation for static optimization problems, there have been speculations that adaptive and self-adaptive mutation can benefit dynamic optimization problems even more since adaptation and self-adaptation are capable of following a dynamic environment. However, few theoretical results are available in analyzing rigorously evolutionary algorithms for dynamic optimization problems. It is unclear when adaptive and self-adaptive mutation rates are likely to be useful for evolutionary algorithms in solving dynamic optimization problems. This paper provides the first rigorous analysis of adaptive mutation and its impact on the computation times of evolutionary algorithms in solving certain dynamic optimizatio...
Grigoryeva, Lyudmila; Henriques, Julie; Larger, Laurent; Ortega, Juan-Pablo
2016-07-01
This letter addresses the reservoir design problem in the context of delay-based reservoir computers for multidimensional input signals, parallel architectures, and real-time multitasking. First, an approximating reservoir model is presented in those frameworks that provides an explicit functional link between the reservoir architecture and its performance in the execution of a specific task. Second, the inference properties of the ridge regression estimator in the multivariate context are used to assess the impact of finite sample training on the decrease of the reservoir capacity. Finally, an empirical study is conducted that shows the adequacy of the theoretical results with the empirical performances exhibited by various reservoir architectures in the execution of several nonlinear tasks with multidimensional inputs. Our results confirm the robustness properties of the parallel reservoir architecture with respect to task misspecification and parameter choice already documented in the literature.
Bousquet, Nicolas
2010-01-01
This article deals with the estimation of a probability p of an undesirable event. Its occurence is formalized by the exceedance of a threshold reliability value by the unidimensional output of a time-consuming computer code G with multivariate probabilistic input X. When G is assumed monotonous with respect to X, the Monotonous Reliability Method was proposed by de Rocquigny (2009) in an engineering context to provide sequentially narrowing 100%-confidence bounds and a crude estimate of p, via deterministic or stochastic designs of experiments. The present article consists in a formalization and technical deepening of this idea, as a large basis for future theoretical and applied studies. Three kinds of results are especially emphasized. First, the bounds themselves remain too crude and conservative estimators of p for a dimension of X upper than 2. Second, a maximum-likelihood estimator of p can be easily built, presenting a high variance reduction with respect to a standard Monte Carlo case, but suffering ...
pyCTQW: A continuous-time quantum walk simulator on distributed memory computers
Izaac, Josh A.; Wang, Jingbo B.
2015-01-01
In the general field of quantum information and computation, quantum walks are playing an increasingly important role in constructing physical models and quantum algorithms. We have recently developed a distributed memory software package pyCTQW, with an object-oriented Python interface, that allows efficient simulation of large multi-particle CTQW (continuous-time quantum walk)-based systems. In this paper, we present an introduction to the Python and Fortran interfaces of pyCTQW, discuss various numerical methods of calculating the matrix exponential, and demonstrate the performance behavior of pyCTQW on a distributed memory cluster. In particular, the Chebyshev and Krylov-subspace methods for calculating the quantum walk propagation are provided, as well as methods for visualization and data analysis.
Guido, Ciro A.; Cortona, Pietro; Adamo, Carlo
2014-03-01
We extend our previous definition of the metric Δr for electronic excitations in the framework of the time-dependent density functional theory [C. A. Guido, P. Cortona, B. Mennucci, and C. Adamo, J. Chem. Theory Comput. 9, 3118 (2013)], by including a measure of the difference of electronic position variances in passing from occupied to virtual orbitals. This new definition, called Γ, permits applications in those situations where the Δr-index is not helpful: transitions in centrosymmetric systems and Rydberg excitations. The Γ-metric is then extended by using the Natural Transition Orbitals, thus providing an intuitive picture of how locally the electron density changes during the electronic transitions. Furthermore, the Γ values give insight about the functional performances in reproducing different type of transitions, and allow one to define a "confidence radius" for GGA and hybrid functionals.
Real-Time Cognitive Computing Architecture for Data Fusion in a Dynamic Environment
Duong, Tuan A.; Duong, Vu A.
2012-01-01
A novel cognitive computing architecture is conceptualized for processing multiple channels of multi-modal sensory data streams simultaneously, and fusing the information in real time to generate intelligent reaction sequences. This unique architecture is capable of assimilating parallel data streams that could be analog, digital, synchronous/asynchronous, and could be programmed to act as a knowledge synthesizer and/or an "intelligent perception" processor. In this architecture, the bio-inspired models of visual pathway and olfactory receptor processing are combined as processing components, to achieve the composite function of "searching for a source of food while avoiding the predator." The architecture is particularly suited for scene analysis from visual data and odorant.
Directory of Open Access Journals (Sweden)
SOUVIK PAL
2016-09-01
Full Text Available Cloud computing is an emerging paradigm of Internet-centric business computing where Cloud Service Providers (CSPs are providing services to the customer according to their needs. The key perception behind cloud computing is on-demand sharing of resources available in the resource pool provided by CSP, which implies new emerging business model. The resources are provisioned when jobs arrive. The job scheduling and minimization of waiting time are the challenging issue in cloud computing. When a large number of jobs are requested, they have to wait for getting allocated to the servers which in turn may increase the queue length and also waiting time. This paper includes system design for implementation which is concerned with Johnson Scheduling Algorithm that provides the optimal sequence. With that sequence, service times can be obtained. The waiting time and queue length can be reduced using queuing model with multi-server and finite capacity which improves the job scheduling model.
Fersht, Alan R
2002-10-29
There are proposals to overcome the current incompatibilities between the time scales of protein folding and molecular dynamics simulation by using a large number of short simulations of only tens of nanoseconds (distributed computing). According to the principles of first-order kinetic processes, a sufficiently large number of short simulations will include, de facto, a small number of long time scale events that have proceeded to completion. But protein folding is not an elementary kinetic step: folding has a series of early conformational steps that lead to lag phases at the beginning of the kinetics. The presence of these lag phases can bias short simulations toward selecting minor pathways that have fewer or faster lag steps and so miss the major folding pathways. Attempts to circumvent the lags by using loosely coupled parallel simulations that search for first-order transitions are also problematic because of the difficulty of detecting transitions in molecular dynamics simulations. Nevertheless, the procedure of using parallel independent simulations is perfectly valid and quite feasible once the time scale of simulation proceeds past the lag phases into a single exponential region.
Directory of Open Access Journals (Sweden)
Ju-Chi Liu
2016-01-01
Full Text Available A high efficient time-shift correlation algorithm was proposed to deal with the peak time uncertainty of P300 evoked potential for a P300-based brain-computer interface (BCI. The time-shift correlation series data were collected as the input nodes of an artificial neural network (ANN, and the classification of four LED visual stimuli was selected as the output node. Two operating modes, including fast-recognition mode (FM and accuracy-recognition mode (AM, were realized. The proposed BCI system was implemented on an embedded system for commanding an adult-size humanoid robot to evaluate the performance from investigating the ground truth trajectories of the humanoid robot. When the humanoid robot walked in a spacious area, the FM was used to control the robot with a higher information transfer rate (ITR. When the robot walked in a crowded area, the AM was used for high accuracy of recognition to reduce the risk of collision. The experimental results showed that, in 100 trials, the accuracy rate of FM was 87.8% and the average ITR was 52.73 bits/min. In addition, the accuracy rate was improved to 92% for the AM, and the average ITR decreased to 31.27 bits/min. due to strict recognition constraints.
Liu, Ju-Chi; Chou, Hung-Chyun; Chen, Chien-Hsiu; Lin, Yi-Tseng; Kuo, Chung-Hsien
2016-01-01
A high efficient time-shift correlation algorithm was proposed to deal with the peak time uncertainty of P300 evoked potential for a P300-based brain-computer interface (BCI). The time-shift correlation series data were collected as the input nodes of an artificial neural network (ANN), and the classification of four LED visual stimuli was selected as the output node. Two operating modes, including fast-recognition mode (FM) and accuracy-recognition mode (AM), were realized. The proposed BCI system was implemented on an embedded system for commanding an adult-size humanoid robot to evaluate the performance from investigating the ground truth trajectories of the humanoid robot. When the humanoid robot walked in a spacious area, the FM was used to control the robot with a higher information transfer rate (ITR). When the robot walked in a crowded area, the AM was used for high accuracy of recognition to reduce the risk of collision. The experimental results showed that, in 100 trials, the accuracy rate of FM was 87.8% and the average ITR was 52.73 bits/min. In addition, the accuracy rate was improved to 92% for the AM, and the average ITR decreased to 31.27 bits/min. due to strict recognition constraints.
Efstratiadis, Andreas; Tsoukalas, Ioannis; Kossieris, Panayiotis; Karavokiros, George; Christofides, Antonis; Siskos, Alexandros; Mamassis, Nikos; Koutsoyiannis, Demetris
2015-04-01
Modelling of large-scale hybrid renewable energy systems (HRES) is a challenging task, for which several open computational issues exist. HRES comprise typical components of hydrosystems (reservoirs, boreholes, conveyance networks, hydropower stations, pumps, water demand nodes, etc.), which are dynamically linked with renewables (e.g., wind turbines, solar parks) and energy demand nodes. In such systems, apart from the well-known shortcomings of water resources modelling (nonlinear dynamics, unknown future inflows, large number of variables and constraints, conflicting criteria, etc.), additional complexities and uncertainties arise due to the introduction of energy components and associated fluxes. A major difficulty is the need for coupling two different temporal scales, given that in hydrosystem modeling, monthly simulation steps are typically adopted, yet for a faithful representation of the energy balance (i.e. energy production vs. demand) a much finer resolution (e.g. hourly) is required. Another drawback is the increase of control variables, constraints and objectives, due to the simultaneous modelling of the two parallel fluxes (i.e. water and energy) and their interactions. Finally, since the driving hydrometeorological processes of the integrated system are inherently uncertain, it is often essential to use synthetically generated input time series of large length, in order to assess the system performance in terms of reliability and risk, with satisfactory accuracy. To address these issues, we propose an effective and efficient modeling framework, key objectives of which are: (a) the substantial reduction of control variables, through parsimonious yet consistent parameterizations; (b) the substantial decrease of computational burden of simulation, by linearizing the combined water and energy allocation problem of each individual time step, and solve each local sub-problem through very fast linear network programming algorithms, and (c) the substantial
Matthias Kasemann
Overview The main focus during the summer was to handle data coming from the detector and to perform Monte Carlo production. The lessons learned during the CCRC and CSA08 challenges in May were addressed by dedicated PADA campaigns lead by the Integration team. Big improvements were achieved in the stability and reliability of the CMS Tier1 and Tier2 centres by regular and systematic follow-up of faults and errors with the help of the Savannah bug tracking system. In preparation for data taking the roles of a Computing Run Coordinator and regular computing shifts monitoring the services and infrastructure as well as interfacing to the data operations tasks are being defined. The shift plan until the end of 2008 is being put together. User support worked on documentation and organized several training sessions. The ECoM task force delivered the report on “Use Cases for Start-up of pp Data-Taking” with recommendations and a set of tests to be performed for trigger rates much higher than the ...
M. Kasemann
Introduction A large fraction of the effort was focused during the last period into the preparation and monitoring of the February tests of Common VO Computing Readiness Challenge 08. CCRC08 is being run by the WLCG collaboration in two phases, between the centres and all experiments. The February test is dedicated to functionality tests, while the May challenge will consist of running at all centres and with full workflows. For this first period, a number of functionality checks of the computing power, data repositories and archives as well as network links are planned. This will help assess the reliability of the systems under a variety of loads, and identifying possible bottlenecks. Many tests are scheduled together with other VOs, allowing the full scale stress test. The data rates (writing, accessing and transfer¬ring) are being checked under a variety of loads and operating conditions, as well as the reliability and transfer rates of the links between Tier-0 and Tier-1s. In addition, the capa...
P. MacBride
The Computing Software and Analysis Challenge CSA07 has been the main focus of the Computing Project for the past few months. Activities began over the summer with the preparation of the Monte Carlo data sets for the challenge and tests of the new production system at the Tier-0 at CERN. The pre-challenge Monte Carlo production was done in several steps: physics generation, detector simulation, digitization, conversion to RAW format and the samples were run through the High Level Trigger (HLT). The data was then merged into three "Soups": Chowder (ALPGEN), Stew (Filtered Pythia) and Gumbo (Pythia). The challenge officially started when the first Chowder events were reconstructed on the Tier-0 on October 3rd. The data operations teams were very busy during the the challenge period. The MC production teams continued with signal production and processing while the Tier-0 and Tier-1 teams worked on splitting the Soups into Primary Data Sets (PDS), reconstruction and skimming. The storage sys...
Contributions from I. Fisk
2012-01-01
Introduction The start of the 2012 run has been busy for Computing. We have reconstructed, archived, and served a larger sample of new data than in 2011, and we are in the process of producing an even larger new sample of simulations at 8 TeV. The running conditions and system performance are largely what was anticipated in the plan, thanks to the hard work and preparation of many people. Heavy ions Heavy Ions has been actively analysing data and preparing for conferences. Operations Office Figure 6: Transfers from all sites in the last 90 days For ICHEP and the Upgrade efforts, we needed to produce and process record amounts of MC samples while supporting the very successful data-taking. This was a large burden, especially on the team members. Nevertheless the last three months were very successful and the total output was phenomenal, thanks to our dedicated site admins who keep the sites operational and the computing project members who spend countless hours nursing the...
I. Fisk
2013-01-01
Computing operation has been lower as the Run 1 samples are completing and smaller samples for upgrades and preparations are ramping up. Much of the computing activity is focusing on preparations for Run 2 and improvements in data access and flexibility of using resources. Operations Office Data processing was slow in the second half of 2013 with only the legacy re-reconstruction pass of 2011 data being processed at the sites. Figure 1: MC production and processing was more in demand with a peak of over 750 Million GEN-SIM events in a single month. Figure 2: The transfer system worked reliably and efficiently and transferred on average close to 520 TB per week with peaks at close to 1.2 PB. Figure 3: The volume of data moved between CMS sites in the last six months The tape utilisation was a focus for the operation teams with frequent deletion campaigns from deprecated 7 TeV MC GEN-SIM samples to INVALID datasets, which could be cleaned up...
Computational time-resolved and resonant x-ray scattering of strongly correlated materials
Energy Technology Data Exchange (ETDEWEB)
Bansil, Arun [Northeastern Univ., Boston, MA (United States)
2016-11-09
Basic-Energy Sciences of the Department of Energy (BES/DOE) has made large investments in x-ray sources in the U.S. (NSLS-II, LCLS, NGLS, ALS, APS) as powerful enabling tools for opening up unprecedented new opportunities for exploring properties of matter at various length and time scales. The coming online of the pulsed photon source, literally allows us to see and follow the dynamics of processes in materials at their natural timescales. There is an urgent need therefore to develop theoretical methodologies and computational models for understanding how x-rays interact with matter and the related spectroscopies of materials. The present project addressed aspects of this grand challenge of x-ray science. In particular, our Collaborative Research Team (CRT) focused on developing viable computational schemes for modeling x-ray scattering and photoemission spectra of strongly correlated materials in the time-domain. The vast arsenal of formal/numerical techniques and approaches encompassed by the members of our CRT were brought to bear through appropriate generalizations and extensions to model the pumped state and the dynamics of this non-equilibrium state, and how it can be probed via x-ray absorption (XAS), emission (XES), resonant and non-resonant x-ray scattering, and photoemission processes. We explored the conceptual connections between the time-domain problems and other second-order spectroscopies, such as resonant inelastic x-ray scattering (RIXS) because RIXS may be effectively thought of as a pump-probe experiment in which the incoming photon acts as the pump, and the fluorescent decay is the probe. Alternatively, when the core-valence interactions are strong, one can view K-edge RIXS for example, as the dynamic response of the material to the transient presence of a strong core-hole potential. Unlike an actual pump-probe experiment, here there is no mechanism for adjusting the time-delay between the pump and the probe. However, the core hole
Takizawa, Kenji; Kostov, Nikolay; Puntel, Anthony; Henicke, Bradley; Tezduyar, Tayfun E.
2012-12-01
We present a detailed computational analysis of bio-inspired flapping-wing aerodynamics of a micro aerial vehicle (MAV). The computational techniques used include the Deforming-Spatial-Domain/Stabilized Space-Time (DSD/SST) formulation, which serves as the core computational technique. The DSD/SST formulation is a moving-mesh technique, and in the computations reported here we use the space-time version of the residual-based variational multiscale (VMS) method, which is called "DSD/ SST-VMST." The motion and deformation of the wings are based on data extracted from the high-speed, multi-camera video recordings of a locust in a wind tunnel. A set of special space-time techniques are also used in the computations in conjunction with the DSD/SST method. The special techniques are based on using, in the space-time flow computations, NURBS basis functions for the temporal representation of the motion and deformation of the wings and for the mesh moving and remeshing. The computational analysis starts with the computation of the base case, and includes computations with increased temporal and spatial resolutions compared to the base case. In increasing the temporal resolution, we separately test increasing the temporal order, the number of temporal subdivisions, and the frequency of remeshing. In terms of the spatial resolution, we separately test increasing the wing-mesh refinement in the normal and tangential directions and changing the way node connectivities are handled at the wingtips. The computational analysis also includes using different combinations of wing configurations for the MAV and investigating the beneficial and disruptive interactions between the wings and the role of wing camber and twist.
Spike-timing computation properties of a feed-forward neural network model
Directory of Open Access Journals (Sweden)
Drew Benjamin Sinha
2014-01-01
Full Text Available Brain function is characterized by dynamical interactions among networks of neurons. These interactions are mediated by network topology at many scales ranging from microcircuits to brain areas. Understanding how networks operate can be aided by understanding how the transformation of inputs depends upon network connectivity patterns, e.g. serial and parallel pathways. To tractably determine how single synapses or groups of synapses in such pathways shape transformations, we modeled feed-forward networks of 7-22 neurons in which synaptic strength changed according to a spike-timing dependent plasticity rule. We investigated how activity varied when dynamics were perturbed by an activity-dependent electrical stimulation protocol (spike-triggered stimulation; STS in networks of different topologies and background input correlations. STS can successfully reorganize functional brain networks in vivo, but with a variability in effectiveness that may derive partially from the underlying network topology. In a simulated network with a single disynaptic pathway driven by uncorrelated background activity, structured spike-timing relationships between polysynaptically connected neurons were not observed. When background activity was correlated or parallel disynaptic pathways were added, however, robust polysynaptic spike timing relationships were observed, and application of STS yielded predictable changes in synaptic strengths and spike-timing relationships. These observations suggest that precise input-related or topologically induced temporal relationships in network activity are necessary for polysynaptic signal propagation. Such constraints for polysynaptic computation suggest potential roles for higher-order topological structure in network organization, such as maintaining polysynaptic correlation in the face of relatively weak synapses.
Computation and analysis of time-dependent sensitivities in Generalized Mass Action systems.
Schwacke, John H; Voit, Eberhard O
2005-09-07
Understanding biochemical system dynamics is becoming increasingly important for insights into the functioning of organisms and for biotechnological manipulations, and additional techniques and methods are needed to facilitate investigations of dynamical properties of systems. Extensions to the method of Ingalls and Sauro, addressing time-dependent sensitivity analysis, provide a new tool for executing such investigations. We present here the results of sample analyses using time-dependent sensitivities for three model systems taken from the literature, namely an anaerobic fermentation pathway in yeast, a negative feedback oscillator modeling cell-cycle phenomena, and the Mitogen Activated Protein (MAP) kinase cascade. The power of time-dependent sensitivities is particularly evident in the case of the MAPK cascade. In this example it is possible to identify the emergence of a concentration of MAPKK that provides the best response with respect to rapid and efficient activation of the cascade, while over- and under-expression of MAPKK relative to this concentration have qualitatively different effects on the transient response of the cascade. Also of interest is the quite general observation that phase-plane representations of sensitivities in oscillating systems provide insights into the manner with which perturbations in the envelope of the oscillation result from small changes in initial concentrations of components of the oscillator. In addition to these applied analyses, we present an algorithm for the efficient computation of time-dependent sensitivities for Generalized Mass Action (GMA) systems, the most general of the canonical system representations of Biochemical Systems Theory (BST). The algorithm is shown to be comparable to, or better than, other methods of solution, as exemplified with three biochemical systems taken from the literature.
Time-Domain Techniques for Computation and Reconstruction of One-Dimensional Profiles
Directory of Open Access Journals (Sweden)
M. Rahman
2005-01-01
Full Text Available This paper presents a time-domain technique to compute the electromagnetic fields and to reconstruct the permittivity profile within a one-dimensional medium of finite length. The medium is characterized by a permittivity as well as conductivity profile which vary only with depth. The discussed scattering problem is thus one-dimensional. The modeling tool is divided into two different schemes which are named as the forward solver and the inverse solver. The task of the forward solver is to compute the internal fields of the specimen which is performed by Green’s function approach. When a known electromagnetic wave is incident normally on the media, the resulting electromagnetic field within the media can be calculated by constructing a Green’s operator. This operator maps the incident field on either side of the medium to the field at an arbitrary observation point. It is nothing but a matrix of integral operators with kernels satisfying known partial differential equations. The reflection and transmission behavior of the medium is also determined from the boundary values of the Green's operator. The inverse solver is responsible for solving an inverse scattering problem by reconstructing the permittivity profile of the medium. Though it is possible to use several algorithms to solve this problem, the invariant embedding method, also known as the layer-stripping method, has been implemented here due to the advantage that it requires a finite time trace of reflection data. Here only one round trip of reflection data is used, where one round trip is defined by the time required by the pulse to propagate through the medium and back again. The inversion process begins by retrieving the reflection kernel from the reflected wave data by simply using a deconvolution technique. The rest of the task can easily be performed by applying a numerical approach to determine different profile parameters. Both the solvers have been found to have the
Computing the total atmospheric refraction for real-time optical imaging sensor simulation
Olson, Richard F.
2015-05-01
Fast and accurate computation of light path deviation due to atmospheric refraction is an important requirement for real-time simulation of optical imaging sensor systems. A large body of existing literature covers various methods for application of Snell's Law to the light path ray tracing problem. This paper provides a discussion of the adaptation to real time simulation of atmospheric refraction ray tracing techniques used in mid-1980's LOWTRAN releases. The refraction ray trace algorithm published in a LOWTRAN-6 technical report by Kneizys (et. al.) has been coded in MATLAB for development, and in C-language for simulation use. To this published algorithm we have added tuning parameters for variable path segment lengths, and extensions for Earth grazing and exoatmospheric "near Earth" ray paths. Model atmosphere properties used to exercise the refraction algorithm were obtained from tables published in another LOWTRAN-6 related report. The LOWTRAN-6 based refraction model is applicable to atmospheric propagation at wavelengths in the IR and visible bands of the electromagnetic spectrum. It has been used during the past two years by engineers at the U.S. Army Aviation and Missile Research, Development and Engineering Center (AMRDEC) in support of several advanced imaging sensor simulations. Recently, a faster (but sufficiently accurate) method using Gauss-Chebyshev Quadrature integration for evaluating the refraction integral was adopted.
Reservoir computing and extreme learning machines for non-linear time-series data analysis.
Butcher, J B; Verstraeten, D; Schrauwen, B; Day, C R; Haycock, P W
2013-02-01
Random projection architectures such as Echo state networks (ESNs) and Extreme Learning Machines (ELMs) use a network containing a randomly connected hidden layer and train only the output weights, overcoming the problems associated with the complex and computationally demanding training algorithms traditionally used to train neural networks, particularly recurrent neural networks. In this study an ESN is shown to contain an antagonistic trade-off between the amount of non-linear mapping and short-term memory it can exhibit when applied to time-series data which are highly non-linear. To overcome this trade-off a new architecture, Reservoir with Random Static Projections (R(2)SP) is investigated, that is shown to offer a significant improvement in performance. A similar approach using an ELM whose input is presented through a time delay (TD-ELM) is shown to further enhance performance where it significantly outperformed the ESN and R(2)SP as well other architectures when applied to a novel task which allows the short-term memory and non-linearity to be varied. The hard-limiting memory of the TD-ELM appears to be best suited for the data investigated in this study, although ESN-based approaches may offer improved performance when processing data which require a longer fading memory.
Fang, You-Qiang; Wu, Jie-Ying; Li, Teng-Cheng; Zheng, Hao-Feng; Liang, Guan-Can; Chen, Yan-Xiong; Hong, Xiao-Bin; Cai, Wei-Zhong; Zang, Zhi-Jun; Di, Jin-Ming
2017-06-01
This study aimed to assess the role of pre-designed route on computer tomography urography (CTU) in the ultrasound-guided percutaneous nephrolithotomy (PCNL) for renal calculus.From August 2013 to May 2016, a total of 100 patients diagnosed with complex renal calculus in our hospital were randomly divided into CTU group and control group (without CTU assistance). CTU was used to design a rational route for puncturing in CTU group. Ultrasound was used in both groups to establish a working trace in the operation areas. Patients' perioperative parameters and postoperative complications were recorded.All operations were successfully performed, without transferring to open surgery. Time of channel establishment in CTU group (6.5 ± 4.3 minutes) was shorter than the control group (10.0 ± 6.7 minutes) (P = .002). In addition, there was shorter operation time, lower rates of blood transfusion, secondary operation, and less establishing channels. The incidence of postoperative complications including residual stones, sepsis, severe hemorrhage, and perirenal hematoma was lower in CTU group than in control group.Pre-designing puncture route on CTU images would improve the puncturing accuracy, lessen establishing channels as well as improve the security in the ultrasound-guided PCNL for complex renal calculus, but at the cost of increased radiation exposure.
Real-Time Evaluation of Breast Self-Examination Using Computer Vision
Directory of Open Access Journals (Sweden)
Eman Mohammadi
2014-01-01
Full Text Available Breast cancer is the most common cancer among women worldwide and breast self-examination (BSE is considered as the most cost-effective approach for early breast cancer detection. The general objective of this paper is to design and develop a computer vision algorithm to evaluate the BSE performance in real-time. The first stage of the algorithm presents a method for detecting and tracking the nipples in frames while a woman performs BSE; the second stage presents a method for localizing the breast region and blocks of pixels related to palpation of the breast, and the third stage focuses on detecting the palpated blocks in the breast region. The palpated blocks are highlighted at the time of BSE performance. In a correct BSE performance, all blocks must be palpated, checked, and highlighted, respectively. If any abnormality, such as masses, is detected, then this must be reported to a doctor to confirm the presence of this abnormality and proceed to perform other confirmatory tests. The experimental results have shown that the BSE evaluation algorithm presented in this paper provides robust performance.
Time-of-flight camera technique for augmented reality in computer-assisted interventions
Mersmann, Sven; Müller, Michael; Seitel, Alexander; Arnegger, Florian; Tetzlaff, Ralf; Dinkel, Julien; Baumhauer, Matthias; Schmied, Bruno; Meinzer, Hans-Peter; Maier-Hein, Lena
2011-03-01
Augmented reality (AR) for enhancement of intra-operative images is gaining increasing interest in the field of navigated medical interventions. In this context, various imaging modalities such as ultrasound (US), C-Arm computed tomography (CT) and endoscopic images have been applied to acquire intra-operative information about the patient's anatomy. The aim of this paper was to evaluate the potential of the novel Time-of-Flight (ToF) camera technique as means for markerless intra-operative registration. For this purpose, ToF range data and corresponding CT images were acquired from a set of explanted non-transplantable human and porcine organs equipped with a set of marker that served as targets. Based on a rigid matching of the surfaces generated from the ToF images with the organ surfaces generated from the CT data, the targets extracted from the planning images were superimposed on the 2D ToF intensity images, and the target visualization error (TVE) was computed as quality measure. Color video data of the same organs were further used to assess the TVE of a previously proposed marker-based registration method. The ToF-based registration showed promising accuracy yielding a mean TVE of 2.5+/-1.1 mm compared to 0.7+/-0.4 mm with the marker-based approach. Furthermore, the target registration error (TRE) was assessed to determine the anisotropy in the localization error of ToF image data. The TRE was 8.9+/- 4.7 mm on average indicating a high localization error in the viewing direction of the camera. Nevertheless, the young ToF technique may become a valuable means for intra-operative surface acquisition. Future work should focus on the calibration of systematic distance errors.
Finite element techniques in computational time series analysis of turbulent flows
Horenko, I.
2009-04-01
In recent years there has been considerable increase of interest in the mathematical modeling and analysis of complex systems that undergo transitions between several phases or regimes. Such systems can be found, e.g., in weather forecast (transitions between weather conditions), climate research (ice and warm ages), computational drug design (conformational transitions) and in econometrics (e.g., transitions between different phases of the market). In all cases, the accumulation of sufficiently detailed time series has led to the formation of huge databases, containing enormous but still undiscovered treasures of information. However, the extraction of essential dynamics and identification of the phases is usually hindered by the multidimensional nature of the signal, i.e., the information is "hidden" in the time series. The standard filtering approaches (like f.~e. wavelets-based spectral methods) have in general unfeasible numerical complexity in high-dimensions, other standard methods (like f.~e. Kalman-filter, MVAR, ARCH/GARCH etc.) impose some strong assumptions about the type of the underlying dynamics. Approach based on optimization of the specially constructed regularized functional (describing the quality of data description in terms of the certain amount of specified models) will be introduced. Based on this approach, several new adaptive mathematical methods for simultaneous EOF/SSA-like data-based dimension reduction and identification of hidden phases in high-dimensional time series will be presented. The methods exploit the topological structure of the analysed data an do not impose severe assumptions on the underlying dynamics. Special emphasis will be done on the mathematical assumptions and numerical cost of the constructed methods. The application of the presented methods will be first demonstrated on a toy example and the results will be compared with the ones obtained by standard approaches. The importance of accounting for the mathematical
Optimizing computed tomography simulation wait times in a busy radiation medicine program.
Roussos, Jerry; Zahedi, Payam; Spence, Tara; Swanson, Lue-Ann; Li-Cheung, Fionna; Cops, Fred; Darcy, Patrick; Chhin, Veng; Moyo, Elen; Warde, Padraig; Foxcroft, Sophie; Liu, Fei-Fei
An audit was conducted of patient schedules for computed tomography simulation (CT-Sim) scans within the Radiation Medicine Program at the Princess Margaret Cancer Centre to investigate opportunities for improved efficiencies, enhancing process, reducing rescanning rates, and decreasing wait times. A 3-phased approach was undertaken to evaluate the current practice in the CT-Sim facility with a view toward implementing improvements. The first phase involved a review and assessment of the validity of current guidelines and protocols associated with 16 different disease sites. The second phase incorporated the use of a patient record and verification program MOSAIQ to capture the duration of each appointment. The last phase allocated additional time for patient-centered care and staff engagement. The audit revealed that efficiency could be achieved through staff training, updating protocols, and improving process coordination. With the exception of sarcoma, pediatric, and palliative patients who require unique management approaches, the duration for each CT-Sim appointment was successfully shortened for all disease sites by 22% to 33%, corresponding to a reduction of 10 to 15 minutes per appointment. Rescanning rates for patients requiring self-administered preparations before CT-Sim procedures were also significantly reduced by enhancing processes to increase patient compliance. Implementation of procedural changes resulted in an overall net gain of 3060 minutes, equivalent to 102 additional 30-minute CT-Sim appointment slots available for each month. This retrospective evaluation, review, and optimization of CT-Sim guidelines and practices identified opportunities to shorten appointment timeslots, and reduce rescanning rates for CT-Sim procedures, thereby significantly shortening wait times and improving access to service for our patients. Crown Copyright © 2016. Published by Elsevier Inc. All rights reserved.
Shallow sediment transport flow computation using time-varying sediment adaptation length
Institute of Scientific and Technical Information of China (English)
Jaan Hui PU; Khalid HUSSAIN; Song-dong SHAO; Yue-fei HUANG
2014-01-01
Based on the common approach, the adaptation length in sediment transport is normally estimated astemporally independent. However, this approach might not be theoretically justified as the process of reaching the sediment transport equilibrium stage is affected by the flow conditions in time, especially for fast moving flows, such as scour-hole developing flows. In this study, the two-dimensional (2D) shallow water formulation together with a sediment continuity-concentration (SCC) model were applied to flow with mobile sediment boundary. A time-varying approach was proposed to determine the sediment transport adaptation length to simulate the sediment erosion-deposition rate. The proposed computational model was based on the Finite Volume (FV) method. The Monotone Upwind Scheme of Conservative Laws (MUSCL)-Hancock scheme was used with the Harten Lax van Leer-contact (HLLC) approximate Riemann solver to discretize the FV model. In the flow applications of this paper, a highly discontinuous dam-break, fast sediment transport flow was used to calibrate the proposed time-varying sediment adaptation length model. Then the calibrated model was further applied to two separate experimental sediment transport flow applications documented in the literature, i.e. a highly concentrated sediment transport flow in a wide alluvial channel and a sediment aggradation flow. Good agreement with the experimental data were obtained with the proposed model simulations. The tests prove that the proposed model, which was calibrated by the discontinuous dam-break bed scouring flow, also performed well to represent rapid bed change and steady sediment mobility conditions.
Computational Fluid Dynamics Study on the Effects of RATO Timing on the Scale Model Acoustic Test
Nielsen, Tanner; Williams, B.; West, Jeff
2015-01-01
The Scale Model Acoustic Test (SMAT) is a 5% scale test of the Space Launch System (SLS), which is currently being designed at Marshall Space Flight Center (MSFC). The purpose of this test is to characterize and understand a variety of acoustic phenomena that occur during the early portions of lift off, one being the overpressure environment that develops shortly after booster ignition. The SLS lift off configuration consists of four RS-25 liquid thrusters on the core stage, with two solid boosters connected to each side. Past experience with scale model testing at MSFC (in ER42), has shown that there is a delay in the ignition of the Rocket Assisted Take Off (RATO) motor, which is used as the 5% scale analog of the solid boosters, after the signal to ignite is given. This delay can range from 0 to 16.5ms. While this small of a delay maybe insignificant in the case of the full scale SLS, it can significantly alter the data obtained during the SMAT due to the much smaller geometry. The speed of sound of the air and combustion gas constituents is not scaled, and therefore the SMAT pressure waves propagate at approximately the same speed as occurs during full scale. However, the SMAT geometry is much smaller allowing the pressure waves to move down the exhaust duct, through the trench, and impact the vehicle model much faster than occurs at full scale. To better understand the effect of the RATO timing simultaneity on the SMAT IOP test data, a computational fluid dynamics (CFD) analysis was performed using the Loci/CHEM CFD software program. Five different timing offsets, based on RATO ignition delay statistics, were simulated. A variety of results and comparisons will be given, assessing the overall effect of RATO timing simultaneity on the SMAT overpressure environment.
A tool for computing time-dependent permeability reduction of fractured volcanic conduit margins.
Farquharson, Jamie; Wadsworth, Fabian; Heap, Michael; Baud, Patrick
2016-04-01
Laterally-oriented fractures within volcanic conduit margins are thought to play an important role in tempering eruption explosivity by allowing magmatic volatiles to outgas. The permeability of a fractured conduit margin - the equivalent permeability - can be modelled as the sum of permeability contributions of the edifice host rock and the fracture(s) within it. We present here a flexible MATLAB® tool which computes the time-dependent equivalent permeability of a volcanic conduit margin containing ash-filled fractures. The tool is designed so that the end-user can define a wide range of input parameters to yield equivalent permeability estimates for their application. The time-dependence of the equivalent permeability is incorporated by considering permeability decrease as a function of porosity loss in the ash-filled fractures due to viscous sintering (after Russell and Quane, 2005), which is in turn dependent on the depth and temperature of each fracture and the crystal-content of the magma (all user-defined variables). The initial viscosity of the granular material filling the fracture is dependent on the water content (Hess and Dingwell, 1996), which is computed assuming equilibrium depth-dependent water content (Liu et al., 2005). Crystallinity is subsequently accounted for by employing the particle-suspension rheological model of Mueller et al. (2010). The user then defines the number of fractures, their widths, and their depths, and the lengthscale of interest (e.g. the length of the conduit). Using these data, the combined influence of transient fractures on the equivalent permeability of the conduit margin is then calculated by adapting a parallel-plate flow model (developed by Baud et al., 2012 for porous sandstones), for host rock permeabilities from 10-11 to 10-22 m2. The calculated values of porosity and equivalent permeability with time for each host rock permeability is then output in text and worksheet file formats. We introduce two dimensionless
Edixhoven, Bas; de Jong, Robin; Bosman, Johan
2011-01-01
Modular forms are tremendously important in various areas of mathematics, from number theory and algebraic geometry to combinatorics and lattices. Their Fourier coefficients, with Ramanujan's tau-function as a typical example, have deep arithmetic significance. Prior to this book, the fastest known algorithms for computing these Fourier coefficients took exponential time, except in some special cases. The case of elliptic curves (Schoof's algorithm) was at the birth of elliptic curve cryptography around 1985. This book gives an algorithm for computing coefficients of modular forms of level on
Evans, Daniel J; Manwaring, Mark L; Soule, Terence
2008-01-01
The technique of inverse computational feedback optimization imaging allows for the imaging of varying tissue without the continuous need of a complex imaging systems such as an MRI or CT. Our method trades complex imaging equipment for computing power. The objective is to use a baseline scan from an imaging system along with finite element method computational software to calculate the physically measurable parameters (such as voltage or temperature). As the physically measurable parameters change the computational model is iteratively run until it matches the measured values. Optimization routines are implemented to accelerate the process of finding the new values. Presented is a computational model demonstrating how the inverse imaging technique would work with a simple homogeneous sample with a circular structure. It demonstrates the ability to locate an object with only a few point measurements. The presented computational model uses swarm optimization techniques to help find the object location from the measured data (which in this case is voltage).
Proton radiography and proton computed tomography based on time-resolved dose measurements
Testa, Mauro; Verburg, Joost M.; Rose, Mark; Min, Chul Hee; Tang, Shikui; Hassane Bentefour, El; Paganetti, Harald; Lu, Hsiao-Ming
2013-11-01
We present a proof of principle study of proton radiography and proton computed tomography (pCT) based on time-resolved dose measurements. We used a prototype, two-dimensional, diode-array detector capable of fast dose rate measurements, to acquire proton radiographic images expressed directly in water equivalent path length (WEPL). The technique is based on the time dependence of the dose distribution delivered by a proton beam traversing a range modulator wheel in passive scattering proton therapy systems. The dose rate produced in the medium by such a system is periodic and has a unique pattern in time at each point along the beam path and thus encodes the WEPL. By measuring the time dose pattern at the point of interest, the WEPL to this point can be decoded. If one measures the time-dose patterns at points on a plane behind the patient for a beam with sufficient energy to penetrate the patient, the obtained 2D distribution of the WEPL forms an image. The technique requires only a 2D dosimeter array and it uses only the clinical beam for a fraction of second with negligible dose to patient. We first evaluated the accuracy of the technique in determining the WEPL for static phantoms aiming at beam range verification of the brain fields of medulloblastoma patients. Accurate beam ranges for these fields can significantly reduce the dose to the cranial skin of the patient and thus the risk of permanent alopecia. Second, we investigated the potential features of the technique for real-time imaging of a moving phantom. Real-time tumor tracking by proton radiography could provide more accurate validations of tumor motion models due to the more sensitive dependence of proton beam on tissue density compared to x-rays. Our radiographic technique is rapid (˜100 ms) and simultaneous over the whole field, it can image mobile tumors without the problem of interplay effect inherently challenging for methods based on pencil beams. Third, we present the reconstructed p
Time- and Computation-Efficient Calibration of MEMS 3D Accelerometers and Gyroscopes
Directory of Open Access Journals (Sweden)
Sara Stančin
2014-08-01
Full Text Available We propose calibration methods for microelectromechanical system (MEMS 3D accelerometers and gyroscopes that are efficient in terms of time and computational complexity. The calibration process for both sensors is simple, does not require additional expensive equipment, and can be performed in the field before or between motion measurements. The methods rely on a small number of defined calibration measurements that are used to obtain the values of 12 calibration parameters. This process enables the static compensation of sensor inaccuracies. The values detected by the 3D sensor are interpreted using a generalized 3D sensor model. The model assumes that the values detected by the sensor are equal to the projections of the measured value on the sensor sensitivity axes. Although this finding is trivial for 3D accelerometers, its validity for 3D gyroscopes is not immediately apparent; thus, this paper elaborates on this latter topic. For an example sensor device, calibration parameters were established using calibration measurements of approximately 1.5 min in duration for the 3D accelerometer and 2.5 min in duration for the 3D gyroscope. Correction of each detected 3D value using the established calibration parameters in further measurements requires only nine addition and nine multiplication operations.
Time- and computation-efficient calibration of MEMS 3D accelerometers and gyroscopes.
Stančin, Sara; Tomažič, Sašo
2014-08-13
We propose calibration methods for microelectromechanical system (MEMS) 3D accelerometers and gyroscopes that are efficient in terms of time and computational complexity. The calibration process for both sensors is simple, does not require additional expensive equipment, and can be performed in the field before or between motion measurements. The methods rely on a small number of defined calibration measurements that are used to obtain the values of 12 calibration parameters. This process enables the static compensation of sensor inaccuracies. The values detected by the 3D sensor are interpreted using a generalized 3D sensor model. The model assumes that the values detected by the sensor are equal to the projections of the measured value on the sensor sensitivity axes. Although this finding is trivial for 3D accelerometers, its validity for 3D gyroscopes is not immediately apparent; thus, this paper elaborates on this latter topic. For an example sensor device, calibration parameters were established using calibration measurements of approximately 1.5 min in duration for the 3D accelerometer and 2.5 min in duration for the 3D gyroscope. Correction of each detected 3D value using the established calibration parameters in further measurements requires only nine addition and nine multiplication operations.
Yamaguchi, Yusaku; Kojima, Takeshi; Yoshinaga, Tetsuya
2016-03-01
In clinical X-ray computed tomography (CT), filtered back-projection as a transform method and iterative reconstruction such as the maximum-likelihood expectation-maximization (ML-EM) method are known methods to reconstruct tomographic images. As the other reconstruction method, we have presented a continuous-time image reconstruction (CIR) system described by a nonlinear dynamical system, based on the idea of continuous methods for solving tomographic inverse problems. Recently, we have also proposed a multiplicative CIR system described by differential equations based on the minimization of a weighted Kullback-Leibler divergence. We prove theoretically that the divergence measure decreases along the solution to the CIR system, for consistent inverse problems. In consideration of the noisy nature of projections in clinical CT, the inverse problem belongs to the category of ill-posed problems. The performance of a noise-reduction scheme for a new (previously developed) CIR system was investigated by means of numerical experiments using a circular phantom image. Compared to the conventional CIR and the ML-EM methods, the proposed CIR method has an advantage on noisy projection with lower signal-to-noise ratios in terms of the divergence measure on the actual image under the same common measure observed via the projection data. The results lead to the conclusion that the multiplicative CIR method is more effective and robust for noise reduction in CT compared to the ML-EM as well as conventional CIR methods.
Real-time 3D computed tomographic reconstruction using commodity graphics hardware
Xu, Fang; Mueller, Klaus
2007-07-01
The recent emergence of various types of flat-panel x-ray detectors and C-arm gantries now enables the construction of novel imaging platforms for a wide variety of clinical applications. Many of these applications require interactive 3D image generation, which cannot be satisfied with inexpensive PC-based solutions using the CPU. We present a solution based on commodity graphics hardware (GPUs) to provide these capabilities. While GPUs have been employed for CT reconstruction before, our approach provides significant speedups by exploiting the various built-in hardwired graphics pipeline components for the most expensive CT reconstruction task, backprojection. We show that the timings so achieved are superior to those obtained when using the GPU merely as a multi-processor, without a drop in reconstruction quality. In addition, we also show how the data flow across the graphics pipeline can be optimized, by balancing the load among the pipeline components. The result is a novel streaming CT framework that conceptualizes the reconstruction process as a steady flow of data across a computing pipeline, updating the reconstruction result immediately after the projections have been acquired. Using a single PC equipped with a single high-end commodity graphics board (the Nvidia 8800 GTX), our system is able to process clinically-sized projection data at speeds meeting and exceeding the typical flat-panel detector data production rates, enabling throughput rates of 40-50 projections s-1 for the reconstruction of 5123 volumes.
Real-time 3D computed tomographic reconstruction using commodity graphics hardware
Energy Technology Data Exchange (ETDEWEB)
Xu Fang; Mueller, Klaus [Center for Visual Computing, Computer Science Department, Stony Brook University, Stony Brook, NY 11794-4400 (United States)
2007-07-21
The recent emergence of various types of flat-panel x-ray detectors and C-arm gantries now enables the construction of novel imaging platforms for a wide variety of clinical applications. Many of these applications require interactive 3D image generation, which cannot be satisfied with inexpensive PC-based solutions using the CPU. We present a solution based on commodity graphics hardware (GPUs) to provide these capabilities. While GPUs have been employed for CT reconstruction before, our approach provides significant speedups by exploiting the various built-in hardwired graphics pipeline components for the most expensive CT reconstruction task, backprojection. We show that the timings so achieved are superior to those obtained when using the GPU merely as a multi-processor, without a drop in reconstruction quality. In addition, we also show how the data flow across the graphics pipeline can be optimized, by balancing the load among the pipeline components. The result is a novel streaming CT framework that conceptualizes the reconstruction process as a steady flow of data across a computing pipeline, updating the reconstruction result immediately after the projections have been acquired. Using a single PC equipped with a single high-end commodity graphics board (the Nvidia 8800 GTX), our system is able to process clinically-sized projection data at speeds meeting and exceeding the typical flat-panel detector data production rates, enabling throughput rates of 40-50 projections s{sup -1} for the reconstruction of 512{sup 3} volumes.
Near real-time, on-the-move multisensor integration and computing framework
Burnette, Chris; Schneider, Matt; Agarwal, Sanjeev; Deterline, Diane; Geyer, Chris; Phan, Chung D.; Lydic, Richard M.; Green, Kevin; Swett, Bruce
2015-05-01
Implanted mines and improvised devices are a persistent threat to Warfighters. Current Army countermine missions for route clearance need on-the-move standoff detection to improve the rate of advance. Vehicle-based forward looking sensors such as electro-optical and infrared (EO/IR) devices can be used to identify potential threats in near real-time (NRT) at safe standoff distance to support route clearance missions. The MOVERS (Micro-Cloud for Operational, Vehicle-Based EO-IR Reconnaissance System) is a vehicle-based multi-sensor integration and exploitation system that ingests and processes video and imagery data captured from forward-looking EO/IR and thermal sensors, and also generates target/feature alerts, using the Video Processing and Exploitation Framework (VPEF) "plug and play" video processing toolset. The MOVERS Framework provides an extensible, flexible, and scalable computing and multi-sensor integration GOTS framework that enables the capability to add more vehicles, sensors, processors or displays, and a service architecture that provides low-latency raw video and metadata streams as well as a command and control interface. Functionality in the framework is exposed through the MOVERS SDK which decouples the implementation of the service and client from the specific communication protocols.
Covey, Jason
2008-01-01
We provide deterministic, polynomial-time computable voting rules that approximate Dodgson's and (the ``minimization version'' of) Young's scoring rules to within a logarithmic factor. Our approximation of Dodgson's rule is tight up to a constant factor, as Dodgson's rule is $\\NP$-hard to approximate to within some logarithmic factor. The ``maximization version'' of Young's rule is known to be $\\NP$-hard to approximate by any constant factor. Both approximations are simple, and natural as rules in their own right: Given a candidate we wish to score, we can regard either its Dodgson or Young score as the edit distance between a given set of voter preferences and one in which the candidate to be scored is the Condorcet winner. (The difference between the two scoring rules is the type of edits allowed.) We regard the marginal cost of a sequence of edits to be the number of edits divided by the number of reductions (in the candidate's deficit against any of its opponents in the pairwise race against that opponent...
Interactive and Audience Adaptive Digital Signage Using Real-Time Computer Vision
Directory of Open Access Journals (Sweden)
Robert Ravnik
2013-02-01
Full Text Available In this paper we present the development of an interactive, content‐aware and cost‐effective digital signage system. Using a monocular camera installed within the frame of a digital signage display, we employ real‐time computer vision algorithms to extract temporal, spatial and demographic features of the observers, which are further used for observer‐specific broadcasting of digital signage content. The number of observers is obtained by the Viola and Jones face detection algorithm, whilst facial images are registered using multi‐view Active Appearance Models. The distance of the observers from the system is estimated from the interpupillary distance of registered faces. Demographic features, including gender and age group, are determined using SVM classifiers to achieve individual observer‐specific selection and adaption of the digital signage broadcasting content. The developed system was evaluated at the laboratory study level and in a field study performed for audience measurement research. Comparison of our monocular localization module with the Kinect stereo‐system reveals a comparable level of accuracy. The facial characterization module is evaluated on the FERET database with 95% accuracy for gender classification and 92% for age group. Finally, the field study demonstrates the applicability of the developed system in real‐life environments.
Directory of Open Access Journals (Sweden)
Regnier D.
2017-01-01
Full Text Available Accurate knowledge of fission fragment yields is an essential ingredient of numerous applications ranging from the formation of elements in the r-process to fuel cycle optimization in nuclear energy. The need for a predictive theory applicable where no data is available, together with the variety of potential applications, is an incentive to develop a fully microscopic approach to fission dynamics. One of the most promising theoretical frameworks is the time dependent generator coordinate method (TDGCM applied under the Gaussian overlap approximation (GOA. However, the computational cost of this method makes it difficult to perform calculations with more than two collective degree of freedom. Meanwhile, it is well-known from both semi-phenomenological and fully microscopic approaches that at least four or five dimensions may play a role in the dynamics of fission. To overcome this limitation, we develop the code FELIX aiming to solve the TDGCM+GOA equation for an arbitrary number of collective variables. In this talk, we report the recent progress toward this enriched description of fission dynamics. We will briefly present the numerical methods adopted as well as the status of the latest version of FELIX. Finally, we will discuss fragments yields obtained within this approach for the low energy fission of major actinides.
Regnier, D.; Dubray, N.; Schunck, N.; Verrière, M.
2017-09-01
Accurate knowledge of fission fragment yields is an essential ingredient of numerous applications ranging from the formation of elements in the r-process to fuel cycle optimization in nuclear energy. The need for a predictive theory applicable where no data is available, together with the variety of potential applications, is an incentive to develop a fully microscopic approach to fission dynamics. One of the most promising theoretical frameworks is the time dependent generator coordinate method (TDGCM) applied under the Gaussian overlap approximation (GOA). However, the computational cost of this method makes it difficult to perform calculations with more than two collective degree of freedom. Meanwhile, it is well-known from both semi-phenomenological and fully microscopic approaches that at least four or five dimensions may play a role in the dynamics of fission. To overcome this limitation, we develop the code FELIX aiming to solve the TDGCM+GOA equation for an arbitrary number of collective variables. In this talk, we report the recent progress toward this enriched description of fission dynamics. We will briefly present the numerical methods adopted as well as the status of the latest version of FELIX. Finally, we will discuss fragments yields obtained within this approach for the low energy fission of major actinides.
FALCON or how to compute measures time efficiently on dynamically evolving dense complex networks?
Franke, R; Ivanova, G
2014-02-01
A large number of topics in biology, medicine, neuroscience, psychology and sociology can be generally described via complex networks in order to investigate fundamental questions of structure, connectivity, information exchange and causality. Especially, research on biological networks like functional spatiotemporal brain activations and changes, caused by neuropsychiatric pathologies, is promising. Analyzing those so-called complex networks, the calculation of meaningful measures can be very long-winded depending on their size and structure. Even worse, in many labs only standard desktop computers are accessible to perform those calculations. Numerous investigations on complex networks regard huge but sparsely connected network structures, where most network nodes are connected to only a few others. Currently, there are several libraries available to tackle this kind of networks. A problem arises when not only a few big and sparse networks have to be analyzed, but hundreds or thousands of smaller and conceivably dense networks (e.g. in measuring brain activation over time). Then every minute per network is crucial. For these cases there several possibilities to use standard hardware more efficiently. It is not sufficient to apply just standard algorithms for dense graph characteristics. This article introduces the new library FALCON developed especially for the exploration of dense complex networks. Currently, it offers 12 different measures (like clustering coefficients), each for undirected-unweighted, undirected-weighted and directed-unweighted networks. It uses a multi-core approach in combination with comprehensive code and hardware optimizations. There is an alternative massively parallel GPU implementation for the most time-consuming measures, too. Finally, a comparing benchmark is integrated to support the choice of the most suitable library for a particular network issue. Copyright © 2013 Elsevier Inc. All rights reserved.
Houshmand, Sina; Salavati, Ali; Segtnan, Eivind Antonsen; Grupe, Peter; Høilund-Carlsen, Poul Flemming; Alavi, Abass
2016-01-01
The techniques of dual-time-point imaging (DTPI) and delayed-time-point imaging, which are mostly being used for distinction between inflammatory and malignant diseases, has increased the specificity of fluorodeoxyglucose (FDG)-PET for diagnosis and prognosis of certain diseases. A gradually increasing trend of FDG uptake over time has been shown in malignant cells, and a decreasing or constant trend has been shown in inflammatory/infectious processes. Tumor heterogeneity can be assessed by using early and delayed imaging because differences between primary versus metastatic sites become more detectable compared with single time points. This article discusses the applications of DTPI and delayed-time-point imaging.
Design and Development of a Run-Time Monitor for Multi-Core Architectures in Cloud Computing
Directory of Open Access Journals (Sweden)
Junghoon Lee
2011-03-01
Full Text Available Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.
Design and development of a run-time monitor for multi-core architectures in cloud computing.
Kang, Mikyung; Kang, Dong-In; Crago, Stephen P; Park, Gyung-Leen; Lee, Junghoon
2011-01-01
Cloud computing is a new information technology trend that moves computing and data away from desktops and portable PCs into large data centers. The basic principle of cloud computing is to deliver applications as services over the Internet as well as infrastructure. A cloud is a type of parallel and distributed system consisting of a collection of inter-connected and virtualized computers that are dynamically provisioned and presented as one or more unified computing resources. The large-scale distributed applications on a cloud require adaptive service-based software, which has the capability of monitoring system status changes, analyzing the monitored information, and adapting its service configuration while considering tradeoffs among multiple QoS features simultaneously. In this paper, we design and develop a Run-Time Monitor (RTM) which is a system software to monitor the application behavior at run-time, analyze the collected information, and optimize cloud computing resources for multi-core architectures. RTM monitors application software through library instrumentation as well as underlying hardware through a performance counter optimizing its computing configuration based on the analyzed data.
2016-05-09
Nagaraja, Jaques Reifman*, Alexander Y. Mitrophanov Department of Defense Biotechnology High Performance Computing Software Applications Institute...demonstrated the utility of mathematical models in the study of inflammation in specific disease scenarios and in the identification of crucial...of our simulations reflected a 20-day period after inflammation initiation. We performed all computations in the software suite MATLAB R2012a
A practical O(n log2 n) time algorithm for computing the triplet distance on binary trees
DEFF Research Database (Denmark)
Sand, Andreas; Pedersen, Christian Nørgaard Storm; Mailund, Thomas
2013-01-01
The triplet distance is a distance measure that compares two rooted trees on the same set of leaves by enumerating all sub-sets of three leaves and counting how often the induced topologies of the tree are equal or different. We present an algorithm that computes the triplet distance between two...... rooted binary trees in time O (n log2 n). The algorithm is related to an algorithm for computing the quartet distance between two unrooted binary trees in time O (n log n). While the quartet distance algorithm has a very severe overhead in the asymptotic time complexity that makes it impractical compared...... to O (n2) time algorithms, we show through experiments that the triplet distance algorithm can be implemented to give a competitive wall-time running time....
Fischer, E A J; De Vlas, S J; Richardus, J H; Habbema, J D F
2008-09-01
Microsimulation of infectious diseases requires simulation of many life histories of interacting individuals. In particular, relatively rare infections such as leprosy need to be studied in very large populations. Computation time increases disproportionally with the size of the simulated population. We present a novel method, MUSIDH, an acronym for multiple use of simulated demographic histories, to reduce computation time. Demographic history refers to the processes of birth, death and all other demographic events that should be unrelated to the natural course of an infection, thus non-fatal infections. MUSIDH attaches a fixed number of infection histories to each demographic history, and these infection histories interact as if being the infection history of separate individuals. With two examples, mumps and leprosy, we show that the method can give a factor 50 reduction in computation time at the cost of a small loss in precision. The largest reductions are obtained for rare infections with complex demographic histories.
Directory of Open Access Journals (Sweden)
Beatriz Vaz de Melo Mendes
2005-12-01
Full Text Available It is now widespread the use of Value-at-Risk (VaR as a canonical measure at risk. Most accurate VaR measures make use of some volatility model such as GARCH-type models. However, the pattern of volatility dynamic of a portfolio follows from the (univariate behavior of the risk assets, as well as from the type and strength of the associations among them. Moreover, the dependence structure among the components may change conditionally t past observations. Some papers have attempted to model this characteristic by assuming a multivariate GARCH model, or by considering the conditional correlation coefficient, or by incorporating some possibility for switches in regimes. In this paper we address this problem using time-varying copulas. Our modeling strategy allows for the margins to follow some FIGARCH type model while the copula dependence structure changes over time.
DEFF Research Database (Denmark)
Houshmand, Sina; Salavati, Ali; Antonsen Segtnan, Eivind;
2016-01-01
The techniques of dual-time-point imaging (DTPI) and delayed-time-point imaging, which are mostly being used for distinction between inflammatory and malignant diseases, has increased the specificity of fluorodeoxyglucose (FDG)-PET for diagnosis and prognosis of certain diseases. A gradually...
Hydrologic Response to Climate Change: Missing Precipitation Data Matters for Computed Timing Trends
Daniels, B.
2016-12-01
This work demonstrates the derivation of climate timing statistics and applying them to determine resulting hydroclimate impacts. Long-term daily precipitation observations from 50 California stations were used to compute climate trends of precipitation event Intensity, event Duration and Pause between events. Each precipitation event trend was then applied as input to a PRMS hydrology model which showed hydrology changes to recharge, baseflow, streamflow, etc. An important concern was precipitation uncertainty induced by missing observation values and causing errors in quantification of precipitation trends. Many standard statistical techniques such as ARIMA and simple endogenous or even exogenous imputation were applied but failed to help resolve these uncertainties. What helped resolve these uncertainties was use of multiple imputation techniques. This involved fitting of Weibull probability distributions to multiple imputed values for the three precipitation trends.Permutation resampling techniques using Monte Carlo processing were then applied to the multiple imputation values to derive significance p-values for each trend. Significance at the 95% level for Intensity was found for 11 of the 50 stations, Duration from 16 of the 50, and Pause from 19, of which 12 were 99% significant. The significance weighted trends for California are Intensity -4.61% per decade, Duration +3.49% per decade, and Pause +3.58% per decade. Two California basins with PRMS hydrologic models were studied: Feather River in the northern Sierra Nevada mountains and the central coast Soquel-Aptos. Each local trend was changed without changing the other trends or the total precipitation. Feather River Basin's critical supply to Lake Oroville and the State Water Project benefited from a total streamflow increase of 1.5%. The Soquel-Aptos Basin water supply was impacted by a total groundwater recharge decrease of -7.5% and streamflow decrease of -3.2%.
Mellem, Daniel; Fischer, Frank; Jaspers, Sören; Wenck, Horst; Rübhausen, Michael
2016-01-01
Mitochondria are essential for the energy production of eukaryotic cells. During aging mitochondria run through various processes which change their quality in terms of activity, health and metabolic supply. In recent years, many of these processes such as fission and fusion of mitochondria, mitophagy, mitochondrial biogenesis and energy consumption have been subject of research. Based on numerous experimental insights, it was possible to qualify mitochondrial behaviour in computational simulations. Here, we present a new biophysical model based on the approach of Figge et al. in 2012. We introduce exponential decay and growth laws for each mitochondrial process to derive its time-dependent probability during the aging of cells. All mitochondrial processes of the original model are mathematically and biophysically redefined and additional processes are implemented: Mitochondrial fission and fusion is separated into a metabolic outer-membrane part and a protein-related inner-membrane part, a quality-dependent threshold for mitophagy and mitochondrial biogenesis is introduced and processes for activity-dependent internal oxidative stress as well as mitochondrial repair mechanisms are newly included. Our findings reveal a decrease of mitochondrial quality and a fragmentation of the mitochondrial network during aging. Additionally, the model discloses a quality increasing mechanism due to the interplay of the mitophagy and biogenesis cycle and the fission and fusion cycle of mitochondria. It is revealed that decreased mitochondrial repair can be a quality saving process in aged cells. Furthermore, the model finds strategies to sustain the quality of the mitochondrial network in cells with high production rates of reactive oxygen species due to large energy demands. Hence, the model adds new insights to biophysical mechanisms of mitochondrial aging and provides novel understandings of the interdependency of mitochondrial processes.
Mellem, Daniel; Fischer, Frank; Jaspers, Sören; Wenck, Horst; Rübhausen, Michael
2016-01-01
Mitochondria are essential for the energy production of eukaryotic cells. During aging mitochondria run through various processes which change their quality in terms of activity, health and metabolic supply. In recent years, many of these processes such as fission and fusion of mitochondria, mitophagy, mitochondrial biogenesis and energy consumption have been subject of research. Based on numerous experimental insights, it was possible to qualify mitochondrial behaviour in computational simulations. Here, we present a new biophysical model based on the approach of Figge et al. in 2012. We introduce exponential decay and growth laws for each mitochondrial process to derive its time-dependent probability during the aging of cells. All mitochondrial processes of the original model are mathematically and biophysically redefined and additional processes are implemented: Mitochondrial fission and fusion is separated into a metabolic outer-membrane part and a protein-related inner-membrane part, a quality-dependent threshold for mitophagy and mitochondrial biogenesis is introduced and processes for activity-dependent internal oxidative stress as well as mitochondrial repair mechanisms are newly included. Our findings reveal a decrease of mitochondrial quality and a fragmentation of the mitochondrial network during aging. Additionally, the model discloses a quality increasing mechanism due to the interplay of the mitophagy and biogenesis cycle and the fission and fusion cycle of mitochondria. It is revealed that decreased mitochondrial repair can be a quality saving process in aged cells. Furthermore, the model finds strategies to sustain the quality of the mitochondrial network in cells with high production rates of reactive oxygen species due to large energy demands. Hence, the model adds new insights to biophysical mechanisms of mitochondrial aging and provides novel understandings of the interdependency of mitochondrial processes. PMID:26771181
Directory of Open Access Journals (Sweden)
Daniel Mellem
Full Text Available Mitochondria are essential for the energy production of eukaryotic cells. During aging mitochondria run through various processes which change their quality in terms of activity, health and metabolic supply. In recent years, many of these processes such as fission and fusion of mitochondria, mitophagy, mitochondrial biogenesis and energy consumption have been subject of research. Based on numerous experimental insights, it was possible to qualify mitochondrial behaviour in computational simulations. Here, we present a new biophysical model based on the approach of Figge et al. in 2012. We introduce exponential decay and growth laws for each mitochondrial process to derive its time-dependent probability during the aging of cells. All mitochondrial processes of the original model are mathematically and biophysically redefined and additional processes are implemented: Mitochondrial fission and fusion is separated into a metabolic outer-membrane part and a protein-related inner-membrane part, a quality-dependent threshold for mitophagy and mitochondrial biogenesis is introduced and processes for activity-dependent internal oxidative stress as well as mitochondrial repair mechanisms are newly included. Our findings reveal a decrease of mitochondrial quality and a fragmentation of the mitochondrial network during aging. Additionally, the model discloses a quality increasing mechanism due to the interplay of the mitophagy and biogenesis cycle and the fission and fusion cycle of mitochondria. It is revealed that decreased mitochondrial repair can be a quality saving process in aged cells. Furthermore, the model finds strategies to sustain the quality of the mitochondrial network in cells with high production rates of reactive oxygen species due to large energy demands. Hence, the model adds new insights to biophysical mechanisms of mitochondrial aging and provides novel understandings of the interdependency of mitochondrial processes.
FlowMax: A Computational Tool for Maximum Likelihood Deconvolution of CFSE Time Courses.
Directory of Open Access Journals (Sweden)
Maxim Nikolaievich Shokhirev
Full Text Available The immune response is a concerted dynamic multi-cellular process. Upon infection, the dynamics of lymphocyte populations are an aggregate of molecular processes that determine the activation, division, and longevity of individual cells. The timing of these single-cell processes is remarkably widely distributed with some cells undergoing their third division while others undergo their first. High cell-to-cell variability and technical noise pose challenges for interpreting popular dye-dilution experiments objectively. It remains an unresolved challenge to avoid under- or over-interpretation of such data when phenotyping gene-targeted mouse models or patient samples. Here we develop and characterize a computational methodology to parameterize a cell population model in the context of noisy dye-dilution data. To enable objective interpretation of model fits, our method estimates fit sensitivity and redundancy by stochastically sampling the solution landscape, calculating parameter sensitivities, and clustering to determine the maximum-likelihood solution ranges. Our methodology accounts for both technical and biological variability by using a cell fluorescence model as an adaptor during population model fitting, resulting in improved fit accuracy without the need for ad hoc objective functions. We have incorporated our methodology into an integrated phenotyping tool, FlowMax, and used it to analyze B cells from two NFκB knockout mice with distinct phenotypes; we not only confirm previously published findings at a fraction of the expended effort and cost, but reveal a novel phenotype of nfkb1/p105/50 in limiting the proliferative capacity of B cells following B-cell receptor stimulation. In addition to complementing experimental work, FlowMax is suitable for high throughput analysis of dye dilution studies within clinical and pharmacological screens with objective and quantitative conclusions.
Implementation of a Real-Time, Distributed Operating System for a Multiple Computer System.
1982-06-01
Segmentation Registers ------ 31 4. The iSBC8O86/12A( Single Board Computer ) ------ 35 C. PHYSICAL ADDRESS GENERATION ---------------------- 36 1. General...iSBC86/12A single board computer (SBC). It is based on the INTEL 8086 16 bit micro-processor. Detailed descriptions of all the components of the SBC and...space. The extra segment register (ES) is typically used for external or shared data, and data storage. 4. The iSBC86/12A( Single Board Computer ) The
Ichikawa, Tsubasa; Sakamoto, Yuji; Subagyo, Agus; Sueoka, Kazuhisa
2011-12-01
The research on reflectance distributions in computer-generated holograms (CGHs) is particularly sparse, and the textures of materials are not expressed. Thus, we propose a method for calculating reflectance distributions in CGHs that uses the finite-difference time-domain method. In this method, reflected light from an uneven surface made on a computer is analyzed by finite-difference time-domain simulation, and the reflected light distribution is applied to the CGH as an object light. We report the relations between the surface roughness of the objects and the reflectance distributions, and show that the reflectance distributions are given to CGHs by imaging simulation.
Institute of Scientific and Technical Information of China (English)
WANG Pei-dong; WEI Zhen-hua
2008-01-01
To make the on-board computer system more dependable and real-time in a satellite, an algorithm of the fault-tolerant scheduling in the on-board computer system with high priority recovery is proposed in this paper. This algorithm can schedule the on-board fault-tolerant tasks in real time. Due to the use of dependability cost, the overhead of scheduling the fault-tolerant tasks can be reduced. The mechanism of the high priority recovery will improve the response to recovery tasks. The fault-tolerant scheduling model is presented simulation results validate the correctness and feasibility of the proposed algorithm.
Introducing SummerTime: a package for high-precision computation of sums appearing in DRA method
Lee, Roman N
2015-01-01
We introduce the Mathematica package SummerTime for arbitrary-precision computation of sums appearing in the results of DRA method. So far these results include the following families of the integrals: 3-loop onshell massless vertices, 3-loop onshell mass operator type integrals, 4-loop QED-type tadpoles, 4-loop massless propagators. The package can be used for high-precision numerical computation of the expansion coefficients of the integrals from the above families around arbitrary space-time dimension. In addition, this package can also be used for calculation of multiple zeta values, harmonic polylogarithms and other transcendental numbers expressed in terms of nested sums with factorized summand.
Nang, Ei Ei Khaing; Salim, Agus; Wu, Yi; Tai, E Shyong; Lee, Jeannette; Van Dam, Rob M
2013-05-30
Recent evidence shows that sedentary behaviour may be an independent risk factor for cardiovascular diseases, diabetes, cancers and all-cause mortality. However, results are not consistent and different types of sedentary behaviour might have different effects on health. Thus the aim of this study was to evaluate the association between television screen time, computer/reading time and cardio-metabolic biomarkers in a multiethnic urban Asian population. We also sought to understand the potential mediators of this association. The Singapore Prospective Study Program (2004-2007), was a cross-sectional population-based study in a multiethnic population in Singapore. We studied 3305 Singaporean adults of Chinese, Malay and Indian ethnicity who did not have pre-existing diseases and conditions that could affect their physical activity. Multiple linear regression analysis was used to assess the association of television screen time and computer/reading time with cardio-metabolic biomarkers [blood pressure, lipids, glucose, adiponectin, C reactive protein and homeostasis model assessment of insulin resistance (HOMA-IR)]. Path analysis was used to examine the role of mediators of the observed association. Longer television screen time was significantly associated with higher systolic blood pressure, total cholesterol, triglycerides, C reactive protein, HOMA-IR, and lower adiponectin after adjustment for potential socio-demographic and lifestyle confounders. Dietary factors and body mass index, but not physical activity, were potential mediators that explained most of these associations between television screen time and cardio-metabolic biomarkers. The associations of television screen time with triglycerides and HOMA-IR were only partly explained by dietary factors and body mass index. No association was observed between computer/ reading time and worse levels of cardio-metabolic biomarkers. In this urban Asian population, television screen time was associated with worse
Energy Technology Data Exchange (ETDEWEB)
Kim, Byung Ki; Lee, Hui Joong; Lee, Jong Min; Kim, Yong Joo; Kang, Duck Sik [Kyungpook National Univ., Kyungpook National Univ. Hospital, College of Medicine, Taegu (Korea, Republic of)
1999-12-01
To determine the usefulness of MgSO{sub 4} for measuring the systemic circulation time. Systemic circulation time, defined as elapsed time from the injection of MgSO{sub 4} solution to the point of pharyngeal burning sensation, was measured in 63 volunteers. MgSO{sub 4} was injected into a superficial vein of an upper extremity. Using dynamic electron beam computed tomography at the level of the abdominal aorta and celiac axis, a time-intensity curve was plotted, and for these two locations, maximal enhancement time was compared. For 60 of the 63 subjects, both systemic circulation time and maximal enhancement time were determined. Average systemic circulation time was 17.4 (SD:3.6) secs. and average maximal enhancement times at the level of the abdominal aorta and celiac axis were 17.5 (SD:3.0) secs. and 18.5 (SD:3.2) secs., respectively. Correlation coefficients between systemic circulation time and maximal enhancement time for the abdominal aorta and celiac axis were 0.73 (p<0.01) and 0.73 (p<0.05) respectively. The systemic circulation time demonstrated by MgSO{sub 4} injection and maximal enhancement time for the abdominal aorta showed significant correlation. Thus, to determine the appropriate scanning time in contrast-enhanced radiological studies, MgSO{sub 4} can be used instead of a test bolus study.
Ziegler, C.; Schilling, D. L.
1977-01-01
Two networks consisting of single server queues, each with a constant service time, are considered. The external inputs to each network are assumed to follow some general probability distribution. Several interesting equivalencies that exist between the two networks considered are derived. This leads to the introduction of an important concept in delay decomposition. It is shown that the waiting time experienced by a customer can be decomposed into two basic components called self delay and interference delay.
The COMICS Tool - Computing Minimal Counterexamples for Discrete-time Markov Chains
Jansen, Nils; Scheffler, Maik; Volk, Matthias; Vorpahl, Andreas; Wimmer, Ralf; Katoen, Joost-Pieter; Becker, Bernd
2012-01-01
This report presents the tool COMICS, which performs model checking and generates counterexamples for DTMCs. For an input DTMC, COMICS computes an abstract system that carries the model checking information and uses this result to compute a critical subsystem, which induces a counterexample. This abstract subsystem can be refined and concretized hierarchically. The tool comes with a command-line version as well as a graphical user interface that allows the user to interactively influence the refinement process of the counterexample.
Directory of Open Access Journals (Sweden)
Alexander V. Baranov
2015-01-01
Full Text Available Taking part in the organized project activities students of the technical University create virtual physics laboratories. The article gives an example of the student’s project-computer modeling and visualization one of the most wonderful manifestations of reality-quantum interference of particles. The real experiment with heavy organic fluorescent molecules is used as a prototype for this computer simulation. The student’s software product can be used in informational space of the system of open education.
Space-Time Fluid-Structure Interaction Computation of Flapping-Wing Aerodynamics
2013-12-01
action, volume 53 of Lecture Notes in Computational Science and Engineering, 82–100, Springer , 2006. [24] K.-U. Bletzinger, R. Wuchner, and A. Kupzok...Lecture Notes in Computational Sci- ence and Engineering, 336–355, Springer , 2006. [25] Y. Bazilevs, V.M. Calo, T.J.R. Hughes, and Y. Zhang...structure interaction: Methods and applica- tion to cerebral aneurysms”, Biomechanics and Modeling in Mechanobiology, 9 (2010) 481–498. [32] Y. Bazilevs, M.-C
Space-Time Interface-Tracking Computations with Contact Between Solid Surfaces
2014-04-01
aneurysms. Biomechanics and Modeling in Mechanobiology, 9:481– 498, 2010. [14] Y. Bazilevs, M.-C. Hsu, Y. Zhang, W. Wang, X. Liang, T. Kvamsdal, R...and M. Schafer, editors, Fluid–Structure Interaction, volume 53 of Lecture Notes in Computational Sci- ence and Engineering, pages 336–355. Springer ... Springer , 2006. [34] M. Manguoglu, A. H. Sameh, T. E. Tezduyar, and S. Sathe. A nested iterative scheme for computation of incompressible flows in long
New multi-DSP parallel computing architecture for real-time image processing
Institute of Scientific and Technical Information of China (English)
Hu Junhong; Zhang Tianxu; Jiang Haoyang
2006-01-01
The flexibility of traditional image processing system is limited because those system are designed for specific applications. In this paper, a new TMS320C64x-based multi-DSP parallel computing architecture is presented. It has many promising characteristics such as powerful computing capability, broad I/O bandwidth, topology flexibility, and expansibility. The parallel system performance is evaluated by practical experiment.
Amsallem, David; Tezaur, Radek; Farhat, Charbel
2016-12-01
A comprehensive approach for real-time computations using a database of parametric, linear, projection-based reduced-order models (ROMs) based on arbitrary underlying meshes is proposed. In the offline phase of this approach, the parameter space is sampled and linear ROMs defined by linear reduced operators are pre-computed at the sampled parameter points and stored. Then, these operators and associated ROMs are transformed into counterparts that satisfy a certain notion of consistency. In the online phase of this approach, a linear ROM is constructed in real-time at a queried but unsampled parameter point by interpolating the pre-computed linear reduced operators on matrix manifolds and therefore computing an interpolated linear ROM. The proposed overall model reduction framework is illustrated with two applications: a parametric inverse acoustic scattering problem associated with a mockup submarine, and a parametric flutter prediction problem associated with a wing-tank system. The second application is implemented on a mobile device, illustrating the capability of the proposed computational framework to operate in real-time.
Zhu, Lingyun; Li, Lianjie; Meng, Chunyan
2014-12-01
There have been problems in the existing multiple physiological parameter real-time monitoring system, such as insufficient server capacity for physiological data storage and analysis so that data consistency can not be guaranteed, poor performance in real-time, and other issues caused by the growing scale of data. We therefore pro posed a new solution which was with multiple physiological parameters and could calculate clustered background data storage and processing based on cloud computing. Through our studies, a batch processing for longitudinal analysis of patients' historical data was introduced. The process included the resource virtualization of IaaS layer for cloud platform, the construction of real-time computing platform of PaaS layer, the reception and analysis of data stream of SaaS layer, and the bottleneck problem of multi-parameter data transmission, etc. The results were to achieve in real-time physiological information transmission, storage and analysis of a large amount of data. The simulation test results showed that the remote multiple physiological parameter monitoring system based on cloud platform had obvious advantages in processing time and load balancing over the traditional server model. This architecture solved the problems including long turnaround time, poor performance of real-time analysis, lack of extensibility and other issues, which exist in the traditional remote medical services. Technical support was provided in order to facilitate a "wearable wireless sensor plus mobile wireless transmission plus cloud computing service" mode moving towards home health monitoring for multiple physiological parameter wireless monitoring.
Moxon, Bruce C.; Green, John A.
1990-01-01
A high-performance platform for development of real-time helicopter flight simulations based on a simulation development and analysis platform combining a parallel simulation development and analysis environment with a scalable multiprocessor computer system is described. Simulation functional decomposition is covered, including the sequencing and data dependency of simulation modules and simulation functional mapping to multiple processors. The multiprocessor-based implementation of a blade-element simulation of the UH-60 helicopter is presented, and a prototype developed for a TC2000 computer is generalized in order to arrive at a portable multiprocessor software architecture. It is pointed out that the proposed approach coupled with a pilot's station creates a setting in which simulation engineers, computer scientists, and pilots can work together in the design and evaluation of advanced real-time helicopter simulations.
Hakvoort, W.B.J.; Aarts, R.G.K.M.; Dijk, van J.; Jonker, J.B.
2009-01-01
Iterative Learning Control (ILC) improves the tracking accuracy of systems that repetitively perform the same task. This paper considers model-based ILC for linear time-varying (LTV) systems. The applied feedforward iteratively minimises a quadratic norm of the feedforward update and the error in th
DEFF Research Database (Denmark)
Vedsted, Pernille; Søgaard, Karen; Blangsted, Anne Katrine
2011-01-01
Continuous electromyographic (EMG) activity level is considered a risk factor in developing muscle disorders. EMG biofeedback is known to be useful in reducing EMG activity in working muscles during computer work. The purpose was to test the following hypotheses: (1) unilateral biofeedback from...... trapezius (TRA) can reduce bilateral TRA activity but not extensor digitorum communis (EDC) activity; (2) biofeedback from EDC can reduce activity in EDC but not in TRA; (3) biofeedback is more effective in no time constraint than in the time constraint working condition. Eleven healthy women performed...... computer work during two different working conditions (time constraint/no time constraint) while receiving biofeedback. Biofeedback was given from right TRA or EDC through two modes (visual/auditory) by the use of EMG or mechanomyography as biofeedback source. During control sessions (no biofeedback), EMG...
DEFF Research Database (Denmark)
Vedsted, Pernille; Søgaard, Karen; Blangsted, Anne Katrine
2011-01-01
Continuous electromyographic (EMG) activity level is considered a risk factor in developing muscle disorders. EMG biofeedback is known to be useful in reducing EMG activity in working muscles during computer work. The purpose was to test the following hypotheses: (1) unilateral biofeedback from...... trapezius (TRA) can reduce bilateral TRA activity but not extensor digitorum communis (EDC) activity; (2) biofeedback from EDC can reduce activity in EDC but not in TRA; (3) biofeedback is more effective in no time constraint than in the time constraint working condition. Eleven healthy women performed...... computer work during two different working conditions (time constraint/no time constraint) while receiving biofeedback. Biofeedback was given from right TRA or EDC through two modes (visual/auditory) by the use of EMG or mechanomyography as biofeedback source. During control sessions (no biofeedback), EMG...
White, Timothy C.; Sauter, Edward A.; Stewart, Duff C.
2014-01-01
Intermagnet is an international oversight group which exists to establish a global network for geomagnetic observatories. This group establishes data standards and standard operating procedures for members and prospective members. Intermagnet has proposed a new One-Second Data Standard, for that emerging geomagnetic product. The standard specifies that all data collected must have a time stamp accuracy of ±10 milliseconds of the top-of-the-second Coordinated Universal Time. Therefore, the U.S. Geological Survey Geomagnetism Program has designed and executed several tests on its current data collection system, the Personal Computer Data Collection Platform. Tests are designed to measure the time shifts introduced by individual components within the data collection system, as well as to measure the time shift introduced by the entire Personal Computer Data Collection Platform. Additional testing designed for Intermagnet will be used to validate further such measurements. Current results of the measurements showed a 5.0–19.9 millisecond lag for the vertical channel (Z) of the Personal Computer Data Collection Platform and a 13.0–25.8 millisecond lag for horizontal channels (H and D) of the collection system. These measurements represent a dynamically changing delay introduced within the U.S. Geological Survey Personal Computer Data Collection Platform.
Dunn, H. J.
1981-01-01
A computer program for performing frequency analysis of time history data is presented. The program uses circular convolution and the fast Fourier transform to calculate power density spectrum (PDS) of time history data. The program interfaces with the advanced continuous simulation language (ACSL) so that a frequency analysis may be performed on ACSL generated simulation variables. An example of the calculation of the PDS of a Van de Pol oscillator is presented.
Particle Swarm Optimization with Time Varying Parameters for Scheduling in Cloud Computing
Directory of Open Access Journals (Sweden)
Shuang Zhao
2015-01-01
Full Text Available Task resource management is important in cloud computing system. It's necessary to find the efficient way to optimize scheduling in cloud computing. In this paper, an optimized particle swarm optimization (PSO algorithms with adaptive change of parameter (viz., inertial weight and acceleration coefficients according to the evolution state evaluation is presented. This adaptation helps to avoid premature convergence and explore the search space more efficiently. Simulations are carried out to test proposed algorithm, test reveal that the algorithm can achieving significant optimization of makespan.
Ortman, Robert L.; Carr, Domenic A.; James, Ryan; Long, Daniel; O'Shaughnessy, Matthew R.; Valenta, Christopher R.; Tuell, Grady H.
2016-05-01
We have developed a prototype real-time computer for a bathymetric lidar capable of producing point clouds attributed with total propagated uncertainty (TPU). This real-time computer employs a "mixed-mode" architecture comprised of an FPGA, CPU, and GPU. Noise reduction and ranging are performed in the digitizer's user-programmable FPGA, and coordinates and TPU are calculated on the GPU. A Keysight M9703A digitizer with user-programmable Xilinx Virtex 6 FPGAs digitizes as many as eight channels of lidar data, performs ranging, and delivers the data to the CPU via PCIe. The floating-point-intensive coordinate and TPU calculations are performed on an NVIDIA Tesla K20 GPU. Raw data and computed products are written to an SSD RAID, and an attributed point cloud is displayed to the user. This prototype computer has been tested using 7m-deep waveforms measured at a water tank on the Georgia Tech campus, and with simulated waveforms to a depth of 20m. Preliminary results show the system can compute, store, and display about 20 million points per second.
Real-time task recognition based on knowledge workers’ computer activities
Koldijk, S.J.; Staalduinen, M. van; Neerincx, M.A.; Kraaij, W.
2012-01-01
Motivation – Supporting knowledge workers in their self-management by providing them overviews of performed tasks. Research approach – Computer interaction data of knowledge workers was logged during their work. For each user different classifiers were trained and compared on their performance on re
Oral Computer-Mediated Interaction between L2 Learners: It's about Time!
Yanguas, Inigo
2010-01-01
This study explores task-based, synchronous oral computer-mediated communication (CMC) among intermediate-level learners of Spanish. In particular, this paper examines (a) how learners in video and audio CMC groups negotiate for meaning during task-based interaction, (b) possible differences between both oral CMC modes and traditional face-to-face…
Knibbe, H.P.
2015-01-01
The oil and gas industry makes use of computational intensive algorithms to provide an image of the subsurface. The image is obtained by sending wave energy into the subsurface and recording the signal required for a seismic wave to reflect back to the surface from the Earth interfaces that may have
A Prototype System for Real Time Computer Animation of Slow Traffic in a Driving Simulator
Roerdink, Jos B.T.M.; Delden, Mattijs J.B. van; Hin, Andrea J.S.; Wolffelaar, Peter C. van
1997-01-01
The Traffic Research Centre (TRC) of the University of Groningen in the Netherlands has developed a driving simulator with ‘intelligent’ computer-controlled traffic, consisting at the moment only of saloon cars. The range of possible applications would be greatly enhanced if other traffic
Real-Time Computer Animation of Bicyclists and Pedestrians in a Driving Simulator
Roerdink, Jos B.T.M.; Delden, Mattijs J.B. van; Hin, Andrea J.S.; Wolffelaar, Peter C. van
1996-01-01
The Traffic Research Centre (TRC) of the University of Groningen in the Netherlands has developed a driving simulator with ‘intelligent’ computer-controlled traffic, consisting at the moment only of saloon cars. The range of possible applications would be greatly enhanced if other traffic
A prototype system for real time computer animation of slow traffic in a driving simulator
Roerdink, JBTM; van Delden, MJB; Hin, AJS; van Wolffelaar, PC; Thalmann, NM; Skala,
1997-01-01
The Traffic Research Centre (TRC) of the University of Groningen in the Netherlands has developed a driving simulator with 'intelligent' computer-controlled traffic, consisting at the moment only of saloon cars. The range of possible applications would be greatly enhanced if other traffic
A Real-time Strategy Agent Framework and Strategy Classifier for Computer Generated Forces
2012-06-01
Python [46] scripts were written that leverage the SciPy [15] and NumPy [26] libraries. NumPy is a numerical package which provides efficient...208.pdf. [26] Numpy developers. “Scientific Computing Tools For Python — Numpy ”, 2011. URL http://numpy.scipy.org. [Online; accessed 24-Oct-2011]. [27
Visuospatial attention based brain-computer interfacing using real-time fMRI
Andersson, J.P.
2012-01-01
Despite all the literature showing that brain-computer interfacing (BCI) is achievable in a research setting, reports of paralyzed people utilizing the proposed systems are few. The performances of the BCI systems must most likely improve before they will move from research and become an assistive t
1980-02-15
Summary, Conclusions, and Recommendations, ESD-TR-66-671, December 1966. AD-646-867 103. Gradwohl, Alan J., Wootan, Wolford 0., Jr., Phase II Final... Larry R., Computer Resources Acquisition and Support for Air Force Weapons System, Defense Systems Management School, Report No. DSMS-PMC-750-1, Vol
Knibbe, H.P.
2015-01-01
The oil and gas industry makes use of computational intensive algorithms to provide an image of the subsurface. The image is obtained by sending wave energy into the subsurface and recording the signal required for a seismic wave to reflect back to the surface from the Earth interfaces that may have
Lipman, Yaron; Daubechies, Ingrid
2011-01-01
This paper is a companion paper to [Lipman and Daubechies 2011]. We provide numerical procedures and algorithms for computing the alignment of and distance between two disk type surfaces. We furthermore generalize the framework to support sphere-type surfaces, prove a result connecting this distance to geodesic distortion, and provide convergence analysis on the discrete approximation to the arising mass-transportation problems.
Mesh and Time-Step Independent Computational Fluid Dynamics (CFD) Solutions
Nijdam, Justin J.
2013-01-01
A homework assignment is outlined in which students learn Computational Fluid Dynamics (CFD) concepts of discretization, numerical stability and accuracy, and verification in a hands-on manner by solving physically realistic problems of practical interest to engineers. The students solve a transient-diffusion problem numerically using the common…
Teufel, Julian; Bardins, S; Spiegel, Rainer; Kremmyda, O.; Schneider, E.; Strupp, M; Kalla, Roger
2016-01-01
BACKGROUND Patients with downbeat nystagmus syndrome suffer from oscillopsia, which leads to an unstable visual perception and therefore impaired visual acuity. The aim of this study was to use real-time computer-based visual feedback to compensate for the destabilizing slow phase eye movements. METHODS The patients were sitting in front of a computer screen with the head fixed on a chin rest. The eye movements were recorded by an eye tracking system (EyeSeeCam®). We tested the vi...
Institute of Scientific and Technical Information of China (English)
LI; Zicheng; SUN; Yukun
2006-01-01
Considering the detection principle that "when load current is periodic current, the integral in a cycle for absolute value of load current subtracting fundamental active current is the least", harmonic current real-time detection methods for power active filter are proposed based on direct computation, simple iterative algorithm and optimal iterative algorithm. According to the direct computation method, the amplitude of the fundamental active current can be accurately calculated when load current is placed in stable state. The simple iterative algorithm and the optimal iterative algorithm provide an idea about judging the state of load current. On the basis of the direct computation method, the simple iterative algorithm, the optimal iterative algorithm and precise definition of the basic concepts such as the true amplitude of the fundamental active current when load current is placed in varying state, etc., the double linear construction idea is proposed in which the amplitude of the fundamental active current at the moment of the sample is accurately calculated by using the first linear construction and the condition which disposes the next sample is created by using the second linear construction. On the basis of the double linear construction idea, a harmonic current real-time detection method for power active filter is proposed based on the double linear construction algorithm. This method has the characteristics of small computing quantity, fine real-time performance, being capable of accurately calculating the amplitude of the fundamental active current and so on.
Geiger, Martin Josef
2008-01-01
The article describes an investigation of the effectiveness of genetic algorithms for multi-objective combinatorial optimization (MOCO) by presenting an application for the vehicle routing problem with soft time windows. The work is motivated by the question, if and how the problem structure influences the effectiveness of different configurations of the genetic algorithm. Computational results are presented for different classes of vehicle routing problems, varying in their coverage with time windows, time window size, distribution and number of customers. The results are compared with a simple, but effective local search approach for multi-objective combinatorial optimization problems.
Real-Time Computation of Parameter Fitting and Image Reconstruction Using Graphical Processing Units
Locans, Uldis; Suter, Andreas; Fischer, Jannis; Lustermann, Werner; Dissertori, Gunther; Wang, Qiulin
2016-01-01
In recent years graphical processing units (GPUs) have become a powerful tool in scientific computing. Their potential to speed up highly parallel applications brings the power of high performance computing to a wider range of users. However, programming these devices and integrating their use in existing applications is still a challenging task. In this paper we examined the potential of GPUs for two different applications. The first application, created at Paul Scherrer Institut (PSI), is used for parameter fitting during data analysis of muSR (muon spin rotation, relaxation and resonance) experiments. The second application, developed at ETH, is used for PET (Positron Emission Tomography) image reconstruction and analysis. Applications currently in use were examined to identify parts of the algorithms in need of optimization. Efficient GPU kernels were created in order to allow applications to use a GPU, to speed up the previously identified parts. Benchmarking tests were performed in order to measure the ...
Using the CPU and GPU for real-time video enhancement on a mobile computer
CSIR Research Space (South Africa)
Bachoo, AK
2010-09-01
Full Text Available Language (2nd Edition), Addison- Wesley Professional, 2006. [6] D. Shreiner, M. Woo, J. Neider, and T. Davis, OpenGL Programming Guide, Addison-Wesley Professional, 2007. [7] SM Pizer, EP Amburn, JD Austin, R Cromartie, A Geselowitz, T Greer, BM ter... computer,? in The Twentieth Annual Symposium of the Pattern Recognition Association of South Africa (PRASA), 2009. [11] RC Gonzalez and RE Woods, Digital image processing, Addison-Wesley Publishing Company, 2002. ...
Blake, Douglas Clifton
A new methodology is presented for conducting numerical simulations of electromagnetic scattering and wave-propagation phenomena on massively parallel computing platforms. A process is constructed which is rooted in the Finite-Volume Time-Domain (FVTD) technique to create a simulation capability that is both versatile and practical. In terms of versatility, the method is platform independent, is easily modifiable, and is capable of solving a large number of problems with no alterations. In terms of practicality, the method is sophisticated enough to solve problems of engineering significance and is not limited to mere academic exercises. In order to achieve this capability, techniques are integrated from several scientific disciplines including computational fluid dynamics, computational electromagnetics, and parallel computing. The end result is the first FVTD solver capable of utilizing the highly flexible overset-gridding process in a distributed-memory computing environment. In the process of creating this capability, work is accomplished to conduct the first study designed to quantify the effects of domain-decomposition dimensionality on the parallel performance of hyperbolic partial differential equations solvers; to develop a new method of partitioning a computational domain comprised of overset grids; and to provide the first detailed assessment of the applicability of overset grids to the field of computational electromagnetics. Using these new methods and capabilities, results from a large number of wave propagation and scattering simulations are presented. The overset-grid FVTD algorithm is demonstrated to produce results of comparable accuracy to single-grid simulations while simultaneously shortening the grid-generation process and increasing the flexibility and utility of the FVTD technique. Furthermore, the new domain-decomposition approaches developed for overset grids are shown to be capable of producing partitions that are better load balanced and
Real-time C Code Generation in Ptolemy II for the Giotto Model of Computation
2009-05-20
BAUER . The time-triggered architecture. In Proceedings of the IEEE, volume 91, pages 112–126, January 2003. [18] N. V. Krishnan. Real-Time Systems Design...http://www.freertos.org. [23] The Ptolemy Project. http://ptolemy.eecs.berkeley.edu/. [24] a. J. T. Wolfgang Pree. The Timing Definition Language, October 21 2005. 38
Reymond, Dominique
2017-04-01
We present a tool for computing the complete arrival times of the dispersed wave-train of a tsunami. The calculus is made using the exact formulation of the tsunami dispersion (and without approximations), at any desired periods between one hour or more (concerning the gravity waves propagation) until 10s (the highly dispersed mode). The computation of the travel times is based on the a summation of the necessary time for a tsunami to cross all the elementary blocs of a grid of bathymetry following a path between the source and receiver at a given period. In addition the source dimensions and the focal mechanism are taken into account to adjust the minimum travel time to the different possible points of emission of the source. A possible application of this tool is to forecast the arrival time of late arrivals of tsunami waves that could produce the resonnance of some bays and sites at higher frequencies than the gravity mode. The theoretical arrival times are compared to the observed ones and to the results obtained by TTT (P. Wessel, 2009) and the ones obtained by numerical simulations. References: Wessel, P. (2009). Analysis of oberved and predicted tsunami travel times for the Pacic and Indian oceans. Pure Appl. Geophys., 166:301-324.
Rasmussen, Patrick P.; Gray, John R.; Glysson, G. Doug; Ziegler, Andrew C.
2010-01-01
Over the last decade, use of a method for computing suspended-sediment concentration and loads using turbidity sensors—primarily nephelometry, but also optical backscatter—has proliferated. Because an in- itu turbidity sensor is capa le of measuring turbidity instantaneously, a turbidity time series can be recorded and related directly to time-varying suspended-sediment concentrations. Depending on the suspended-sediment characteristics of the measurement site, this method can be more reliable and, in many cases, a more accurate means for computing suspended-sediment concentrations and loads than traditional U.S. Geological Survey computational methods. Guidelines and procedures for estimating time s ries of suspended-sediment concentration and loading as a function of turbidity and streamflow data have been published in a U.S. Geological Survey Techniques and Methods Report, Book 3, Chapter C4. This paper is a summary of these guidelines and discusses some of the concepts, s atistical procedures, and techniques used to maintain a multiyear suspended sediment time series.
Directory of Open Access Journals (Sweden)
Saffet Ayasun
2014-01-01
Full Text Available This paper investigates the effect of time delays on the stability of a generator excitation control system compensated with a stabilizing transformer known as rate feedback stabilizer to damp out oscillations. The time delays are due to the use of measurement devices and communication links for data transfer. An analytical method is presented to compute the delay margin for stability. The delay margin is the maximum amount of time delay that the system can tolerate before it becomes unstable. First, without using any approximation, the transcendental characteristic equation is converted into a polynomial without the transcendentality such that its real roots coincide with the imaginary roots of the characteristic equation exactly. The resulting polynomial also enables us to easily determine the delay dependency of the system stability and the sensitivities of crossing roots with respect to the time delay. Then, an expression in terms of system parameters and imaginary root of the characteristic equation is derived for computing the delay margin. Theoretical delay margins are computed for a wide range of controller gains and their accuracy is verified by performing simulation studies. Results indicate that the addition of a stabilizing transformer to the excitation system increases the delay margin and improves the system damping significantly.
Hui, Xiaonan; Ye, Taihang; Zheng, Shilie; Zhou, Jinhai; Chi, Hao; Jin, Xiaofeng; Zhang, Xianmin
2014-10-01
For a phase-sensitive optical time-domain reflectometer (ϕ-OTDR) distributed sensor system, space-frequency analysis can reduce the false alarm by analyzing the frequency distribution compared with the traditional difference value method. We propose a graphics processing unit (GPU)-based parallel computing method to perform multichannel fast Fourier transform (FFT) and realize the real-time space-frequency analysis. The experiment results show that the time taken by the multichannel FFT decreased considerably based on this GPU parallel computing. The method can be completed with a sensing fiber up to 16 km long and an entry-level GPU. Meanwhile, the GPU can reduce the computing load of the central processing unit from 70% down to less than 20%. We carried out an experiment on a two-point space-frequency analysis, and the results clearly and simultaneously show the vibration point locations and frequency components. The sensor system outputs the real-time space-frequency spectra continuously with a spatial resolution of 16.3 m and frequency resolution of 2.25 Hz.
Cohen, Stephanie A; McIlvried, Dawn E
2011-06-01
Cancer genetic counseling sessions traditionally encompass collecting medical and family history information, evaluating that information for the likelihood of a genetic predisposition for a hereditary cancer syndrome, conveying that information to the patient, offering genetic testing when appropriate, obtaining consent and subsequently documenting the encounter with a clinic note and pedigree. Software programs exist to collect family and medical history information electronically, intending to improve efficiency and simplicity of collecting, managing and storing this data. This study compares the genetic counselor's time spent in cancer genetic counseling tasks in a traditional model and one using computer-assisted data collection, which is then used to generate a pedigree, risk assessment and consult note. Genetic counselor time spent collecting family and medical history and providing face-to-face counseling for a new patient session decreased from an average of 85-69 min when using the computer-assisted data collection. However, there was no statistically significant change in overall genetic counselor time on all aspects of the genetic counseling process, due to an increased amount of time spent generating an electronic pedigree and consult note. Improvements in the computer program's technical design would potentially minimize data manipulation. Certain aspects of this program, such as electronic collection of family history and risk assessment, appear effective in improving cancer genetic counseling efficiency while others, such as generating an electronic pedigree and consult note, do not.
Directory of Open Access Journals (Sweden)
Wei Huang
2017-03-01
Full Text Available Geospatial big data analysis (GBDA is extremely significant for time-constraint applications such as disaster response. However, the time-constraint analysis is not yet a trivial task in the cloud computing environment. Spatial query processing (SQP is typical computation-intensive and indispensable for GBDA, and the spatial range query, join query, and the nearest neighbor query algorithms are not scalable without using MapReduce-liked frameworks. Parallel SQP algorithms (PSQPAs are trapped in screw-processing, which is a known issue in Geoscience. To satisfy time-constrained GBDA, we propose an elastic SQP approach in this paper. First, Spark is used to implement PSQPAs. Second, Kubernetes-managed Core Operation System (CoreOS clusters provide self-healing Docker containers for running Spark clusters in the cloud. Spark-based PSQPAs are submitted to Docker containers, where Spark master instances reside. Finally, the horizontal pod auto-scaler (HPA would scale-out and scale-in Docker containers for supporting on-demand computing resources. Combined with an auto-scaling group of virtual instances, HPA helps to find each of the five nearest neighbors for 46,139,532 query objects from 834,158 spatial data objects in less than 300 s. The experiments conducted on an OpenStack cloud demonstrate that auto-scaling containers can satisfy time-constraint GBDA in clouds.
Directory of Open Access Journals (Sweden)
J. Adam Wilson
2009-07-01
Full Text Available The clock speeds of modern computer processors have nearly plateaued in the past five years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a graphics card (GPU was developed for real-time neural signal processing of a brain-computer interface (BCI. The NVIDIA CUDA system was used to offload processing to the GPU, which is capable of running many operations in parallel, potentially greatly increasing the speed of existing algorithms. The BCI system records many channels of data, which are processed and translated into a control signal, such as the movement of a computer cursor. This signal processing chain involves computing a matrix-matrix multiplication (i.e., a spatial filter, followed by calculating the power spectral density on every channel using an auto-regressive method, and finally classifying appropriate features for control. In this study, the first two computationally-intensive steps were implemented on the GPU, and the speed was compared to both the current implementation and a CPU-based implementation that uses multi-threading. Significant performance gains were obtained with GPU processing: the current implementation processed 1000 channels in 933 ms, while the new GPU method took only 27 ms, an improvement of nearly 35 times.
Wilson, J Adam; Williams, Justin C
2009-01-01
The clock speeds of modern computer processors have nearly plateaued in the past 5 years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a graphics card [graphics processing unit (GPU)] was developed for real-time neural signal processing of a brain-computer interface (BCI). The NVIDIA CUDA system was used to offload processing to the GPU, which is capable of running many operations in parallel, potentially greatly increasing the speed of existing algorithms. The BCI system records many channels of data, which are processed and translated into a control signal, such as the movement of a computer cursor. This signal processing chain involves computing a matrix-matrix multiplication (i.e., a spatial filter), followed by calculating the power spectral density on every channel using an auto-regressive method, and finally classifying appropriate features for control. In this study, the first two computationally intensive steps were implemented on the GPU, and the speed was compared to both the current implementation and a central processing unit-based implementation that uses multi-threading. Significant performance gains were obtained with GPU processing: the current implementation processed 1000 channels of 250 ms in 933 ms, while the new GPU method took only 27 ms, an improvement of nearly 35 times.
Directory of Open Access Journals (Sweden)
Her-Tyan Yeh
2013-01-01
Full Text Available Mobile devices such as personal digital assistants (PDAs, smartphones, and tablets have increased in popularity and are extremely efficient for work-related, social, and entertainment uses. Popular entertainment services have also attracted substantial attention. Thus, relevant industries have exerted considerable efforts in establishing a method by which mobile devices can be used to develop excellent and convenient entertainment services. Because cloud-computing technology is mature and possesses a strong computing processing capacity, integrating this technology into the entertainment service function in mobile devices can reduce the data load on a system and maintain mobile device performances. This study combines cloud computing with a mobile device to design a karaoke system that contains real-time media merging and sharing functions. This system enables users to download music videos (MVs from their mobile device and sing and record their singing by using the device. They can upload the recorded song to the cloud server where it is merged with real-time media. Subsequently, by employing a media streaming technology, users can store their personal MVs in their mobile device or computer and instantaneously share these videos with others on the Internet. Through this process, people can instantly watch shared videos, enjoy the leisure and entertainment effects of mobile devices, and satisfy their desire for singing.
Directory of Open Access Journals (Sweden)
René Riedl
2013-01-01
Full Text Available In today’s society, as computers, the Internet, and mobile phones pervade almost every corner of life, the impact of Information and Communication Technologies (ICT on humans is dramatic. The use of ICT, however, may also have a negative side. Human interaction with technology may lead to notable stress perceptions, a phenomenon referred to as technostress. An investigation of the literature reveals that computer users’ gender has largely been ignored in technostress research, treating users as “gender-neutral.” To close this significant research gap, we conducted a laboratory experiment in which we investigated users’ physiological reaction to the malfunctioning of technology. Based on theories which explain that men, in contrast to women, are more sensitive to “achievement stress,” we predicted that male users would exhibit higher levels of stress than women in cases of system breakdown during the execution of a human-computer interaction task under time pressure, if compared to a breakdown situation without time pressure. Using skin conductance as a stress indicator, the hypothesis was confirmed. Thus, this study shows that user gender is crucial to better understanding the influence of stress factors such as computer malfunctions on physiological stress reactions.
Investigate Methods to Decrease Compilation Time-AX-Program Code Group Computer Science R& D Project
Energy Technology Data Exchange (ETDEWEB)
Cottom, T
2003-06-11
Large simulation codes can take on the order of hours to compile from scratch. In Kull, which uses generic programming techniques, a significant portion of the time is spent generating and compiling template instantiations. I would like to investigate methods that would decrease the overall compilation time for large codes. These would be methods which could then be applied, hopefully, as standard practice to any large code. Success is measured by the overall decrease in wall clock time a developer spends waiting for an executable. Analyzing the make system of a slow to build project can benefit all developers on the project. Taking the time to analyze the number of processors used over the life of the build and restructuring the system to maximize the parallelization can significantly reduce build times. Distributing the build across multiple machines with the same configuration can increase the number of available processors for building and can help evenly balance the load. Becoming familiar with compiler options can have its benefits as well. The time improvements of the sum can be significant. Initial compilation time for Kull on OSF1 was {approx} 3 hours. Final time on OSF1 after completion is 16 minutes. Initial compilation time for Kull on AIX was {approx} 2 hours. Final time on AIX after completion is 25 minutes. Developers now spend 3 hours less waiting for a Kull executable on OSF1, and 2 hours less on AIX platforms. In the eyes of many Kull code developers, the project was a huge success.
Plant, Richard R
2016-03-01
There is an ongoing 'replication crisis' across the field of psychology in which researchers, funders, and members of the public are questioning the results of some scientific studies and the validity of the data they are based upon. However, few have considered that a growing proportion of research in modern psychology is conducted using a computer. Could it simply be that the hardware and software, or experiment generator, being used to run the experiment itself be a cause of millisecond timing error and subsequent replication failure? This article serves as a reminder that millisecond timing accuracy in psychology studies remains an important issue and that care needs to be taken to ensure that studies can be replicated on current computer hardware and software.
Directory of Open Access Journals (Sweden)
Harish S. Venkatarama
2010-10-01
Full Text Available Ecommerce is an area where an Autonomic Computing system could be very effectively deployed. Ecommerce has created demand for high quality information technology services and businesses are seeking quality of service guarantees from their service providers. These guarantees are expressed as part of service level agreements. Properly adjusting tuning parameters for enforcement of the service level agreement is time-consuming and skills-intensive. Moreover, in case of changes to the workload, the setting of the parameters may no longer be optimum. In an ecommerce system, where the workload changes frequently, there is a need to update the parameters at regular intervals. This paper describes two approaches, one, using a proportional controller and two, using a fuzzy controller, to automate the tuning of MaxClients parameter of Apache web server based on the required response time and the current workload. This is an illustration of the self-optimizing characteristic of an autonomic computing system.
Energy Technology Data Exchange (ETDEWEB)
Nieto, J., E-mail: jnieto@sec.upm.es [Grupo de Investigacion en Instrumentacion y Acustica Aplicada. Universidad Politecnica de Madrid, Crta. Valencia Km-7, Madrid 28031 Spain (Spain); Arcas, G. de; Ruiz, M. [Grupo de Investigacion en Instrumentacion y Acustica Aplicada. Universidad Politecnica de Madrid, Crta. Valencia Km-7, Madrid 28031 Spain (Spain); Vega, J. [Asociacion EURATOM/CIEMAT para Fusion, Madrid (Spain); Lopez, J.M.; Barrera, E. [Grupo de Investigacion en Instrumentacion y Acustica Aplicada. Universidad Politecnica de Madrid, Crta. Valencia Km-7, Madrid 28031 Spain (Spain); Castro, R. [Asociacion EURATOM/CIEMAT para Fusion, Madrid (Spain); Sanz, D. [Grupo de Investigacion en Instrumentacion y Acustica Aplicada. Universidad Politecnica de Madrid, Crta. Valencia Km-7, Madrid 28031 Spain (Spain); Utzel, N.; Makijarvi, P.; Zabeo, L. [ITER Organization, CS 90 046, 13067 St. Paul lez Durance Cedex (France)
2012-12-15
Highlights: Black-Right-Pointing-Pointer Implementation of fast plant system controller (FPSC) for ITER CODAC. Black-Right-Pointing-Pointer GPU-based real time high performance computing service. Black-Right-Pointing-Pointer Performance evaluation with respect to other solutions based in multi-core processors. - Abstract: EURATOM/CIEMAT and the Technical University of Madrid UPM are involved in the development of a FPSC (fast plant system control) prototype for ITER based on PXIe form factor. The FPSC architecture includes a GPU-based real time high performance computing service which has been integrated under EPICS (experimental physics and industrial control system). In this work we present the design of this service and its performance evaluation with respect to other solutions based in multi-core processors. Plasma pre-processing algorithms, illustrative of the type of tasks that could be required for both control and diagnostics, are used during the performance evaluation.
Introducing SummerTime: A package for high-precision computation of sums appearing in DRA1 method
Lee, Roman N.; Mingulov, Kirill T.
2016-06-01
We present the Mathematica package SummerTime for arbitrary-precision computation of sums appearing in the results of DRA method (Lee, 2010). So far these results include the following families of the integrals: 3-loop onshell massless vertices, 3-loop onshell mass operator type integrals, 4-loop QED-type tadpoles, 4-loop massless propagators (Lee et al., 2010; Lee and Smirnov, 2011; Lee et al., 2011, 2012). The package can be used for high-precision numerical computation of the expansion of the integrals from the above families around arbitrary space-time dimension. In addition, this package contains convenient tools for the calculation of multiple zeta values, harmonic polylogarithms and other transcendental numbers expressed in terms of nested sums with factorized summand.
Institute of Scientific and Technical Information of China (English)
CHEN Bai-lian; LI En-dong; Kazunobu Yamawuchi; Ken Kato; Shinji Naganawa; MIAO Wei-jun
2010-01-01
Background As an important determinant of patient satisfaction, waiting time, has gained increasing attention in the field of health care services. The present study aimed to illustrate the distribution characteristics of waiting time in a community hospital and explore the impact of potential measures to reduce outpatient waiting time based on a computer simulation approach. Methods During a one-month study period in 2006, a cross-sectional study was conducted in a community hospital located in Shanghai, China. Baseline data of outpatient waiting time were calculated according to the records of registration time and payment time. A simulation technique was adopted to investigate the impact of perspective reform methods on reducing waiting time. Results Data from a total of 10 092 patients and 26 816 medical consultations were collected in the study and 19 947 medical consultations were included. The average of the total visit time for outpatients in this hospital was 43.6 minutes in the morning, 19.1 minutes in the afternoon, and 34.3 minutes for the whole day studied period. The simulation results suggested that waiting time for outpatients could be greatly reduced through the introduction of appointment system and flexible demand-orientated doctor scheduling according to the numbers of patients waiting at different time of the workday. Conclusion Adoption of an appointment system and flexible management of doctor scheduling may be effective way to achieve decreased waiting time.
McCarthy, Peter M.
2006-01-01
The Yellowstone River is very important in a variety of ways to the residents of southeastern Montana; however, it is especially vulnerable to spilled contaminants. In 2004, the U.S. Geological Survey, in cooperation with Montana Department of Environmental Quality, initiated a study to develop a computer program to rapidly estimate instream travel times and concentrations of a potential contaminant in the Yellowstone River using regression equations developed in 1999 by the U.S. Geological Survey. The purpose of this report is to describe these equations and their limitations, describe the development of a computer program to apply the equations to the Yellowstone River, and provide detailed instructions on how to use the program. This program is available online at [http://pubs.water.usgs.gov/sir2006-5057/includes/ytot.xls]. The regression equations provide estimates of instream travel times and concentrations in rivers where little or no contaminant-transport data are available. Equations were developed and presented for the most probable flow velocity and the maximum probable flow velocity. These velocity estimates can then be used to calculate instream travel times and concentrations of a potential contaminant. The computer program was developed so estimation equations for instream travel times and concentrations can be solved quickly for sites along the Yellowstone River between Corwin Springs and Sidney, Montana. The basic types of data needed to run the program are spill data, streamflow data, and data for locations of interest along the Yellowstone River. Data output from the program includes spill location, river mileage at specified locations, instantaneous discharge, mean-annual discharge, drainage area, and channel slope. Travel times and concentrations are provided for estimates of the most probable velocity of the peak concentration and the maximum probable velocity of the peak concentration. Verification of estimates of instream travel times and
Energy Technology Data Exchange (ETDEWEB)
SO,I.; SIDDONS, D.P.; CALIEBE, W.A.; KHALID, S.
2007-04-25
We describe here the data acquisition subsystem of the Quick EXAFS (QEXAFS) experiment at the National Synchrotron Light Source of Brookhaven National Laboratory. For ease of future growth and flexibility, almost all software components are open source with very active maintainers. Among them, Linux running on x86 desktop computer, RTAI for real-time response, COMEDI driver for the data acquisition hardware, Qt and PyQt for graphical user interface, PyQwt for plotting, and Python for scripting. The signal (A/D) and energy-reading (IK220 encoder) devices in the PCI computer are also EPICS enabled. The control system scans the monochromator energy through a networked EPICS motor. With the real-time kernel, the system is capable of deterministic data-sampling period of tens of micro-seconds with typical timing-jitter of several micro-seconds. At the same time, Linux is running in other non-real-time processes handling the user-interface. A modern Qt-based controls-front end enhances productivity. The fast plotting and zooming of data in time or energy coordinates let the experimenters verify the quality of the data before detailed analysis. Python scripting is built-in for automation. The typical data-rate for continuous runs are around ten mega-bytes per minute.
Energy Technology Data Exchange (ETDEWEB)
SO,I.; SIDDONS, D.P.; CALIEBE, W.A.; KHALID, S.
2007-04-25
We describe here the data acquisition subsystem of the Quick EXAFS (QEXAFS) experiment at the National Synchrotron Light Source of Brookhaven National Laboratory. For ease of future growth and flexibility, almost all software components are open source with very active maintainers. Among them, Linux running on x86 desktop computer, RTAI for real-time response, COMEDI driver for the data acquisition hardware, Qt and PyQt for graphical user interface, PyQwt for plotting, and Python for scripting. The signal (A/D) and energy-reading (IK220 encoder) devices in the PCI computer are also EPICS enabled. The control system scans the monochromator energy through a networked EPICS motor. With the real-time kernel, the system is capable of deterministic data-sampling period of tens of micro-seconds with typical timing-jitter of several micro-seconds. At the same time, Linux is running in other non-real-time processes handling the user-interface. A modern Qt-based controls-front end enhances productivity. The fast plotting and zooming of data in time or energy coordinates let the experimenters verify the quality of the data before detailed analysis. Python scripting is built-in for automation. The typical data-rate for continuous runs are around ten mega-bytes per minute.
Knibbe, H.P.
2015-01-01
The oil and gas industry makes use of computational intensive algorithms to provide an image of the subsurface. The image is obtained by sending wave energy into the subsurface and recording the signal required for a seismic wave to reflect back to the surface from the Earth interfaces that may have different physical properties. A seismic wave is usually generated by shots of known frequencies, placed close to the surface on land or close to the water surface in the sea. Returning waves are ...
Cone-beam computed tomography: Time to move from ALARA to ALADA
Energy Technology Data Exchange (ETDEWEB)
Jaju, Prashant P.; Jaju, Sushma P. [Rishiraj College of Dental Sciences and Research Centre, Bhopa(Indonesia)
2015-12-15
Cone-beam computed tomography (CBCT) is routinely recommended for dental diagnosis and treatment planning. CBCT exposes patients to less radiation than does conventional CT. Still, lack of proper education among dentists and specialists is resulting in improper referral for CBCT. In addition, aiming to generate high-quality images, operators may increase the radiation dose, which can expose the patient to unnecessary risk. This letter advocates appropriate radiation dosing during CBCT to the benefit of both patients and dentists, and supports moving from the concept of 'as low as reasonably achievable' (ALARA) to 'as low as diagnostically acceptable' (ALADA.
Toward a web-based real-time radiation treatment planning system in a cloud computing environment
Hum Na, Yong; Suh, Tae-Suk; Kapp, Daniel S.; Xing, Lei
2013-09-01
To exploit the potential dosimetric advantages of intensity modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT), an in-depth approach is required to provide efficient computing methods. This needs to incorporate clinically related organ specific constraints, Monte Carlo (MC) dose calculations, and large-scale plan optimization. This paper describes our first steps toward a web-based real-time radiation treatment planning system in a cloud computing environment (CCE). The Amazon Elastic Compute Cloud (EC2) with a master node (named m2.xlarge containing 17.1 GB of memory, two virtual cores with 3.25 EC2 Compute Units each, 420 GB of instance storage, 64-bit platform) is used as the backbone of cloud computing for dose calculation and plan optimization. The master node is able to scale the workers on an ‘on-demand’ basis. MC dose calculation is employed to generate accurate beamlet dose kernels by parallel tasks. The intensity modulation optimization uses total-variation regularization (TVR) and generates piecewise constant fluence maps for each initial beam direction in a distributed manner over the CCE. The optimized fluence maps are segmented into deliverable apertures. The shape of each aperture is iteratively rectified to be a sequence of arcs using the manufacture’s constraints. The output plan file from the EC2 is sent to the simple storage service. Three de-identified clinical cancer treatment plans have been studied for evaluating the performance of the new planning platform with 6 MV flattening filter free beams (40 × 40 cm2) from the Varian TrueBeamTM STx linear accelerator. A CCE leads to speed-ups of up to 14-fold for both dose kernel calculations and plan optimizations in the head and neck, lung, and prostate cancer cases considered in this study. The proposed system relies on a CCE that is able to provide an infrastructure for parallel and distributed computing. The resultant plans from the cloud computing are identical
Toward a web-based real-time radiation treatment planning system in a cloud computing environment.
Na, Yong Hum; Suh, Tae-Suk; Kapp, Daniel S; Xing, Lei
2013-09-21
To exploit the potential dosimetric advantages of intensity modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT), an in-depth approach is required to provide efficient computing methods. This needs to incorporate clinically related organ specific constraints, Monte Carlo (MC) dose calculations, and large-scale plan optimization. This paper describes our first steps toward a web-based real-time radiation treatment planning system in a cloud computing environment (CCE). The Amazon Elastic Compute Cloud (EC2) with a master node (named m2.xlarge containing 17.1 GB of memory, two virtual cores with 3.25 EC2 Compute Units each, 420 GB of instance storage, 64-bit platform) is used as the backbone of cloud computing for dose calculation and plan optimization. The master node is able to scale the workers on an 'on-demand' basis. MC dose calculation is employed to generate accurate beamlet dose kernels by parallel tasks. The intensity modulation optimization uses total-variation regularization (TVR) and generates piecewise constant fluence maps for each initial beam direction in a distributed manner over the CCE. The optimized fluence maps are segmented into deliverable apertures. The shape of each aperture is iteratively rectified to be a sequence of arcs using the manufacture's constraints. The output plan file from the EC2 is sent to the simple storage service. Three de-identified clinical cancer treatment plans have been studied for evaluating the performance of the new planning platform with 6 MV flattening filter free beams (40 × 40 cm(2)) from the Varian TrueBeam(TM) STx linear accelerator. A CCE leads to speed-ups of up to 14-fold for both dose kernel calculations and plan optimizations in the head and neck, lung, and prostate cancer cases considered in this study. The proposed system relies on a CCE that is able to provide an infrastructure for parallel and distributed computing. The resultant plans from the cloud computing are
J. Adam Wilson; Williams, Justin C.
2009-01-01
The clock speeds of modern computer processors have nearly plateaued in the past 5 years. Consequently, neural prosthetic systems that rely on processing large quantities of data in a short period of time face a bottleneck, in that it may not be possible to process all of the data recorded from an electrode array with high channel counts and bandwidth, such as electrocorticographic grids or other implantable systems. Therefore, in this study a method of using the processing capabilities of a ...
Directory of Open Access Journals (Sweden)
Julien eVitay
2014-01-01
Full Text Available Neural activity in dopaminergic areas such as the ventral tegmental area is influenced by timing processes, in particular by the temporal expectation of rewards during Pavlovian conditioning. Receipt of a reward at the expected time allows to compute reward-prediction errors which can drive learning in motor or cognitive structures. Reciprocally, dopamine plays an important role in the timing of external events. Several models of the dopaminergic system exist, but the substrate of temporal learning is rather unclear. In this article, we propose a neuro-computational model of the afferent network to the ventral tegmental area, including the lateral hypothalamus, the pedunculopontine nucleus, the amygdala, the ventromedial prefrontal cortex, the ventral basal ganglia (including the nucleus accumbens and the ventral pallidum, as well as the lateral habenula and the rostromedial tegmental nucleus. Based on a plausible connectivity and realistic learning rules, this neuro-computational model reproduces several experimental observations, such as the progressive cancellation of dopaminergic bursts at reward delivery, the appearance of bursts at the onset of reward-predicting cues or the influence of reward magnitude on activity in the amygdala and ventral tegmental area. While associative learning occurs primarily in the amygdala, learning of the temporal relationship between the cue and the associated reward is implemented as a dopamine-modulated coincidence detection mechanism in the nucleus accumbens.
Hao, Stephanie; Subramanian, Sandya; Jordan, Austin; Santaniello, Sabato; Yaffe, Robert; Jouny, Christophe C; Bergey, Gregory K; Anderson, William S; Sarma, Sridevi V
2014-01-01
The surgical resection of the epileptogenic zone (EZ) is the only effective treatment for many drug-resistant epilepsy (DRE) patients, but the pre-surgical identification of the EZ is challenging. This study investigates whether the EZ exhibits a computationally identifiable signature during seizures. In particular, we compute statistics of the brain network from intracranial EEG (iEEG) recordings and track the evolution of network connectivity before, during, and after seizures. We define each node in the network as an electrode and weight each edge connecting a pair of nodes by the gamma band cross power of the corresponding iEEG signals. The eigenvector centrality (EVC) of each node is tracked over two seizures per patient and the electrodes are ranked according to the corresponding EVC value. We hypothesize that electrodes covering the EZ have a signature EVC rank evolution during seizure that differs from electrodes outside the EZ. We tested this hypothesis on multi-channel iEEG recordings from 2 DRE patients who had successful surgery (i.e., seizures were under control with or without medications) and 1 patient who had unsuccessful surgery. In the successful cases, we assumed that the resected region contained the EZ and found that the EVC rank evolution of the electrodes within the resected region had a distinct "arc" signature, i.e., the EZ ranks first rose together shortly after seizure onset and then fell later during seizure.
Tight omega(loglogn)-bound on the time for parallel RAMs to compute nondegenerated Boolean functions
Energy Technology Data Exchange (ETDEWEB)
Simon, H.U.
1982-10-01
A function f:(0,1)/sup n/=(0,1) is said to depend on dimension i iff there exists an input vector x such that f(x) differs from f(x/sup i/), where x/sup i/ agrees with x in every dimension except i. In this case x is said to be critical for f with respect to i. Function f is called nondegenerated iff it depends on all n dimensions. The main result of this paper is that for each nondegenerated function f:(0,1)/sup n/=(0,1) there exists an input vector x which is critical with respect to at least omega(log n) dimensions. A function achieving this bound is presented. Together with earlier results from Cook and Dwork (Proceedings, 14th ACM Symp. On Theory of Computing, 1982) and Reischuk (IBM Research Report, no.Rj 3431, 1982) it can be concluded that a parallel RAM requires at least omega(loglog n) steps to compute f. 3 references.
Directory of Open Access Journals (Sweden)
B. Raja Singh
2015-01-01
Full Text Available Pulverised coal preparation system (Coal mills is the heart of coal-fired power plants. The complex nature of a milling process, together with the complex interactions between coal quality and mill conditions, would lead to immense difficulties for obtaining an effective mathematical model of the milling process. In this paper, vertical spindle coal mills (bowl mill that are widely used in coal-fired power plants, is considered for the model development and its pulverised fuel flow rate is computed using the model. For the steady state coal mill model development, plant measurements such as air-flow rate, differential pressure across mill etc., are considered as inputs/outputs. The mathematical model is derived from analysis of energy, heat and mass balances. An Evolutionary computation technique is adopted to identify the unknown model parameters using on-line plant data. Validation results indicate that this model is accurate enough to represent the whole process of steady state coal mill dynamics. This coal mill model is being implemented on-line in a 210 MW thermal power plant and the results obtained are compared with plant data. The model is found accurate and robust that will work better in power plants for system monitoring. Therefore, the model can be used for online monitoring, fault detection, and control to improve the efficiency of combustion.
Fabien-Ouellet, Gabriel; Gloaguen, Erwan; Giroux, Bernard
2017-03-01
Full Waveform Inversion (FWI) aims at recovering the elastic parameters of the Earth by matching recordings of the ground motion with the direct solution of the wave equation. Modeling the wave propagation for realistic scenarios is computationally intensive, which limits the applicability of FWI. The current hardware evolution brings increasing parallel computing power that can speed up the computations in FWI. However, to take advantage of the diversity of parallel architectures presently available, new programming approaches are required. In this work, we explore the use of OpenCL to develop a portable code that can take advantage of the many parallel processor architectures now available. We present a program called SeisCL for 2D and 3D viscoelastic FWI in the time domain. The code computes the forward and adjoint wavefields using finite-difference and outputs the gradient of the misfit function given by the adjoint state method. To demonstrate the code portability on different architectures, the performance of SeisCL is tested on three different devices: Intel CPUs, NVidia GPUs and Intel Xeon PHI. Results show that the use of GPUs with OpenCL can speed up the computations by nearly two orders of magnitudes over a single threaded application on the CPU. Although OpenCL allows code portability, we show that some device-specific optimization is still required to get the best performance out of a specific architecture. Using OpenCL in conjunction with MPI allows the domain decomposition of large models on several devices located on different nodes of a cluster. For large enough models, the speedup of the domain decomposition varies quasi-linearly with the number of devices. Finally, we investigate two different approaches to compute the gradient by the adjoint state method and show the significant advantages of using OpenCL for FWI.
Computer simulation of time-resolved optical imaging of objects hidden in turbid media
Michielsen, K.; Raedt, H. De; Przeslawski, J.; Garcia, N.
1998-01-01
We review research on time-resolved optical imaging of objects hidden in strongly scattering media, with emphasis on the application to breast cancer detection. A method is presented to simulate the propagation of light in turbid media. Based on a numerical algorithm to solve the time-dependent diff
Chu, Hui-Chun; Yang, Kai-Hsiang; Chen, Jing-Hong
2015-01-01
Concept maps have been recognized as an effective tool for students to organize their knowledge; however, in history courses, it is important for students to learn and organize historical events according to the time of their occurrence. Therefore, in this study, a time sequence-oriented concept map approach is proposed for developing a game-based…
Chu, Hui-Chun; Yang, Kai-Hsiang; Chen, Jing-Hong
2015-01-01
Concept maps have been recognized as an effective tool for students to organize their knowledge; however, in history courses, it is important for students to learn and organize historical events according to the time of their occurrence. Therefore, in this study, a time sequence-oriented concept map approach is proposed for developing a game-based…
Energy Technology Data Exchange (ETDEWEB)
Son, Soon Yong [Wonkwang Health Science University, Iksan (Korea, Republic of); Choi, Kwan Woo [Asan Medical Center, Seoul (Korea, Republic of); Jeong, Hoi Woun [Baekseok Culture University College, Cheonan (Korea, Republic of); Jang, Seo Goo [Soonchunhyang University, Asan (Korea, Republic of); Jung, Jae Young [Sanggye Paik Hospital, Seoul (Korea, Republic of); Yun, Jung Soo [Samsung Medical Center, Seoul (Korea, Republic of); Kim, Ki Won [Kyung Hee University Hospital at Gang-dong, Seoul (Korea, Republic of); Lee, Young Ah; Son, Jin Hyun; Min, Jung Whan [Shingu University College, Sungnam (Korea, Republic of)
2016-03-15
The objective of this study was to analyze the factors influencing integrated bolus peak timing in contrast- enhanced computed tomographic angiography (CTA) and to determine a method of calculating personal peak time. The optimal time was calculated by performing multiple linear regression analysis, after finding the influence factors through correlation analysis between integrated peak time of contrast medium and personal measured value by monitoring CTA scans. The radiation exposure dose in CTA was 716.53 mGy·cm and the radiation exposure dose in monitoring scan was 15.52 mGy (2 - 34 mGy). The results were statistically significant (p < .01). Regression analysis revealed, a -0.160 times decrease with a one-step increase in heart rate in male, and -0.004, -0.174, and 0.006 times decrease with one-step in DBP, heart rate, and blood sugar, respectively, in female. In a consistency test of peak time by calculating measured peak time and peak time by using the regression equation, the consistency was determined to be very high for male and female. This study could prevent unnecessary dose exposure by encouraging in clinic calculation of personal integrated peak time of contrast medium prior to examination.
Simulation of a real-time brain computer interface for detecting a self-paced hitting task
DEFF Research Database (Denmark)
Hammad, Sofyan H.; Kamavuako, Ernest N.; Farina, Dario;
2016-01-01
on a single feature. However, the duration of the window length was not statistically significant (p = 0.5). CONCLUSION: Our results showed the feasibility of detecting a motor task in real time in a less restricted environment compared to environments commonly applied within invasive BCI research......OBJECTIVES: An invasive brain-computer interface (BCI) is a promising neurorehabilitation device for severely disabled patients. Although some systems have been shown to work well in restricted laboratory settings, their utility must be tested in less controlled, real-time environments. Our...
Zu, Penghe; Chen, Long; Xin, Jack
2015-09-01
The minimal speeds (c∗) of the Kolmogorov-Petrovsky-Piskunov (KPP) fronts at small diffusion (ɛ ≪ 1) in a class of time-periodic cellular flows with chaotic streamlines is investigated in this paper. The variational principle of c∗ reduces the computation to that of a principle eigenvalue problem on a periodic domain of a linear advection-diffusion operator with space-time periodic coefficients and small diffusion. To solve the advection dominated time-dependent eigenvalue problem efficiently over large time, a combination of spectral methods and finite element, as well as the associated fast solvers, are utilized to accelerate computation. In contrast to the scaling c∗ = O(ɛ 1 / 4) in steady cellular flows, a new relation c∗ = O(1) as ɛ ≪ 1 is revealed in the time-periodic cellular flows due to the presence of chaotic streamlines. Residual propagation speed emerges from the Lagrangian chaos which is quantified as a sub-diffusion process.
Energy Technology Data Exchange (ETDEWEB)
Gosnell, T.B.; Wood, R.E.; Anzelon, G.A.
1981-12-01
An electronic interface between a Digital Equipment Corporation (DEC) LSI-11 microcomputer and a LeCroy Research Systems model 3001 qVt multichannel analyzer is described in detail. This interface provides for 16-bit parallel data transfer from the memory of the analyzer to the memory of the computer. An unusual feature of the interface is a provision that allows storage of counts of logic pulses (e.g., from radiation detector discriminators) in the first 16 channels of the analyzer's memory. A further provision allows use of a LeCroy printer and display interface that is designed specifically as a companion module to the qVt analyzer.
Role of computed tomography enterography/magnetic resonance enterography: is it in prime time?
Kim, Ah Young
2012-09-01
Today, cross-sectional imaging modalities, such as computed tomography enterography (CTE) and magnetic resonance enterography (MRE), are particularly suited to evaluate small bowel diseases, especially Crohn's disease (CD). It is well known that CTE/MRE can provide excellent assessment of disease activity as well as the macroscopic features, extramural abnormalities, and complications of the small intestine in patients with CD. In general, CTE is considered as the first-line modality for the evaluation of suspected inflammatory bowel disease and for long-term assessment or follow-up of these patients. Because of the advantage of lack of radiation, MRE is being used more frequently, especially in children or young patients with CD.
High-fidelity gate operations for quantum computing beyond dephasing time limits
Souza, Alexandre M.; Sarthour, Roberto S.; Oliveira, Ivan S.; Suter, Dieter
2015-12-01
The implementation of quantum gates with fidelities that exceed the threshold for reliable quantum computing requires robust gates whose performance is not limited by the precision of the available control fields. The performance of these gates also should not be affected by the noisy environment of the quantum register. Here we use randomized benchmarking of quantum gate operations to compare the performance of different families of gates that compensate errors in the control field amplitudes and decouple the system from the environmental noise. We obtain average fidelities of up to 99.8%, which exceeds the threshold value for some quantum error correction schemes as well as the expected limit from the dephasing induced by the environment.
Building a computer-aided design capability using a standard time share operating system
Sobieszczanski, J.
1975-01-01
The paper describes how an integrated system of engineering computer programs can be built using a standard commercially available operating system. The discussion opens with an outline of the auxiliary functions that an operating system can perform for a team of engineers involved in a large and complex task. An example of a specific integrated system is provided to explain how the standard operating system features can be used to organize the programs into a simple and inexpensive but effective system. Applications to an aircraft structural design study are discussed to illustrate the use of an integrated system as a flexible and efficient engineering tool. The discussion concludes with an engineer's assessment of an operating system's capabilities and desirable improvements.
Computer vision for real-time orbital operations. Center directors discretionary fund
Vinz, F. L.; Brewster, L. L.; Thomas, L. D.
1984-01-01
Machine vision research is examined as it relates to the NASA Space Station program and its associated Orbital Maneuvering Vehicle (OMV). Initial operation of OMV for orbital assembly, docking, and servicing are manually controlled from the ground by means of an on board TV camera. These orbital operations may be accomplished autonomously by machine vision techniques which use the TV camera as a sensing device. Classical machine vision techniques are described. An alternate method is developed and described which employs a syntactic pattern recognition scheme. It has the potential for substantial reduction of computing and data storage requirements in comparison to the Two-Dimensional Fast Fourier Transform (2D FFT) image analysis. The method embodies powerful heuristic pattern recognition capability by identifying image shapes such as elongation, symmetry, number of appendages, and the relative length of appendages.
Qin, Ke; Oommen, B John
2009-11-01
The Adachi neural network (AdNN) is a fascinating neural network (NN) which has been shown to possess chaotic properties, and to also demonstrate associative memory (AM) and pattern recognition (PR) characteristics. Variants of the AdNN have also been used to obtain other PR phenomena, and even blurring. An unsurmountable problem associated with the AdNN and the variants referred to above is that all of them require a quadratic number of computations. This is essentially because the NNs in each case are completely connected graphs. In this paper, we consider how the computations can be significantly reduced by merely using a linear number of computations. To achieves this, we extract from the original completely connected graph one of its spanning trees. We then address the problem of computing the weights for this spanning tree. This is done in such a manner that the modified tree-based NN has approximately the same input-output characteristics, and thus the new weights are themselves calculated using a gradient-based algorithm. By a detailed experimental analysis, we show that the new linear-time AdNN-like network possesses chaotic and PR properties for different settings. As far as we know, such a tree-based AdNN has not been reported, and the results given here are novel.
Mendez, E.; Widdowson, M.; Brauner, S.; Chapelle, F.; Casey, C.; ,
2004-01-01
This paper describes the development and application of a modeling system called Natural Attenuation Software (NAS). NAS was designed as a screening tool to estimate times of remediation (TORs), associated with monitored natural attenuation (MNA), to lower groundwater contaminant concentrations to regulatory limits. Natural attenuation processes that NAS models include advection, dispersion, sorption, biodegradation, and non-aqueous phase liquid (NAPL) dissolution. This paper discusses the three main interactive components of NAS: 1) estimation of the target source concentration required for a plume extent to contract to regulatory limits, 2) estimation of the time required for NAFL contaminants in the source area to attenuate to a predetermined target source concentration, and 3) estimation of the time required for a plume extent to contract to regulatory limits after source reduction. The model's capability is illustrated by results from a case study at a MNA site, where NAS time of remediation estimates compared well with observed monitoring data over multiple years.
Yan, Su; Arslanbekov, Robert R; Kolobov, Vladimir I; Jin, Jian-Ming
2016-01-01
A discontinuous Galerkin time-domain (DGTD) method based on dynamically adaptive Cartesian meshes (ACM) is developed for a full-wave analysis of electromagnetic fields in dispersive media. Hierarchical Cartesian grids offer simplicity close to that of structured grids and the flexibility of unstructured grids while being highly suited for adaptive mesh refinement (AMR). The developed DGTD-ACM achieves a desired accuracy by refining non-conformal meshes near material interfaces to reduce stair-casing errors without sacrificing the high efficiency afforded with uniform Cartesian meshes. Moreover, DGTD-ACM can dynamically refine the mesh to resolve the local variation of the fields during propagation of electromagnetic pulses. A local time-stepping scheme is adopted to alleviate the constraint on the time-step size due to the stability condition of the explicit time integration. Simulations of electromagnetic wave diffraction over conducting and dielectric cylinders and spheres demonstrate that the proposed meth...
Prediction of Target Travel During Missile Time of Flight. A Computer Simulation
1978-06-01
the second program, the subject had to decide if a gap between two covered areas was wide enough for a missile to be fired successfully. Functional...Program Display During "Time to Decide"....... . ............ 7 4. Gap Program Display ,howing Covered Areas and Dot Track ... .......... 7 5. Gap...18 17. SD of Decision Time As a Funcion of Target Speed and Relative Gap Size ....... .............................. 19 18. Mean Decision
Kazemi, M.; Afarideh, H.; Riazi, Z.
2015-11-01
The aim of this research work is to use a better parallel software structure to improve the performance of the Monte Carlo Geant4 code in proton treatment planning. The hadron therapy simulation is rewritten to parallelize the shared memory multiprocessor systems by using the Message-Passing Interface (MPI). The speedup performance of the code has been studied by using two MPI-compliant libraries including Open MPI and the MPICH2, separately. Despite the speedup, the results are almost linear for both the Open MPI and MPICH2; the latter was chosen because of its better characteristics and lower computation time. The Geant4 parameters, including the step limiter and the set cut, have been analyzed to minimize the simulation time as much as possible. For a reasonable compromise between the spatial dose distribution and the calculation time, the improvement in time reduction coefficient reaches about 157.
Pandey, Parul; Lee, Eun Kyung; Pompili, Dario
2016-11-01
Stress is one of the key factor that impacts the quality of our daily life: From the productivity and efficiency in the production processes to the ability of (civilian and military) individuals in making rational decisions. Also, stress can propagate from one individual to other working in a close proximity or toward a common goal, e.g., in a military operation or workforce. Real-time assessment of the stress of individuals alone is, however, not sufficient, as understanding its source and direction in which it propagates in a group of people is equally-if not more-important. A continuous near real-time in situ personal stress monitoring system to quantify level of stress of individuals and its direction of propagation in a team is envisioned. However, stress monitoring of an individual via his/her mobile device may not always be possible for extended periods of time due to limited battery capacity of these devices. To overcome this challenge a novel distributed mobile computing framework is proposed to organize the resources in the vicinity and form a mobile device cloud that enables offloading of computation tasks in stress detection algorithm from resource constrained devices (low residual battery, limited CPU cycles) to resource rich devices. Our framework also supports computing parallelization and workflows, defining how the data and tasks divided/assigned among the entities of the framework are designed. The direction of propagation and magnitude of influence of stress in a group of individuals are studied by applying real-time, in situ analysis of Granger Causality. Tangible benefits (in terms of energy expenditure and execution time) of the proposed framework in comparison to a centralized framework are presented via thorough simulations and real experiments.
Moore, Candace R; Johnson, Logan S; Kwak, Il-Youp; Livny, Miron; Broman, Karl W; Spalding, Edgar P
2013-11-01
Automated image acquisition, a custom analysis algorithm, and a distributed computing resource were used to add time as a third dimension to a quantitative trait locus (QTL) map for plant root gravitropism, a model growth response to an environmental cue. Digital images of Arabidopsis thaliana seedling roots from two independently reared sets of 162 recombinant inbred lines (RILs) and one set of 92 near isogenic lines (NILs) derived from a Cape Verde Islands (Cvi) × Landsberg erecta (Ler) cross were collected automatically every 2 min for 8 hr following induction of gravitropism by 90° reorientation of the sample. High-throughput computing (HTC) was used to measure root tip angle in each of the 1.1 million images acquired and perform statistical regression of tip angle against the genotype at each of the 234 RIL or 102 NIL DNA markers independently at each time point using a standard stepwise procedure. Time-dependent QTL were detected on chromosomes 1, 3, and 4 by this mapping method and by an approach developed to treat the phenotype time course as a function-valued trait. The QTL on chromosome 4 was earliest, appearing at 0.5 hr and remaining significant for 5 hr, while the QTL on chromosome 1 appeared at 3 hr and thereafter remained significant. The Cvi allele generally had a negative effect of 2.6-4.0%. Heritability due to the QTL approached 25%. This study shows how computer vision and statistical genetic analysis by HTC can characterize the developmental timing of genetic architectures.
Andrade, Xavier; Alberdi-Rodriguez, Joseba; Strubbe, David A.; Oliveira, Micael J. T.; Nogueira, Fernando; Castro, Alberto; Muguerza, Javier; Arruabarrena, Agustin; Louie, Steven G.; Aspuru-Guzik, Alán; Rubio, Angel; Marques, Miguel A. L.
2012-06-01
Octopus is a general-purpose density-functional theory (DFT) code, with a particular emphasis on the time-dependent version of DFT (TDDFT). In this paper we present the ongoing efforts to achieve the parallelization of octopus. We focus on the real-time variant of TDDFT, where the time-dependent Kohn-Sham equations are directly propagated in time. This approach has great potential for execution in massively parallel systems such as modern supercomputers with thousands of processors and graphics processing units (GPUs). For harvesting the potential of conventional supercomputers, the main strategy is a multi-level parallelization scheme that combines the inherent scalability of real-time TDDFT with a real-space grid domain-partitioning approach. A scalable Poisson solver is critical for the efficiency of this scheme. For GPUs, we show how using blocks of Kohn-Sham states provides the required level of data parallelism and that this strategy is also applicable for code optimization on standard processors. Our results show that real-time TDDFT, as implemented in octopus, can be the method of choice for studying the excited states of large molecular systems in modern parallel architectures.
Susskind, Joel; Molnar, Gyula; Iredell, Lena; Loeb, Norman G.
2011-01-01
This paper compares recent spatial and temporal anomaly time series of OLR as observed by CERES and computed based on AIRS retrieved surface and atmospheric geophysical parameters over the 7 year time period September 2002 through February 2010. This time period is marked by a substantial decrease of OLR, on the order of +/-0.1 W/sq m/yr, averaged over the globe, and very large spatial variations of changes in OLR in the tropics, with local values ranging from -2.8 W/sq m/yr to +3.1 W/sq m/yr. Global and Tropical OLR both began to decrease significantly at the onset of a strong La Ni a in mid-2007. Late 2009 is characterized by a strong El Ni o, with a corresponding change in sign of both Tropical and Global OLR anomalies. The spatial patterns of the 7 year short term changes in AIRS and CERES OLR have a spatial correlation of 0.97 and slopes of the linear least squares fits of anomaly time series averaged over different spatial regions agree on the order of +/-0.01 W/sq m/yr. This essentially perfect agreement of OLR anomaly time series derived from observations by two different instruments, determined in totally independent and different manners, implies that both sets of results must be highly stable. This agreement also validates the anomaly time series of the AIRS derived products used to compute OLR and furthermore indicates that anomaly time series of AIRS derived products can be used to explain the factors contributing to anomaly time series of OLR.
Directory of Open Access Journals (Sweden)
Othman M. K. Alsmadi
2015-01-01
Full Text Available A robust computational technique for model order reduction (MOR of multi-time-scale discrete systems (single input single output (SISO and multi-input multioutput (MIMO is presented in this paper. This work is motivated by the singular perturbation of multi-time-scale systems where some specific dynamics may not have significant influence on the overall system behavior. The new approach is proposed using genetic algorithms (GA with the advantage of obtaining a reduced order model, maintaining the exact dominant dynamics in the reduced order, and minimizing the steady state error. The reduction process is performed by obtaining an upper triangular transformed matrix of the system state matrix defined in state space representation along with the elements of B, C, and D matrices. The GA computational procedure is based on maximizing the fitness function corresponding to the response deviation between the full and reduced order models. The proposed computational intelligence MOR method is compared to recently published work on MOR techniques where simulation results show the potential and advantages of the new approach.
Alsmadi, Othman M K; Abo-Hammour, Zaer S
2015-01-01
A robust computational technique for model order reduction (MOR) of multi-time-scale discrete systems (single input single output (SISO) and multi-input multioutput (MIMO)) is presented in this paper. This work is motivated by the singular perturbation of multi-time-scale systems where some specific dynamics may not have significant influence on the overall system behavior. The new approach is proposed using genetic algorithms (GA) with the advantage of obtaining a reduced order model, maintaining the exact dominant dynamics in the reduced order, and minimizing the steady state error. The reduction process is performed by obtaining an upper triangular transformed matrix of the system state matrix defined in state space representation along with the elements of B, C, and D matrices. The GA computational procedure is based on maximizing the fitness function corresponding to the response deviation between the full and reduced order models. The proposed computational intelligence MOR method is compared to recently published work on MOR techniques where simulation results show the potential and advantages of the new approach.
Real-time EO/IR sensor fusion on a portable computer and head-mounted display
Yue, Zhanfeng; Topiwala, Pankaj
2007-04-01
Multi-sensor platforms are widely used in surveillance video systems for both military and civilian applications. The complimentary nature of different types of sensors (e.g. EO and IR sensors) makes it possible to observe the scene under almost any condition (day/night/fog/smoke). In this paper, we propose an innovative EO/IR sensor registration and fusion algorithm which runs real-time on a portable computing unit with head-mounted display. The EO/IR sensor suite is mounted on a helmet for a dismounted soldier and the fused scene is shown in the goggle display upon the processing on a portable computing unit. The linear homography transformation between images from the two sensors is precomputed for the mid-to-far scene, which reduces the computational cost for the online calibration of the sensors. The system is implemented in a highly optimized C++ code, with MMX/SSE, and performing a real-time registration. The experimental results on real captured video show the system works very well both in speed and in performance.
Accurate Time-Dependent Traveling-Wave Tube Model Developed for Computational Bit-Error-Rate Testing
Kory, Carol L.
2001-01-01
The phenomenal growth of the satellite communications industry has created a large demand for traveling-wave tubes (TWT's) operating with unprecedented specifications requiring the design and production of many novel devices in record time. To achieve this, the TWT industry heavily relies on computational modeling. However, the TWT industry's computational modeling capabilities need to be improved because there are often discrepancies between measured TWT data and that predicted by conventional two-dimensional helical TWT interaction codes. This limits the analysis and design of novel devices or TWT's with parameters differing from what is conventionally manufactured. In addition, the inaccuracy of current computational tools limits achievable TWT performance because optimized designs require highly accurate models. To address these concerns, a fully three-dimensional, time-dependent, helical TWT interaction model was developed using the electromagnetic particle-in-cell code MAFIA (Solution of MAxwell's equations by the Finite-Integration-Algorithm). The model includes a short section of helical slow-wave circuit with excitation fed by radiofrequency input/output couplers, and an electron beam contained by periodic permanent magnet focusing. A cutaway view of several turns of the three-dimensional helical slow-wave circuit with input/output couplers is shown. This has been shown to be more accurate than conventionally used two-dimensional models. The growth of the communications industry has also imposed a demand for increased data rates for the transmission of large volumes of data. To achieve increased data rates, complex modulation and multiple access techniques are employed requiring minimum distortion of the signal as it is passed through the TWT. Thus, intersymbol interference (ISI) becomes a major consideration, as well as suspected causes such as reflections within the TWT. To experimentally investigate effects of the physical TWT on ISI would be
Observability conditions of linear time-varying systems and its computational complexity aspects
Directory of Open Access Journals (Sweden)
Starkov Konstantin E.
2002-01-01
Full Text Available We propose necessary and sufficient Observability conditions for linear time-varying systems with coefficients being time polynomials. These conditions are deduced from the Gabrielov–Khovansky theorem on multiplicity of a zero of a Noetherian function and the Wei–Norman formula for the representation of a solution of a linear time-varying system as a product of matrix exponentials. We define a Noetherian chain consisted of some finite number of usual exponentials corresponding to this system. Our results are formulated in terms of a Noetherian chain generated by these exponential functions and an upper bound of multiplicity of zero of one locally analytic function which is defined with help of the Wei–Norman formula. Relations with Observability conditions of bilinear systems are discussed. The case of two-dimensional systems is examined.
WAVE MAKING COMPUTATION IN TIME DOMAIN FOR MULTI-HULL SHIPS
Institute of Scientific and Technical Information of China (English)
LI Guo-an; YE Heng-kui
2006-01-01
A method of three-dimensional time domain Green function satisfying linear conditions at free surface and body surface boundary was employed to analyze the wave resistance and wave profile of a displacement multi-hull ship. The wave profile induced by a moving time domain point source was compared with those by a Havelock source, and satisfactory results were obtained. The panel method based on the time domain source distribution on the ship mean wetted hull surface was used to perform the wave making com- putations for mono-hull ships, catamaran and trimaran. Reasonable results were also obtained. Using the numerical method the wave profile simulations of multi-hull ships for a given Froude number were conducted.
Computationally effective solution of the inverse problem in time-of-flight spectroscopy
DEFF Research Database (Denmark)
Kamran, Faisal; Abildgaard, Otto Højager Attermann; Subash, Arman Ahamed
2015-01-01
Photon time-of-flight (PTOF) spectroscopy enables the estimation of absorption and reduced scattering coefficients of turbid media by measuring the propagation time of short light pulses through turbid medium. The present investigation provides a comparison of the assessed absorption and reduced...... scattering coefficients from PTOF measurements of intralipid 20% and India ink-based optical phantoms covering a wide range of optical properties relevant for biological tissues and dairy products. Three different models are used to obtain the optical properties by fitting to measured temporal profiles...
Computationally effective solution of the inverse problem in time-of-flight spectroscopy
DEFF Research Database (Denmark)
Kamran, Faisal; Abildgaard, Otto Højager Attermann; Subash, Arman Ahamed;
2015-01-01
Photon time-of-flight (PTOF) spectroscopy enables the estimation of absorption and reduced scattering coefficients of turbid media by measuring the propagation time of short light pulses through turbid medium. The present investigation provides a comparison of the assessed absorption and reduced...... scattering coefficients from PTOF measurements of intralipid 20% and India ink-based optical phantoms covering a wide range of optical properties relevant for biological tissues and dairy products. Three different models are used to obtain the optical properties by fitting to measured temporal profiles...
Adaptive real-time methodology for optimizing energy-efficient computing
Hsu, Chung-Hsing; Feng, Wu-Chun
2013-01-29
Dynamic voltage and frequency scaling (DVFS) is an effective way to reduce energy and power consumption in microprocessor units. Current implementations of DVFS suffer from inaccurate modeling of power requirements and usage, and from inaccurate characterization of the relationships between the applicable variables. A system and method is proposed that adjusts CPU frequency and voltage based on run-time calculations of the workload processing time, as well as a calculation of performance sensitivity with respect to CPU frequency. The system and method are processor independent, and can be applied to either an entire system as a unit, or individually to each process running on a system.
Computing the real-time Green's Functions of large Hamiltonian matrices
Iitaka, T
1996-01-01
A numerical method is developed for calculating the real time Green's functions of very large sparse Hamiltonian matrices, which exploits the numerical solution of the inhomogeneous time-dependent Schroedinger equation. The method has a clear-cut structure reflecting the most naive definition of the Green's functions, and is very suitable to parallel and vector supercomputers. The effectiveness of the method is illustrated by applying it to simple lattice models. An application of this method to condensed matter physics will be found in H. Tanaka, Phys. PRB 57, 2168 (1998).
Computing and visualizing time-varying merge trees for high-dimensional data
Energy Technology Data Exchange (ETDEWEB)
Oesterling, Patrick [Univ. of Leipzig (Germany); Heine, Christian [Univ. of Kaiserslautern (Germany); Weber, Gunther H. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Morozov, Dmitry [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Scheuermann, Gerik [Univ. of Leipzig (Germany)
2017-06-03
We introduce a new method that identifies and tracks features in arbitrary dimensions using the merge tree -- a structure for identifying topological features based on thresholding in scalar fields. This method analyzes the evolution of features of the function by tracking changes in the merge tree and relates features by matching subtrees between consecutive time steps. Using the time-varying merge tree, we present a structural visualization of the changing function that illustrates both features and their temporal evolution. We demonstrate the utility of our approach by applying it to temporal cluster analysis of high-dimensional point clouds.
Directory of Open Access Journals (Sweden)
Tatsuya Nishii
2016-02-01
Full Text Available Backgrounds. This study examines the hypothesis that four-dimensional noise reduction (4DNR with short interval times reduces noise in cardiac computed tomography (CCT using “padding” phases. Furthermore, the capability of reducing the reduction dose in CCT using this post-processing technique was assessed. Methods. Using base and quarter radiation doses for CCT (456 and 114 mAs/rot with 120 kVp, a static phantom was scanned ten times with retrospective electrocardiogram gating, and 4DNR with short interval times (50 ms was performed using a post-processing technique. Differences in the computed tomography (CT attenuation, contrast-to-noise ratio (CNR and spatial resolution with modulation transfer function in each dose image obtained with and without 4DNR were assessed by conducting a Tukey–Kramer’s test and non-inferiority test. Results. For the base dose, by using 4DNR, the CNR was improved from 1.18 ± 0.15 to 2.08 ± 0.20 (P = 0.001, while the CT attenuation and spatial resolution of the image of 4DNR did not were significantly inferior to those of reference image (P < 0.001. CNRs of the quarter-dose image in 4DNR also improved to 1.28 ± 0.11, and were not inferior to those of the non-4DNR images of the base dose (P < 0.001. Conclusions. 4DNR with short interval times significantly reduced noise. Furthermore, applying this method to CCT would have the potential of reducing the radiation dose by 75%, while maintaining a similar image noise level.
Costa Ferrer, Raquel; Serrano Rosa, Miguel Ángel; Zornoza Abad, Ana; Salvador Fernández-Montejo, Alicia
2010-11-01
The cardiovascular (CV) response to social challenge and stress is associated with the etiology of cardiovascular diseases. New ways of communication, time pressure and different types of information are common in our society. In this study, the cardiovascular response to two different tasks (open vs. closed information) was examined employing different communication channels (computer-mediated vs. face-to-face) and with different pace control (self vs. external). Our results indicate that there was a higher CV response in the computer-mediated condition, on the closed information task and in the externally paced condition. These role of these factors should be considered when studying the consequences of social stress and their underlying mechanisms.
Directory of Open Access Journals (Sweden)
Anthony Gasperin
2013-09-01
Full Text Available To study groups with small Dehn's function, Olshanskii and Sapir developed a new invariant of bipartite chords diagrams and applied it to hub-free realization of S-machines. In this paper we consider this new invariant together with groups constructed from S-machines containing the hub relation. The idea is to study the links between the topology of the asymptotic cones and polynomial time computations. Indeed it is known that the topology of such metric space depends on diagrams without hubs that do not correspond to the computations of the considered S-machine. This work gives sufficient conditions that avoid this misbehaviour, but as we shall see the method has a significant drawback.
Teufel, Julian; Bardins, S; Spiegel, Rainer; Kremmyda, O; Schneider, E; Strupp, M; Kalla, R
2016-01-04
Patients with downbeat nystagmus syndrome suffer from oscillopsia, which leads to an unstable visual perception and therefore impaired visual acuity. The aim of this study was to use real-time computer-based visual feedback to compensate for the destabilizing slow phase eye movements. The patients were sitting in front of a computer screen with the head fixed on a chin rest. The eye movements were recorded by an eye tracking system (EyeSeeCam®). We tested the visual acuity with a fixed Landolt C (static) and during real-time feedback driven condition (dynamic) in gaze straight ahead and (20°) sideward gaze. In the dynamic condition, the Landolt C moved according to the slow phase eye velocity of the downbeat nystagmus. The Shapiro-Wilk test was used to test for normal distribution and one-way ANOVA for comparison. Ten patients with downbeat nystagmus were included in the study. Median age was 76 years and the median duration of symptoms was 6.3 years (SD +/- 3.1y). The mean slow phase velocity was moderate during gaze straight ahead (1.44°/s, SD +/- 1.18°/s) and increased significantly in sideward gaze (mean left 3.36°/s; right 3.58°/s). In gaze straight ahead, we found no difference between the static and feedback driven condition. In sideward gaze, visual acuity improved in five out of ten subjects during the feedback-driven condition (p = 0.043). This study provides proof of concept that non-invasive real-time computer-based visual feedback compensates for the SPV in DBN. Therefore, real-time visual feedback may be a promising aid for patients suffering from oscillopsia and impaired text reading on screen. Recent technological advances in the area of virtual reality displays might soon render this approach feasible in fully mobile settings.
Coe, Ryan L; Seibel, Eric J
2012-12-01
We present a method for modeling image formation in optical projection tomographic microscopy (OPTM) using high numerical aperture (NA) condensers and objectives. Similar to techniques used in computed tomography, OPTM produces three-dimensional, reconstructed images of single cells from two-dimensional projections. The model is capable of simulating axial scanning of a microscope objective to produce projections, which are reconstructed using filtered backprojection. Simulation of optical scattering in transmission optical microscopy is designed to analyze all aspects of OPTM image formation, such as degree of specimen staining, refractive-index matching, and objective scanning. In this preliminary work, a set of simulations is performed to examine the effect of changing the condenser NA, objective scan range, and complex refractive index on the final reconstruction of a microshell with an outer radius of 1.5 μm and an inner radius of 0.9 μm. The model lays the groundwork for optimizing OPTM imaging parameters and triaging efforts to further improve the overall system design. As the model is expanded in the future, it will be used to simulate a more realistic cell, which could lead to even greater impact.
Measurement of image quality according to the time of computed radiography system
Energy Technology Data Exchange (ETDEWEB)
Son, Soon Yong; Choi, Kwan Woo [Dept. of Radiology, Asan Medical Center, Seoul (Korea, Republic of); Kim, Jung Min [Dept. of College of Health Science, Radiologic Science, Korea University, Seoul (Korea, Republic of); and others
2015-12-15
The regular quality assurance (RQA) of X-ray images is essential for maintaining a high accuracy of diagnosis. This study was to evaluate the modulation transfer function (MTF), the noise power spectrum (NPS), and the detective quantum efficiency (DQE) of a computed radiography (CR) system for various periods of use from 2006 to 2015. We measured the pre-sampling MTF using the edge method and RQA 5 based on commission standard international electro-technical commission (IEC). The spatial frequencies corresponding to the 50% MTF for the CR systems in 2006, 2009, 2012 and 2015 were 1.54, 1.14, 1.12, and 1.38 mm-1, respectively and the10 % MTF for 2006, 2009, 2012, and 2015 were 2.68, 2.44, 2 .44, and 2.46mm-1, respectively. In the NPS results, the CR systems showed the best noise distribution in 2006, and with the quality of distributions in the order of 2015, 2009, and 2012. At peak DQE and DQE at 1 mm-1, the CR systems showed the best efficiency in 2006, and showed better efficiency in order of 2015, 2009, and 2012. Because the eraser lamp in the CR systems was replaced, the image quality in 2015 was superior to those in 2009 and 2012. This study can be incorporated into used in clinical QA requiring performance and evaluation of the performance of the CR systems.
Screening for lung cancer: time for large-scale screening by chest computed tomography.
Shlomi, Dekel; Ben-Avi, Ronny; Balmor, Gingy Ronen; Onn, Amir; Peled, Nir
2014-07-01
Lung cancer is the leading cause of cancer death worldwide. Age and smoking are the primary risk factors for lung cancer. Treatment based on surgical removal in the early stages of the disease results in better survival. Screening programmes for early detection that used chest radiography and sputum cytology failed to attain reduction of lung cancer mortality. Screening by low-dose computed tomography (CT) demonstrated high rates of early-stage lung cancer detection in a high-risk population. Nevertheless, no mortality advantage was manifested in small randomised control trials. A large randomised control trial in the U.S.A., the National Lung Screening Trial (NLST), showed a significant relative reduction of 20% in lung cancer mortality and 6.7% reduction in total mortality, yet no reduction was evidenced in the late-stage prevalence. Screening for lung cancer by low-dose CT reveals a high level of false-positive lesions, which necessitates further noninvasive and invasive evaluations. Based primarily on the NLST eligible criteria, new guidelines have recently been developed by major relevant organisations. The overall recommendation coming out of this collective work calls for lung cancer screening by low-dose CT to be performed in medical centres manned by specialised multidisciplinary teams, as well as for a mandatory, pre-screening, comprehensive discussion with the patient about the risks and advantages involved in the process. Lung cancer screening is on the threshold of a new era, with ever more questions still left open to challenge future studies.
20 CFR 725.311 - Communications with respect to claims; time computations.
2010-04-01
... legal holiday. “Legal holiday” includes New Year's Day, Birthday of Martin Luther King, Jr., Washington... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Communications with respect to claims; time... TITLE IV OF THE FEDERAL MINE SAFETY AND HEALTH ACT, AS AMENDED Filing of Claims § 725.311...
Real-Time Assessment of Problem-Solving of Physics Students Using Computer-Based Technology
Gok, Tolga
2012-01-01
The change in students' problem solving ability in upper-level course through the application of a technological interactive environment--Tablet PC running InkSurvey--was investigated in present study. Tablet PC/InkSurvey interactive technology allowing the instructor to receive real-time formative assessment as the class works through the problem…
Energy Technology Data Exchange (ETDEWEB)
Kasperl, St.; Franz, M. [Fraunhofer Development Center for X-ray Technology EZRT, a cooperative dept. of the Fraunhofer Institutes IZFP Saarbrucken and IIS Erlangen, Furth (Germany)
2007-07-01
This article deals with the enhancement of accuracy in CT by just-in-time correction of artefacts (beam hardening, scattered radiation) caused by the interaction of X-rays with matter. The so called EAR method needs for simulation a registration of the object. Therefore the article presents two different registration concepts. (authors)
Comparison of Methods for Demonstrating Passage of Time When Using Computer-Based Video Prompting
Mechling, Linda C.; Bryant, Kathryn J.; Spencer, Galen P.; Ayres, Kevin M.
2015-01-01
Two different video-based procedures for presenting the passage of time (how long a step lasts) were examined. The two procedures were presented within the framework of video prompting to promote independent multi-step task completion across four young adults with moderate intellectual disability. The two procedures demonstrating passage of the…
Using Biometric Measurement in Real-Time as a Sympathetic System in Computer Games
Charij, Stephanie; Oikonomou, Andreas
2013-01-01
With the increasing potential for gaming hardware and peripherals to support biometrics, their application within the games industry for software and design should be considered. This paper assesses the ability to use a form of biometric measurement, heart rate, in real-time to improve the challenge and enjoyment of a game by catering it to…
Kale, Nimish; Lee, Jaeseong; Lotfian, Reza; Jafari, Roozbeh
2012-10-01
Daily living activity monitoring is important for early detection of the onset of many diseases and for improving quality of life especially in elderly. A wireless wearable network of inertial sensor nodes can be used to observe daily motions. Continuous stream of data generated by these sensor networks can be used to recognize the movements of interest. Dynamic Time Warping (DTW) is a widely used signal processing method for time-series pattern matching because of its robustness to variations in time and speed as opposed to other template matching methods. Despite this flexibility, for the application of activity recognition, DTW can only find the similarity between the template of a movement and the incoming samples, when the location and orientation of the sensor remains unchanged. Due to this restriction, small sensor misplacements can lead to a decrease in the classification accuracy. In this work, we adopt DTW distance as a feature for real-time detection of human daily activities like sit to stand in the presence of sensor misplacement. To measure this performance of DTW, we need to create a large number of sensor configurations while the sensors are rotated or misplaced. Creating a large number of closely spaced sensors is impractical. To address this problem, we use the marker based optical motion capture system and generate simulated inertial sensor data for different locations and orientations on the body. We study the performance of the DTW under these conditions to determine the worst-case sensor location variations that the algorithm can accommodate.
High Performance Computing (HPC) for Real-Time Course of Action (COA) Analysis
2008-01-01
architectures including Beowulf clusters . Installing FSS is discussed in the Installation Guide FSS supports a force on force simulation which...workstations, clusters , or combinations of architectures. FSS is a discrete event simulation meaning there is no ‘time step’ to the simulation
Pistorius, M.; Stolte, J.
2012-01-01
We present a new numerical method to price vanilla options quickly in time-changed Brownian motion models. The method is based on rational function approximations of the Black-Scholes formula. Detailed numerical results are given for a number of widely used models. In particular, we use the variance
Rossi, R.; Tarim, S.A.; Hnich, B.; Prestwich, S.
2010-01-01
In this paper we address the general multi-period production/inventory problem with non-stationary stochastic demand and supplier lead-time under service level constraints. A replenishment cycle policy (Rn,Sn) is modeled, where Rn is the nth replenishment cycle length and Sn is the respective order-
3D time-dependent flow computations using a molecular stress function model with constraint release
DEFF Research Database (Denmark)
Rasmussen, Henrik Koblitz
2002-01-01
The numerical simulation of time dependent viscoelastic flow (in three dimensions) is of interest in connection with a variety of polymer processing operations. The application of the numerical simulation techniques is in the analysis and design of polymer processing problems. This is operations,...
Real-time Evolutionary Computation in the Control of Mobile Cyber-Physics System
Directory of Open Access Journals (Sweden)
Rogachev Gennady
2017-01-01
Full Text Available The procedure of the time-optimal controller synthesis in the mobile cyber-physical system control problems in the form of the Dubins car is considered. In the decision quality assessment, the total time spent on executing the task of transferring an object to a given point is taken into account, including the time for the control program calculation and the time of the object movement to the target. Both online and offline variants of the production model numerical determination of the controller as a system of rules “condition” – “action” are considered. As an online variant, it is suggested to use an anytime genetic algorithm. A comparison between online and offline methods of calculating a rule system for the regulator performance is made. The advantages of the online method in solving the problem of transferring the mobile cyber physical system in a homogeneous environment and the advantages of an offline method in a heterogeneous environment are demonstrated.
Postmortem time estimation using body temperature and a finite-element computer model
Hartog, E.A. den; Lotens, W.A.
2004-01-01
In the Netherlands most murder victims are found 2-24 h after the crime. During this period, body temperature decrease is the most reliable method to estimate the postmortem time (PMT). Recently, two murder cases were analysed in which currently available methods did not provide a su.ciently reliabl
Wang, P.; Li, P.
1998-01-01
A high-resolution numerical study on parallel systems is reported on three-dimensional, time-dependent, thermal convective flows. A parallel implentation on the finite volume method with a multigrid scheme is discussed, and a parallel visualization systemm is developed on distributed systems for visualizing the flow.
Video and Computer Games in the '90s: Children's Time Commitment and Game Preference.
Buchman, Debra D.; Funk, Jeanne B.
1996-01-01
Examined electronic game-playing habits of 900 children. Found that time commitment to game-playing decreased from fourth to eighth grade. Boys played more than girls. Preference for general entertainment games increased across grades while educational games preference decreased. Violent game popularity remained consistent; fantasy violence was…
Witt, Annette; Palmigiano, Agostina; Neef, Andreas; El Hady, Ahmed; Wolf, Fred; Battaglia, Demian
2013-01-01
Dynamic oscillatory coherence is believed to play a central role in flexible communication between brain circuits. To test this communication-through-coherence hypothesis, experimental protocols that allow a reliable control of phase-relations between neuronal populations are needed. In this modeling study, we explore the potential of closed-loop optogenetic stimulation for the control of functional interactions mediated by oscillatory coherence. The theory of non-linear oscillators predicts that the efficacy of local stimulation will depend not only on the stimulation intensity but also on its timing relative to the ongoing oscillation in the target area. Induced phase-shifts are expected to be stronger when the stimulation is applied within specific narrow phase intervals. Conversely, stimulations with the same or even stronger intensity are less effective when timed randomly. Stimulation should thus be properly phased with respect to ongoing oscillations (in order to optimally perturb them) and the timing of the stimulation onset must be determined by a real-time phase analysis of simultaneously recorded local field potentials (LFPs). Here, we introduce an electrophysiologically calibrated model of Channelrhodopsin 2 (ChR2)-induced photocurrents, based on fits holding over two decades of light intensity. Through simulations of a neural population which undergoes coherent gamma oscillations—either spontaneously or as an effect of continuous optogenetic driving—we show that precisely-timed photostimulation pulses can be used to shift the phase of oscillation, even at transduction rates smaller than 25%. We consider then a canonic circuit with two inter-connected neural populations oscillating with gamma frequency in a phase-locked manner. We demonstrate that photostimulation pulses applied locally to a single population can induce, if precisely phased, a lasting reorganization of the phase-locking pattern and hence modify functional interactions between the
Directory of Open Access Journals (Sweden)
Annette eWitt
2013-04-01
Full Text Available Dynamic oscillatory coherence is believed to play a central role in flexible communication between brain circuits. To test this communication-through-coherence hypothesis, experimental protocols that allow a reliable control of phase-relations between neuronal populations are needed. In this modeling study, we explore the potential of closed-loop optogenetic stimulation for the control of functional interactions mediated by oscillatory coherence. The theory of nonlinear oscillators predicts that the efficacy of local stimulation will depend not only on the stimulation intensity but also on its timing relative to the ongoing oscillation in the target area. Induced phase-shifts are expected to be stronger when the stimulation is applied within specific narrow phase intervals. Conversely, stimulations with the same or even stronger intensity are less effective when timed randomly. Stimulation should thus be properly phased with respect to ongoing oscillations (in order to optimally perturb them and the timing of the stimulation onset must be determined by a real-time phase analysis of simultaneously recorded local field potentials (LFPs.Here, we introduce an electrophysiologically calibrated model of Channelrhodopsin 2 (ChR2-induced photocurrents, based on fits holding over two decades of light intensity. Through simulations of a neural population which undergoes coherent gamma oscillations —either spontaneously or as an effect of continuous optogenetic driving—, we show that precisely-timed photostimulation pulses can be used to shift phase, even at transduction rates smaller than 25%. We consider then a canonic circuit with two inter-connected neural populations oscillating with gamma frequency in a phase-locked manner. We demonstrate that photostimulation pulses applied locally to a single population can induce, if precisely phased, a lasting reorganization of the phase-locking pattern and hence modify functional interactions between the two
Directory of Open Access Journals (Sweden)
K. Ide
2002-01-01
Full Text Available In this paper we develop analytical and numerical methods for finding special hyperbolic trajectories that govern geometry of Lagrangian structures in time-dependent vector fields. The vector fields (or velocity fields may have arbitrary time dependence and be realized only as data sets over finite time intervals, where space and time are discretized. While the notion of a hyperbolic trajectory is central to dynamical systems theory, much of the theoretical developments for Lagrangian transport proceed under the assumption that such a special hyperbolic trajectory exists. This brings in new mathematical issues that must be addressed in order for Lagrangian transport theory to be applicable in practice, i.e. how to determine whether or not such a trajectory exists and, if it does exist, how to identify it in a sequence of instantaneous velocity fields. We address these issues by developing the notion of a distinguished hyperbolic trajectory (DHT. We develop an existence criteria for certain classes of DHTs in general time-dependent velocity fields, based on the time evolution of Eulerian structures that are observed in individual instantaneous fields over the entire time interval of the data set. We demonstrate the concept of DHTs in inhomogeneous (or "forced" time-dependent linear systems and develop a theory and analytical formula for computing DHTs. Throughout this work the notion of linearization is very important. This is not surprising since hyperbolicity is a "linearized" notion. To extend the analytical formula to more general nonlinear time-dependent velocity fields, we develop a series of coordinate transforms including a type of linearization that is not typically used in dynamical systems theory. We refer to it as Eulerian linearization, which is related to the frame independence of DHTs, as opposed to the Lagrangian linearization, which is typical in dynamical systems theory, which is used in the computation of Lyapunov exponents. We
Institute of Scientific and Technical Information of China (English)
无
2002-01-01
This paper discussed the current of works on computerisation of all problems related to mining subsidence, including the time factor,carried out in the Division of Mining Geodesy of Technical University of Silesia, Poland. First, the formulas implemented in the programs were presented. These formulas considerably increase the description accuracy of final deformations by taking into uncaved strip along extraction rib (extraction margin). They also improve the deformation description of areas located far from the extraction place. Then, the research results aiming to improving the description of deformation with time were introduced. Finally, the Windows-based version of the program for the creation of mining-geological opinions were presented in the form accepted by Mining Offices of Poland.
Sublinear Time Algorithm for PageRank Computations and Related Applications
Borgs, Christian; Chayes, Jennifer; Teng, Shang-Hua
2012-01-01
In a network, identifying all vertices whose PageRank is more than a given threshold value $\\Delta$ is a basic problem that has arisen in Web and social network analyses. In this paper, we develop a nearly optimal, sublinear time, randomized algorithm for a close variant of this problem. When given a network \\graph, a threshold value $\\Delta$, and a positive constant $c>1$, with probability $1-o(1)$, our algorithm will return a subset $S\\subseteq V$ with the property that $S$ contains all vertices of PageRank at least $\\Delta$ and no vertex with PageRank less than $\\Delta/c$. The running time of our algorithm is always $\\tilde{O}(\\frac{n}{\\Delta})$. In addition, our algorithm can be efficiently implemented in various network access models including the Jump and Crawl query model recently studied by \\cite{brautbar_kearns10}, making it suitable for dealing with large social and information networks. As part of our analysis, we show that any algorithm for solving this problem must have expected time complexity o...
Thong, Patricia S. P.; Tandjung, Stephanus S.; Movania, Muhammad Mobeen; Chiew, Wei-Ming; Olivo, Malini; Bhuvaneswari, Ramaswamy; Seah, Hock-Soon; Lin, Feng; Qian, Kemao; Soo, Khee-Chee
2012-05-01
Oral lesions are conventionally diagnosed using white light endoscopy and histopathology. This can pose a challenge because the lesions may be difficult to visualise under white light illumination. Confocal laser endomicroscopy can be used for confocal fluorescence imaging of surface and subsurface cellular and tissue structures. To move toward real-time "virtual" biopsy of oral lesions, we interfaced an embedded computing system to a confocal laser endomicroscope to achieve a prototype three-dimensional (3-D) fluorescence imaging system. A field-programmable gated array computing platform was programmed to enable synchronization of cross-sectional image grabbing and Z-depth scanning, automate the acquisition of confocal image stacks and perform volume rendering. Fluorescence imaging of the human and murine oral cavities was carried out using the fluorescent dyes fluorescein sodium and hypericin. Volume rendering of cellular and tissue structures from the oral cavity demonstrate the potential of the system for 3-D fluorescence visualization of the oral cavity in real-time. We aim toward achieving a real-time virtual biopsy technique that can complement current diagnostic techniques and aid in targeted biopsy for better clinical outcomes.
Pramod Kumar, K.; Mahendra, P.; Ramakrishna rReddy, V.; Tirupathi, T.; Akilan, A.; Usha Devi, R.; Anuradha, R.; Ravi, N.; Solanki, S. S.; Achary, K. K.; Satish, A. L.; Anshu, C.
2014-11-01
In the last decade, the remote sensing community has observed a significant growth in number of satellites, sensors and their resolutions, thereby increasing the volume of data to be processed each day. Satellite data processing is a complex and time consuming activity. It consists of various tasks, such as decode, decrypt, decompress, radiometric normalization, stagger corrections, ephemeris data processing for geometric corrections etc., and finally writing of the product in the form of an image file. Each task in the processing chain is sequential in nature and has different computing needs. Conventionally the processes are cascaded in a well organized workflow to produce the data products, which are executed on general purpose high-end servers / workstations in an offline mode. Hence, these systems are considered to be ineffective for real-time applications that require quick response and just-intime decision making such as disaster management, home land security and so on. This paper discusses anovel approach to processthe data online (as the data is being acquired) using a heterogeneous computing platform namely XSTREAM which has COTS hardware of CPUs, GPUs and FPGA. This paper focuses on the process architecture, re-engineering aspects and mapping of tasks to the right computing devicewithin the XSTREAM system, which makes it an ideal cost-effective platform for acquiring, processing satellite payload data in real-time and displaying the products in original resolution for quick response. The system has been tested for IRS CARTOSAT and RESOURCESAT series of satellites which have maximum data downlink speed of 210 Mbps.
Parra, M. A.; Santiago, J. L.; Martín, F.; Martilli, A.; Santamaría, J. M.
2010-06-01
The representativeness of point measurements in urban areas is limited due to the strong heterogeneity of the atmospheric flows in cities. To get information on air quality in the gaps between measurement points, and have a 3D field of pollutant concentration, Computational Fluid Dynamic (CFD) models can be used. However, unsteady simulations during time periods of the order of months, often required for regulatory purposes, are not possible for computational reasons. The main objective of this study is to develop a methodology to evaluate the air quality in a real urban area during large time periods by means of steady CFD simulations. One steady simulation for each inlet wind direction was performed and factors like the number of cars inside each street, the length of streets and the wind speed and direction were taken into account to compute the pollutant concentration. This approach is only valid in winter time when the pollutant concentrations are less affected by atmospheric chemistry. A model based on the steady-state Reynolds-Averaged Navier-Stokes equations (RANS) and standard k-ɛ turbulence model was used to simulate a set of 16 different inlet wind directions over a real urban area (downtown Pamplona, Spain). The temporal series of NO x and PM 10 and the spatial differences in pollutant concentration of NO 2 and BTEX obtained were in agreement with experimental data. Inside urban canopy, an important influence of urban boundary layer dynamics on the pollutant concentration patterns was observed. Large concentration differences between different zones of the same square were found. This showed that concentration levels measured by an automatic monitoring station depend on its location in the street or square, and a modelling methodology like this is useful to complement the experimental information. On the other hand, this methodology can also be applied to evaluate abatement strategies by redistributing traffic emissions.
Mahmood, T.; Shahzad, A.; Iqbal, Z.; Ahmed, J.; Khan, M.
A study is presented for the flow and heat transfer of Sisko fluid model over an unsteady stretching sheet in the presence of uniform magnetic field. While taking newly developed similarity transformations, the governing time dependent partial differential equations are reduced to nonlinear ordinary differential equations. Numerical solutions of the reduced nonlinear differential equations are found by employing Shooting method. The influence of physical parameters of interest on the velocity and temperature profiles are highlighted graphically and examined in detail. Moreover, the skin friction coefficient and Nusselt number are tabulated against influential parameters. Skin friction coefficient increases with unsteadiness parameter, magnetic field and suction parameter.
Time-Delay Artificial Neural Network Computing Models for Predicting Shelf Life of Processed Cheese
Directory of Open Access Journals (Sweden)
Sumit Goyal
2012-04-01
Full Text Available This paper presents the capability of Time–delay artificial neural network models for predicting shelf life of processed cheese. Datasets were divided into two subsets (30 for training and 6 for validation. Models with single and multi layers were developed and compared with each other. Mean Square Error, Root Mean Square Error, Coefficient of Determination and Nash -
Sutcliffo Coefficient were used as performance evaluators, Time- delay model predicted the shelf life of processed cheese as 28.25 days, which is very close to experimental shelf life of 30 days.
Computing the Lyapunov spectrum of a dynamical system from an observed time series
Brown, Reggie; Bryant, Paul; Abarbanel, Henry D. I.
1991-01-01
The paper examines the problem of accurately determining, from an observed time series, the Liapunov exponents for the dynamical system generating the data. It is shown that, even with very large data sets, it is clearly advantageous to utilize local neighborhood-to-neighborhood mappings with higher-order Taylor series rather than just local linear maps. This procedure is demonstrated on the Henon and Ikeda maps of the plane itself, the Lorenz system of three ordinary differential equations, and the Mackey-Glass delay differential equation.