WorldWideScience

Sample records for total computational time

  1. Television viewing, computer use and total screen time in Canadian youth.

    Science.gov (United States)

    Mark, Amy E; Boyce, William F; Janssen, Ian

    2006-11-01

    Research has linked excessive television viewing and computer use in children and adolescents to a variety of health and social problems. Current recommendations are that screen time in children and adolescents should be limited to no more than 2 h per day. To determine the percentage of Canadian youth meeting the screen time guideline recommendations. The representative study sample consisted of 6942 Canadian youth in grades 6 to 10 who participated in the 2001/2002 World Health Organization Health Behaviour in School-Aged Children survey. Only 41% of girls and 34% of boys in grades 6 to 10 watched 2 h or less of television per day. Once the time of leisure computer use was included and total daily screen time was examined, only 18% of girls and 14% of boys met the guidelines. The prevalence of those meeting the screen time guidelines was higher in girls than boys. Fewer than 20% of Canadian youth in grades 6 to 10 met the total screen time guidelines, suggesting that increased public health interventions are needed to reduce the number of leisure time hours that Canadian youth spend watching television and using the computer.

  2. A new system of computer-assisted navigation leading to reduction in operating time in uncemented total hip replacement in a matched population.

    Science.gov (United States)

    Chaudhry, Fouad A; Ismail, Sanaa Z; Davis, Edward T

    2018-05-01

    Computer-assisted navigation techniques are used to optimise component placement and alignment in total hip replacement. It has developed in the last 10 years but despite its advantages only 0.3% of all total hip replacements in England and Wales are done using computer navigation. One of the reasons for this is that computer-assisted technology increases operative time. A new method of pelvic registration has been developed without the need to register the anterior pelvic plane (BrainLab hip 6.0) which has shown to improve the accuracy of THR. The purpose of this study was to find out if the new method reduces the operating time. This was a retrospective analysis of comparing operating time in computer navigated primary uncemented total hip replacement using two methods of registration. Group 1 included 128 cases that were performed using BrainLab versions 2.1-5.1. This version relied on the acquisition of the anterior pelvic plane for registration. Group 2 included 128 cases that were performed using the newest navigation software, BrainLab hip 6.0 (registration possible with the patient in the lateral decubitus position). The operating time was 65.79 (40-98) minutes using the old method of registration and was 50.87 (33-74) minutes using the new method of registration. This difference was statistically significant. The body mass index (BMI) was comparable in both groups. The study supports the use of new method of registration in improving the operating time in computer navigated primary uncemented total hip replacements.

  3. Can You Depend Totally on Computers?

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 3; Issue 2. Can You Depend Totally on Computers? Computer Security, Availability and Correctness. H N Mahabala. General Article Volume 3 Issue 2 February 1998 pp 35-44 ...

  4. Effects on mortality, treatment, and time management as a result of routine use of total body computed tomography in blunt high-energy trauma patients.

    Science.gov (United States)

    van Vugt, Raoul; Kool, Digna R; Deunk, Jaap; Edwards, Michael J R

    2012-03-01

    Currently, total body computed tomography (TBCT) is rapidly implemented in the evaluation of trauma patients. With this review, we aim to evaluate the clinical implications-mortality, change in treatment, and time management-of the routine use of TBCT in adult blunt high-energy trauma patients compared with a conservative approach with the use of conventional radiography, ultrasound, and selective computed tomography. A literature search for original studies on TBCT in blunt high-energy trauma patients was performed. Two independent observers included studies concerning mortality, change of treatment, and/or time management as outcome measures. For each article, relevant data were extracted and analyzed. In addition, the quality according to the Oxford levels of evidence was assessed. From 183 articles initially identified, the observers included nine original studies in consensus. One of three studies described a significant difference in mortality; four described a change of treatment in 2% to 27% of patients because of the use of TBCT. Five studies found a gain in time with the use of immediate routine TBCT. Eight studies scored a level of evidence of 2b and one of 3b. Current literature has predominantly suboptimal design to prove terminally that the routine use of TBCT results in improved survival of blunt high-energy trauma patients. TBCT can give a change of treatment and improves time intervals in the emergency department as compared with its selective use.

  5. Total variation-based neutron computed tomography

    Science.gov (United States)

    Barnard, Richard C.; Bilheux, Hassina; Toops, Todd; Nafziger, Eric; Finney, Charles; Splitter, Derek; Archibald, Rick

    2018-05-01

    We perform the neutron computed tomography reconstruction problem via an inverse problem formulation with a total variation penalty. In the case of highly under-resolved angular measurements, the total variation penalty suppresses high-frequency artifacts which appear in filtered back projections. In order to efficiently compute solutions for this problem, we implement a variation of the split Bregman algorithm; due to the error-forgetting nature of the algorithm, the computational cost of updating can be significantly reduced via very inexact approximate linear solvers. We present the effectiveness of the algorithm in the significantly low-angular sampling case using synthetic test problems as well as data obtained from a high flux neutron source. The algorithm removes artifacts and can even roughly capture small features when an extremely low number of angles are used.

  6. Timing the total reflection of light

    International Nuclear Information System (INIS)

    Chauvat, Dominique; Bonnet, Christophe; Dunseath, Kevin; Emile, Olivier; Le Floch, Albert

    2005-01-01

    We have identified for the first time the absolute delay at total reflection, envisioned by Newton. We show that there are in fact two divergent Wigner delays, depending on the polarisation of the incident light. These measurements give a new insight on the passage from total reflection to refraction

  7. Trends in television and computer/videogame use and total screen time in high school students from Caruaru city, Pernambuco, Brazil: A repeated panel study between 2007 and 2012

    Directory of Open Access Journals (Sweden)

    Luis José Lagos Aros

    2018-01-01

    Full Text Available Abstract Aim: to analyze the pattern and trends of use of screen-based devices and associated factors from two surveys conducted on public high school students in Caruaru-PE. Methods: two representative school-based cross-sectional surveys conducted in 2007 (n=600 and 2012 (n=715 on high school students (15-20 years old. The time of exposure to television (TV and computer/videogames PC/VG was obtained through a validated questionnaire, and ≥3 hours/day was considered as being excessive exposure. The independent variables were socioeconomic status, school related, and physical activity. Crude and adjusted binary logistic regression were employed to examine the factors associated with screen time. The statistical significance was set at p<0.05. Results: There was a significant reduction in TV time on weekdays and total weekly, but no change in the prevalence of excessive exposure. The proportion of exposure to PC/VG of ≥3 hours/day increased 182.5% on weekdays and 69.5% on weekends (p <0.05. In 2007, being physically active was the only protection factor for excessive exposure to total screen time. In 2012, girls presented less chance of excessive exposure to all screen-based devices and total screen time. Other protective factors were studying at night and being physically active (PC/VG time, while residing in an urban area [OR 5.03(2.77-7.41] and having higher family income [OR 1.55(1.04-2.30] were risk factors. Conclusion: Significant and important changes in the time trends and pattern of use PC/VG were observed during the interval of 5 years. This rapid increase could be associated with increased family income and improved access to these devices, driven by technological developments.

  8. Total sleep time severely drops during adolescence.

    Directory of Open Access Journals (Sweden)

    Damien Leger

    Full Text Available UNLABELLED: Restricted sleep duration among young adults and adolescents has been shown to increase the risk of morbidities such as obesity, diabetes or accidents. However there are few epidemiological studies on normal total sleep time (TST in representative groups of teen-agers which allow to get normative data. PURPOSE: To explore perceived total sleep time on schooldays (TSTS and non schooldays (TSTN and the prevalence of sleep initiating insomnia among a nationally representative sample of teenagers. METHODS: Data from 9,251 children aged 11 to 15 years-old, 50.7% of which were boys, as part of the cross-national study 2011 HBSC were analyzed. Self-completion questionnaires were administered in classrooms. An estimate of TSTS and TSTN (week-ends and vacations was calculated based on specifically designed sleep habits report. Sleep deprivation was estimated by a TSTN - TSTS difference >2 hours. Sleep initiating nsomnia was assessed according to International classification of sleep disorders (ICSD 2. Children who reported sleeping 7 hours or less per night were considered as short sleepers. RESULTS: A serious drop of TST was observed between 11 yo and 15 yo, both during the schooldays (9 hours 26 minutes vs. 7 h 55 min.; p<0.001 and at a lesser extent during week-ends (10 h 17 min. vs. 9 h 44 min.; p<0.001. Sleep deprivation concerned 16.0% of chidren aged of 11 yo vs. 40.5% of those of 15 yo (p<0.001. Too short sleep was reported by 2.6% of the 11 yo vs. 24.6% of the 15 yo (p<0.001. CONCLUSION: Despite the obvious need for sleep in adolescence, TST drastically decreases with age among children from 11 to 15 yo which creates significant sleep debt increasing with age.

  9. Time-Predictable Computer Architecture

    Directory of Open Access Journals (Sweden)

    Schoeberl Martin

    2009-01-01

    Full Text Available Today's general-purpose processors are optimized for maximum throughput. Real-time systems need a processor with both a reasonable and a known worst-case execution time (WCET. Features such as pipelines with instruction dependencies, caches, branch prediction, and out-of-order execution complicate WCET analysis and lead to very conservative estimates. In this paper, we evaluate the issues of current architectures with respect to WCET analysis. Then, we propose solutions for a time-predictable computer architecture. The proposed architecture is evaluated with implementation of some features in a Java processor. The resulting processor is a good target for WCET analysis and still performs well in the average case.

  10. 12 CFR 1102.27 - Computing time.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 1102.27 Section 1102.27 Banks... for Proceedings § 1102.27 Computing time. (a) General rule. In computing any period of time prescribed... time begins to run is not included. The last day so computed is included, unless it is a Saturday...

  11. Novel crystal timing calibration method based on total variation

    Science.gov (United States)

    Yu, Xingjian; Isobe, Takashi; Watanabe, Mitsuo; Liu, Huafeng

    2016-11-01

    A novel crystal timing calibration method based on total variation (TV), abbreviated as ‘TV merge’, has been developed for a high-resolution positron emission tomography (PET) system. The proposed method was developed for a system with a large number of crystals, it can provide timing calibration at the crystal level. In the proposed method, the timing calibration process was formulated as a linear problem. To robustly optimize the timing resolution, a TV constraint was added to the linear equation. Moreover, to solve the computer memory problem associated with the calculation of the timing calibration factors for systems with a large number of crystals, the merge component was used for obtaining the crystal level timing calibration values. Compared with other conventional methods, the data measured from a standard cylindrical phantom filled with a radioisotope solution was sufficient for performing a high-precision crystal-level timing calibration. In this paper, both simulation and experimental studies were performed to demonstrate the effectiveness and robustness of the TV merge method. We compare the timing resolutions of a 22Na point source, which was located in the field of view (FOV) of the brain PET system, with various calibration techniques. After implementing the TV merge method, the timing resolution improved from 3.34 ns at full width at half maximum (FWHM) to 2.31 ns FWHM.

  12. Total and Partial Computation in Categorical Quantum Foundations

    Directory of Open Access Journals (Sweden)

    Kenta Cho

    2015-11-01

    Full Text Available This paper uncovers the fundamental relationship between total and partial computation in the form of an equivalence of certain categories. This equivalence involves on the one hand effectuses, which are categories for total computation, introduced by Jacobs for the study of quantum/effect logic. On the other hand, it involves what we call FinPACs with effects; they are finitely partially additive categories equipped with effect algebra structures, serving as categories for partial computation. It turns out that the Kleisli category of the lift monad (-+1 on an effectus is always a FinPAC with effects, and this construction gives rise to the equivalence. Additionally, state-and-effect triangles over FinPACs with effects are presented.

  13. 12 CFR 622.21 - Computing time.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 6 2010-01-01 2010-01-01 false Computing time. 622.21 Section 622.21 Banks and... Formal Hearings § 622.21 Computing time. (a) General rule. In computing any period of time prescribed or... run is not to be included. The last day so computed shall be included, unless it is a Saturday, Sunday...

  14. Current Role of Computer Navigation in Total Knee Arthroplasty.

    Science.gov (United States)

    Jones, Christopher W; Jerabek, Seth A

    2018-01-31

    Computer-assisted surgical (CAS) navigation has been developed with the aim of improving the accuracy and precision of total knee arthroplasty (TKA) component positioning and therefore overall limb alignment. The historical goal of knee arthroplasty has been to restore the mechanical alignment of the lower limb by aligning the femoral and tibial components perpendicular to the mechanical axis of the femur and tibia. Despite over 4 decades of TKA component development and nearly 2 decades of interest in CAS, the fundamental question remains; does the alignment goal and/or the method of achieving that goal affect the outcome of the TKA in terms of patient-reported outcome measures and/or overall survivorship? The quest for reliable and reproducible achievement of the intraoperative alignment goal has been the primary motivator for the introduction, development, and refinement of CAS navigation. Numerous proprietary systems now exist, and rapid technological advancements in computer processing power are stimulating further development of robotic surgical systems. Three categories of CAS can be defined: image-based large-console navigation; imageless large-console navigation, and more recently, accelerometer-based handheld navigation systems have been developed. A review of the current literature demonstrates that there are enough well-designed studies to conclude that both large-console CAS and handheld navigation systems improve the accuracy and precision of component alignment in TKA. However, missing from the evidence base, other than the subgroup analysis provided by the Australian Orthopaedic Association National Joint Replacement Registry, are any conclusive demonstrations of a clinical superiority in terms of improved patient-reported outcome measures and/or decreased cumulative revision rates in the long term. Few authors would argue that accuracy of alignment is a goal to ignore; therefore, in the absence of clinical evidence, many of the arguments against

  15. A Matter of Computer Time

    Science.gov (United States)

    Celano, Donna; Neuman, Susan B.

    2010-01-01

    Many low-income children do not have the opportunity to develop the computer skills necessary to succeed in our technological economy. Their only access to computers and the Internet--school, afterschool programs, and community organizations--is woefully inadequate. Educators must work to close this knowledge gap and to ensure that low-income…

  16. 12 CFR 908.27 - Computing time.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 908.27 Section 908.27 Banks and... PRACTICE AND PROCEDURE IN HEARINGS ON THE RECORD General Rules § 908.27 Computing time. (a) General rule. In computing any period of time prescribed or allowed by this subpart, the date of the act or event...

  17. Total quality through computer integrated manufacturing in the pharmaceutical industry.

    Science.gov (United States)

    Ufret, C M

    1995-01-01

    The role of Computer Integrated Manufacturing (CIM) in the pursue of total quality in pharmaceutical manufacturing is assessed. CIM key objectives, design criteria, and performance measurements, in addition to its scope and implementation in a hierarchical structure, are explored in detail. Key elements for the success of each phase in a CIM project and a brief status of current CIM implementations in the pharmaceutical industry are presented. The role of World Class Manufacturing performance standards and other key issues to achieve full CIM benefits are also addressed.

  18. 12 CFR 1780.11 - Computing time.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 7 2010-01-01 2010-01-01 false Computing time. 1780.11 Section 1780.11 Banks... time. (a) General rule. In computing any period of time prescribed or allowed by this subpart, the date of the act or event that commences the designated period of time is not included. The last day so...

  19. Computer Package for Graphite Total Cross-Section Calculations

    International Nuclear Information System (INIS)

    Adib, M.; Fathalla, M.

    2008-01-01

    An additive formula is given which allows calculating the contribution of the total neut.>neutron transmission through crystalline graphite. The formula takes into account the graphite form of poly or pyrolytic crystals and its parameters. Computer package Graphite has been designed in order to provide the required calculations in the neutron energy range from 0.1 MeV to 10 eV. The package includes three codes: PCG (Polycrystalline Graphite), PG (Pyrolytic Graphite) and HOPG (Highly Oriented Pyrolytic Graphite) for calculating neutron transmission through fine graphite powder (polycrystalline), neutron transmission and removal coefficient of PG crystal in terms of its mosaic spread for neutrons incident along its c-axis and the transmission of neutrons incident on HOPG crystal at different angles, respectively. For comparison of the experimental neutron transmission data with the calculated values, the program takes into consideration the effect of both wavelength and neutron beam divergence in either 2 constant wavelength spread mode (δλ=constant) or constant wavelength resolution mode (δλ/λ=constant). In order to check the validity for application of computer package Graphite in cross-section calculations, a comparison between calculated values with the available experimental data were carried out. An overall agreement is indicated with an accuracy sufficient for determine the neutron transmission characteristics

  20. Optical computation based on nonlinear total reflectional optical ...

    Indian Academy of Sciences (India)

    2School of Education Science, South China Normal University, Guangzhou, 510631, China. *Corresponding ... Before the computation, all the inputs are prepared in the polarization state. The key .... The all-optical computing system described.

  1. Optical computation based on nonlinear total reflectional optical ...

    Indian Academy of Sciences (India)

    Optical computing; beam splitter; optical switch; polarized beams. ... main research direction called quantum information and quantum computation is .... above has several advantages: Firstly, it is easy to be integrated with appropriate.

  2. Artroplastia total do joelho assistida por computador Computer-assisted knee total arthroplasty

    Directory of Open Access Journals (Sweden)

    Roberto Freire da Mota e Albuquerque

    2006-01-01

    Full Text Available Um dos avanços tecnológicos mais significativos da medicina atual é a cirurgia assistida por computador, sendo que na ortopedia uma das aplicações mais importantes dessa tecnologia é na artroplastia do joelho. A principal contribuição da cirurgia ortopédica assistida por computador (Computer Aided Orthopaedic Surgery’s - CAOS na artroplastia do joelho é o seu potencial em aprimorar a precisão da implantação das próteses e do alinhamento do membro operado contribuindo para a otimização e longevidade dos resultados. A navegação independente de imagens, baseada em referências anatômicas adquiridas durante o ato cirúrgico através de transmissores de raios infra-vermelho, tem sido a técnica preponderante na artroplastia do joelho. Utilizamos o sistema de navegação para artroplastia total do joelho "OrthoPilot" versão 2.2 para a colocação de 72 próteses de joelho "Search Evolution" da "Aesculap AG CO. KG" com ou sem estabilização posterior em uma série contínua. O objetivo foi aferir a precisão do alinhamento obtido com a navegação através de radiografias panorâmicas obtidas no período pós-operatório. Obtivemos um desvio médio do eixo mecânico nulo de 0,66º com desvio padrão de 0,7º, sendo que 98,6% dos joelhos ficaram dentro de uma margem de erro de 3º e 79,2% com erro inferior a 1º. Concluímos que o sistema é seguro e preciso, não adicionando morbidade à cirurgia convencional.One of the most significant technological advancements in current medicine is the computer-assisted surgery, which, for orthopaedics, one of the major uses of this technology is in knee arthroplasty. The main contribution provided by computer-assisted orthopaedic surgery (CAOS to knee arthroplasty is its potential to improve prosthesis implant precision and the operated limb alignment, contributing to results optimization and longevity. The image-independent navigation, based on anatomical references acquired during surgical

  3. General purpose computers in real time

    International Nuclear Information System (INIS)

    Biel, J.R.

    1989-01-01

    I see three main trends in the use of general purpose computers in real time. The first is more processing power. The second is the use of higher speed interconnects between computers (allowing more data to be delivered to the processors). The third is the use of larger programs running in the computers. Although there is still work that needs to be done, I believe that all indications are that the online need for general purpose computers should be available for the SCC and LHC machines. 2 figs

  4. Applied time series analysis and innovative computing

    CERN Document Server

    Ao, Sio-Iong

    2010-01-01

    This text is a systematic, state-of-the-art introduction to the use of innovative computing paradigms as an investigative tool for applications in time series analysis. It includes frontier case studies based on recent research.

  5. PROCESS INNOVATION: HOLISTIC SCENARIOS TO REDUCE TOTAL LEAD TIME

    Directory of Open Access Journals (Sweden)

    Alin POSTEUCĂ

    2015-11-01

    Full Text Available The globalization of markets requires continuous development of business holistic scenarios to ensure acceptable flexibility to satisfy customers. Continuous improvement of supply chain supposes continuous improvement of materials and products lead time and flow, material stocks and finished products stocks and increasing the number of suppliers close by as possible. The contribution of our study is to present holistic scenarios of total lead time improvement and innovation by implementing supply chain policy.

  6. Whole blood coagulation time, haematocrit, haemoglobin and total ...

    African Journals Online (AJOL)

    The study was carried out to determine the values of whole blood coagulation time (WBCT), haematocrit (HM), haemaglobin (HB) and total protein (TP) of one hundred and eighteen apparently healthy turkeys reared under an extensive management system in Zaria. The mean values for WBCT, HM, HB and TP were 1.12 ...

  7. Instruction timing for the CDC 7600 computer

    International Nuclear Information System (INIS)

    Lipps, H.

    1975-01-01

    This report provides timing information for all instructions of the Control Data 7600 computer, except for instructions of type 01X, to enable the optimization of 7600 programs. The timing rules serve as background information for timing charts which are produced by a program (TIME76) of the CERN Program Library. The rules that co-ordinate the different sections of the CPU are stated in as much detail as is necessary to time the flow of instructions for a given sequence of code. Instruction fetch, instruction issue, and access to small core memory are treated at length, since details are not available from the computer manuals. Annotated timing charts are given for 24 examples, chosen to display the full range of timing considerations. (Author)

  8. Total reduction of distorted echelle spectrograms - An automatic procedure. [for computer controlled microdensitometer

    Science.gov (United States)

    Peterson, R. C.; Title, A. M.

    1975-01-01

    A total reduction procedure, notable for its use of a computer-controlled microdensitometer for semi-automatically tracing curved spectra, is applied to distorted high-dispersion echelle spectra recorded by an image tube. Microdensitometer specifications are presented and the FORTRAN, TRACEN and SPOTS programs are outlined. The intensity spectrum of the photographic or electrographic plate is plotted on a graphic display. The time requirements are discussed in detail.

  9. Fast algorithms for computing phylogenetic divergence time.

    Science.gov (United States)

    Crosby, Ralph W; Williams, Tiffani L

    2017-12-06

    The inference of species divergence time is a key step in most phylogenetic studies. Methods have been available for the last ten years to perform the inference, but the performance of the methods does not yet scale well to studies with hundreds of taxa and thousands of DNA base pairs. For example a study of 349 primate taxa was estimated to require over 9 months of processing time. In this work, we present a new algorithm, AncestralAge, that significantly improves the performance of the divergence time process. As part of AncestralAge, we demonstrate a new method for the computation of phylogenetic likelihood and our experiments show a 90% improvement in likelihood computation time on the aforementioned dataset of 349 primates taxa with over 60,000 DNA base pairs. Additionally, we show that our new method for the computation of the Bayesian prior on node ages reduces the running time for this computation on the 349 taxa dataset by 99%. Through the use of these new algorithms we open up the ability to perform divergence time inference on large phylogenetic studies.

  10. Computer network time synchronization the network time protocol

    CERN Document Server

    Mills, David L

    2006-01-01

    What started with the sundial has, thus far, been refined to a level of precision based on atomic resonance: Time. Our obsession with time is evident in this continued scaling down to nanosecond resolution and beyond. But this obsession is not without warrant. Precision and time synchronization are critical in many applications, such as air traffic control and stock trading, and pose complex and important challenges in modern information networks.Penned by David L. Mills, the original developer of the Network Time Protocol (NTP), Computer Network Time Synchronization: The Network Time Protocol

  11. Total system for manufacture of nuclear vessels by computer: VECTRON

    International Nuclear Information System (INIS)

    Inagawa, Jin; Ueno, Osamu; Hanai, Yoshiharu; Ohkawa, Isao; Washizu, Hideyuki

    1980-01-01

    VECTRON (Vessel Engineering by Computer Tool and Rapid Operating for the N/C System) is a CAM (Computer Aided Manufacturing) system that has been developed to produce high quality and highly accurate vessels for nuclear power plants and other industrial plants. Outputs of this system are design drawings, manufacturing information and magnetic tapes of the N/C marking machine for vessel shell plates including their attachments. And it can also output information at each stage of designing, marking, cutting, forming and assembling by treating the vessels in three dimensions and by using data filing systems and plotting program for general use. The data filing systems consist of functional and manufacturing data of each part of vessels. This system not only realizes a change from manual work to computer work, but also leads us to improve production engineering and production jigs for safety and high quality. At present, VECTRON is being applied to the manufacture of the shell plates of primary containment vessels in the Kashiwazaki-Kariwa Nuclear Power Station Unit 1 (K-1) and the Fukushima Daini Nuclear Power Station Unit 3 (2F-3), to realize increased productivity. (author)

  12. Computer Assisted Surgery and Current Trends in Orthopaedics Research and Total Joint Replacements

    Science.gov (United States)

    Amirouche, Farid

    2008-06-01

    Musculoskeletal research has brought about revolutionary changes in our ability to perform high precision surgery in joint replacement procedures. Recent advances in computer assisted surgery as well better materials have lead to reduced wear and greatly enhanced the quality of life of patients. The new surgical techniques to reduce the size of the incision and damage to underlying structures have been the primary advance toward this goal. These new techniques are known as MIS or Minimally Invasive Surgery. Total hip and knee Arthoplasties are at all time high reaching 1.2 million surgeries per year in the USA. Primary joint failures are usually due to osteoarthristis, rheumatoid arthritis, osteocronis and other inflammatory arthritis conditions. The methods for THR and TKA are critical to initial stability and longevity of the prostheses. This research aims at understanding the fundamental mechanics of the joint Arthoplasty and providing an insight into current challenges in patient specific fitting, fixing, and stability. Both experimental and analytical work will be presented. We will examine Cementless total hip arthroplasty success in the last 10 years and how computer assisted navigation is playing in the follow up studies. Cementless total hip arthroplasty attains permanent fixation by the ingrowth of bone into a porous coated surface. Loosening of an ingrown total hip arthroplasty occurs as a result of osteolysis of the periprosthetic bone and degradation of the bone prosthetic interface. The osteolytic process occurs as a result of polyethylene wear particles produced by the metal polyethylene articulation of the prosthesis. The total hip arthroplasty is a congruent joint and the submicron wear particles produced are phagocytized by macrophages initiating an inflammatory cascade. This cascade produces cytokines ultimately implicated in osteolysis. Resulting bone loss both on the acetabular and femoral sides eventually leads to component instability. As

  13. BIOMECHANICAL INDICES OF STANDING AND GAIT IN PATIENTS AFTER TOTAL KNEE REPLACEMENT USING COMPUTER NAVIGATION

    Directory of Open Access Journals (Sweden)

    Y. A. Bezgodkov

    2011-01-01

    Full Text Available Several biomechanical parameters of standing and walking in 50 patients with osteoarthrosis after total knee arthroplasty were evaluated. The patients were randomly divided in two equal groups: in the first group the surgery was performed with computer navigation system and in the second - with traditional instruments. After TKA with computer navigation centers of common body pressure and legs pressure during standing phase improved significantly better than in traditional group. Walking parameters like step length, ground contact time and rhythm coefficient improved in both groups of patients but without significant difference. Thereby more precise orientation of implant that achieved during computer assisted TKA leads to better functional performance at 6 and 12 month after surgery.

  14. Single machine total completion time minimization scheduling with a time-dependent learning effect and deteriorating jobs

    Science.gov (United States)

    Wang, Ji-Bo; Wang, Ming-Zheng; Ji, Ping

    2012-05-01

    In this article, we consider a single machine scheduling problem with a time-dependent learning effect and deteriorating jobs. By the effects of time-dependent learning and deterioration, we mean that the job processing time is defined by a function of its starting time and total normal processing time of jobs in front of it in the sequence. The objective is to determine an optimal schedule so as to minimize the total completion time. This problem remains open for the case of -1 < a < 0, where a denotes the learning index; we show that an optimal schedule of the problem is V-shaped with respect to job normal processing times. Three heuristic algorithms utilising the V-shaped property are proposed, and computational experiments show that the last heuristic algorithm performs effectively and efficiently in obtaining near-optimal solutions.

  15. TimeSet: A computer program that accesses five atomic time services on two continents

    Science.gov (United States)

    Petrakis, P. L.

    1993-01-01

    TimeSet is a shareware program for accessing digital time services by telephone. At its initial release, it was capable of capturing time signals only from the U.S. Naval Observatory to set a computer's clock. Later the ability to synchronize with the National Institute of Standards and Technology was added. Now, in Version 7.10, TimeSet is able to access three additional telephone time services in Europe - in Sweden, Austria, and Italy - making a total of five official services addressable by the program. A companion program, TimeGen, allows yet another source of telephone time data strings for callers equipped with TimeSet version 7.10. TimeGen synthesizes UTC time data strings in the Naval Observatory's format from an accurately set and maintained DOS computer clock, and transmits them to callers. This allows an unlimited number of 'freelance' time generating stations to be created. Timesetting from TimeGen is made feasible by the advent of Becker's RighTime, a shareware program that learns the drift characteristics of a computer's clock and continuously applies a correction to keep it accurate, and also brings .01 second resolution to the DOS clock. With clock regulation by RighTime and periodic update calls by the TimeGen station to an official time source via TimeSet, TimeGen offers the same degree of accuracy within the resolution of the computer clock as any official atomic time source.

  16. Time series modeling, computation, and inference

    CERN Document Server

    Prado, Raquel

    2010-01-01

    The authors systematically develop a state-of-the-art analysis and modeling of time series. … this book is well organized and well written. The authors present various statistical models for engineers to solve problems in time series analysis. Readers no doubt will learn state-of-the-art techniques from this book.-Hsun-Hsien Chang, Computing Reviews, March 2012My favorite chapters were on dynamic linear models and vector AR and vector ARMA models.-William Seaver, Technometrics, August 2011… a very modern entry to the field of time-series modelling, with a rich reference list of the current lit

  17. Computer-based anthropometrical system for total body irradiation.

    Science.gov (United States)

    Sánchez-Nieto, B; Sánchez-Doblado, F; Terrón, J A; Arráns, R; Errazquin, L

    1997-05-01

    For total body irradiation (TBI) dose calculation requirements, anatomical information about the whole body is needed. Despite the fact that video image grabbing techniques are used by some treatment planning systems for standard radiotherapy, there are no such systems designed to generate anatomical parameters for TBI planning. The paper describes an anthropometrical computerised system based on video image grabbing which was purpose-built to provide anatomical data for a PC-based TBI planning system. Using software, the system controls the acquisition and digitalisation of the images (external images of the patient in treatment position) and the measurement procedure itself (on the external images or the digital CT information). An ASCII file, readable by the TBI planning system, is generated to store the required parameters of the dose calculation points, i.e. depth, backscatter tissue thickness, thickness of inhomogeneity, off-axis distance (OAD) and source to skin distance (SSD).

  18. Primary total knee arthroplasty assisted by computer (orthopilot)

    International Nuclear Information System (INIS)

    Naquira, Luis Felipe; Restrepo Tello, Fabio; Pineda Acero, Gustavo; Clavijo, Edgar; Buitrago, Mario

    2004-01-01

    Between December 2001 and December 2003, 21 patients suffering from osteoarthrosis of the knee were treated using computer assisted complete primary arthroplasty of the knee (Orthopilot) at the Department of Orthopedics and Traumatology in the Hospital Militar Central. Results were followed up for an average of 18 months. The criteria evaluated were function and mechanical axis, using ortho-radiography, and length of surgery. The average value obtained was 4.3 degrades with a mechanical axis passing through the centre of the knee joint in 95% of cases, confirmed by a mechanical femoral-tibial angle of -2.5 degrades. Length of surgery was increased by an average of 1 hour compared to the conventional technique. Complications were as follows: One patient experienced an infection which required the withdrawal of the prosthesis and in four of the operations technical problems with the navigator made it necessary to proceed using conventional techniques and these patients were excluded from the study. Radiological results ranged from Excellent to Good in 100% of cases

  19. TV time but not computer time is associated with cardiometabolic risk in Dutch young adults.

    Science.gov (United States)

    Altenburg, Teatske M; de Kroon, Marlou L A; Renders, Carry M; Hirasing, Remy; Chinapaw, Mai J M

    2013-01-01

    TV time and total sedentary time have been positively related to biomarkers of cardiometabolic risk in adults. We aim to examine the association of TV time and computer time separately with cardiometabolic biomarkers in young adults. Additionally, the mediating role of waist circumference (WC) is studied. Data of 634 Dutch young adults (18-28 years; 39% male) were used. Cardiometabolic biomarkers included indicators of overweight, blood pressure, blood levels of fasting plasma insulin, cholesterol, glucose, triglycerides and a clustered cardiometabolic risk score. Linear regression analyses were used to assess the cross-sectional association of self-reported TV and computer time with cardiometabolic biomarkers, adjusting for demographic and lifestyle factors. Mediation by WC was checked using the product-of-coefficient method. TV time was significantly associated with triglycerides (B = 0.004; CI = [0.001;0.05]) and insulin (B = 0.10; CI = [0.01;0.20]). Computer time was not significantly associated with any of the cardiometabolic biomarkers. We found no evidence for WC to mediate the association of TV time or computer time with cardiometabolic biomarkers. We found a significantly positive association of TV time with cardiometabolic biomarkers. In addition, we found no evidence for WC as a mediator of this association. Our findings suggest a need to distinguish between TV time and computer time within future guidelines for screen time.

  20. Real-Time Thevenin Impedance Computation

    DEFF Research Database (Denmark)

    Sommer, Stefan Horst; Jóhannsson, Hjörtur

    2013-01-01

    operating state, and strict time constraints are difficult to adhere to as the complexity of the grid increases. Several suggested approaches for real-time stability assessment require Thevenin impedances to be determined for the observed system conditions. By combining matrix factorization, graph reduction......, and parallelization, we develop an algorithm for computing Thevenin impedances an order of magnitude faster than previous approaches. We test the factor-and-solve algorithm with data from several power grids of varying complexity, and we show how the algorithm allows realtime stability assessment of complex power...

  1. Time reversibility, computer simulation, algorithms, chaos

    CERN Document Server

    Hoover, William Graham

    2012-01-01

    A small army of physicists, chemists, mathematicians, and engineers has joined forces to attack a classic problem, the "reversibility paradox", with modern tools. This book describes their work from the perspective of computer simulation, emphasizing the author's approach to the problem of understanding the compatibility, and even inevitability, of the irreversible second law of thermodynamics with an underlying time-reversible mechanics. Computer simulation has made it possible to probe reversibility from a variety of directions and "chaos theory" or "nonlinear dynamics" has supplied a useful vocabulary and a set of concepts, which allow a fuller explanation of irreversibility than that available to Boltzmann or to Green, Kubo and Onsager. Clear illustration of concepts is emphasized throughout, and reinforced with a glossary of technical terms from the specialized fields which have been combined here to focus on a common theme. The book begins with a discussion, contrasting the idealized reversibility of ba...

  2. Complications of fixed infrared emitters in computer-assisted total knee arthroplasties

    Directory of Open Access Journals (Sweden)

    Suárez-Vázquez Abelardo

    2007-07-01

    Full Text Available Abstract Background The first stage in the implant of a total knee arthroplasty with computer-assisted surgery is to fasten the emitters to the femur and the tibia. These trackers must be hard-fixed to the bone. The objectives of our study are to evaluate the technical problems and complications of these tracker-pins, the necessary time to fix them to the bone and the possible advantages of a new femoral-fixed tracker-pin. Methods Three hundred and sixty seven tracker-pins were used in one hundred and fifty one computer-assisted total knee replacements. A bicortical screw was used to fix the tracker to the tibia in all cases; in the femur, however, a bicortical tracker was used in 112 cases, while a new device (OrthoLock with percutaneous fixation pins was employed in the remaining 39. Results Technical problems related to the fixing of the trackers appeared in nine cases (2.5%. The mean surgery time to fix the tracker pin to the tibia was 3 minutes (range 2–7, and 5 minutes in the case of the femoral pin (range: 4–11, although with the new tool it was only three minutes (range 2–4 (p Conclusion The incidence of problems and complications with the fixing systems used in knee navigation is very small. The use of a new device with percutaneous pins facilitates the fixing of femoral trackers and decreases the time needed to place them.

  3. Total sitting time, leisure time physical activity and risk of hospitalization due to low back pain

    DEFF Research Database (Denmark)

    Balling, Mie; Holmberg, Teresa; Petersen, Christina B

    2018-01-01

    AIMS: This study aimed to test the hypotheses that a high total sitting time and vigorous physical activity in leisure time increase the risk of low back pain and herniated lumbar disc disease. METHODS: A total of 76,438 adults answered questions regarding their total sitting time and physical...... activity during leisure time in the Danish Health Examination Survey 2007-2008. Information on low back pain diagnoses up to 10 September 2015 was obtained from The National Patient Register. The mean follow-up time was 7.4 years. Data were analysed using Cox regression analysis with adjustment...... disc disease. However, moderate or vigorous physical activity, as compared to light physical activity, was associated with increased risk of low back pain (HR = 1.16, 95% CI: 1.03-1.30 and HR = 1.45, 95% CI: 1.15-1.83). Moderate, but not vigorous physical activity was associated with increased risk...

  4. Optimizing Ship Speed to Minimize Total Fuel Consumption with Multiple Time Windows

    Directory of Open Access Journals (Sweden)

    Jae-Gon Kim

    2016-01-01

    Full Text Available We study the ship speed optimization problem with the objective of minimizing the total fuel consumption. We consider multiple time windows for each port call as constraints and formulate the problem as a nonlinear mixed integer program. We derive intrinsic properties of the problem and develop an exact algorithm based on the properties. Computational experiments show that the suggested algorithm is very efficient in finding an optimal solution.

  5. Time related total lactic acid bacteria population diversity and ...

    African Journals Online (AJOL)

    The total lactic acid bacterial community involved in the spontaneous fermentation of malted cowpea fortified cereal weaning food was investigated by phenotypically and cultivation independent method. A total of 74 out of the isolated 178 strains were Lactobacillus plantarum, 32 were Pediococcus acidilactici and over 60% ...

  6. Computing Refined Buneman Trees in Cubic Time

    DEFF Research Database (Denmark)

    Brodal, G.S.; Fagerberg, R.; Östlin, A.

    2003-01-01

    Reconstructing the evolutionary tree for a set of n species based on pairwise distances between the species is a fundamental problem in bioinformatics. Neighbor joining is a popular distance based tree reconstruction method. It always proposes fully resolved binary trees despite missing evidence...... in the underlying distance data. Distance based methods based on the theory of Buneman trees and refined Buneman trees avoid this problem by only proposing evolutionary trees whose edges satisfy a number of constraints. These trees might not be fully resolved but there is strong combinatorial evidence for each...... proposed edge. The currently best algorithm for computing the refined Buneman tree from a given distance measure has a running time of O(n 5) and a space consumption of O(n 4). In this paper, we present an algorithm with running time O(n 3) and space consumption O(n 2). The improved complexity of our...

  7. Real-Time Accumulative Computation Motion Detectors

    Directory of Open Access Journals (Sweden)

    Saturnino Maldonado-Bascón

    2009-12-01

    Full Text Available The neurally inspired accumulative computation (AC method and its application to motion detection have been introduced in the past years. This paper revisits the fact that many researchers have explored the relationship between neural networks and finite state machines. Indeed, finite state machines constitute the best characterized computational model, whereas artificial neural networks have become a very successful tool for modeling and problem solving. The article shows how to reach real-time performance after using a model described as a finite state machine. This paper introduces two steps towards that direction: (a A simplification of the general AC method is performed by formally transforming it into a finite state machine. (b A hardware implementation in FPGA of such a designed AC module, as well as an 8-AC motion detector, providing promising performance results. We also offer two case studies of the use of AC motion detectors in surveillance applications, namely infrared-based people segmentation and color-based people tracking, respectively.

  8. Cluster Computing for Embedded/Real-Time Systems

    Science.gov (United States)

    Katz, D.; Kepner, J.

    1999-01-01

    Embedded and real-time systems, like other computing systems, seek to maximize computing power for a given price, and thus can significantly benefit from the advancing capabilities of cluster computing.

  9. Criterion-based laparoscopic training reduces total training time

    NARCIS (Netherlands)

    Brinkman, W.M.; Buzink, S.N.; Alevizos, L.; De Hingh, I.H.J.T.; Jakimowicz, J.J.

    2011-01-01

    The benefits of criterion-based laparoscopic training over time-oriented training are unclear. The purpose of this study is to compare these types of training based on training outcome and time efficiency. Methods During four training sessions within 1 week (one session per day) 34 medical interns

  10. A general method for computing the total solar radiation force on complex spacecraft structures

    Science.gov (United States)

    Chan, F. K.

    1981-01-01

    The method circumvents many of the existing difficulties in computational logic presently encountered in the direct analytical or numerical evaluation of the appropriate surface integral. It may be applied to complex spacecraft structures for computing the total force arising from either specular or diffuse reflection or even from non-Lambertian reflection and re-radiation.

  11. CONTEMPORARY VIEW ON COMPUTER NAVIGATION USING AT PRIMARY KNEE TOTAL REPLACEMENT (REVIEW

    Directory of Open Access Journals (Sweden)

    A. I. Petukhov

    2010-01-01

    Full Text Available The topical questions of optical computer navigation at knee total arthroplasty are widely covered. The indications, contraindication to use, using features and possible complications of this technique are listed. The analysis of literature data makes it clear that computer navigation assists in the accuracy of endoprosthesis implantation that may to decrease the rate of revision surgeries in future.

  12. Real-time analysis of total, elemental, and total speciated mercury

    International Nuclear Information System (INIS)

    Schlager, R.J.; Wilson, K.G.; Sappey, A.D.

    1995-01-01

    ADA Technologies, Inc., is developing a continuous emissions monitoring system that measures the concentrations of mercury in flue gas. Mercury is emitted as an air pollutant from a number of industrial processes. The largest contributors of these emissions are coal and oil combustion, municipal waste combustion, medical waste combustion, and the thermal treatment of hazardous materials. It is difficult, time consuming, and expensive to measure mercury emissions using current testing methods. Part of the difficulty lies in the fact that mercury is emitted from sources in several different forms, such as elemental mercury and mercuric chloride. The ADA analyzer measures these emissions in real time, thus providing a number of advantages over existing test methods: (1) it will provide a real-time measure of emission rates, (2) it will assure facility operators, regulators, and the public that emissions control systems are working at peak efficiency, and (3) it will provide information as to the nature of the emitted mercury (elemental mercury or speciated compounds). This update presents an overview of the CEM and describes features of key components of the monitoring system--the mercury detector, a mercury species converter, and the analyzer calibration system

  13. Real-time analysis of total, elemental, and total speciated mercury

    Energy Technology Data Exchange (ETDEWEB)

    Schlager, R.J.; Wilson, K.G.; Sappey, A.D. [ADA Technologies, Inc., Englewood, CO (United States)

    1995-11-01

    ADA Technologies, Inc., is developing a continuous emissions monitoring system that measures the concentrations of mercury in flue gas. Mercury is emitted as an air pollutant from a number of industrial processes. The largest contributors of these emissions are coal and oil combustion, municipal waste combustion, medical waste combustion, and the thermal treatment of hazardous materials. It is difficult, time consuming, and expensive to measure mercury emissions using current testing methods. Part of the difficulty lies in the fact that mercury is emitted from sources in several different forms, such as elemental mercury and mercuric chloride. The ADA analyzer measures these emissions in real time, thus providing a number of advantages over existing test methods: (1) it will provide a real-time measure of emission rates, (2) it will assure facility operators, regulators, and the public that emissions control systems are working at peak efficiency, and (3) it will provide information as to the nature of the emitted mercury (elemental mercury or speciated compounds). This update presents an overview of the CEM and describes features of key components of the monitoring system--the mercury detector, a mercury species converter, and the analyzer calibration system.

  14. RURAL EXTENSION EPISTEMOLOGY AND THE TIME OF TOTAL EXTENSION

    Directory of Open Access Journals (Sweden)

    Silvio Calgaro Neto

    2016-09-01

    Full Text Available This article is dedicated to explore the field of knowledge related to rural extension. In general, a three complementary perspective is used as theoretical strategy to present this epistemological study. The first perspective, seeks to accomplish a brief archeology of rural extension, identifying the remarkable historical passages. At the second, we look to some theoretical models through the modern epistemological platform. Finally, the third perspective, aims to present a methodological proposal that contemplate this epistemic characteristics, relating with the contemporary transformations observed in the knowledge construction and technological transference for a rural development. Keywords: Total institutions. University.

  15. Criterion-based laparoscopic training reduces total training time

    OpenAIRE

    Brinkman, Willem M.; Buzink, Sonja N.; Alevizos, Leonidas; de Hingh, Ignace H. J. T.; Jakimowicz, Jack J.

    2011-01-01

    Introduction The benefits of criterion-based laparoscopic training over time-oriented training are unclear. The purpose of this study is to compare these types of training based on training outcome and time efficiency. Methods During four training sessions within 1 week (one session per day) 34 medical interns (no laparoscopic experience) practiced on two basic tasks on the Simbionix LAP Mentor virtual-reality (VR) simulator: ‘clipping and grasping’ and ‘cutting’. Group C (criterion-based) (N...

  16. Software Accelerates Computing Time for Complex Math

    Science.gov (United States)

    2014-01-01

    Ames Research Center awarded Newark, Delaware-based EM Photonics Inc. SBIR funding to utilize graphic processing unit (GPU) technology- traditionally used for computer video games-to develop high-computing software called CULA. The software gives users the ability to run complex algorithms on personal computers with greater speed. As a result of the NASA collaboration, the number of employees at the company has increased 10 percent.

  17. Minimizing total weighted completion time in a proportionate flow shop

    NARCIS (Netherlands)

    Shakhlevich, N.V.; Hoogeveen, J.A.; Pinedo, M.L.

    1998-01-01

    We study the special case of the m machine flow shop problem in which the processing time of each operation of job j is equal to pj; this variant of the flow shop problem is known as the proportionate flow shop problem. We show that for any number of machines and for any regular performance

  18. 7 CFR 1.603 - How are time periods computed?

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 1 2010-01-01 2010-01-01 false How are time periods computed? 1.603 Section 1.603... Licenses General Provisions § 1.603 How are time periods computed? (a) General. Time periods are computed as follows: (1) The day of the act or event from which the period begins to run is not included. (2...

  19. 50 CFR 221.3 - How are time periods computed?

    Science.gov (United States)

    2010-10-01

    ... 50 Wildlife and Fisheries 7 2010-10-01 2010-10-01 false How are time periods computed? 221.3... Provisions § 221.3 How are time periods computed? (a) General. Time periods are computed as follows: (1) The day of the act or event from which the period begins to run is not included. (2) The last day of the...

  20. 6 CFR 13.27 - Computation of time.

    Science.gov (United States)

    2010-01-01

    ... 6 Domestic Security 1 2010-01-01 2010-01-01 false Computation of time. 13.27 Section 13.27 Domestic Security DEPARTMENT OF HOMELAND SECURITY, OFFICE OF THE SECRETARY PROGRAM FRAUD CIVIL REMEDIES § 13.27 Computation of time. (a) In computing any period of time under this part or in an order issued...

  1. Real-time Tsunami Inundation Prediction Using High Performance Computers

    Science.gov (United States)

    Oishi, Y.; Imamura, F.; Sugawara, D.

    2014-12-01

    Recently off-shore tsunami observation stations based on cabled ocean bottom pressure gauges are actively being deployed especially in Japan. These cabled systems are designed to provide real-time tsunami data before tsunamis reach coastlines for disaster mitigation purposes. To receive real benefits of these observations, real-time analysis techniques to make an effective use of these data are necessary. A representative study was made by Tsushima et al. (2009) that proposed a method to provide instant tsunami source prediction based on achieving tsunami waveform data. As time passes, the prediction is improved by using updated waveform data. After a tsunami source is predicted, tsunami waveforms are synthesized from pre-computed tsunami Green functions of linear long wave equations. Tsushima et al. (2014) updated the method by combining the tsunami waveform inversion with an instant inversion of coseismic crustal deformation and improved the prediction accuracy and speed in the early stages. For disaster mitigation purposes, real-time predictions of tsunami inundation are also important. In this study, we discuss the possibility of real-time tsunami inundation predictions, which require faster-than-real-time tsunami inundation simulation in addition to instant tsunami source analysis. Although the computational amount is large to solve non-linear shallow water equations for inundation predictions, it has become executable through the recent developments of high performance computing technologies. We conducted parallel computations of tsunami inundation and achieved 6.0 TFLOPS by using 19,000 CPU cores. We employed a leap-frog finite difference method with nested staggered grids of which resolution range from 405 m to 5 m. The resolution ratio of each nested domain was 1/3. Total number of grid points were 13 million, and the time step was 0.1 seconds. Tsunami sources of 2011 Tohoku-oki earthquake were tested. The inundation prediction up to 2 hours after the

  2. Working time intervals and total work time on nursing positions in Poland

    Directory of Open Access Journals (Sweden)

    Danuta Kunecka

    2015-06-01

    Full Text Available Background: For the last few years a topic of overwork on nursing posts has given rise to strong discussions. The author has set herself a goal of answering the question if it is a result of real overwork of this particular profession or rather commonly assumed frustration of this professional group. The aim of this paper is to conduct the analysis of working time on chosen nursing positions in relation to measures of time being used as intervals in the course of conducting standard professional activities during one working day. Material and Methods: Research material consisted of documentation of work time on chosen nursing workplaces, compiled between 2007–2012 within the framework of a nursing course at the Pomeranian Medical University in Szczecin. As a method of measurement a photograph of a working day has been used. Measurements were performed in institutions located in 6 voivodeships in Poland. Results: Results suggest that only 6.5% of total of surveyed representatives of nurse profession spends proper amount of time (meaning: a time set by the applicable standards on work intervals during a working day. Conclusions: The scale of the phenomenon indicates excessive workload for nursing positions, which along with a longer period of time, longer working hours may cause decrease in efficiency of work and cause a drop in quality of provided services. Med Pr 2015;66,(2:165–172

  3. Radiologic total lung capacity measurement. Development and evaluation of a computer-based system

    Energy Technology Data Exchange (ETDEWEB)

    Seeley, G.W.; Mazzeo, J.; Borgstrom, M.; Hunter, T.B.; Newell, J.D.; Bjelland, J.C.

    1986-11-01

    The development of a computer-based radiologic total lung capacity (TLC) measurement system designed to be used by non-physician personnel is detailed. Four operators tested the reliability and validity of the system by measuring inspiratory PA and lateral pediatric chest radiographs with a Graf spark pen interfaced to a DEC VAX 11/780 computer. First results suggest that the ultimate goal of developing an accurate and easy to use TLC measurement system for non-physician personnel is attainable.

  4. Computer Aided Continuous Time Stochastic Process Modelling

    DEFF Research Database (Denmark)

    Kristensen, N.R.; Madsen, Henrik; Jørgensen, Sten Bay

    2001-01-01

    A grey-box approach to process modelling that combines deterministic and stochastic modelling is advocated for identification of models for model-based control of batch and semi-batch processes. A computer-aided tool designed for supporting decision-making within the corresponding modelling cycle...

  5. Real-time data-intensive computing

    Energy Technology Data Exchange (ETDEWEB)

    Parkinson, Dilworth Y., E-mail: dyparkinson@lbl.gov; Chen, Xian; Hexemer, Alexander; MacDowell, Alastair A.; Padmore, Howard A.; Shapiro, David; Tamura, Nobumichi [Advanced Light Source, Lawrence Berkeley National Laboratory, Berkeley, CA 94720 (United States); Beattie, Keith; Krishnan, Harinarayan; Patton, Simon J.; Perciano, Talita; Stromsness, Rune; Tull, Craig E.; Ushizima, Daniela [Computational Research Division, Lawrence Berkeley National Laboratory Berkeley CA 94720 (United States); Correa, Joaquin; Deslippe, Jack R. [National Energy Research Scientific Computing Center, Berkeley, CA 94720 (United States); Dart, Eli; Tierney, Brian L. [Energy Sciences Network, Berkeley, CA 94720 (United States); Daurer, Benedikt J.; Maia, Filipe R. N. C. [Uppsala University, Uppsala (Sweden); and others

    2016-07-27

    Today users visit synchrotrons as sources of understanding and discovery—not as sources of just light, and not as sources of data. To achieve this, the synchrotron facilities frequently provide not just light but often the entire end station and increasingly, advanced computational facilities that can reduce terabytes of data into a form that can reveal a new key insight. The Advanced Light Source (ALS) has partnered with high performance computing, fast networking, and applied mathematics groups to create a “super-facility”, giving users simultaneous access to the experimental, computational, and algorithmic resources to make this possible. This combination forms an efficient closed loop, where data—despite its high rate and volume—is transferred and processed immediately and automatically on appropriate computing resources, and results are extracted, visualized, and presented to users or to the experimental control system, both to provide immediate insight and to guide decisions about subsequent experiments during beamtime. We will describe our work at the ALS ptychography, scattering, micro-diffraction, and micro-tomography beamlines.

  6. Differentiation of acute total occlusion of coronary artery from chronic total occlusion in coronary computed tomography angiography

    Energy Technology Data Exchange (ETDEWEB)

    Kwag, Hyon Joo [Sungkyunkwan Univ. School of Medicine, Seoul (Korea, Republic of)

    2012-08-15

    To compare the features of coronary computed tomography angiography (CCTA) imaging of the patients with acute total occlusion (ATO) of coronary artery with those of chronic total occlusion (CTO). CCTA of 26 patients with complete interruption of the coronary artery in CCTA and occlusion in conventional coronary angiography, were retrospectively analyzed. Discrimination between the ATO group (n = 11, patients with non ST elevation myocardial infarction or unstable angina) and the CTO group (n = 15, patients with stable angina or nonspecific symptom) was arbitrarily determined by clinical diagnosis. Lesion length, remodeling index (RI), plaque density measured by Hounsfield units (HU), plaque composition, percentage attenuation drop across the lesion, and presence of myocardial thinning were evaluated. Comparisons between the ATO and CTO groups revealed significantly shorter lesion length in the ATO group (0.40 cm vs. 1.87 cm, respectively; p = 0.001), and significantly higher RI (1.56 vs. 1.10, respectively; p = 0.004). Plaque density of the ATO group was lower (37.0 HU vs. 104.7 HU, respectively; p < 0.001) and non calcified plaque was frequently seen in the ATO group (72.7% vs. 26.7%, respectively; p = 0.02). Percentage attenuation drop across the lesion was lower for the ATO group (10.92% vs. 25.44%, respectively; p = 0.005). Myocardial thinning was exclusively observed in the CTO group (seven of 15 patients, p = 0.01). CCTA shows various statistically significant differences between the ATO and CTO groups.

  7. Differentiation of acute total occlusion of coronary artery from chronic total occlusion in coronary computed tomography angiography

    International Nuclear Information System (INIS)

    Kwag, Hyon Joo

    2012-01-01

    To compare the features of coronary computed tomography angiography (CCTA) imaging of the patients with acute total occlusion (ATO) of coronary artery with those of chronic total occlusion (CTO). CCTA of 26 patients with complete interruption of the coronary artery in CCTA and occlusion in conventional coronary angiography, were retrospectively analyzed. Discrimination between the ATO group (n = 11, patients with non ST elevation myocardial infarction or unstable angina) and the CTO group (n = 15, patients with stable angina or nonspecific symptom) was arbitrarily determined by clinical diagnosis. Lesion length, remodeling index (RI), plaque density measured by Hounsfield units (HU), plaque composition, percentage attenuation drop across the lesion, and presence of myocardial thinning were evaluated. Comparisons between the ATO and CTO groups revealed significantly shorter lesion length in the ATO group (0.40 cm vs. 1.87 cm, respectively; p = 0.001), and significantly higher RI (1.56 vs. 1.10, respectively; p = 0.004). Plaque density of the ATO group was lower (37.0 HU vs. 104.7 HU, respectively; p < 0.001) and non calcified plaque was frequently seen in the ATO group (72.7% vs. 26.7%, respectively; p = 0.02). Percentage attenuation drop across the lesion was lower for the ATO group (10.92% vs. 25.44%, respectively; p = 0.005). Myocardial thinning was exclusively observed in the CTO group (seven of 15 patients, p = 0.01). CCTA shows various statistically significant differences between the ATO and CTO groups

  8. 29 CFR 779.253 - What is included in computing the total annual inflow volume.

    Science.gov (United States)

    2010-07-01

    ... FAIR LABOR STANDARDS ACT AS APPLIED TO RETAILERS OF GOODS OR SERVICES Employment to Which the Act May... taxes and other charges which the enterprise must pay for such goods. Generally, all charges will be... computing the total annual inflow volume. The goods which the establishment purchases or receives for resale...

  9. Real time computer system with distributed microprocessors

    International Nuclear Information System (INIS)

    Heger, D.; Steusloff, H.; Syrbe, M.

    1979-01-01

    The usual centralized structure of computer systems, especially of process computer systems, cannot sufficiently use the progress of very large-scale integrated semiconductor technology with respect to increasing the reliability and performance and to decreasing the expenses especially of the external periphery. This and the increasing demands on process control systems has led the authors to generally examine the structure of such systems and to adapt it to the new surroundings. Computer systems with distributed, optical fibre-coupled microprocessors allow a very favourable problem-solving with decentralized controlled buslines and functional redundancy with automatic fault diagnosis and reconfiguration. A fit programming system supports these hardware properties: PEARL for multicomputer systems, dynamic loader, processor and network operating system. The necessary design principles for this are proved mainly theoretically and by value analysis. An optimal overall system of this new generation of process control systems was established, supported by results of 2 PDV projects (modular operating systems, input/output colour screen system as control panel), for the purpose of testing by apllying the system for the control of 28 pit furnaces of a steel work. (orig.) [de

  10. Spying on real-time computers to improve performance

    International Nuclear Information System (INIS)

    Taff, L.M.

    1975-01-01

    The sampled program-counter histogram, an established technique for shortening the execution times of programs, is described for a real-time computer. The use of a real-time clock allows particularly easy implementation. (Auth.)

  11. Accessible high performance computing solutions for near real-time image processing for time critical applications

    Science.gov (United States)

    Bielski, Conrad; Lemoine, Guido; Syryczynski, Jacek

    2009-09-01

    High Performance Computing (HPC) hardware solutions such as grid computing and General Processing on a Graphics Processing Unit (GPGPU) are now accessible to users with general computing needs. Grid computing infrastructures in the form of computing clusters or blades are becoming common place and GPGPU solutions that leverage the processing power of the video card are quickly being integrated into personal workstations. Our interest in these HPC technologies stems from the need to produce near real-time maps from a combination of pre- and post-event satellite imagery in support of post-disaster management. Faster processing provides a twofold gain in this situation: 1. critical information can be provided faster and 2. more elaborate automated processing can be performed prior to providing the critical information. In our particular case, we test the use of the PANTEX index which is based on analysis of image textural measures extracted using anisotropic, rotation-invariant GLCM statistics. The use of this index, applied in a moving window, has been shown to successfully identify built-up areas in remotely sensed imagery. Built-up index image masks are important input to the structuring of damage assessment interpretation because they help optimise the workload. The performance of computing the PANTEX workflow is compared on two different HPC hardware architectures: (1) a blade server with 4 blades, each having dual quad-core CPUs and (2) a CUDA enabled GPU workstation. The reference platform is a dual CPU-quad core workstation and the PANTEX workflow total computing time is measured. Furthermore, as part of a qualitative evaluation, the differences in setting up and configuring various hardware solutions and the related software coding effort is presented.

  12. Computed tomography for preoperative planning in total hip arthroplasty: what radiologists need to know

    Energy Technology Data Exchange (ETDEWEB)

    Huppertz, Alexander [Charite - University Hospitals Berlin, Department of Radiology, Berlin (Germany); Imaging Science Institute Charite, Berlin (Germany); Radmer, Sebastian [Proendo, Orthopedic Surgery, Berlin (Germany); Wagner, Moritz; Hamm, Bernd [Charite - University Hospitals Berlin, Department of Radiology, Berlin (Germany); Roessler, Torsten [Klinikum Ernst von Bergmann, Department of Trauma and Orthopedic Surgery, Potsdam (Germany); Sparmann, Martin [Proendo, Orthopedic Surgery, Berlin (Germany); Charite - University Hospital, Berlin (Germany)

    2014-08-15

    The number of total hip arthroplasties is continuously rising. Although less invasive surgical techniques, sophisticated component design, and intraoperative navigation techniques have been introduced, the rate of peri- and postoperative complications, including dislocations, fractures, nerve palsies, and infections, is still a major clinical problem. Better patient outcome, faster recovery and rehabilitation, and shorter operation times therefore remain to be accomplished. A promising strategy is to use minimally invasive techniques in conjunction with modular implants, aimed at independently reconstructing femoral offset and leg length on the basis of highly accurate preoperative planning. Plain radiographs have clear limitations for the correct estimation of hip joint geometry and bone quality. Three-dimensional assessment based on computed tomography (CT) allows optimizing the choice and positions of implants and anticipating difficulties to be encountered during surgery. Postoperative CT is used to monitor operative translation and plays a role in arthroplastic quality management. Radiologists should be familiar with the needs of orthopedic surgeons in terms of CT acquisition, post-processing, and data transfer. The CT protocol should be optimized to enhance image quality and reduce radiation exposure. When dedicated orthopedic CT protocols and state-of-the-art scanner hardware are used, radiation exposure can be decreased to a level just marginally higher than that of conventional preoperative radiography. Surgeons and radiologists should use similar terminology to avoid misunderstanding and inaccuracies in the transfer of preoperative planning. (orig.)

  13. Results of computer assisted mini-incision subvastus approach for total knee arthroplasty.

    Science.gov (United States)

    Turajane, Thana; Larbpaiboonpong, Viroj; Kongtharvonskul, Jatupon; Maungsiri, Samart

    2009-12-01

    Mini-incision subvastus approach is soft tissue preservation of the knee. Advantages of the mini-incision subvastus approach included reduced blood loss, reduced pain, self rehabilitation and faster recovery. However, the improved visualization, component alignment, and more blood preservation have been debatable to achieve the better outcome and preventing early failure of the Total Knee Arthroplasty (TKA). The computer navigation has been introduced to improve alignment and blood loss. The purpose of this study was to evaluate the short term outcomes of the combination of computer assisted mini-incision subvastus approach for Total Knee Arthroplasty (CMS-TKA). A prospective case series of the initial 80 patients who underwent computer assisted mini-incision subvastus approach for CMS-TKA from January 2007 to October 2008 was carried out. The patients' conditions were classified into 2 groups, the simple OA knee (varus deformity was less than 15 degree, BMI was less than 20%, no associated deformities) and the complex deformity (varus deformity was more than 15 degrees, BMI more was than 20%, associated with flexion contractor). There were 59 patients in group 1 and 21 patients in group 2. Of the 80 knees, 38 were on the left and 42 on the right. The results of CMS-TKA [the mean (range)] in group 1: group 2 were respectively shown as the incision length [10.88 (8-13): 11.92 (10-14], the operation time [118 (111.88-125.12): 131 (119.29-143.71) minutes, lateral releases (0 in both groups), postoperative range of motion in flexion [94.5 (90-100): 95.25 (90-105) degree] and extension [1.75 (0-5): 1.5 (0-5) degree] Blood loss in 24 hours [489.09 (414.7-563.48): 520 (503.46-636.54) ml] and blood transfusion [1 (0-1) unit? in both groups], Tibiofemoral angle preoperative [Varus = 4 (varus 0-10): Varus = 17.14 (varus 15.7-18.5) degree, Tibiofemoral angle postoperative [Valgus = 1.38 (Valgus 0-4): Valgus = 2.85 (valgus 2.1-3.5) degree], Tibiofemoral angle outlier (85% both

  14. Simplified neural networks for solving linear least squares and total least squares problems in real time.

    Science.gov (United States)

    Cichocki, A; Unbehauen, R

    1994-01-01

    In this paper a new class of simplified low-cost analog artificial neural networks with on chip adaptive learning algorithms are proposed for solving linear systems of algebraic equations in real time. The proposed learning algorithms for linear least squares (LS), total least squares (TLS) and data least squares (DLS) problems can be considered as modifications and extensions of well known algorithms: the row-action projection-Kaczmarz algorithm and/or the LMS (Adaline) Widrow-Hoff algorithms. The algorithms can be applied to any problem which can be formulated as a linear regression problem. The correctness and high performance of the proposed neural networks are illustrated by extensive computer simulation results.

  15. Exact and Heuristic Solutions to Minimize Total Waiting Time in the Blood Products Distribution Problem

    Directory of Open Access Journals (Sweden)

    Amir Salehipour

    2012-01-01

    Full Text Available This paper presents a novel application of operations research to support decision making in blood distribution management. The rapid and dynamic increasing demand, criticality of the product, storage, handling, and distribution requirements, and the different geographical locations of hospitals and medical centers have made blood distribution a complex and important problem. In this study, a real blood distribution problem containing 24 hospitals was tackled by the authors, and an exact approach was presented. The objective of the problem is to distribute blood and its products among hospitals and medical centers such that the total waiting time of those requiring the product is minimized. Following the exact solution, a hybrid heuristic algorithm is proposed. Computational experiments showed the optimal solutions could be obtained for medium size instances, while for larger instances the proposed hybrid heuristic is very competitive.

  16. Improving the Prediction of Total Surgical Procedure Time Using Linear Regression Modeling.

    Science.gov (United States)

    Edelman, Eric R; van Kuijk, Sander M J; Hamaekers, Ankie E W; de Korte, Marcel J M; van Merode, Godefridus G; Buhre, Wolfgang F F A

    2017-01-01

    For efficient utilization of operating rooms (ORs), accurate schedules of assigned block time and sequences of patient cases need to be made. The quality of these planning tools is dependent on the accurate prediction of total procedure time (TPT) per case. In this paper, we attempt to improve the accuracy of TPT predictions by using linear regression models based on estimated surgeon-controlled time (eSCT) and other variables relevant to TPT. We extracted data from a Dutch benchmarking database of all surgeries performed in six academic hospitals in The Netherlands from 2012 till 2016. The final dataset consisted of 79,983 records, describing 199,772 h of total OR time. Potential predictors of TPT that were included in the subsequent analysis were eSCT, patient age, type of operation, American Society of Anesthesiologists (ASA) physical status classification, and type of anesthesia used. First, we computed the predicted TPT based on a previously described fixed ratio model for each record, multiplying eSCT by 1.33. This number is based on the research performed by van Veen-Berkx et al., which showed that 33% of SCT is generally a good approximation of anesthesia-controlled time (ACT). We then systematically tested all possible linear regression models to predict TPT using eSCT in combination with the other available independent variables. In addition, all regression models were again tested without eSCT as a predictor to predict ACT separately (which leads to TPT by adding SCT). TPT was most accurately predicted using a linear regression model based on the independent variables eSCT, type of operation, ASA classification, and type of anesthesia. This model performed significantly better than the fixed ratio model and the method of predicting ACT separately. Making use of these more accurate predictions in planning and sequencing algorithms may enable an increase in utilization of ORs, leading to significant financial and productivity related benefits.

  17. Improving the Prediction of Total Surgical Procedure Time Using Linear Regression Modeling

    Directory of Open Access Journals (Sweden)

    Eric R. Edelman

    2017-06-01

    Full Text Available For efficient utilization of operating rooms (ORs, accurate schedules of assigned block time and sequences of patient cases need to be made. The quality of these planning tools is dependent on the accurate prediction of total procedure time (TPT per case. In this paper, we attempt to improve the accuracy of TPT predictions by using linear regression models based on estimated surgeon-controlled time (eSCT and other variables relevant to TPT. We extracted data from a Dutch benchmarking database of all surgeries performed in six academic hospitals in The Netherlands from 2012 till 2016. The final dataset consisted of 79,983 records, describing 199,772 h of total OR time. Potential predictors of TPT that were included in the subsequent analysis were eSCT, patient age, type of operation, American Society of Anesthesiologists (ASA physical status classification, and type of anesthesia used. First, we computed the predicted TPT based on a previously described fixed ratio model for each record, multiplying eSCT by 1.33. This number is based on the research performed by van Veen-Berkx et al., which showed that 33% of SCT is generally a good approximation of anesthesia-controlled time (ACT. We then systematically tested all possible linear regression models to predict TPT using eSCT in combination with the other available independent variables. In addition, all regression models were again tested without eSCT as a predictor to predict ACT separately (which leads to TPT by adding SCT. TPT was most accurately predicted using a linear regression model based on the independent variables eSCT, type of operation, ASA classification, and type of anesthesia. This model performed significantly better than the fixed ratio model and the method of predicting ACT separately. Making use of these more accurate predictions in planning and sequencing algorithms may enable an increase in utilization of ORs, leading to significant financial and productivity related

  18. Hybrid Cloud Computing Architecture Optimization by Total Cost of Ownership Criterion

    Directory of Open Access Journals (Sweden)

    Elena Valeryevna Makarenko

    2014-12-01

    Full Text Available Achieving the goals of information security is a key factor in the decision to outsource information technology and, in particular, to decide on the migration of organizational data, applications, and other resources to the infrastructure, based on cloud computing. And the key issue in the selection of optimal architecture and the subsequent migration of business applications and data to the cloud organization information environment is the question of the total cost of ownership of IT infrastructure. This paper focuses on solving the problem of minimizing the total cost of ownership cloud.

  19. 29 CFR 1921.22 - Computation of time.

    Science.gov (United States)

    2010-07-01

    ... 29 Labor 7 2010-07-01 2010-07-01 false Computation of time. 1921.22 Section 1921.22 Labor Regulations Relating to Labor (Continued) OCCUPATIONAL SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR... WORKERS' COMPENSATION ACT Miscellaneous § 1921.22 Computation of time. Sundays and holidays shall be...

  20. 43 CFR 45.3 - How are time periods computed?

    Science.gov (United States)

    2010-10-01

    ... 43 Public Lands: Interior 1 2010-10-01 2010-10-01 false How are time periods computed? 45.3... IN FERC HYDROPOWER LICENSES General Provisions § 45.3 How are time periods computed? (a) General... run is not included. (2) The last day of the period is included. (i) If that day is a Saturday, Sunday...

  1. 5 CFR 890.101 - Definitions; time computations.

    Science.gov (United States)

    2010-01-01

    ... 5 Administrative Personnel 2 2010-01-01 2010-01-01 false Definitions; time computations. 890.101....101 Definitions; time computations. (a) In this part, the terms annuitant, carrier, employee, employee... in section 8901 of title 5, United States Code, and supplement the following definitions: Appropriate...

  2. Development of a totally computer-controlled triple quadrupole mass spectrometer system

    International Nuclear Information System (INIS)

    Wong, C.M.; Crawford, R.W.; Barton, V.C.; Brand, H.R.; Neufeld, K.W.; Bowman, J.E.

    1983-01-01

    A totally computer-controlled triple quadrupole mass spectrometer (TQMS) is described. It has a number of unique features not available on current commercial instruments, including: complete computer control of source and all ion axial potentials; use of dual computers for data acquisition and data processing; and capability for self-adaptive control of experiments. Furthermore, it has been possible to produce this instrument at a cost significantly below that of commercial instruments. This triple quadrupole mass spectrometer has been constructed using components commercially available from several different manufacturers. The source is a standard Hewlett-Packard 5985B GC/MS source. The two quadrupole analyzers and the quadrupole CAD region contain Balzers QMA 150 rods with Balzers QMG 511 rf controllers for the analyzers and a Balzers QHS-511 controller for the CAD region. The pulsed-positive-ion-negative-ion-chemical ionization (PPINICI) detector is made by Finnigan Corporation. The mechanical and electronics design were developed at LLNL for linking these diverse elements into a functional TQMS as described. The computer design for total control of the system is unique in that two separate LSI-11/23 minicomputers and assorted I/O peripherals and interfaces from several manufacturers are used. The evolution of this design concept from totally computer-controlled instrumentation into future self-adaptive or ''expert'' systems for instrumental analysis is described. Operational characteristics of the instrument and initial results from experiments involving the analysis of the high explosive HMX (1,3,5,7-Tetranitro-1,3,5,7-Tetrazacyclooctane) are presented

  3. Time computations in anuran auditory systems

    Directory of Open Access Journals (Sweden)

    Gary J Rose

    2014-05-01

    Full Text Available Temporal computations are important in the acoustic communication of anurans. In many cases, calls between closely related species are nearly identical spectrally but differ markedly in temporal structure. Depending on the species, calls can differ in pulse duration, shape and/or rate (i.e., amplitude modulation, direction and rate of frequency modulation, and overall call duration. Also, behavioral studies have shown that anurans are able to discriminate between calls that differ in temporal structure. In the peripheral auditory system, temporal information is coded primarily in the spatiotemporal patterns of activity of auditory-nerve fibers. However, major transformations in the representation of temporal information occur in the central auditory system. In this review I summarize recent advances in understanding how temporal information is represented in the anuran midbrain, with particular emphasis on mechanisms that underlie selectivity for pulse duration and pulse rate (i.e., intervals between onsets of successive pulses. Two types of neurons have been identified that show selectivity for pulse rate: long-interval cells respond well to slow pulse rates but fail to spike or respond phasically to fast pulse rates; conversely, interval-counting neurons respond to intermediate or fast pulse rates, but only after a threshold number of pulses, presented at optimal intervals, have occurred. Duration-selectivity is manifest as short-pass, band-pass or long-pass tuning. Whole-cell patch recordings, in vivo, suggest that excitation and inhibition are integrated in diverse ways to generate temporal selectivity. In many cases, activity-related enhancement or depression of excitatory or inhibitory processes appear to contribute to selective responses.

  4. Advanced real time radioscopy and computed tomography

    International Nuclear Information System (INIS)

    Sauerwein, Ch.; Nuding, W.; Grimm, R.; Wiacker, H.

    1996-01-01

    The paper describes three x-ray inspection systems. One radioscopic system is designed for the inspection of castings. The next integrates a radioscopic and a tomographic mode. The radioscopy has a high resolution camera and real time image processor. Radiation sources are a 450 kV industrial and a 200 kV microfocus tube. The third system is a tomographic system with 30 scintillation detectors for the inspection of nuclear waste containers. (author)

  5. Real time computer controlled weld skate

    Science.gov (United States)

    Wall, W. A., Jr.

    1977-01-01

    A real time, adaptive control, automatic welding system was developed. This system utilizes the general case geometrical relationships between a weldment and a weld skate to precisely maintain constant weld speed and torch angle along a contoured workplace. The system is compatible with the gas tungsten arc weld process or can be adapted to other weld processes. Heli-arc cutting and machine tool routing operations are possible applications.

  6. Application of total care time and payment per unit time model for physician reimbursement for common general surgery operations.

    Science.gov (United States)

    Chatterjee, Abhishek; Holubar, Stefan D; Figy, Sean; Chen, Lilian; Montagne, Shirley A; Rosen, Joseph M; Desimone, Joseph P

    2012-06-01

    The relative value unit system relies on subjective measures of physician input in the care of patients. A payment per unit time model incorporates surgeon reimbursement to the total care time spent in the operating room, postoperative in-house, and clinic time to define payment per unit time. We aimed to compare common general surgery operations by using the total care time and payment per unit time method in order to demonstrate a more objective measurement for physician reimbursement. Average total physician payment per case was obtained for 5 outpatient operations and 4 inpatient operations in general surgery. Total care time was defined as the sum of operative time, 30 minutes per hospital day, and 30 minutes per office visit for each operation. Payment per unit time was calculated by dividing the physician reimbursement per case by the total care time. Total care time, physician payment per case, and payment per unit time for each type of operation demonstrated that an average payment per time spent for inpatient operations was $455.73 and slightly more at $467.51 for outpatient operations. Partial colectomy with primary anastomosis had the longest total care time (8.98 hours) and the least payment per unit time ($188.52). Laparoscopic gastric bypass had the highest payment per time ($707.30). The total care time and payment per unit time method can be used as an adjunct to compare reimbursement among different operations on an institutional level as well as on a national level. Although many operations have similar payment trends based on time spent by the surgeon, payment differences using this methodology are seen and may be in need of further review. Copyright © 2012 American College of Surgeons. Published by Elsevier Inc. All rights reserved.

  7. An empirical method for peak-to-total ratio computation of a gamma-ray detector

    International Nuclear Information System (INIS)

    Cesana, A.; Terrani, M.

    1989-01-01

    A simple expression for peak-to-total ratio evaluation of gamma-ray detectors in the energy range 0.3-10 MeV is proposed. The quantities one needs to know for the computation are: Detector dimensions and chemical composition, photon corss sections and an empirical energy dependent function which is valid for all the detector materials considered. This procedure seems able to produce peak-to-total values with an accuracy comparable with the most sophisticated Monte Carlo calculations. It has been tested using experimental peak-to-total values of Ge, NaI, CsI and BGO detectors but it is reasonable to suppose that it is valid for any detector material. (orig.)

  8. Recent achievements in real-time computational seismology in Taiwan

    Science.gov (United States)

    Lee, S.; Liang, W.; Huang, B.

    2012-12-01

    Real-time computational seismology is currently possible to be achieved which needs highly connection between seismic database and high performance computing. We have developed a real-time moment tensor monitoring system (RMT) by using continuous BATS records and moment tensor inversion (CMT) technique. The real-time online earthquake simulation service is also ready to open for researchers and public earthquake science education (ROS). Combine RMT with ROS, the earthquake report based on computational seismology can provide within 5 minutes after an earthquake occurred (RMT obtains point source information ROS completes a 3D simulation real-time now. For more information, welcome to visit real-time computational seismology earthquake report webpage (RCS).

  9. Total knee arthroplasty with computer-assisted navigation: an analysis of 200 cases,

    Directory of Open Access Journals (Sweden)

    Marcus Vinicius Malheiros Luzo

    2014-04-01

    Full Text Available OBJECTIVE: to evaluate the results from surgery with computer-assisted navigation in cases of total knee arthroplasty.METHOD: a total of 196 patients who underwent total knee arthroplasty with computer-assisted navigation were evaluated. The extension and flexion spaces (gaps were evaluated during the operation and the alignment after the operation was assessed. The Knee Society Score (KSS questionnaire for assessing patient's function was applied preoperatively and postoperatively after a mean follow-up of 22 months.RESULTS: in all, 86.7% of the patients presented good alignment of the mechanical axis (less than 3◦ of varus or valgus in relation to the mechanical axis and 96.4% of the patients presented balanced flexion and extension gaps. Before the operation, 97% of the patients presented poor or insufficient KSS, but after the operation, 77.6% presented good or excellent KSS.CONCLUSION: the navigation system made it possible to achieve aligned and balanced implants, with notable functional improvement among the patients. It was found to be useful in assessing, understanding and improving knowledge in relation to performing arthroplasty procedures.

  10. Time-of-Flight Cameras in Computer Graphics

    DEFF Research Database (Denmark)

    Kolb, Andreas; Barth, Erhardt; Koch, Reinhard

    2010-01-01

    Computer Graphics, Computer Vision and Human Machine Interaction (HMI). These technologies are starting to have an impact on research and commercial applications. The upcoming generation of ToF sensors, however, will be even more powerful and will have the potential to become “ubiquitous real-time geometry...

  11. 29 CFR 4245.8 - Computation of time.

    Science.gov (United States)

    2010-07-01

    ... 29 Labor 9 2010-07-01 2010-07-01 false Computation of time. 4245.8 Section 4245.8 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION INSOLVENCY, REORGANIZATION, TERMINATION, AND OTHER RULES APPLICABLE TO MULTIEMPLOYER PLANS NOTICE OF INSOLVENCY § 4245.8 Computation of...

  12. Cross-sectional associations of total sitting and leisure screen time with cardiometabolic risk in adults. Results from the HUNT Study, Norway

    NARCIS (Netherlands)

    Chau, J.Y.; Grunseit, A.; Midthjell, K.; Holmen, J.; Holmen, T.L.; Bauman, A.E.; van der Ploeg, H.P.

    2014-01-01

    Objectives: To examine associations of total sitting time, TV-viewing and leisure-time computer use with cardiometabolic risk biomarkers in adults. Design: Population based cross-sectional study. Methods: Waist circumference, BMI, total cholesterol, HDL cholesterol, blood pressure, non-fasting

  13. Total variation regularization for a backward time-fractional diffusion problem

    International Nuclear Information System (INIS)

    Wang, Liyan; Liu, Jijun

    2013-01-01

    Consider a two-dimensional backward problem for a time-fractional diffusion process, which can be considered as image de-blurring where the blurring process is assumed to be slow diffusion. In order to avoid the over-smoothing effect for object image with edges and to construct a fast reconstruction scheme, the total variation regularizing term and the data residual error in the frequency domain are coupled to construct the cost functional. The well posedness of this optimization problem is studied. The minimizer is sought approximately using the iteration process for a series of optimization problems with Bregman distance as a penalty term. This iteration reconstruction scheme is essentially a new regularizing scheme with coupling parameter in the cost functional and the iteration stopping times as two regularizing parameters. We give the choice strategy for the regularizing parameters in terms of the noise level of measurement data, which yields the optimal error estimate on the iterative solution. The series optimization problems are solved by alternative iteration with explicit exact solution and therefore the amount of computation is much weakened. Numerical implementations are given to support our theoretical analysis on the convergence rate and to show the significant reconstruction improvements. (paper)

  14. Noise-constrained switching times for heteroclinic computing

    Science.gov (United States)

    Neves, Fabio Schittler; Voit, Maximilian; Timme, Marc

    2017-03-01

    Heteroclinic computing offers a novel paradigm for universal computation by collective system dynamics. In such a paradigm, input signals are encoded as complex periodic orbits approaching specific sequences of saddle states. Without inputs, the relevant states together with the heteroclinic connections between them form a network of states—the heteroclinic network. Systems of pulse-coupled oscillators or spiking neurons naturally exhibit such heteroclinic networks of saddles, thereby providing a substrate for general analog computations. Several challenges need to be resolved before it becomes possible to effectively realize heteroclinic computing in hardware. The time scales on which computations are performed crucially depend on the switching times between saddles, which in turn are jointly controlled by the system's intrinsic dynamics and the level of external and measurement noise. The nonlinear dynamics of pulse-coupled systems often strongly deviate from that of time-continuously coupled (e.g., phase-coupled) systems. The factors impacting switching times in pulse-coupled systems are still not well understood. Here we systematically investigate switching times in dependence of the levels of noise and intrinsic dissipation in the system. We specifically reveal how local responses to pulses coact with external noise. Our findings confirm that, like in time-continuous phase-coupled systems, piecewise-continuous pulse-coupled systems exhibit switching times that transiently increase exponentially with the number of switches up to some order of magnitude set by the noise level. Complementarily, we show that switching times may constitute a good predictor for the computation reliability, indicating how often an input signal must be reiterated. By characterizing switching times between two saddles in conjunction with the reliability of a computation, our results provide a first step beyond the coding of input signal identities toward a complementary coding for

  15. Relativistic Photoionization Computations with the Time Dependent Dirac Equation

    Science.gov (United States)

    2016-10-12

    Naval Research Laboratory Washington, DC 20375-5320 NRL/MR/6795--16-9698 Relativistic Photoionization Computations with the Time Dependent Dirac... Photoionization Computations with the Time Dependent Dirac Equation Daniel F. Gordon and Bahman Hafizi Naval Research Laboratory 4555 Overlook Avenue, SW...Unclassified Unlimited Unclassified Unlimited 22 Daniel Gordon (202) 767-5036 Tunneling Photoionization Ionization of inner shell electrons by laser

  16. Cross-sectional associations of total sitting and leisure screen time with cardiometabolic risk in adults. Results from the HUNT Study, Norway.

    Science.gov (United States)

    Chau, Josephine Y; Grunseit, Anne; Midthjell, Kristian; Holmen, Jostein; Holmen, Turid L; Bauman, Adrian E; van der Ploeg, Hidde P

    2014-01-01

    To examine associations of total sitting time, TV-viewing and leisure-time computer use with cardiometabolic risk biomarkers in adults. Population based cross-sectional study. Waist circumference, BMI, total cholesterol, HDL cholesterol, blood pressure, non-fasting glucose, gamma glutamyltransferase (GGT) and triglycerides were measured in 48,882 adults aged 20 years or older from the Nord-Trøndelag Health Study 2006-2008 (HUNT3). Adjusted multiple regression models were used to test for associations between these biomarkers and self-reported total sitting time, TV-viewing and leisure-time computer use in the whole sample and by cardiometabolic disease status sub-groups. In the whole sample, reporting total sitting time ≥10 h/day was associated with poorer BMI, waist circumference, total cholesterol, HDL cholesterol, diastolic blood pressure, systolic blood pressure, non-fasting glucose, GGT and triglyceride levels compared to those reporting total sitting time Leisure-time computer use ≥1 h/day was associated with poorer BMI, total cholesterol, diastolic blood pressure, GGT and triglycerides compared with those reporting no leisure-time computing. Sub-group analyses by cardiometabolic disease status showed similar patterns in participants free of cardiometabolic disease, while similar albeit non-significant patterns were observed in those with cardiometabolic disease. Total sitting time, TV-viewing and leisure-time computer use are associated with poorer cardiometabolic risk profiles in adults. Reducing sedentary behaviour throughout the day and limiting TV-viewing and leisure-time computer use may have health benefits. Copyright © 2013 Sports Medicine Australia. Published by Elsevier Ltd. All rights reserved.

  17. Use of inpatient continuous passive motion versus no CPM in computer-assisted total knee arthroplasty.

    Science.gov (United States)

    Alkire, Martha R; Swank, Michael L

    2010-01-01

    Continuous passive motion (CPM) has shown positive effects on tissue healing, edema, hemarthrosis, and joint function (L. Brosseau et al., 2004). CPM has also been shown to increase short-term early flexion and decrease length of stay (LOS) ( L. Brosseau et al., 2004; C. M. Chiarello, C. M. S. Gundersen, & T. O'Halloran, 2004). The benefits of CPM for the population of patients undergoing computer-assisted total knee arthroplasty (TKA) have not been examined. The primary objective of this study was to determine whether the use of CPM following computer-assisted TKA resulted in differences in range of motion, edema/drainage, functional ability, and pain. This was an experimental, prospective, randomized study of patients undergoing unilateral, computer-assisted TKA. The experimental group received CPM thrice daily and physical therapy (PT) twice daily during their hospitalization. The control group received PT twice daily and no CPM during the hospital stay. Both groups received PT after discharge. Measurement included Knee Society scores, Western Ontario McMaster Osteoarthritis Index values, range of motion, knee circumference, and HemoVac drainage. Data were collected at various intervals from preoperatively through 3 months. Although the control group was found to be higher functioning preoperatively, there was no statistically significant difference in flexion, edema or drainage, function, or pain between groups through the 3-month study period.

  18. A Distributed Computing Network for Real-Time Systems.

    Science.gov (United States)

    1980-11-03

    7 ) AU2 o NAVA TUNDEWATER SY$TEMS CENTER NEWPORT RI F/G 9/2 UIS RIBUT E 0 COMPUTIN G N LTWORK FOR REAL - TIME SYSTEMS .(U) UASSIFIED NOV Al 6 1...MORAIS - UT 92 dLEVEL c A Distributed Computing Network for Real - Time Systems . 11 𔃺-1 Gordon E/Morson I7 y tm- ,r - t "en t As J 2 -p .. - 7 I’ cNaval...NUMBER TD 5932 / N 4. TITLE mand SubotI. S. TYPE OF REPORT & PERIOD COVERED A DISTRIBUTED COMPUTING NETWORK FOR REAL - TIME SYSTEMS 6. PERFORMING ORG

  19. Computer simulations of long-time tails: what's new?

    NARCIS (Netherlands)

    Hoef, van der M.A.; Frenkel, D.

    1995-01-01

    Twenty five years ago Alder and Wainwright discovered, by simulation, the 'long-time tails' in the velocity autocorrelation function of a single particle in fluid [1]. Since then, few qualitatively new results on long-time tails have been obtained by computer simulations. However, within the

  20. Computation of reactor control rod drop time under accident conditions

    International Nuclear Information System (INIS)

    Dou Yikang; Yao Weida; Yang Renan; Jiang Nanyan

    1998-01-01

    The computational method of reactor control rod drop time under accident conditions lies mainly in establishing forced vibration equations for the components under action of outside forces on control rod driven line and motion equation for the control rod moving in vertical direction. The above two kinds of equations are connected by considering the impact effects between control rod and its outside components. Finite difference method is adopted to make discretization of the vibration equations and Wilson-θ method is applied to deal with the time history problem. The non-linearity caused by impact is iteratively treated with modified Newton method. Some experimental results are used to validate the validity and reliability of the computational method. Theoretical and experimental testing problems show that the computer program based on the computational method is applicable and reliable. The program can act as an effective tool of design by analysis and safety analysis for the relevant components

  1. Scalable space-time adaptive simulation tools for computational electrocardiology

    OpenAIRE

    Krause, Dorian; Krause, Rolf

    2013-01-01

    This work is concerned with the development of computational tools for the solution of reaction-diffusion equations from the field of computational electrocardiology. We designed lightweight spatially and space-time adaptive schemes for large-scale parallel simulations. We propose two different adaptive schemes based on locally structured meshes, managed either via a conforming coarse tessellation or a forest of shallow trees. A crucial ingredient of our approach is a non-conforming morta...

  2. Continuous-Time Symmetric Hopfield Nets are Computationally Universal

    Czech Academy of Sciences Publication Activity Database

    Šíma, Jiří; Orponen, P.

    2003-01-01

    Roč. 15, č. 3 (2003), s. 693-733 ISSN 0899-7667 R&D Projects: GA AV ČR IAB2030007; GA ČR GA201/02/1456 Institutional research plan: AV0Z1030915 Keywords : continuous-time Hopfield network * Liapunov function * analog computation * computational power * Turing universality Subject RIV: BA - General Mathematics Impact factor: 2.747, year: 2003

  3. Parallel algorithm of real-time infrared image restoration based on total variation theory

    Science.gov (United States)

    Zhu, Ran; Li, Miao; Long, Yunli; Zeng, Yaoyuan; An, Wei

    2015-10-01

    Image restoration is a necessary preprocessing step for infrared remote sensing applications. Traditional methods allow us to remove the noise but penalize too much the gradients corresponding to edges. Image restoration techniques based on variational approaches can solve this over-smoothing problem for the merits of their well-defined mathematical modeling of the restore procedure. The total variation (TV) of infrared image is introduced as a L1 regularization term added to the objective energy functional. It converts the restoration process to an optimization problem of functional involving a fidelity term to the image data plus a regularization term. Infrared image restoration technology with TV-L1 model exploits the remote sensing data obtained sufficiently and preserves information at edges caused by clouds. Numerical implementation algorithm is presented in detail. Analysis indicates that the structure of this algorithm can be easily implemented in parallelization. Therefore a parallel implementation of the TV-L1 filter based on multicore architecture with shared memory is proposed for infrared real-time remote sensing systems. Massive computation of image data is performed in parallel by cooperating threads running simultaneously on multiple cores. Several groups of synthetic infrared image data are used to validate the feasibility and effectiveness of the proposed parallel algorithm. Quantitative analysis of measuring the restored image quality compared to input image is presented. Experiment results show that the TV-L1 filter can restore the varying background image reasonably, and that its performance can achieve the requirement of real-time image processing.

  4. Heterogeneous real-time computing in radio astronomy

    Science.gov (United States)

    Ford, John M.; Demorest, Paul; Ransom, Scott

    2010-07-01

    Modern computer architectures suited for general purpose computing are often not the best choice for either I/O-bound or compute-bound problems. Sometimes the best choice is not to choose a single architecture, but to take advantage of the best characteristics of different computer architectures to solve your problems. This paper examines the tradeoffs between using computer systems based on the ubiquitous X86 Central Processing Units (CPU's), Field Programmable Gate Array (FPGA) based signal processors, and Graphical Processing Units (GPU's). We will show how a heterogeneous system can be produced that blends the best of each of these technologies into a real-time signal processing system. FPGA's tightly coupled to analog-to-digital converters connect the instrument to the telescope and supply the first level of computing to the system. These FPGA's are coupled to other FPGA's to continue to provide highly efficient processing power. Data is then packaged up and shipped over fast networks to a cluster of general purpose computers equipped with GPU's, which are used for floating-point intensive computation. Finally, the data is handled by the CPU and written to disk, or further processed. Each of the elements in the system has been chosen for its specific characteristics and the role it can play in creating a system that does the most for the least, in terms of power, space, and money.

  5. Highly reliable computer network for real time system

    International Nuclear Information System (INIS)

    Mohammed, F.A.; Omar, A.A.; Ayad, N.M.A.; Madkour, M.A.I.; Ibrahim, M.K.

    1988-01-01

    Many of computer networks have been studied different trends regarding the network architecture and the various protocols that govern data transfers and guarantee a reliable communication among all a hierarchical network structure has been proposed to provide a simple and inexpensive way for the realization of a reliable real-time computer network. In such architecture all computers in the same level are connected to a common serial channel through intelligent nodes that collectively control data transfers over the serial channel. This level of computer network can be considered as a local area computer network (LACN) that can be used in nuclear power plant control system since it has geographically dispersed subsystems. network expansion would be straight the common channel for each added computer (HOST). All the nodes are designed around a microprocessor chip to provide the required intelligence. The node can be divided into two sections namely a common section that interfaces with serial data channel and a private section to interface with the host computer. This part would naturally tend to have some variations in the hardware details to match the requirements of individual host computers. fig 7

  6. Minimizing total weighted tardiness for the single machine scheduling problem with dependent setup time and precedence constraints

    Directory of Open Access Journals (Sweden)

    Hamidreza Haddad

    2012-04-01

    Full Text Available This paper tackles the single machine scheduling problem with dependent setup time and precedence constraints. The primary objective of this paper is minimization of total weighted tardiness. Since the complexity of the resulted problem is NP-hard we use metaheuristics method to solve the resulted model. The proposed model of this paper uses genetic algorithm to solve the problem in reasonable amount of time. Because of high sensitivity of GA to its initial values of parameters, a Taguchi approach is presented to calibrate its parameters. Computational experiments validate the effectiveness and capability of proposed method.

  7. Computing return times or return periods with rare event algorithms

    Science.gov (United States)

    Lestang, Thibault; Ragone, Francesco; Bréhier, Charles-Edouard; Herbert, Corentin; Bouchet, Freddy

    2018-04-01

    The average time between two occurrences of the same event, referred to as its return time (or return period), is a useful statistical concept for practical applications. For instance insurances or public agencies may be interested by the return time of a 10 m flood of the Seine river in Paris. However, due to their scarcity, reliably estimating return times for rare events is very difficult using either observational data or direct numerical simulations. For rare events, an estimator for return times can be built from the extrema of the observable on trajectory blocks. Here, we show that this estimator can be improved to remain accurate for return times of the order of the block size. More importantly, we show that this approach can be generalised to estimate return times from numerical algorithms specifically designed to sample rare events. So far those algorithms often compute probabilities, rather than return times. The approach we propose provides a computationally extremely efficient way to estimate numerically the return times of rare events for a dynamical system, gaining several orders of magnitude of computational costs. We illustrate the method on two kinds of observables, instantaneous and time-averaged, using two different rare event algorithms, for a simple stochastic process, the Ornstein–Uhlenbeck process. As an example of realistic applications to complex systems, we finally discuss extreme values of the drag on an object in a turbulent flow.

  8. A computational parametric study on edge loading in ceramic-on-ceramic total hip joint replacements.

    Science.gov (United States)

    Liu, Feng; Feng, Li; Wang, Junyuan

    2018-07-01

    Edge loading in ceramic-on-ceramic total hip joint replacement is an adverse condition that occurs as the result of a direct contact between the head and the cup rim. It has been associated with translational mismatch in the centres of rotation of the cup and head, and found to cause severe wear and early failure of the implants. Edge loading has been considered in particular in relation to dynamic separation of the cup and head centres during a gait cycle. Research has been carried out both experimentally and computationally to understand the mechanism including the influence of bearing component positioning on the occurrence and severity of edge loading. However, it is experimentally difficult to measure both the load magnitude and duration of edge loading as it occurs as a short impact within the tight space of hip joints. Computationally, a dynamic contact model, for example, developed using the MSC ADAMS software for a multi-body dynamics simulation can be particularly useful for calculating the loads and characterising the edge loading. The aim of the present study was to further develop the computational model, and improve the predictions of contact force and the understanding of mechanism in order to provide guidance on design and surgical factors to avoid or to reduce edge loading and wear. The results have shown that edge loading can be avoided for a low range of translational mismatch in the centres of rotation of the cup and head during gait at the level of approximately 1.0 mm for a cup at 45° inclination, keeping a correct cup inclination at 45° is important to reduce the edge loading severity, and edge loading can be avoided for a certain range of translational mismatch of the cup and head centres with an increased swing phase load. Copyright © 2018 Elsevier Ltd. All rights reserved.

  9. Total knee arthroplasty with a computer-navigated saw: a pilot study.

    Science.gov (United States)

    Garvin, Kevin L; Barrera, Andres; Mahoney, Craig R; Hartman, Curtis W; Haider, Hani

    2013-01-01

    Computer-aided surgery aims to improve implant alignment in TKA but has only been adopted by a minority for routine use. A novel approach, navigated freehand bone cutting (NFC), is intended to achieve wider acceptance by eliminating the need for cumbersome, implant-specific mechanical jigs and avoiding the expense of navigation. We determined cutting time, surface quality, implant fit, and implant alignment after NFC of synthetic femoral specimens and the feasibility and alignment of a complete TKA performed with NFC technology in cadaveric specimens. Seven surgeons prepared six synthetic femoral specimens each, using our custom NFC system. Cutting times, quality of bone cuts, and implant fit and alignment were assessed quantitatively by CT surface scanning and computational measurements. Additionally, a single surgeon performed a complete TKA on two cadaveric specimens using the NFC system, with cutting time and implant alignment analyzed through plain radiographs and CT. For the synthetic specimens, femoral coronal alignment was within ± 2° of neutral in 94% of the specimens. Sagittal alignment was within 0° to 5° of flexion in all specimens. Rotation was within ± 1° of the epicondylar axis in 97% of the specimens. The mean time to make cuts improved from 13 minutes for the first specimen to 9 minutes for the fourth specimen. TKA was performed in two cadaveric specimens without complications and implants were well aligned. TKA is feasible with NFC, which eliminates the need for implant-specific instruments. We observed a fast learning curve. NFC has the potential to improve TKA alignment, reduce operative time, and reduce the number of instruments in surgery. Fewer instruments and less sterilization could reduce costs associated with TKA.

  10. Development of real-time visualization system for Computational Fluid Dynamics on parallel computers

    International Nuclear Information System (INIS)

    Muramatsu, Kazuhiro; Otani, Takayuki; Matsumoto, Hideki; Takei, Toshifumi; Doi, Shun

    1998-03-01

    A real-time visualization system for computational fluid dynamics in a network connecting between a parallel computing server and the client terminal was developed. Using the system, a user can visualize the results of a CFD (Computational Fluid Dynamics) simulation on the parallel computer as a client terminal during the actual computation on a server. Using GUI (Graphical User Interface) on the client terminal, to user is also able to change parameters of the analysis and visualization during the real-time of the calculation. The system carries out both of CFD simulation and generation of a pixel image data on the parallel computer, and compresses the data. Therefore, the amount of data from the parallel computer to the client is so small in comparison with no compression that the user can enjoy the swift image appearance comfortably. Parallelization of image data generation is based on Owner Computation Rule. GUI on the client is built on Java applet. A real-time visualization is thus possible on the client PC only if Web browser is implemented on it. (author)

  11. Asymptotic behavior of total times For jobs that must start over if a failure occurs

    DEFF Research Database (Denmark)

    Asmussen, Søren; Fiorini, Pierre; Lipsky, Lester

    the ready queue, or it may restart the task. The behavior of systems under the first two scenarios is well documented, but the third (RESTART) has resisted detailed analysis. In this paper we derive tight asymptotic relations between the distribution of task times without failures to the total time when...... including failures, for any failure distribution. In particular, we show that if the task time distribution has an unbounded support then the total time distribution H is always heavy-tailed. Asymptotic expressions are given for the tail of H in various scenarios. The key ingredients of the analysis...

  12. Asymptotic behaviour of total times for jobs that must start over if a failure occurs

    DEFF Research Database (Denmark)

    Asmussen, Søren; Fiorini, Pierre; Lipsky, Lester

    2008-01-01

    the ready queue, or it may restart the task. The behavior of systems under the first two scenarios is well documented, but the third (RESTART) has resisted detailed analysis. In this paper we derive tight asymptotic relations between the distribution of task times without failures and the total time when...... including failures, for any failure distribution. In particular, we show that if the task-time distribution has an unbounded support, then the total-time distribution H is always heavy tailed. Asymptotic expressions are given for the tail of H in various scenarios. The key ingredients of the analysis...

  13. Associations of Total and Domain-Specific Sedentary Time With Type 2 Diabetes in Taiwanese Older Adults

    Directory of Open Access Journals (Sweden)

    Ming-Chun Hsueh

    2016-07-01

    Full Text Available Background: The increasing prevalence of type 2 diabetes in older adults has become a public health concern. We investigated the associations of total and domain-specific sedentary time with risk of type 2 diabetes in older adults. Methods: The sample comprised 1046 older people (aged ≥65 years. Analyses were performed using crosssectional data collected via computer-assisted telephone-based interviews in 2014. Data on six self-reported domains of sedentary time (Measure of Older Adults’ Sedentary Time, type 2 diabetes status, and sociodemographic variables were included in the study. Binary logistic regression analysis was performed to calculate the adjusted odds ratios (ORs and 95% confidence intervals (CIs for total and individual sedentary behavior components and likelihood of type 2 diabetes. Results: A total of 17.5% of the participants reported type 2 diabetes. No significant associations were found between total sitting time and risk of type 2 diabetes, after controlling for confounding factors. After total sedentary behavior was stratified into six domains, only watching television for more than 2 hours per day was associated with higher odds of type 2 diabetes (OR 1.56; 95% CI, 1.10–2.21, but no significant associations were found between other domains of sedentary behavior (computer use, reading, socializing, transport, and hobbies and risk of type 2 diabetes. Conclusions: These findings suggest that, among domain-specific sedentary behavior, excessive television viewing might increase the risk of type 2 diabetes among older adults more than other forms of sedentary behavior.

  14. Total Work, Gender and Social Norms in EU and US Time Use

    OpenAIRE

    Burda , Michael C; Hamermesh , Daniel S; Weil , Philippe

    2008-01-01

    Using time-diary data from 27 countries, we demonstrate a negative relationship between real GDP per capita and the female-male difference in total work time--the sum of work for pay and work at home. We also show that in rich non-Catholic countries on four continents men and women do the same amount of total work on average. Our survey results demonstrate that labor economists, macroeconomists, sociologists and the general public consistently believe that women perform more total work. The f...

  15. New Flutter Analysis Technique for Time-Domain Computational Aeroelasticity

    Science.gov (United States)

    Pak, Chan-Gi; Lung, Shun-Fat

    2017-01-01

    A new time-domain approach for computing flutter speed is presented. Based on the time-history result of aeroelastic simulation, the unknown unsteady aerodynamics model is estimated using a system identification technique. The full aeroelastic model is generated via coupling the estimated unsteady aerodynamic model with the known linear structure model. The critical dynamic pressure is computed and used in the subsequent simulation until the convergence of the critical dynamic pressure is achieved. The proposed method is applied to a benchmark cantilevered rectangular wing.

  16. Effect of the MCNP model definition on the computation time

    International Nuclear Information System (INIS)

    Šunka, Michal

    2017-01-01

    The presented work studies the influence of the method of defining the geometry in the MCNP transport code and its impact on the computational time, including the difficulty of preparing an input file describing the given geometry. Cases using different geometric definitions including the use of basic 2-dimensional and 3-dimensional objects and theirs combinations were studied. The results indicate that an inappropriate definition can increase the computational time by up to 59% (a more realistic case indicates 37%) for the same results and the same statistical uncertainty. (orig.)

  17. Real-time computational photon-counting LiDAR

    Science.gov (United States)

    Edgar, Matthew; Johnson, Steven; Phillips, David; Padgett, Miles

    2018-03-01

    The availability of compact, low-cost, and high-speed MEMS-based spatial light modulators has generated widespread interest in alternative sampling strategies for imaging systems utilizing single-pixel detectors. The development of compressed sensing schemes for real-time computational imaging may have promising commercial applications for high-performance detectors, where the availability of focal plane arrays is expensive or otherwise limited. We discuss the research and development of a prototype light detection and ranging (LiDAR) system via direct time of flight, which utilizes a single high-sensitivity photon-counting detector and fast-timing electronics to recover millimeter accuracy three-dimensional images in real time. The development of low-cost real time computational LiDAR systems could have importance for applications in security, defense, and autonomous vehicles.

  18. Lot-Order Assignment Applying Priority Rules for the Single-Machine Total Tardiness Scheduling with Nonnegative Time-Dependent Processing Times

    Directory of Open Access Journals (Sweden)

    Jae-Gon Kim

    2015-01-01

    Full Text Available Lot-order assignment is to assign items in lots being processed to orders to fulfill the orders. It is usually performed periodically for meeting the due dates of orders especially in a manufacturing industry with a long production cycle time such as the semiconductor manufacturing industry. In this paper, we consider the lot-order assignment problem (LOAP with the objective of minimizing the total tardiness of the orders with distinct due dates. We show that we can solve the LOAP optimally by finding an optimal sequence for the single-machine total tardiness scheduling problem with nonnegative time-dependent processing times (SMTTSP-NNTDPT. Also, we address how the priority rules for the SMTTSP can be modified to those for the SMTTSP-NNTDPT to solve the LOAP. In computational experiments, we discuss the performances of the suggested priority rules and show the result of the proposed approach outperforms that of the commercial optimization software package.

  19. Imprecise results: Utilizing partial computations in real-time systems

    Science.gov (United States)

    Lin, Kwei-Jay; Natarajan, Swaminathan; Liu, Jane W.-S.

    1987-01-01

    In real-time systems, a computation may not have time to complete its execution because of deadline requirements. In such cases, no result except the approximate results produced by the computations up to that point will be available. It is desirable to utilize these imprecise results if possible. Two approaches are proposed to enable computations to return imprecise results when executions cannot be completed normally. The milestone approach records results periodically, and if a deadline is reached, returns the last recorded result. The sieve approach demarcates sections of code which can be skipped if the time available is insufficient. By using these approaches, the system is able to produce imprecise results when deadlines are reached. The design of the Concord project is described which supports imprecise computations using these techniques. Also presented is a general model of imprecise computations using these techniques, as well as one which takes into account the influence of the environment, showing where the latter approach fits into this model.

  20. Total sleep time, alcohol consumption, and the duration and severity of alcohol hangover

    NARCIS (Netherlands)

    van Schrojenstein Lantman, Marith; Mackus, Marlou; Roth, Thomas; Verster, Joris C|info:eu-repo/dai/nl/241442702

    2017-01-01

    INTRODUCTION: An evening of alcohol consumption often occurs at the expense of sleep time. The aim of this study was to determine the relationship between total sleep time and the duration and severity of the alcohol hangover. METHODS: A survey was conducted among Dutch University students to

  1. Real-time brain computer interface using imaginary movements

    DEFF Research Database (Denmark)

    El-Madani, Ahmad; Sørensen, Helge Bjarup Dissing; Kjær, Troels W.

    2015-01-01

    Background: Brain Computer Interface (BCI) is the method of transforming mental thoughts and imagination into actions. A real-time BCI system can improve the quality of life of patients with severe neuromuscular disorders by enabling them to communicate with the outside world. In this paper...

  2. GRAPHIC, time-sharing magnet design computer programs at Argonne

    International Nuclear Information System (INIS)

    Lari, R.J.

    1974-01-01

    This paper describes three magnet design computer programs in use at the Zero Gradient Synchrotron of Argonne National Laboratory. These programs are used in the time sharing mode in conjunction with a Tektronix model 4012 graphic display terminal. The first program in called TRIM, the second MAGNET, and the third GFUN. (U.S.)

  3. Instructional Advice, Time Advice and Learning Questions in Computer Simulations

    Science.gov (United States)

    Rey, Gunter Daniel

    2010-01-01

    Undergraduate students (N = 97) used an introductory text and a computer simulation to learn fundamental concepts about statistical analyses (e.g., analysis of variance, regression analysis and General Linear Model). Each learner was randomly assigned to one cell of a 2 (with or without instructional advice) x 2 (with or without time advice) x 2…

  4. Increased Total Anesthetic Time Leads to Higher Rates of Surgical Site Infections in Spinal Fusions.

    Science.gov (United States)

    Puffer, Ross C; Murphy, Meghan; Maloney, Patrick; Kor, Daryl; Nassr, Ahmad; Freedman, Brett; Fogelson, Jeremy; Bydon, Mohamad

    2017-06-01

    A retrospective review of a consecutive series of spinal fusions comparing patient and procedural characteristics of patients who developed surgical site infections (SSIs) after spinal fusion. It is known that increased surgical time (incision to closure) is associated with a higher rate of postoperative SSIs. We sought to determine whether increased total anesthetic time (intubation to extubation) is a factor in the development of SSIs as well. In spine surgery for deformity and degenerative disease, SSI has been associated with operative time, revealing a nearly 10-fold increase in SSI rates in prolonged surgery. Surgical time is associated with infections in other surgical disciplines as well. No studies have reported whether total anesthetic time (intubation to extubation) has an association with SSIs. Surgical records were searched in a retrospective fashion to identify all spine fusion procedures performed between January 2010 and July 2012. All SSIs during that timeframe were recorded and compared with the list of cases performed between 2010 and 2012 in a case-control design. There were 20 (1.7%) SSIs in this fusion cohort. On univariate analyses of operative factors, there was a significant association between total anesthetic time (Infection 7.6 ± 0.5 hrs vs. no infection -6.0 ± 0.1 hrs, P operative time (infection 5.5 ± 0.4 hrs vs. no infection - 4.4 ± 0.06 hrs, P infections, whereas level of pathology and emergent surgery were not significant. On multivariate logistic analysis, BMI and total anesthetic time remained independent predictors of SSI whereas ASA status and operative time did not. Increasing BMI and total anesthetic time were independent predictors of SSIs in this cohort of over 1000 consecutive spinal fusions. 3.

  5. Protocol for concomitant temporomandibular joint custom-fitted total joint reconstruction and orthognathic surgery utilizing computer-assisted surgical simulation.

    Science.gov (United States)

    Movahed, Reza; Teschke, Marcus; Wolford, Larry M

    2013-12-01

    Clinicians who address temporomandibular joint (TMJ) pathology and dentofacial deformities surgically can perform the surgery in 1 stage or 2 separate stages. The 2-stage approach requires the patient to undergo 2 separate operations and anesthesia, significantly prolonging the overall treatment. However, performing concomitant TMJ and orthognathic surgery (CTOS) in these cases requires careful treatment planning and surgical proficiency in the 2 surgical areas. This article presents a new treatment protocol for the application of computer-assisted surgical simulation in CTOS cases requiring reconstruction with patient-fitted total joint prostheses. The traditional and new CTOS protocols are described and compared. The new CTOS protocol helps decrease the preoperative workup time and increase the accuracy of model surgery. Copyright © 2013 American Association of Oral and Maxillofacial Surgeons. Published by Elsevier Inc. All rights reserved.

  6. Prosthetic liner wear in total hip replacement: a longitudinal 13-year study with computed tomography.

    Science.gov (United States)

    Weidenhielm, Lars; Olivecrona, Henrik; Maguire, Gerald Q; Noz, Marilyn E

    2018-06-01

    This case report follows a woman who had a total hip replacement in 1992 when she was 45 years old. Six serial computed tomography (CT) examinations over a period of 13 years provided information that allowed her revision surgery to be limited to liner replacement as opposed to replacement of the entire prosthesis. Additionally, they provided data that ruled out the presence of osteolysis and indeed none was found at surgery. In 2004, when the first CT was performed, the 3D distance the femoral head had penetrated into the cup was determined to be 2.6 mm. By 2017, femoral head penetration had progressed to 5.0 mm. The extracted liner showed wear at the thinnest part to be 5.5 mm, as measured with a micrometer. The use of modern CT techniques can identify problems, while still correctable without major surgery. Furthermore, the ability of CT to assess the direction of wear revealed that the liner wear changed from the cranial to dorsal direction.

  7. Neural Computations in a Dynamical System with Multiple Time Scales.

    Science.gov (United States)

    Mi, Yuanyuan; Lin, Xiaohan; Wu, Si

    2016-01-01

    Neural systems display rich short-term dynamics at various levels, e.g., spike-frequency adaptation (SFA) at the single-neuron level, and short-term facilitation (STF) and depression (STD) at the synapse level. These dynamical features typically cover a broad range of time scales and exhibit large diversity in different brain regions. It remains unclear what is the computational benefit for the brain to have such variability in short-term dynamics. In this study, we propose that the brain can exploit such dynamical features to implement multiple seemingly contradictory computations in a single neural circuit. To demonstrate this idea, we use continuous attractor neural network (CANN) as a working model and include STF, SFA and STD with increasing time constants in its dynamics. Three computational tasks are considered, which are persistent activity, adaptation, and anticipative tracking. These tasks require conflicting neural mechanisms, and hence cannot be implemented by a single dynamical feature or any combination with similar time constants. However, with properly coordinated STF, SFA and STD, we show that the network is able to implement the three computational tasks concurrently. We hope this study will shed light on the understanding of how the brain orchestrates its rich dynamics at various levels to realize diverse cognitive functions.

  8. Does antegrade JJ stenting affect the total operative time during laparoscopic pyeloplasty?

    Science.gov (United States)

    Bolat, Mustafa Suat; Çınar, Önder; Akdeniz, Ekrem

    2017-12-01

    We aimed to show the effect of retrograde JJ stenting and intraoperative antegrade JJ stenting techniques on operative time in patients who underwent laparoscopic pyeloplasty. A total of 34 patients were retrospectively investigated (15 male and 19 female) with ureteropelvic junction obstruction. Of the patients stentized under local anesthesia preoperatively, as a part of surgery, 15 were retrogradely stentized at the beginning of the procedure (Group 1), and 19 were antegradely stentized during the procedure (Group 2). A transperitoneal dismembered pyeloplasty technique was performed in all patients. The two groups were retrospectively compared in terms of complications, the mean total operative time, and the mean stenting times. The mean ages of the patients were 31.5±15.5 and 33.2±15.5 years (p=0.09), and the mean body mass indexes were 25.8±5.6 and 26.2.3±8.4 kg/m 2 in Group 1 and Group 2, respectively. The mean total operative times were 128.9±38.9 min and 112.7±21.9 min (p=0.04); the mean stenting times were 12.6±5.4 min and 3.5±2.4 min (p=0.02); and the mean rates of catheterization-to-total surgery times were 0.1 and 0.03 (p=0.01) in Group 1 and 2, respectively. The mean hospital stays and the mean anastomosis times were similar between the two groups (p>0.05). Antegrade JJ stenting during laparoscopic pyeloplasty significantly decreased the total operative time.

  9. Reduced computational cost in the calculation of worst case response time for real time systems

    OpenAIRE

    Urriza, José M.; Schorb, Lucas; Orozco, Javier D.; Cayssials, Ricardo

    2009-01-01

    Modern Real Time Operating Systems require reducing computational costs even though the microprocessors become more powerful each day. It is usual that Real Time Operating Systems for embedded systems have advance features to administrate the resources of the applications that they support. In order to guarantee either the schedulability of the system or the schedulability of a new task in a dynamic Real Time System, it is necessary to know the Worst Case Response Time of the Real Time tasks ...

  10. 12 CFR Appendix K to Part 226 - Total Annual Loan Cost Rate Computations for Reverse Mortgage Transactions

    Science.gov (United States)

    2010-01-01

    ... Appendix K to Part 226—Total Annual Loan Cost Rate Computations for Reverse Mortgage Transactions (a... loan cost rate for various transactions, as well as instructions, explanations, and examples for.... (2) Term of the transaction. For purposes of total annual loan cost disclosures, the term of a...

  11. Objectively Measured Total and Occupational Sedentary Time in Three Work Settings

    Science.gov (United States)

    van Dommelen, Paula; Coffeng, Jennifer K.; van der Ploeg, Hidde P.; van der Beek, Allard J.; Boot, Cécile R. L.; Hendriksen, Ingrid J. M.

    2016-01-01

    Background Sedentary behaviour increases the risk for morbidity. Our primary aim is to determine the proportion and factors associated with objectively measured total and occupational sedentary time in three work settings. Secondary aim is to study the proportion of physical activity and prolonged sedentary bouts. Methods Data were obtained using ActiGraph accelerometers from employees of: 1) a financial service provider (n = 49 men, 31 women), 2) two research institutes (n = 30 men, 57 women), and 3) a construction company (n = 38 men). Total (over the whole day) and occupational sedentary time, physical activity and prolonged sedentary bouts (lasting ≥30 minutes) were calculated by work setting. Linear regression analyses were performed to examine general, health and work-related factors associated with sedentary time. Results The employees of the financial service provider and the research institutes spent 76–80% of their occupational time in sedentary behaviour, 18–20% in light intensity physical activity and 3–5% in moderate-to-vigorous intensity physical activity. Occupational time in prolonged sedentary bouts was 27–30%. Total time was less sedentary (64–70%), and had more light intensity physical activity (26–33%). The employees of the construction company spent 44% of their occupational time in sedentary behaviour, 49% in light, and 7% in moderate intensity physical activity, and spent 7% in sedentary bouts. Total time spent in sedentary behavior was 56%, 40% in light, and 4% in moderate intensity physical behaviour, and 12% in sedentary bouts. For women, low to intermediate education was the only factor that was negatively associated with occupational sedentary time. Conclusions Sedentary behaviour is high among white-collar employees, especially in highly educated women. A relatively small proportion of sedentary time was accrued in sedentary bouts. It is recommended that worksite health promotion efforts should focus on reducing sedentary

  12. Fast Megavoltage Computed Tomography: A Rapid Imaging Method for Total Body or Marrow Irradiation in Helical Tomotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Magome, Taiki [Department of Radiological Sciences, Faculty of Health Sciences, Komazawa University, Tokyo (Japan); Department of Radiology, The University of Tokyo Hospital, Tokyo (Japan); Masonic Cancer Center, University of Minnesota, Minneapolis, Minnesota (United States); Haga, Akihiro [Department of Radiology, The University of Tokyo Hospital, Tokyo (Japan); Takahashi, Yutaka [Masonic Cancer Center, University of Minnesota, Minneapolis, Minnesota (United States); Department of Radiation Oncology, Osaka University, Osaka (Japan); Nakagawa, Keiichi [Department of Radiology, The University of Tokyo Hospital, Tokyo (Japan); Dusenbery, Kathryn E. [Department of Therapeutic Radiology, University of Minnesota, Minneapolis, Minnesota (United States); Hui, Susanta K., E-mail: shui@coh.org [Masonic Cancer Center, University of Minnesota, Minneapolis, Minnesota (United States); Department of Therapeutic Radiology, University of Minnesota, Minneapolis, Minnesota (United States); Department of Radiation Oncology and Beckman Research Institute, City of Hope, Duarte, California (United States)

    2016-11-01

    Purpose: Megavoltage computed tomographic (MVCT) imaging has been widely used for the 3-dimensional (3-D) setup of patients treated with helical tomotherapy (HT). One drawback of MVCT is its very long imaging time, the result of slow couch speeds of approximately 1 mm/s, which can be difficult for the patient to tolerate. We sought to develop an MVCT imaging method allowing faster couch speeds and to assess its accuracy for image guidance for HT. Methods and Materials: Three cadavers were scanned 4 times with couch speeds of 1, 2, 3, and 4 mm/s. The resulting MVCT images were reconstructed using an iterative reconstruction (IR) algorithm with a penalty term of total variation and with a conventional filtered back projection (FBP) algorithm. The MVCT images were registered with kilovoltage CT images, and the registration errors from the 2 reconstruction algorithms were compared. This fast MVCT imaging was tested in 3 cases of total marrow irradiation as a clinical trial. Results: The 3-D registration errors of the MVCT images reconstructed with the IR algorithm were smaller than the errors of images reconstructed with the FBP algorithm at fast couch speeds (2, 3, 4 mm/s). The scan time and imaging dose at a speed of 4 mm/s were reduced to 30% of those from a conventional coarse mode scan. For the patient imaging, faster MVCT (3 mm/s couch speed) scanning reduced the imaging time and still generated images useful for anatomic registration. Conclusions: Fast MVCT with the IR algorithm is clinically feasible for large 3-D target localization, which may reduce the overall time for the treatment procedure. This technique may also be useful for calculating daily dose distributions or organ motion analyses in HT treatment over a wide area. Automated integration of this imaging is at least needed to further assess its clinical benefits.

  13. Efficient quantum algorithm for computing n-time correlation functions.

    Science.gov (United States)

    Pedernales, J S; Di Candia, R; Egusquiza, I L; Casanova, J; Solano, E

    2014-07-11

    We propose a method for computing n-time correlation functions of arbitrary spinorial, fermionic, and bosonic operators, consisting of an efficient quantum algorithm that encodes these correlations in an initially added ancillary qubit for probe and control tasks. For spinorial and fermionic systems, the reconstruction of arbitrary n-time correlation functions requires the measurement of two ancilla observables, while for bosonic variables time derivatives of the same observables are needed. Finally, we provide examples applicable to different quantum platforms in the frame of the linear response theory.

  14. Minimizing Total Completion Time For Preemptive Scheduling With Release Dates And Deadline Constraints

    Directory of Open Access Journals (Sweden)

    He Cheng

    2014-02-01

    Full Text Available It is known that the single machine preemptive scheduling problem of minimizing total completion time with release date and deadline constraints is NP- hard. Du and Leung solved some special cases by the generalized Baker's algorithm and the generalized Smith's algorithm in O(n2 time. In this paper we give an O(n2 algorithm for the special case where the processing times and deadlines are agreeable. Moreover, for the case where the processing times and deadlines are disagreeable, we present two properties which could enable us to reduce the range of the enumeration algorithm

  15. Distributed computing for real-time petroleum reservoir monitoring

    Energy Technology Data Exchange (ETDEWEB)

    Ayodele, O. R. [University of Alberta, Edmonton, AB (Canada)

    2004-05-01

    Computer software architecture is presented to illustrate how the concept of distributed computing can be applied to real-time reservoir monitoring processes, permitting the continuous monitoring of the dynamic behaviour of petroleum reservoirs at much shorter intervals. The paper describes the fundamental technologies driving distributed computing, namely Java 2 Platform Enterprise edition (J2EE) by Sun Microsystems, and the Microsoft Dot-Net (Microsoft.Net) initiative, and explains the challenges involved in distributed computing. These are: (1) availability of permanently placed downhole equipment to acquire and transmit seismic data; (2) availability of high bandwidth to transmit the data; (3) security considerations; (4) adaptation of existing legacy codes to run on networks as downloads on demand; and (5) credibility issues concerning data security over the Internet. Other applications of distributed computing in the petroleum industry are also considered, specifically MWD, LWD and SWD (measurement-while-drilling, logging-while-drilling, and simulation-while-drilling), and drill-string vibration monitoring. 23 refs., 1 fig.

  16. Soft Real-Time PID Control on a VME Computer

    Science.gov (United States)

    Karayan, Vahag; Sander, Stanley; Cageao, Richard

    2007-01-01

    microPID (uPID) is a computer program for real-time proportional + integral + derivative (PID) control of a translation stage in a Fourier-transform ultraviolet spectrometer. microPID implements a PID control loop over a position profile at sampling rate of 8 kHz (sampling period 125microseconds). The software runs in a strippeddown Linux operating system on a VersaModule Eurocard (VME) computer operating in real-time priority queue using an embedded controller, a 16-bit digital-to-analog converter (D/A) board, and a laser-positioning board (LPB). microPID consists of three main parts: (1) VME device-driver routines, (2) software that administers a custom protocol for serial communication with a control computer, and (3) a loop section that obtains the current position from an LPB-driver routine, calculates the ideal position from the profile, and calculates a new voltage command by use of an embedded PID routine all within each sampling period. The voltage command is sent to the D/A board to control the stage. microPID uses special kernel headers to obtain microsecond timing resolution. Inasmuch as microPID implements a single-threaded process and all other processes are disabled, the Linux operating system acts as a soft real-time system.

  17. Observed and simulated time evolution of HCl, ClONO2, and HF total columns

    Science.gov (United States)

    Ruhnke, Roland; Geomon, Ndacc Infrared, Modelling Working Group

    2010-05-01

    Institute of Technology (KIT), IMK-IFU, Garmisch-Partenkirchen, Germany, (16) University of Denver, Dept. of Physics and Astronomy, Denver, CO, USA, (17) National Center for Atmospheric Research (NCAR), Boulder, CO, USA, (18) NASA Langley Research Center, Hampton, VA, USA, (19) Karlsruhe Institute of Technology (KIT), Steinbuch Centre for Computing, Karlsruhe, Germany Total column abundances of HCl and ClONO2, the primary components of the stratospheric inorganic chlorine (Cly) budget, and of HF have been retrieved from ground-based, high-resolution infrared solar absorption spectra recorded at 17 sites of the Network for the Detection of Atmospheric Composition Change (NDACC) located at latitudes between 80.05°N and 77.82°S. These data extend over more than 20 years (through 2007) during a period when the growth in atmospheric halogen loading has slowed in response to the Montreal Protocol (and ammendments). These observed time series are interpreted with calculations performed with a 2-D model, the 3-D chemistry-transport models (CTMs) KASIMA and SLIMCAT, and the 3-D chemistry-climate models (CCMs) EMAC and SOCOLv2.0. The observed Cly and in particular HCl column abundances decreases significantely since the end of the nineties at all stations, which is consistent with the observed changes in the halocarbon source gases, with an increasing rate in the last years. In contrast to Cly, the trend values for total column HF at the different stations show a less consistent behaviour pointing to the fact that the time development of the HF columns is peaking. There is a good overall qualitative agreement regarding trends between models and data. With respect to the CTMs the agreement improves if simulation results for measurement days only are used in the trend analysis instead of simulation results for each day.

  18. Effects of computing time delay on real-time control systems

    Science.gov (United States)

    Shin, Kang G.; Cui, Xianzhong

    1988-01-01

    The reliability of a real-time digital control system depends not only on the reliability of the hardware and software used, but also on the speed in executing control algorithms. The latter is due to the negative effects of computing time delay on control system performance. For a given sampling interval, the effects of computing time delay are classified into the delay problem and the loss problem. Analysis of these two problems is presented as a means of evaluating real-time control systems. As an example, both the self-tuning predicted (STP) control and Proportional-Integral-Derivative (PID) control are applied to the problem of tracking robot trajectories, and their respective effects of computing time delay on control performance are comparatively evaluated. For this example, the STP (PID) controller is shown to outperform the PID (STP) controller in coping with the delay (loss) problem.

  19. Timing of Re-Transfusion Drain Removal Following Total Knee Replacement

    Science.gov (United States)

    Leeman, MF; Costa, ML; Costello, E; Edwards, D

    2006-01-01

    INTRODUCTION The use of postoperative drains following total knee replacement (TKR) has recently been modified by the use of re-transfusion drains. The aim of our study was to investigate the optimal time for removal of re-transfusion drains following TKR. PATIENTS AND METHODS The medical records of 66 patients who had a TKR performed between October 2003 and October 2004 were reviewed; blood drained before 6 h and the total volume of blood drained was recorded. RESULTS A total of 56 patients had complete records of postoperative drainage. The mean volume of blood collected in the drain in the first 6 h was 442 ml. The mean total volume of blood in the drain was 595 ml. Therefore, of the blood drained, 78% was available for transfusion. CONCLUSION Re-transfusion drains should be removed after 6 h, when no further re-transfusion is permissible. PMID:16551400

  20. Sorting on STAR. [CDC computer algorithm timing comparison

    Science.gov (United States)

    Stone, H. S.

    1978-01-01

    Timing comparisons are given for three sorting algorithms written for the CDC STAR computer. One algorithm is Hoare's (1962) Quicksort, which is the fastest or nearly the fastest sorting algorithm for most computers. A second algorithm is a vector version of Quicksort that takes advantage of the STAR's vector operations. The third algorithm is an adaptation of Batcher's (1968) sorting algorithm, which makes especially good use of vector operations but has a complexity of N(log N)-squared as compared with a complexity of N log N for the Quicksort algorithms. In spite of its worse complexity, Batcher's sorting algorithm is competitive with the serial version of Quicksort for vectors up to the largest that can be treated by STAR. Vector Quicksort outperforms the other two algorithms and is generally preferred. These results indicate that unusual instruction sets can introduce biases in program execution time that counter results predicted by worst-case asymptotic complexity analysis.

  1. Computational intelligence in time series forecasting theory and engineering applications

    CERN Document Server

    Palit, Ajoy K

    2005-01-01

    Foresight in an engineering enterprise can make the difference between success and failure, and can be vital to the effective control of industrial systems. Applying time series analysis in the on-line milieu of most industrial plants has been problematic owing to the time and computational effort required. The advent of soft computing tools offers a solution. The authors harness the power of intelligent technologies individually and in combination. Examples of the particular systems and processes susceptible to each technique are investigated, cultivating a comprehensive exposition of the improvements on offer in quality, model building and predictive control and the selection of appropriate tools from the plethora available. Application-oriented engineers in process control, manufacturing, production industry and research centres will find much to interest them in this book. It is suitable for industrial training purposes, as well as serving as valuable reference material for experimental researchers.

  2. A polynomial time algorithm for checking regularity of totally normed process algebra

    NARCIS (Netherlands)

    Yang, F.; Huang, H.

    2015-01-01

    A polynomial algorithm for the regularity problem of weak and branching bisimilarity on totally normed process algebra (PA) processes is given. Its time complexity is O(n 3 +mn) O(n3+mn), where n is the number of transition rules and m is the maximal length of the rules. The algorithm works for

  3. Complexities of the storm-time characteristics of ionospheric total electron content

    International Nuclear Information System (INIS)

    Kane, R.P.

    1982-01-01

    The complexities of the storm-time variations of the ionospheric total electron content are briefly reviewed. It is suggested that large variations from storm to storm may be due to irregular flows from the auroral region towards equator. A proper study of such flows needs an elaborate network of TEC measuring instruments. The need of planning and organizing such a network is emphasized

  4. Objectively measured total and occupational sedentary time in three work settings

    NARCIS (Netherlands)

    Dommelen, P. van; Coffeng, J. K.; Ploeg, H.P. van der; Beek, A.J. van der; Boot, C.R.; Hendriksen, I.J.

    2016-01-01

    Background. Sedentary behaviour increases the risk for morbidity. Our primary aim is to determine the proportion and factors associated with objectively measured total and occupational sedentary time in three work settings. Secondary aim is to study the proportion of physical activity and prolonged

  5. What are the important manoeuvres for beginners to minimize surgical time in primary total knee arthroplasty?

    Science.gov (United States)

    Harato, Kengo; Maeno, Shinichi; Tanikawa, Hidenori; Kaneda, Kazuya; Morishige, Yutaro; Nomoto, So; Niki, Yasuo

    2016-08-01

    It was hypothesized that surgical time of beginners would be much longer than that of experts. Our purpose was to investigate and clarify the important manoeuvres for beginners to minimize surgical time in primary total knee arthroplasty (TKA) as a multicentre study. A total of 300 knees in 248 patients (averaged 74.6 years) were enrolled. All TKAs were done using the same instruments and the same measured resection technique at 14 facilities by 25 orthopaedic surgeons. Surgeons were divided into three surgeon groups (four experts, nine medium-volume surgeons and 12 beginners). The surgical technique was divided into five phases. Detailed surgical time and ratio of the time in each phase to overall surgical time were recorded and compared among the groups in each phase. A total of 62, 119, and 119 TKAs were done by beginners, medium-volume surgeons, and experts, respectively. Significant differences in surgical time among the groups were seen in each phase. Concerning the ratio of the time, experts and medium-volume surgeons seemed cautious in fixation of the permanent component compared to other phases. Interestingly, even in ratio, beginners and medium-volume surgeons took more time in exposure of soft tissue compared to experts. (0.14 in beginners, 0.13 in medium-volume surgeons, 0.11 in experts, P time in exposure and closure of soft tissue compared to experts. Improvement in basic technique is essential to minimize surgical time among beginners. First of all, surgical instructors should teach basic techniques in primary TKA for beginners. Therapeutic studies, Level IV.

  6. Reduced opiate use after total knee arthroplasty using computer-assisted cryotherapy.

    Science.gov (United States)

    Thijs, Elke; Schotanus, Martijn G M; Bemelmans, Yoeri F L; Kort, Nanne P

    2018-05-03

    Despite multimodal pain management and advances in anesthetic techniques, total knee arthroplasty (TKA) remains painful during the early postoperative phase. This trial investigated whether computer-assisted cryotherapy (CAC) is effective in reduction of pain and consumption of opioids in patients operated for TKA following an outpatient surgery pathway. Sixty patients scheduled for primary TKA were included in this prospective, double-blind, randomized controlled trial receiving CAC at 10-12 °C (Cold-group, n = 30) or at 21 °C (Warm-group, n = 30) during the first 7 days after TKA according to a fixed schedule. All patients received the same pre-, peri- and postoperative care with a multimodal pain protocol. Pain was assessed before and after every session of cryotherapy using the numerical rating scale for pain (NRS-pain). The consumption of opioids was strictly noted during the first 4 postoperative days. Secondary outcomes were knee swelling, visual hematoma and patient reported outcome measures (PROMs). These parameters were measured pre-, 1, 2 and 6 weeks postoperatively. In both study groups, a reduction in NRS-pain after every CAC session were seen during the postoperative period of 7 days. A mean reduction of 0.9 and 0.7 on the NRS-pain was seen for respectively the Cold- (P = 0.008) and Warm-group (n.s.). A significant (P = 0.001) lower number of opioids were used by the Cold-group during the acute postoperative phase of 4 days, 47 and 83 tablets for respectively the Cold and Warm-group. No difference could be observed for secondary outcomes and adverse effects between both study groups. Postoperative CAC can be in added value in patients following an outpatient surgery pathway for TKA, resulting in reduced experienced pain and consumption of opioids during the first postoperative days.

  7. Cerebral perfusion computed tomography deconvolution via structure tensor total variation regularization

    Energy Technology Data Exchange (ETDEWEB)

    Zeng, Dong; Zhang, Xinyu; Bian, Zhaoying, E-mail: zybian@smu.edu.cn, E-mail: jhma@smu.edu.cn; Huang, Jing; Zhang, Hua; Lu, Lijun; Lyu, Wenbing; Feng, Qianjin; Chen, Wufan; Ma, Jianhua, E-mail: zybian@smu.edu.cn, E-mail: jhma@smu.edu.cn [Department of Biomedical Engineering, Southern Medical University, Guangzhou, Guangdong 510515, China and Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong 510515 (China); Zhang, Jing [Department of Radiology, Tianjin Medical University General Hospital, Tianjin 300052 (China)

    2016-05-15

    Purpose: Cerebral perfusion computed tomography (PCT) imaging as an accurate and fast acute ischemic stroke examination has been widely used in clinic. Meanwhile, a major drawback of PCT imaging is the high radiation dose due to its dynamic scan protocol. The purpose of this work is to develop a robust perfusion deconvolution approach via structure tensor total variation (STV) regularization (PD-STV) for estimating an accurate residue function in PCT imaging with the low-milliampere-seconds (low-mAs) data acquisition. Methods: Besides modeling the spatio-temporal structure information of PCT data, the STV regularization of the present PD-STV approach can utilize the higher order derivatives of the residue function to enhance denoising performance. To minimize the objective function, the authors propose an effective iterative algorithm with a shrinkage/thresholding scheme. A simulation study on a digital brain perfusion phantom and a clinical study on an old infarction patient were conducted to validate and evaluate the performance of the present PD-STV approach. Results: In the digital phantom study, visual inspection and quantitative metrics (i.e., the normalized mean square error, the peak signal-to-noise ratio, and the universal quality index) assessments demonstrated that the PD-STV approach outperformed other existing approaches in terms of the performance of noise-induced artifacts reduction and accurate perfusion hemodynamic maps (PHM) estimation. In the patient data study, the present PD-STV approach could yield accurate PHM estimation with several noticeable gains over other existing approaches in terms of visual inspection and correlation analysis. Conclusions: This study demonstrated the feasibility and efficacy of the present PD-STV approach in utilizing STV regularization to improve the accuracy of residue function estimation of cerebral PCT imaging in the case of low-mAs.

  8. Spike-timing-based computation in sound localization.

    Directory of Open Access Journals (Sweden)

    Dan F M Goodman

    2010-11-01

    Full Text Available Spike timing is precise in the auditory system and it has been argued that it conveys information about auditory stimuli, in particular about the location of a sound source. However, beyond simple time differences, the way in which neurons might extract this information is unclear and the potential computational advantages are unknown. The computational difficulty of this task for an animal is to locate the source of an unexpected sound from two monaural signals that are highly dependent on the unknown source signal. In neuron models consisting of spectro-temporal filtering and spiking nonlinearity, we found that the binaural structure induced by spatialized sounds is mapped to synchrony patterns that depend on source location rather than on source signal. Location-specific synchrony patterns would then result in the activation of location-specific assemblies of postsynaptic neurons. We designed a spiking neuron model which exploited this principle to locate a variety of sound sources in a virtual acoustic environment using measured human head-related transfer functions. The model was able to accurately estimate the location of previously unknown sounds in both azimuth and elevation (including front/back discrimination in a known acoustic environment. We found that multiple representations of different acoustic environments could coexist as sets of overlapping neural assemblies which could be associated with spatial locations by Hebbian learning. The model demonstrates the computational relevance of relative spike timing to extract spatial information about sources independently of the source signal.

  9. A heterogeneous hierarchical architecture for real-time computing

    Energy Technology Data Exchange (ETDEWEB)

    Skroch, D.A.; Fornaro, R.J.

    1988-12-01

    The need for high-speed data acquisition and control algorithms has prompted continued research in the area of multiprocessor systems and related programming techniques. The result presented here is a unique hardware and software architecture for high-speed real-time computer systems. The implementation of a prototype of this architecture has required the integration of architecture, operating systems and programming languages into a cohesive unit. This report describes a Heterogeneous Hierarchial Architecture for Real-Time (H{sup 2} ART) and system software for program loading and interprocessor communication.

  10. Modern EMC analysis I time-domain computational schemes

    CERN Document Server

    Kantartzis, Nikolaos V

    2008-01-01

    The objective of this two-volume book is the systematic and comprehensive description of the most competitive time-domain computational methods for the efficient modeling and accurate solution of contemporary real-world EMC problems. Intended to be self-contained, it performs a detailed presentation of all well-known algorithms, elucidating on their merits or weaknesses, and accompanies the theoretical content with a variety of applications. Outlining the present volume, the analysis covers the theory of the finite-difference time-domain, the transmission-line matrix/modeling, and the finite i

  11. Stochastic nonlinear time series forecasting using time-delay reservoir computers: performance and universality.

    Science.gov (United States)

    Grigoryeva, Lyudmila; Henriques, Julie; Larger, Laurent; Ortega, Juan-Pablo

    2014-07-01

    Reservoir computing is a recently introduced machine learning paradigm that has already shown excellent performances in the processing of empirical data. We study a particular kind of reservoir computers called time-delay reservoirs that are constructed out of the sampling of the solution of a time-delay differential equation and show their good performance in the forecasting of the conditional covariances associated to multivariate discrete-time nonlinear stochastic processes of VEC-GARCH type as well as in the prediction of factual daily market realized volatilities computed with intraday quotes, using as training input daily log-return series of moderate size. We tackle some problems associated to the lack of task-universality for individually operating reservoirs and propose a solution based on the use of parallel arrays of time-delay reservoirs. Copyright © 2014 Elsevier Ltd. All rights reserved.

  12. STICK: Spike Time Interval Computational Kernel, a Framework for General Purpose Computation Using Neurons, Precise Timing, Delays, and Synchrony.

    Science.gov (United States)

    Lagorce, Xavier; Benosman, Ryad

    2015-11-01

    There has been significant research over the past two decades in developing new platforms for spiking neural computation. Current neural computers are primarily developed to mimic biology. They use neural networks, which can be trained to perform specific tasks to mainly solve pattern recognition problems. These machines can do more than simulate biology; they allow us to rethink our current paradigm of computation. The ultimate goal is to develop brain-inspired general purpose computation architectures that can breach the current bottleneck introduced by the von Neumann architecture. This work proposes a new framework for such a machine. We show that the use of neuron-like units with precise timing representation, synaptic diversity, and temporal delays allows us to set a complete, scalable compact computation framework. The framework provides both linear and nonlinear operations, allowing us to represent and solve any function. We show usability in solving real use cases from simple differential equations to sets of nonlinear differential equations leading to chaotic attractors.

  13. Batch Scheduling for Hybrid Assembly Differentiation Flow Shop to Minimize Total Actual Flow Time

    Science.gov (United States)

    Maulidya, R.; Suprayogi; Wangsaputra, R.; Halim, A. H.

    2018-03-01

    A hybrid assembly differentiation flow shop is a three-stage flow shop consisting of Machining, Assembly and Differentiation Stages and producing different types of products. In the machining stage, parts are processed in batches on different (unrelated) machines. In the assembly stage, each part of the different parts is assembled into an assembly product. Finally, the assembled products will further be processed into different types of final products in the differentiation stage. In this paper, we develop a batch scheduling model for a hybrid assembly differentiation flow shop to minimize the total actual flow time defined as the total times part spent in the shop floor from the arrival times until its due date. We also proposed a heuristic algorithm for solving the problems. The proposed algorithm is tested using a set of hypothetic data. The solution shows that the algorithm can solve the problems effectively.

  14. Climate Data Provenance Tracking for Just-In-Time Computation

    Science.gov (United States)

    Fries, S.; Nadeau, D.; Doutriaux, C.; Williams, D. N.

    2016-12-01

    The "Climate Data Management System" (CDMS) was created in 1996 as part of the Climate Data Analysis Tools suite of software. It provides a simple interface into a wide variety of climate data formats, and creates NetCDF CF-Compliant files. It leverages the NumPy framework for high performance computation, and is an all-in-one IO and computation package. CDMS has been extended to track manipulations of data, and trace that data all the way to the original raw data. This extension tracks provenance about data, and enables just-in-time (JIT) computation. The provenance for each variable is packaged as part of the variable's metadata, and can be used to validate data processing and computations (by repeating the analysis on the original data). It also allows for an alternate solution for sharing analyzed data; if the bandwidth for a transfer is prohibitively expensive, the provenance serialization can be passed in a much more compact format and the analysis rerun on the input data. Data provenance tracking in CDMS enables far-reaching and impactful functionalities, permitting implementation of many analytical paradigms.

  15. Different but Equal: Total Work, Gender and Social Norms in EU and US Time Use

    OpenAIRE

    Daniel S Hamermesh; Michael C Burda; Philippe Weil

    2008-01-01

    Using time-diary data from 27 countries, we demonstrate a negative relationship between real GDP per capita and the female-male difference in total work time—the sum of work for pay and work at home. We also show that in rich non-Catholic countries on four continents men and women do the same amount of total work on average. Our survey results demonstrate that labor economists, macroeconomists, sociologists and the general public consistently believe that women perform more tot...

  16. A note on computing average state occupation times

    Directory of Open Access Journals (Sweden)

    Jan Beyersmann

    2014-05-01

    Full Text Available Objective: This review discusses how biometricians would probably compute or estimate expected waiting times, if they had the data. Methods: Our framework is a time-inhomogeneous Markov multistate model, where all transition hazards are allowed to be time-varying. We assume that the cumulative transition hazards are given. That is, they are either known, as in a simulation, determined by expert guesses, or obtained via some method of statistical estimation. Our basic tool is product integration, which transforms the transition hazards into the matrix of transition probabilities. Product integration enjoys a rich mathematical theory, which has successfully been used to study probabilistic and statistical aspects of multistate models. Our emphasis will be on practical implementation of product integration, which allows us to numerically approximate the transition probabilities. Average state occupation times and other quantities of interest may then be derived from the transition probabilities.

  17. Reducing the throughput time of the diagnostic track involving CT scanning with computer simulation

    International Nuclear Information System (INIS)

    Lent, Wineke A.M. van; Deetman, Joost W.; Teertstra, H. Jelle; Muller, Sara H.; Hans, Erwin W.; Harten, Wim H. van

    2012-01-01

    Introduction: To examine the use of computer simulation to reduce the time between the CT request and the consult in which the CT report is discussed (diagnostic track) while restricting idle time and overtime. Methods: After a pre implementation analysis in our case study hospital, by computer simulation three scenarios were evaluated on access time, overtime and idle time of the CT; after implementation these same aspects were evaluated again. Effects on throughput time were measured for outpatient short-term and urgent requests only. Conclusion: The pre implementation analysis showed an average CT access time of 9.8 operating days and an average diagnostic track of 14.5 operating days. Based on the outcomes of the simulation, management changed the capacity for the different patient groups to facilitate a diagnostic track of 10 operating days, with a CT access time of 7 days. After the implementation of changes, the average diagnostic track duration was 12.6 days with an average CT access time of 7.3 days. The fraction of patients with a total throughput time within 10 days increased from 29% to 44% while the utilization remained equal with 82%, the idle time increased by 11% and the overtime decreased by 82%. The fraction of patients that completed the diagnostic track within 10 days improved with 52%. Computer simulation proved useful for studying the effects of proposed scenarios in radiology management. Besides the tangible effects, the simulation increased the awareness that optimizing capacity allocation can reduce access times.

  18. Reducing the throughput time of the diagnostic track involving CT scanning with computer simulation

    Energy Technology Data Exchange (ETDEWEB)

    Lent, Wineke A.M. van, E-mail: w.v.lent@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); University of Twente, IGS Institute for Innovation and Governance Studies, Department of Health Technology Services Research (HTSR), Enschede (Netherlands); Deetman, Joost W., E-mail: j.deetman@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Teertstra, H. Jelle, E-mail: h.teertstra@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Muller, Sara H., E-mail: s.muller@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); Hans, Erwin W., E-mail: e.w.hans@utwente.nl [University of Twente, School of Management and Governance, Dept. of Industrial Engineering and Business Intelligence Systems, Enschede (Netherlands); Harten, Wim H. van, E-mail: w.v.harten@nki.nl [Netherlands Cancer Institute - Antoni van Leeuwenhoek Hospital (NKI-AVL), P.O. Box 90203, 1006 BE Amsterdam (Netherlands); University of Twente, IGS Institute for Innovation and Governance Studies, Department of Health Technology Services Research (HTSR), Enschede (Netherlands)

    2012-11-15

    Introduction: To examine the use of computer simulation to reduce the time between the CT request and the consult in which the CT report is discussed (diagnostic track) while restricting idle time and overtime. Methods: After a pre implementation analysis in our case study hospital, by computer simulation three scenarios were evaluated on access time, overtime and idle time of the CT; after implementation these same aspects were evaluated again. Effects on throughput time were measured for outpatient short-term and urgent requests only. Conclusion: The pre implementation analysis showed an average CT access time of 9.8 operating days and an average diagnostic track of 14.5 operating days. Based on the outcomes of the simulation, management changed the capacity for the different patient groups to facilitate a diagnostic track of 10 operating days, with a CT access time of 7 days. After the implementation of changes, the average diagnostic track duration was 12.6 days with an average CT access time of 7.3 days. The fraction of patients with a total throughput time within 10 days increased from 29% to 44% while the utilization remained equal with 82%, the idle time increased by 11% and the overtime decreased by 82%. The fraction of patients that completed the diagnostic track within 10 days improved with 52%. Computer simulation proved useful for studying the effects of proposed scenarios in radiology management. Besides the tangible effects, the simulation increased the awareness that optimizing capacity allocation can reduce access times.

  19. Computer network time synchronization the network time protocol on earth and in space

    CERN Document Server

    Mills, David L

    2010-01-01

    Carefully coordinated, reliable, and accurate time synchronization is vital to a wide spectrum of fields-from air and ground traffic control, to buying and selling goods and services, to TV network programming. Ill-gotten time could even lead to the unimaginable and cause DNS caches to expire, leaving the entire Internet to implode on the root servers.Written by the original developer of the Network Time Protocol (NTP), Computer Network Time Synchronization: The Network Time Protocol on Earth and in Space, Second Edition addresses the technological infrastructure of time dissemination, distrib

  20. Computer-navigated minimally invasive total knee arthroplasty for patients with retained implants in the femur

    Directory of Open Access Journals (Sweden)

    Sung-Yen Lin

    2014-08-01

    Full Text Available Total knee arthroplasty (TKA in patients with knee arthritis and retained implants in the ipsilateral femur is a challenge for knee surgeons. Use of a conventional intramedullary femoral cutting guide is not practical because of the obstruction of the medullary canal by implants. Previous studies have shown that computer-assisted surgery (CAS can help restore alignment in conventional TKA for patients with knee arthritis with retained femoral implants or extra-articular deformity, without the need for implant removal or osteotomy. However, little has been published regarding outcomes with the use of navigation in minimally invasive surgery (MIS-TKA for patients with this complex knee arthritis. MIS has been proven to provide less postoperative pain and faster recovery than conventional TKA, but MIS-TKA in patients with retained femoral implants poses a greater risk in limb malalignment. The purpose of this study is to report the outcome of CAS-MIS-TKA in patients with knee arthritis and retained femoral implants. Between April 2006 and March 2008, eight patients with knee arthritis and retained femoral implants who underwent the CAS-MIS-TKA were retrospectively reviewed. Three of the eight patients had extra-articular deformity, including two femur bones and one tibia bone, in the preoperative examination. The anteroposterior, lateral, and long-leg weight-bearing radiographs carried out at 3-month follow-up was used to determine the mechanical axis of lower limb and the position of components. The mean preoperative femorotibial angle in patients without extra-articular deformity was 3.8° of varus and was corrected to 4.6° of valgus. With the use of navigation in MIS-TKA, the two patients in this study with extra-articular femoral deformity also obtained an ideal postoperative mechanical axis within 2° of normal alignment. Overall, there was a good restoration of postoperative mechanical alignment in all cases, with a mean angle of 0.4° of

  1. Physiotherapy Exercise After Fast-Track Total Hip and Knee Arthroplasty: Time for Reconsideration?

    DEFF Research Database (Denmark)

    Bandholm, Thomas; Kehlet, Henrik

    2012-01-01

    Bandholm T, Kehlet H. Physiotherapy exercise after fast-track total hip and knee arthroplasty: time for reconsideration? Major surgery, including total hip arthroplasty (THA) and total knee arthroplasty (TKA), is followed by a convalescence period, during which the loss of muscle strength......-track methodology or enhanced recovery programs. It is the nature of this methodology to systematically and scientifically optimize all perioperative care components, with the overall goal of enhancing recovery. This is also the case for the care component "physiotherapy exercise" after THA and TKA. The 2 latest...... meta-analyses on the effectiveness of physiotherapy exercise after THA and TKA generally conclude that physiotherapy exercise after THA and TKA either does not work or is not very effective. The reason for this may be that the "pill" of physiotherapy exercise typically offered after THA and TKA does...

  2. Computational electrodynamics the finite-difference time-domain method

    CERN Document Server

    Taflove, Allen

    2005-01-01

    This extensively revised and expanded third edition of the Artech House bestseller, Computational Electrodynamics: The Finite-Difference Time-Domain Method, offers engineers the most up-to-date and definitive resource on this critical method for solving Maxwell's equations. The method helps practitioners design antennas, wireless communications devices, high-speed digital and microwave circuits, and integrated optical devices with unsurpassed efficiency. There has been considerable advancement in FDTD computational technology over the past few years, and the third edition brings professionals the very latest details with entirely new chapters on important techniques, major updates on key topics, and new discussions on emerging areas such as nanophotonics. What's more, to supplement the third edition, the authors have created a Web site with solutions to problems, downloadable graphics and videos, and updates, making this new edition the ideal textbook on the subject as well.

  3. In this issue: Time to replace doctors’ judgement with computers

    Directory of Open Access Journals (Sweden)

    Simon de Lusignan

    2015-11-01

    Full Text Available Informaticians continue to rise to the challenge, set by the English Health Minister, of trying to replace doctors’ judgement with computers. This issue describes successes and where there are barriers. However, whilst there is progress this tends to be incremental and there are grand challenges to be overcome before computers can replace clinician. These grand challenges include: (1 improving usability so it is possible to more readily incorporate technology into clinical workflow; (2 rigorous new analytic methods that make use of the mass of available data, ‘Big data’, to create real-world evidence; (3 faster ways of meeting regulatory and legal requirements including ensuring privacy; (4 provision of reimbursement models to fund innovative technology that can substitute for clinical time and (5 recognition that innovations that improve quality also often increase cost. Informatics more is likely to support and augment clinical decision making rather than replace clinicians.

  4. Computation Offloading for Frame-Based Real-Time Tasks under Given Server Response Time Guarantees

    Directory of Open Access Journals (Sweden)

    Anas S. M. Toma

    2014-11-01

    Full Text Available Computation offloading has been adopted to improve the performance of embedded systems by offloading the computation of some tasks, especially computation-intensive tasks, to servers or clouds. This paper explores computation offloading for real-time tasks in embedded systems, provided given response time guarantees from the servers, to decide which tasks should be offloaded to get the results in time. We consider frame-based real-time tasks with the same period and relative deadline. When the execution order of the tasks is given, the problem can be solved in linear time. However, when the execution order is not specified, we prove that the problem is NP-complete. We develop a pseudo-polynomial-time algorithm for deriving feasible schedules, if they exist.  An approximation scheme is also developed to trade the error made from the algorithm and the complexity. Our algorithms are extended to minimize the period/relative deadline of the tasks for performance maximization. The algorithms are evaluated with a case study for a surveillance system and synthesized benchmarks.

  5. Time-based analysis of total cost of patient episodes: a case study of hip replacement.

    Science.gov (United States)

    Peltokorpi, Antti; Kujala, Jaakko

    2006-01-01

    Healthcare in the public and private sectors is facing increasing pressure to become more cost-effective. Time-based competition and work-in-progress have been used successfully to measure and improve the efficiency of industrial manufacturing. Seeks to address this issue. Presents a framework for time based management of the total cost of a patient episode and apply it to the six sigma DMAIC-process development approach. The framework is used to analyse hip replacement patient episodes in Päijät-Häme Hospital District in Finland, which has a catchment area of 210,000 inhabitants and performs an average of 230 hip replacements per year. The work-in-progress concept is applicable to healthcare--notably that the DMAIC-process development approach can be used to analyse the total cost of patient episodes. Concludes that a framework, which combines the patient-in-process and the DMAIC development approach, can be used not only to analyse the total cost of patient episode but also to improve patient process efficiency. Presents a framework that combines patient-in-process and DMAIC-process development approaches, which can be used to analyse the total cost of a patient episode in order to improve patient process efficiency.

  6. Just In Time Value Chain Total Quality Management Part Of Technical Strategic Management Accounting

    Directory of Open Access Journals (Sweden)

    Lesi Hertati

    2015-08-01

    Full Text Available This article aims to determine Just In Time Value Chain Total Quality Management tqm as a technique in management accounting stategis.Tujuan Just In Time value chain or value chain Total Quality Management TQM is strategic for customer satisfaction in the long term obtained from the information. Quality information is the way to continuous improvement in order to increase the companys financial performance in the long term to increase competitive advantage. Strategic Management Accounting process gather competitor information explore opportunities to reduce costs integrate accounting with emphasis on the strategic position of the competition is a great plan. An overall strategic plan interrelated and serves as the basis for achieving targets or goals ahead.

  7. Estimation of total bacteria by real-time PCR in patients with periodontal disease.

    Science.gov (United States)

    Brajović, Gavrilo; Popović, Branka; Puletić, Miljan; Kostić, Marija; Milasin, Jelena

    2016-01-01

    Periodontal diseases are associated with the presence of elevated levels of bacteria within the gingival crevice. The aim of this study was to evaluate a total amount of bacteria in subgingival plaque samples in patients with a periodontal disease. A quantitative evaluation of total bacteria amount using quantitative real-time polymerase chain reaction (qRT-PCR) was performed on 20 samples of patients with ulceronecrotic periodontitis and on 10 samples of healthy subjects. The estimation of total bacterial amount was based on gene copy number for 16S rRNA that was determined by comparing to Ct values/gene copy number of the standard curve. A statistically significant difference between average gene copy number of total bacteria in periodontal patients (2.55 x 10⁷) and healthy control (2.37 x 10⁶) was found (p = 0.01). Also, a trend of higher numbers of the gene copy in deeper periodontal lesions (> 7 mm) was confirmed by a positive value of coefficient of correlation (r = 0.073). The quantitative estimation of total bacteria based on gene copy number could be an important additional tool in diagnosing periodontitis.

  8. Real time simulation of large systems on mini-computer

    International Nuclear Information System (INIS)

    Nakhle, Michel; Roux, Pierre.

    1979-01-01

    Most simulation languages will only accept an explicit formulation of differential equations, and logical variables hold no special status therein. The pace of the suggested methods of integration is limited by the smallest time constant of the model submitted. The NEPTUNIX 2 simulation software has a language that will take implicit equations and an integration method of which the variable pace is not limited by the time constants of the model. This, together with high time and memory ressources optimization of the code generated, makes NEPTUNIX 2 a basic tool for simulation on mini-computers. Since the logical variables are specific entities under centralized control, correct processing of discontinuities and synchronization with a real process are feasible. The NEPTUNIX 2 is the industrial version of NEPTUNIX 1 [fr

  9. Objectively Measured Total and Occupational Sedentary Time in Three Work Settings

    OpenAIRE

    van Dommelen, Paula; Coffeng, Jennifer K.; van der Ploeg, Hidde P.; van der Beek, Allard J.; Boot, C?cile R. L.; Hendriksen, Ingrid J. M.

    2016-01-01

    Background. Sedentary behaviour increases the risk for morbidity. Our primary aim is to determine the proportion and factors associated with objectively measured total and occupational sedentary time in three work settings. Secondary aim is to study the proportion of physical activity and prolonged sedentary bouts. Methods. Data were obtained using ActiGraph accelerometers from employees of: 1) a financial service provider (n = 49 men, 31 women), 2) two research institutes (n = 30 men, 57 wom...

  10. NNSA?s Computing Strategy, Acquisition Plan, and Basis for Computing Time Allocation

    Energy Technology Data Exchange (ETDEWEB)

    Nikkel, D J

    2009-07-21

    This report is in response to the Omnibus Appropriations Act, 2009 (H.R. 1105; Public Law 111-8) in its funding of the National Nuclear Security Administration's (NNSA) Advanced Simulation and Computing (ASC) Program. This bill called for a report on ASC's plans for computing and platform acquisition strategy in support of stockpile stewardship. Computer simulation is essential to the stewardship of the nation's nuclear stockpile. Annual certification of the country's stockpile systems, Significant Finding Investigations (SFIs), and execution of Life Extension Programs (LEPs) are dependent on simulations employing the advanced ASC tools developed over the past decade plus; indeed, without these tools, certification would not be possible without a return to nuclear testing. ASC is an integrated program involving investments in computer hardware (platforms and computing centers), software environments, integrated design codes and physical models for these codes, and validation methodologies. The significant progress ASC has made in the past derives from its focus on mission and from its strategy of balancing support across the key investment areas necessary for success. All these investment areas must be sustained for ASC to adequately support current stockpile stewardship mission needs and to meet ever more difficult challenges as the weapons continue to age or undergo refurbishment. The appropriations bill called for this report to address three specific issues, which are responded to briefly here but are expanded upon in the subsequent document: (1) Identify how computing capability at each of the labs will specifically contribute to stockpile stewardship goals, and on what basis computing time will be allocated to achieve the goal of a balanced program among the labs. (2) Explain the NNSA's acquisition strategy for capacity and capability of machines at each of the labs and how it will fit within the existing budget constraints. (3

  11. FRANTIC: a computer code for time dependent unavailability analysis

    International Nuclear Information System (INIS)

    Vesely, W.E.; Goldberg, F.F.

    1977-03-01

    The FRANTIC computer code evaluates the time dependent and average unavailability for any general system model. The code is written in FORTRAN IV for the IBM 370 computer. Non-repairable components, monitored components, and periodically tested components are handled. One unique feature of FRANTIC is the detailed, time dependent modeling of periodic testing which includes the effects of test downtimes, test overrides, detection inefficiencies, and test-caused failures. The exponential distribution is used for the component failure times and periodic equations are developed for the testing and repair contributions. Human errors and common mode failures can be included by assigning an appropriate constant probability for the contributors. The output from FRANTIC consists of tables and plots of the system unavailability along with a breakdown of the unavailability contributions. Sensitivity studies can be simply performed and a wide range of tables and plots can be obtained for reporting purposes. The FRANTIC code represents a first step in the development of an approach that can be of direct value in future system evaluations. Modifications resulting from use of the code, along with the development of reliability data based on operating reactor experience, can be expected to provide increased confidence in its use and potential application to the licensing process

  12. A prospective study of spine fractures diagnosed by total spine computed tomography in high energy trauma patients

    International Nuclear Information System (INIS)

    Takami, Masanari; Nohda, Kazuhiro; Sakanaka, Junya; Nakamura, Masamichi; Yoshida, Munehito

    2011-01-01

    Since it is known to be impossible to identify spinal fractures in high-energy trauma patients the primary trauma evaluation, we have been performing total spine computed tomography (CT) in high-energy trauma cases. We investigated the spinal fractures that it was possible to detect by total spine CT in 179 cases and evaluated the usefulness of total spine CT prospectively. There were 54 (30.2%) spinal fractures among the 179 cases. Six (37.5%) of the 16 cervical spine fractures that were not detected on plain X-ray films were identified by total spine CT. Six (14.0%) of 43 thoracolumbar spine fractures were considered difficult to diagnose based on the clinical findings if total spine CT had not been performed. We therefore concluded that total spine CT is very useful and should be performed during the primary trauma evaluation in high-energy trauma cases. (author)

  13. Variable dead time counters: 2. A computer simulation

    International Nuclear Information System (INIS)

    Hooton, B.W.; Lees, E.W.

    1980-09-01

    A computer model has been developed to give a pulse train which simulates that generated by a variable dead time counter (VDC) used in safeguards determination of Pu mass. The model is applied to two algorithms generally used for VDC analysis. It is used to determine their limitations at high counting rates and to investigate the effects of random neutrons from (α,n) reactions. Both algorithms are found to be deficient for use with masses of 240 Pu greater than 100g and one commonly used algorithm is shown, by use of the model and also by theory, to yield a result which is dependent on the random neutron intensity. (author)

  14. [Determination of total and segmental colonic transit time in constipated children].

    Science.gov (United States)

    Zhang, Shu-cheng; Wang, Wei-lin; Bai, Yu-zuo; Yuan, Zheng-wei; Wang, Wei

    2003-03-01

    To determine the total and segmental colonic transit time of normal Chinese children and to explore its value in constipation in children. The subjects involved in this study were divided into 2 groups. One group was control, which had 33 healthy children (21 males and 12 females) aged 2 - 13 years (mean 5 years). The other was constipation group, which had 25 patients (15 males and 10 females) aged 3 - 14 years (mean 7 years) with constipation according to Benninga's criteria. Written informed consent was obtained from the parents of each subject. In this study the simplified method of radio opaque markers was used to determine the total gastrointestinal transit time and segmental colonic transit time of the normal and constipated children, and in part of these patients X-ray defecography was also used. The total gastrointestinal transit time (TGITT), right colonic transit time (RCTT), left colonic transit time (LCTT) and rectosigmoid colonic transit time (RSTT) of the normal children were 28.7 +/- 7.7 h, 7.5 +/- 3.2 h, 6.5 +/- 3.8 h and 13.4 +/- 5.6 h, respectively. In the constipated children, the TGITT, LCTT and RSTT were significantly longer than those in controls (92.2 +/- 55.5 h vs 28.7 +/- 7.7 h, P < 0.001; 16.9 +/- 12.6 h vs 6.5 +/- 3.8 h, P < 0.01; 61.5 +/- 29.0 h vs 13.4 +/- 5.6 h, P < 0.001), while the RCTT had no significant difference. X-ray defecography demonstrated one rectocele, one perineal descent syndrome and one puborectal muscle syndrome, respectively. The TGITT, RCTT, LCTT and RSTT of the normal children were 28.7 +/- 7.7 h, 7.5 +/- 3.2 h, 6.5 +/- 3.8 h and 13.4 +/- 5.6 h, respectively. With the segmental colonic transit time, constipation can be divided into four types: slow-transit constipation, outlet obstruction, mixed type and normal transit constipation. X-ray defecography can demonstrate the anatomical or dynamic abnormalities within the anorectal area, with which constipation can be further divided into different subtypes, and

  15. Computer-determined assay time based on preset precision

    International Nuclear Information System (INIS)

    Foster, L.A.; Hagan, R.; Martin, E.R.; Wachter, J.R.; Bonner, C.A.; Malcom, J.E.

    1994-01-01

    Most current assay systems for special nuclear materials (SNM) operate on the principle of a fixed assay time which provides acceptable measurement precision without sacrificing the required throughput of the instrument. Waste items to be assayed for SNM content can contain a wide range of nuclear material. Counting all items for the same preset assay time results in a wide range of measurement precision and wastes time at the upper end of the calibration range. A short time sample taken at the beginning of the assay could optimize the analysis time on the basis of the required measurement precision. To illustrate the technique of automatically determining the assay time, measurements were made with a segmented gamma scanner at the Plutonium Facility of Los Alamos National Laboratory with the assay time for each segment determined by counting statistics in that segment. Segments with very little SNM were quickly determined to be below the lower limit of the measurement range and the measurement was stopped. Segments with significant SNM were optimally assays to the preset precision. With this method the total assay time for each item is determined by the desired preset precision. This report describes the precision-based algorithm and presents the results of measurements made to test its validity

  16. First passage times in homogeneous nucleation: Dependence on the total number of particles

    International Nuclear Information System (INIS)

    Yvinec, Romain; Bernard, Samuel; Pujo-Menjouet, Laurent; Hingant, Erwan

    2016-01-01

    Motivated by nucleation and molecular aggregation in physical, chemical, and biological settings, we present an extension to a thorough analysis of the stochastic self-assembly of a fixed number of identical particles in a finite volume. We study the statistics of times required for maximal clusters to be completed, starting from a pure-monomeric particle configuration. For finite volumes, we extend previous analytical approaches to the case of arbitrary size-dependent aggregation and fragmentation kinetic rates. For larger volumes, we develop a scaling framework to study the first assembly time behavior as a function of the total quantity of particles. We find that the mean time to first completion of a maximum-sized cluster may have a surprisingly weak dependence on the total number of particles. We highlight how higher statistics (variance, distribution) of the first passage time may nevertheless help to infer key parameters, such as the size of the maximum cluster. Finally, we present a framework to quantify formation of macroscopic sized clusters, which are (asymptotically) very unlikely and occur as a large deviation phenomenon from the mean-field limit. We argue that this framework is suitable to describe phase transition phenomena, as inherent infrequent stochastic processes, in contrast to classical nucleation theory

  17. 17 CFR 402.2c - Appendix C-Consolidated computations of liquid capital and total haircuts for certain...

    Science.gov (United States)

    2010-04-01

    ... SECURITIES EXCHANGE ACT OF 1934 FINANCIAL RESPONSIBILITY § 402.2c Appendix C—Consolidated computations of liquid capital and total haircuts for certain subsidiaries and affiliates. (a) Consolidation. (1) A...) Principles of consolidation. The following minimum and non-exclusive requirements shall govern the...

  18. Empirical forecast of quiet time ionospheric Total Electron Content maps over Europe

    Science.gov (United States)

    Badeke, Ronny; Borries, Claudia; Hoque, Mainul M.; Minkwitz, David

    2018-06-01

    An accurate forecast of the atmospheric Total Electron Content (TEC) is helpful to investigate space weather influences on the ionosphere and technical applications like satellite-receiver radio links. The purpose of this work is to compare four empirical methods for a 24-h forecast of vertical TEC maps over Europe under geomagnetically quiet conditions. TEC map data are obtained from the Space Weather Application Center Ionosphere (SWACI) and the Universitat Politècnica de Catalunya (UPC). The time-series methods Standard Persistence Model (SPM), a 27 day median model (MediMod) and a Fourier Series Expansion are compared to maps for the entire year of 2015. As a representative of the climatological coefficient models the forecast performance of the Global Neustrelitz TEC model (NTCM-GL) is also investigated. Time periods of magnetic storms, which are identified with the Dst index, are excluded from the validation. By calculating the TEC values with the most recent maps, the time-series methods perform slightly better than the coefficient model NTCM-GL. The benefit of NTCM-GL is its independence on observational TEC data. Amongst the time-series methods mentioned, MediMod delivers the best overall performance regarding accuracy and data gap handling. Quiet-time SWACI maps can be forecasted accurately and in real-time by the MediMod time-series approach.

  19. The position of a standard optical computer mouse affects cardiorespiratory responses during the operation of a computer under time constraints.

    Science.gov (United States)

    Sako, Shunji; Sugiura, Hiromichi; Tanoue, Hironori; Kojima, Makoto; Kono, Mitsunobu; Inaba, Ryoichi

    2014-08-01

    This study investigated the association between task-induced stress and fatigue by examining the cardiovascular responses of subjects using different mouse positions while operating a computer under time constraints. The study was participated by 16 young, healthy men and examined the use of optical mouse devices affixed to laptop computers. Two mouse positions were investigated: (1) the distal position (DP), in which the subjects place their forearms on the desk accompanied by the abduction and flexion of their shoulder joints, and (2) the proximal position (PP), in which the subjects place only their wrists on the desk without using an armrest. The subjects continued each task for 16 min. We assessed differences in several characteristics according to mouse position, including expired gas values, autonomic nerve activities (based on cardiorespiratory responses), operating efficiencies (based on word counts), and fatigue levels (based on the visual analog scale - VAS). Oxygen consumption (VO(2)), the ratio of inspiration time to respiration time (T(i)/T(total)), respiratory rate (RR), minute ventilation (VE), and the ratio of expiration to inspiration (Te/T(i)) were significantly lower when the participants were performing the task in the DP than those obtained in the PP. Tidal volume (VT), carbon dioxide output rates (VCO(2)/VE), and oxygen extraction fractions (VO(2)/VE) were significantly higher for the DP than they were for the PP. No significant difference in VAS was observed between the positions; however, as the task progressed, autonomic nerve activities were lower and operating efficiencies were significantly higher for the DP than they were for the PP. Our results suggest that the DP has fewer effects on cardiorespiratory functions, causes lower levels of sympathetic nerve activity and mental stress, and produces a higher total workload than the PP. This suggests that the DP is preferable to the PP when operating a computer.

  20. Time Series Analysis, Modeling and Applications A Computational Intelligence Perspective

    CERN Document Server

    Chen, Shyi-Ming

    2013-01-01

    Temporal and spatiotemporal data form an inherent fabric of the society as we are faced with streams of data coming from numerous sensors, data feeds, recordings associated with numerous areas of application embracing physical and human-generated phenomena (environmental data, financial markets, Internet activities, etc.). A quest for a thorough analysis, interpretation, modeling and prediction of time series comes with an ongoing challenge for developing models that are both accurate and user-friendly (interpretable). The volume is aimed to exploit the conceptual and algorithmic framework of Computational Intelligence (CI) to form a cohesive and comprehensive environment for building models of time series. The contributions covered in the volume are fully reflective of the wealth of the CI technologies by bringing together ideas, algorithms, and numeric studies, which convincingly demonstrate their relevance, maturity and visible usefulness. It reflects upon the truly remarkable diversity of methodological a...

  1. Computer-guided total synthesis of natural products: Recent examples and future perspectives

    Energy Technology Data Exchange (ETDEWEB)

    Della-Felice, Franco; Pilli, Ronaldo A. [Universidade Estadual de Campinas (UNICAMP), SP (Brazil). Instituto de Química; Sarotti, Ariel M., E-mail: pilli@iqm.unicamp.br, E-mail: sarotti@iquir-conicet.gov.ar [Instituto de Química, Universidad Nacional de Rosario-CONICET (Argentina)

    2018-05-01

    Quantum chemical calculations of nuclear magnetic resonance (NMR) shifts and coupling constants have been extensively employed in recent years mainly to facilitate structural elucidation of organic molecules. When the results of such calculations are used to determine the most likely structure of a natural product in advance, guiding the subsequent synthetic work, the term 'computer-guided synthesis' could be coined. This review article describes the most relevant examples from recent literature, highlighting the scope and limitations of this merged computational/experimental approach as well. (author)

  2. Computer-guided total synthesis of natural products: Recent examples and future perspectives

    International Nuclear Information System (INIS)

    Della-Felice, Franco; Pilli, Ronaldo A.

    2018-01-01

    Quantum chemical calculations of nuclear magnetic resonance (NMR) shifts and coupling constants have been extensively employed in recent years mainly to facilitate structural elucidation of organic molecules. When the results of such calculations are used to determine the most likely structure of a natural product in advance, guiding the subsequent synthetic work, the term 'computer-guided synthesis' could be coined. This review article describes the most relevant examples from recent literature, highlighting the scope and limitations of this merged computational/experimental approach as well. (author)

  3. Minimizing total costs of forest roads with computer-aided design ...

    Indian Academy of Sciences (India)

    imum total road costs, while conforming to design specifications, environmental ..... quality, and enhancing fish and wildlife habitat, an appropriate design ..... Soil, Water and Timber Management: Forest Engineering Solutions in Response to.

  4. Computational modeling of the probability of destructions in total joint endoprosthesis ceramic heads using Weibull's theory

    Czech Academy of Sciences Publication Activity Database

    Janíček, P.; Fuis, Vladimír; Málek, M.

    2010-01-01

    Roč. 14, č. 4 (2010), s. 42-51 ISSN 1335-2393 Institutional research plan: CEZ:AV0Z20760514 Keywords : computational modeling * ceramic head * in vivo destructions * hip joint endoprosthesis * probabily of rupture Subject RIV: BO - Biophysics

  5. Reduced Operating Time but Not Blood Loss With Cruciate Retaining Total Knee Arthroplasty

    Science.gov (United States)

    Vermesan, Dinu; Trocan, Ilie; Prejbeanu, Radu; Poenaru, Dan V; Haragus, Horia; Gratian, Damian; Marrelli, Massimo; Inchingolo, Francesco; Caprio, Monica; Cagiano, Raffaele; Tatullo, Marco

    2015-01-01

    Background There is no consensus regarding the use of retaining or replacing cruciate implants for patients with limited deformity who undergo a total knee replacement. Scope of this paper is to evaluate whether a cruciate sparing total knee replacement could have a reduced operating time compared to a posterior stabilized implant. Methods For this purpose, we performed a randomized study on 50 subjects. All procedures were performed by a single surgeon in the same conditions to minimize bias and only knees with a less than 20 varus deviation and/or maximum 15° fixed flexion contracture were included. Results Surgery time was significantly shorter with the cruciate retaining implant (P = 0.0037). The mean duration for the Vanguard implant was 68.9 (14.7) and for the NexGen II Legacy was 80.2 (11.3). A higher range of motion, but no significant Knee Society Scores at 6 months follow-up, was used as controls. Conclusions In conclusion, both implants had the potential to assure great outcomes. However, if a decision has to be made, choosing a cruciate retaining procedure could significantly reduce the surgical time. When performed under tourniquet, this gain does not lead to reduced blood loss. PMID:25584102

  6. Real-time fusion of coronary CT angiography with x-ray fluoroscopy during chronic total occlusion PCI.

    Science.gov (United States)

    Ghoshhajra, Brian B; Takx, Richard A P; Stone, Luke L; Girard, Erin E; Brilakis, Emmanouil S; Lombardi, William L; Yeh, Robert W; Jaffer, Farouc A

    2017-06-01

    The purpose of this study was to demonstrate the feasibility of real-time fusion of coronary computed tomography angiography (CTA) centreline and arterial wall calcification with x-ray fluoroscopy during chronic total occlusion (CTO) percutaneous coronary intervention (PCI). Patients undergoing CTO PCI were prospectively enrolled. Pre-procedural CT scans were integrated with conventional coronary fluoroscopy using prototype software. We enrolled 24 patients who underwent CTO PCI using the prototype CT fusion software, and 24 consecutive CTO PCI patients without CT guidance served as a control group. Mean age was 66 ± 11 years, and 43/48 patients were men. Real-time CTA fusion during CTO PCI provided additional information regarding coronary arterial calcification and tortuosity that generated new insights into antegrade wiring, antegrade dissection/reentry, and retrograde wiring during CTO PCI. Overall CTO success rates and procedural outcomes remained similar between the two groups, despite a trend toward higher complexity in the fusion CTA group. This study demonstrates that real-time automated co-registration of coronary CTA centreline and calcification onto live fluoroscopic images is feasible and provides new insights into CTO PCI, and in particular, antegrade dissection reentry-based CTO PCI. • Real-time semi-automated fusion of CTA/fluoroscopy is feasible during CTO PCI. • CTA fusion data can be toggled on/off as desired during CTO PCI • Real-time CT calcium and centreline overlay could benefit antegrade dissection/reentry-based CTO PCI.

  7. Computer-assisted total knee arthroplasty marketing and patient education: an evaluation of quality, content and accuracy of related websites.

    Science.gov (United States)

    Shemesh, Shai S; Bronson, Michael J; Moucha, Calin S

    2016-10-01

    The internet is increasingly being used as a resource for health-related information by the general public. We sought to establish the authorship, content and accuracy of the information available online regarding computer-assisted total knee arthroplasty (CA-TKA). One hundred fifty search results from three leading search engines available online (Google, Yahoo!, Bing) from ten different countries worldwide were reviewed. While private physicians/groups authored 50.7 % of the websites, only 17.3 % were authored by a hospital/university. As compared to traditional TKA, 59.3 % of the websites claimed that navigated TKA offers better longevity, 46.6 % claimed accelerated recovery and 26 % claimed fewer complications. Only 11.3 % mentioned the prolonged operating room time required, and only 15.3 % noted the current lack of long-term evidence in support of this technology. Patients seeking information regarding CA-TKA through the major search engines are likely to encounter websites presenting a narrow, unscientific, viewpoint of the present technology, putting emphasis on unsubstantiated benefits while disregarding potential drawbacks. Survey of Materials-Internet.

  8. Time-dependent density functional theory description of total photoabsorption cross sections

    Science.gov (United States)

    Tenorio, Bruno Nunes Cabral; Nascimento, Marco Antonio Chaer; Rocha, Alexandre Braga

    2018-02-01

    The time-dependent version of the density functional theory (TDDFT) has been used to calculate the total photoabsorption cross section of a number of molecules, namely, benzene, pyridine, furan, pyrrole, thiophene, phenol, naphthalene, and anthracene. The discrete electronic pseudo-spectra, obtained in a L2 basis set calculation were used in an analytic continuation procedure to obtain the photoabsorption cross sections. The ammonia molecule was chosen as a model system to compare the results obtained with TDDFT to those obtained with the linear response coupled cluster approach in order to make a link with our previous work and establish benchmarks.

  9. Extending DIII-D Neutral Beam Modulated Operations with a Camac Based Total on Time Interlock

    International Nuclear Information System (INIS)

    Baggest, D.S.; Broesch, J.D.; Phillips, J.C.

    1999-01-01

    A new total-on-time interlock has increased the operational time limits of the Neutral Beam systems at DIII-D. The interlock, called the Neutral Beam On-Time-Limiter (NBOTL), is a custom built CAMAC module utilizing a Xilinx 9572 Complex Programmable Logic Device (CPLD) as its primary circuit. The Neutral Beam Injection Systems are the primary source of auxiliary heating for DIII-D plasma discharges and contain eight sources capable of delivering 20MW of power. The delivered power is typically limited to 3.5 s per source to protect beam-line components, while a DIII-D plasma discharge usually exceeds 5 s. Implemented as a hardware interlock within the neutral beam power supplies, the NBOTL limits the beam injection time. With a continuing emphasis on modulated beam injections, the NBOTL guards against command faults and allows the beam injection to be safely spread over a longer plasma discharge time. The NBOTL design is an example of incorporating modern circuit design techniques (CPLD) within an established format (CAMAC). The CPLD is the heart of the NBOTL and contains 90% of the circuitry, including a loadable, 1 MHz, 28 bit, BCD count down timer, buffers, and CAMAC communication circuitry. This paper discusses the circuit design and implementation. Of particular interest is the melding of flexible modern programmable logic devices with the CAMAC format

  10. Time delay and duration of ionospheric total electron content responses to geomagnetic disturbances

    Directory of Open Access Journals (Sweden)

    J. Liu

    2010-03-01

    Full Text Available Although positive and negative signatures of ionospheric storms have been reported many times, global characteristics such as the time of occurrence, time delay and duration as well as their relations to the intensity of the ionospheric storms have not received enough attention. The 10 years of global ionosphere maps (GIMs of total electron content (TEC retrieved at Jet Propulsion Laboratory (JPL were used to conduct a statistical study of the time delay of the ionospheric responses to geomagnetic disturbances. Our results show that the time delays between geomagnetic disturbances and TEC responses depend on season, magnetic local time and magnetic latitude. In the summer hemisphere at mid- and high latitudes, the negative storm effects can propagate to the low latitudes at post-midnight to the morning sector with a time delay of 4–7 h. As the earth rotates to the sunlight, negative phase retreats to higher latitudes and starts to extend to the lower latitude toward midnight sector. In the winter hemisphere during the daytime and after sunset at mid- and low latitudes, the negative phase appearance time is delayed from 1–10 h depending on the local time, latitude and storm intensity compared to the same area in the summer hemisphere. The quick response of positive phase can be observed at the auroral area in the night-side of the winter hemisphere. At the low latitudes during the dawn-noon sector, the ionospheric negative phase responses quickly with time delays of 5–7 h in both equinoctial and solsticial months.

    Our results also manifest that there is a positive correlation between the intensity of geomagnetic disturbances and the time duration of both the positive phase and negative phase. The durations of both negative phase and positive phase have clear latitudinal, seasonal and magnetic local time (MLT dependence. In the winter hemisphere, long durations for the positive phase are 8–11 h and 12–14 h during the daytime at

  11. Time delay and duration of ionospheric total electron content responses to geomagnetic disturbances

    Directory of Open Access Journals (Sweden)

    J. Liu

    2010-03-01

    Full Text Available Although positive and negative signatures of ionospheric storms have been reported many times, global characteristics such as the time of occurrence, time delay and duration as well as their relations to the intensity of the ionospheric storms have not received enough attention. The 10 years of global ionosphere maps (GIMs of total electron content (TEC retrieved at Jet Propulsion Laboratory (JPL were used to conduct a statistical study of the time delay of the ionospheric responses to geomagnetic disturbances. Our results show that the time delays between geomagnetic disturbances and TEC responses depend on season, magnetic local time and magnetic latitude. In the summer hemisphere at mid- and high latitudes, the negative storm effects can propagate to the low latitudes at post-midnight to the morning sector with a time delay of 4–7 h. As the earth rotates to the sunlight, negative phase retreats to higher latitudes and starts to extend to the lower latitude toward midnight sector. In the winter hemisphere during the daytime and after sunset at mid- and low latitudes, the negative phase appearance time is delayed from 1–10 h depending on the local time, latitude and storm intensity compared to the same area in the summer hemisphere. The quick response of positive phase can be observed at the auroral area in the night-side of the winter hemisphere. At the low latitudes during the dawn-noon sector, the ionospheric negative phase responses quickly with time delays of 5–7 h in both equinoctial and solsticial months. Our results also manifest that there is a positive correlation between the intensity of geomagnetic disturbances and the time duration of both the positive phase and negative phase. The durations of both negative phase and positive phase have clear latitudinal, seasonal and magnetic local time (MLT dependence. In the winter hemisphere, long durations for the positive phase are 8–11 h and 12–14 h during the daytime at middle

  12. Can a surgery-first orthognathic approach reduce the total treatment time?

    Science.gov (United States)

    Jeong, Woo Shik; Choi, Jong Woo; Kim, Do Yeon; Lee, Jang Yeol; Kwon, Soon Man

    2017-04-01

    Although pre-surgical orthodontic treatment has been accepted as a necessary process for stable orthognathic correction in the traditional orthognathic approach, recent advances in the application of miniscrews and in the pre-surgical simulation of orthodontic management using dental models have shown that it is possible to perform a surgery-first orthognathic approach without pre-surgical orthodontic treatment. This prospective study investigated the surgical outcomes of patients with diagnosed skeletal class III dentofacial deformities who underwent orthognathic surgery between December 2007 and December 2014. Cephalometric landmark data for patients undergoing the surgery-first approach were analyzed in terms of postoperative changes in vertical and horizontal skeletal pattern, dental pattern, and soft tissue profile. Forty-five consecutive Asian patients with skeletal class III dentofacial deformities who underwent surgery-first orthognathic surgery and 52 patients who underwent conventional two-jaw orthognathic surgery were included. The analysis revealed that the total treatment period for the surgery-first approach averaged 14.6 months, compared with 22.0 months for the orthodontics-first approach. Comparisons between the immediate postoperative and preoperative and between the postoperative and immediate postoperative cephalometric data revealed factors that correlated with the total treatment duration. The surgery-first orthognathic approach can dramatically reduce the total treatment time, with no major complications. Copyright © 2016 International Association of Oral and Maxillofacial Surgeons. Published by Elsevier Ltd. All rights reserved.

  13. Time-driven Activity-based Cost of Fast-Track Total Hip and Knee Arthroplasty

    DEFF Research Database (Denmark)

    Andreasen, Signe E; Holm, Henriette B; Jørgensen, Mira

    2017-01-01

    this between 2 departments with different logistical set-ups. METHODS: Prospective data collection was analyzed using the time-driven activity-based costing method (TDABC) on time consumed by different staff members involved in patient treatment in the perioperative period of fast-track THA and TKA in 2 Danish...... orthopedic departments with standardized fast-track settings, but different logistical set-ups. RESULTS: Length of stay was median 2 days in both departments. TDABC revealed minor differences in the perioperative settings between departments, but the total cost excluding the prosthesis was similar at USD......-track methodology, the result could be a more cost-effective pathway altogether. As THA and TKA are potentially costly procedures and the numbers are increasing in an economical limited environment, the aim of this study is to present baseline detailed economical calculations of fast-track THA and TKA and compare...

  14. Minimizing the Total Service Time of Discrete Dynamic Berth Allocation Problem by an Iterated Greedy Heuristic

    Science.gov (United States)

    2014-01-01

    Berth allocation is the forefront operation performed when ships arrive at a port and is a critical task in container port optimization. Minimizing the time ships spend at berths constitutes an important objective of berth allocation problems. This study focuses on the discrete dynamic berth allocation problem (discrete DBAP), which aims to minimize total service time, and proposes an iterated greedy (IG) algorithm to solve it. The proposed IG algorithm is tested on three benchmark problem sets. Experimental results show that the proposed IG algorithm can obtain optimal solutions for all test instances of the first and second problem sets and outperforms the best-known solutions for 35 out of 90 test instances of the third problem set. PMID:25295295

  15. Minimizing the Total Service Time of Discrete Dynamic Berth Allocation Problem by an Iterated Greedy Heuristic

    Directory of Open Access Journals (Sweden)

    Shih-Wei Lin

    2014-01-01

    Full Text Available Berth allocation is the forefront operation performed when ships arrive at a port and is a critical task in container port optimization. Minimizing the time ships spend at berths constitutes an important objective of berth allocation problems. This study focuses on the discrete dynamic berth allocation problem (discrete DBAP, which aims to minimize total service time, and proposes an iterated greedy (IG algorithm to solve it. The proposed IG algorithm is tested on three benchmark problem sets. Experimental results show that the proposed IG algorithm can obtain optimal solutions for all test instances of the first and second problem sets and outperforms the best-known solutions for 35 out of 90 test instances of the third problem set.

  16. Decreasing Postanesthesia Care Unit to Floor Transfer Times to Facilitate Short Stay Total Joint Replacements.

    Science.gov (United States)

    Sibia, Udai S; Grover, Jennifer; Turcotte, Justin J; Seanger, Michelle L; England, Kimberly A; King, Jennifer L; King, Paul J

    2018-04-01

    We describe a process for studying and improving baseline postanesthesia care unit (PACU)-to-floor transfer times after total joint replacements. Quality improvement project using lean methodology. Phase I of the investigational process involved collection of baseline data. Phase II involved developing targeted solutions to improve throughput. Phase III involved measured project sustainability. Phase I investigations revealed that patients spent an additional 62 minutes waiting in the PACU after being designated ready for transfer. Five to 16 telephone calls were needed between the PACU and the unit to facilitate each patient transfer. The most common reason for delay was unavailability of the unit nurse who was attending to another patient (58%). Phase II interventions resulted in transfer times decreasing to 13 minutes (79% reduction, P care at other institutions. Copyright © 2016 American Society of PeriAnesthesia Nurses. Published by Elsevier Inc. All rights reserved.

  17. Multiscale Methods, Parallel Computation, and Neural Networks for Real-Time Computer Vision.

    Science.gov (United States)

    Battiti, Roberto

    1990-01-01

    This thesis presents new algorithms for low and intermediate level computer vision. The guiding ideas in the presented approach are those of hierarchical and adaptive processing, concurrent computation, and supervised learning. Processing of the visual data at different resolutions is used not only to reduce the amount of computation necessary to reach the fixed point, but also to produce a more accurate estimation of the desired parameters. The presented adaptive multiple scale technique is applied to the problem of motion field estimation. Different parts of the image are analyzed at a resolution that is chosen in order to minimize the error in the coefficients of the differential equations to be solved. Tests with video-acquired images show that velocity estimation is more accurate over a wide range of motion with respect to the homogeneous scheme. In some cases introduction of explicit discontinuities coupled to the continuous variables can be used to avoid propagation of visual information from areas corresponding to objects with different physical and/or kinematic properties. The human visual system uses concurrent computation in order to process the vast amount of visual data in "real -time." Although with different technological constraints, parallel computation can be used efficiently for computer vision. All the presented algorithms have been implemented on medium grain distributed memory multicomputers with a speed-up approximately proportional to the number of processors used. A simple two-dimensional domain decomposition assigns regions of the multiresolution pyramid to the different processors. The inter-processor communication needed during the solution process is proportional to the linear dimension of the assigned domain, so that efficiency is close to 100% if a large region is assigned to each processor. Finally, learning algorithms are shown to be a viable technique to engineer computer vision systems for different applications starting from

  18. Neural Computations in a Dynamical System with Multiple Time Scales

    Directory of Open Access Journals (Sweden)

    Yuanyuan Mi

    2016-09-01

    Full Text Available Neural systems display rich short-term dynamics at various levels, e.g., spike-frequencyadaptation (SFA at single neurons, and short-term facilitation (STF and depression (STDat neuronal synapses. These dynamical features typically covers a broad range of time scalesand exhibit large diversity in different brain regions. It remains unclear what the computationalbenefit for the brain to have such variability in short-term dynamics is. In this study, we proposethat the brain can exploit such dynamical features to implement multiple seemingly contradictorycomputations in a single neural circuit. To demonstrate this idea, we use continuous attractorneural network (CANN as a working model and include STF, SFA and STD with increasing timeconstants in their dynamics. Three computational tasks are considered, which are persistent activity,adaptation, and anticipative tracking. These tasks require conflicting neural mechanisms, andhence cannot be implemented by a single dynamical feature or any combination with similar timeconstants. However, with properly coordinated STF, SFA and STD, we show that the network isable to implement the three computational tasks concurrently. We hope this study will shed lighton the understanding of how the brain orchestrates its rich dynamics at various levels to realizediverse cognitive functions.

  19. Chemistry, physics and time: the computer modelling of glassmaking.

    Science.gov (United States)

    Martlew, David

    2003-01-01

    A decade or so ago the remains of an early flat glass furnace were discovered in St Helens. Continuous glass production only became feasible after the Siemens Brothers demonstrated their continuous tank furnace at Dresden in 1870. One manufacturer of flat glass enthusiastically adopted the new technology and secretly explored many variations on this theme during the next fifteen years. Study of the surviving furnace remains using today's computer simulation techniques showed how, in 1887, that technology was adapted to the special demands of window glass making. Heterogeneous chemical reactions at high temperatures are required to convert the mixture of granular raw materials into the homogeneous glass needed for windows. Kinetics (and therefore the economics) of glassmaking is dominated by heat transfer and chemical diffusion as refractory grains are converted to highly viscous molten glass. Removal of gas bubbles in a sufficiently short period of time is vital for profitability, but the glassmaker must achieve this in a reaction vessel which is itself being dissolved by the molten glass. Design and operational studies of today's continuous tank furnaces need to take account of these factors, and good use is made of computer simulation techniques to shed light on the way furnaces behave and how improvements may be made. This paper seeks to show how those same techniques can be used to understand how the early Siemens continuous tank furnaces were designed and operated, and how the Victorian entrepreneurs succeeded in managing the thorny problems of what was, in effect, a vulnerable high temperature continuous chemical reactor.

  20. Time-gated scintillator imaging for real-time optical surface dosimetry in total skin electron therapy

    Science.gov (United States)

    Bruza, Petr; Gollub, Sarah L.; Andreozzi, Jacqueline M.; Tendler, Irwin I.; Williams, Benjamin B.; Jarvis, Lesley A.; Gladstone, David J.; Pogue, Brian W.

    2018-05-01

    The purpose of this study was to measure surface dose by remote time-gated imaging of plastic scintillators. A novel technique for time-gated, intensified camera imaging of scintillator emission was demonstrated, and key parameters influencing the signal were analyzed, including distance, angle and thickness. A set of scintillator samples was calibrated by using thermo-luminescence detector response as reference. Examples of use in total skin electron therapy are described. The data showed excellent room light rejection (signal-to-noise ratio of scintillation SNR  ≈  470), ideal scintillation dose response linearity, and 2% dose rate error. Individual sample scintillation response varied by 7% due to sample preparation. Inverse square distance dependence correction and lens throughput error (8% per meter) correction were needed. At scintillator-to-source angle and observation angle  <50°, the radiant energy fluence error was smaller than 1%. The achieved standard error of the scintillator cumulative dose measurement compared to the TLD dose was 5%. The results from this proof-of-concept study documented the first use of small scintillator targets for remote surface dosimetry in ambient room lighting. The measured dose accuracy renders our method to be comparable to thermo-luminescent detector dosimetry, with the ultimate realization of accuracy likely to be better than shown here. Once optimized, this approach to remote dosimetry may substantially reduce the time and effort required for surface dosimetry.

  1. Near-real-time Estimation and Forecast of Total Precipitable Water in Europe

    Science.gov (United States)

    Bartholy, J.; Kern, A.; Barcza, Z.; Pongracz, R.; Ihasz, I.; Kovacs, R.; Ferencz, C.

    2013-12-01

    Information about the amount and spatial distribution of atmospheric water vapor (or total precipitable water) is essential for understanding weather and the environment including the greenhouse effect, the climate system with its feedbacks and the hydrological cycle. Numerical weather prediction (NWP) models need accurate estimations of water vapor content to provide realistic forecasts including representation of clouds and precipitation. In the present study we introduce our research activity for the estimation and forecast of atmospheric water vapor in Central Europe using both observations and models. The Eötvös Loránd University (Hungary) operates a polar orbiting satellite receiving station in Budapest since 2002. This station receives Earth observation data from polar orbiting satellites including MODerate resolution Imaging Spectroradiometer (MODIS) Direct Broadcast (DB) data stream from satellites Terra and Aqua. The received DB MODIS data are automatically processed using freely distributed software packages. Using the IMAPP Level2 software total precipitable water is calculated operationally using two different methods. Quality of the TPW estimations is a crucial question for further application of the results, thus validation of the remotely sensed total precipitable water fields is presented using radiosonde data. In a current research project in Hungary we aim to compare different estimations of atmospheric water vapor content. Within the frame of the project we use a NWP model (DBCRAS; Direct Broadcast CIMSS Regional Assimilation System numerical weather prediction software developed by the University of Wisconsin, Madison) to forecast TPW. DBCRAS uses near real time Level2 products from the MODIS data processing chain. From the wide range of the derived Level2 products the MODIS TPW parameter found within the so-called mod07 results (Atmospheric Profiles Product) and the cloud top pressure and cloud effective emissivity parameters from the so

  2. Smoking is associated with earlier time to revision of total knee arthroplasty.

    Science.gov (United States)

    Lim, Chin Tat; Goodman, Stuart B; Huddleston, James I; Harris, Alex H S; Bhowmick, Subhrojyoti; Maloney, William J; Amanatullah, Derek F

    2017-10-01

    Smoking is associated with early postoperative complications, increased length of hospital stay, and an increased risk of revision after total knee arthroplasty (TKA). However, the effect of smoking on time to revision TKA is unknown. A total of 619 primary TKAs referred to an academic tertiary center for revision TKA were retrospectively stratified according to the patient smoking status. Smoking status was then analyzed for associations with time to revision TKA using a Chi square test. The association was also analyzed according to the indication for revision TKA. Smokers (37/41, 90%) have an increased risk of earlier revision for any reason compared to non-smokers (274/357, 77%, p=0.031). Smokers (37/41, 90%) have an increased risk of earlier revision for any reason compared to ex-smokers (168/221, 76%, p=0.028). Subgroup analysis did not reveal a difference in indication for revision TKA (p>0.05). Smokers are at increased risk of earlier revision TKA when compared to non-smokers and ex-smokers. The risk for ex-smokers was similar to that of non-smokers. Smoking appears to have an all-or-none effect on earlier revision TKA as patients who smoked more did not have higher risk of early revision TKA. These results highlight the need for clinicians to urge patients not to begin smoking and encourage smokers to quit smoking prior to primary TKA. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Time-driven activity based costing of total knee replacement surgery at a London teaching hospital.

    Science.gov (United States)

    Chen, Alvin; Sabharwal, Sanjeeve; Akhtar, Kashif; Makaram, Navnit; Gupte, Chinmay M

    2015-12-01

    The aim of this study was to conduct a time-driven activity based costing (TDABC) analysis of the clinical pathway for total knee replacement (TKR) and to determine where the major cost drivers lay. The in-patient pathway was prospectively mapped utilising a TDABC model, following 20 TKRs. The mean age for these patients was 73.4 years. All patients were ASA grade I or II and their mean BMI was 30.4. The 14 varus knees had a mean deformity of 5.32° and the six valgus knee had a mean deformity of 10.83°. Timings were prospectively collected as each patient was followed through the TKR pathway. Pre-operative costs including pre-assessment and joint school were £ 163. Total staff costs for admission and the operating theatre were £ 658. Consumables cost for the operating theatre were £ 1862. The average length of stay was 5.25 days at a total cost of £ 910. Trust overheads contributed £ 1651. The overall institutional cost of a 'noncomplex' TKR in patients without substantial medical co-morbidities was estimated to be £ 5422, representing a profit of £ 1065 based on a best practice tariff of £ 6487. The major cost drivers in the TKR pathway were determined to be theatre consumables, corporate overheads, overall ward cost and operating theatre staffing costs. Appropriate discounting of implant costs, reduction in length of stay by adopting an enhanced recovery programme and control of corporate overheads through the use of elective orthopaedic treatment centres are proposed approaches for reducing the overall cost of treatment. Copyright © 2015 Elsevier B.V. All rights reserved.

  4. Computations of concentration of radon and its decay products against time. Computer program; Obliczanie koncentracji radonu i jego produktow rozpadu w funkcji czasu. Program komputerowy

    Energy Technology Data Exchange (ETDEWEB)

    Machaj, B. [Institute of Nuclear Chemistry and Technology, Warsaw (Poland)

    1996-12-31

    This research is aimed to develop a device for continuous monitoring of radon in the air, by measuring alpha activity of radon and its short lived decay products. The influence of alpha activity variation of radon and its daughters on the measured results is of importance and requires a knowledge of this variation with time. Employing the measurement of alpha radiation of radon and of its short lived decay products, require knowledge of radon concentration variation and its decay products against the time. A computer program in Turbo Pascal language was therefore developed performing the computations employing the known relations involved, the program being adapted for IBM PC computers. The presented program enables computation of activity of {sup 222}Rn and its daughter products: {sup 218}Po, {sup 214}Pb, {sup 214}Bi and {sup 214}Po every 1 min within the period of 0-255 min for any state of radiation equilibrium between the radon and its daughter products. The program permits also to compute alpha activity of {sup 222}Rn + {sup 218}Po + {sup 214}Po against time and the total alpha activity at selected interval of time. The results of computations are stored on the computer hard disk in ASCII format and are used a graphic program e.g. by DrawPerfect program to make diagrams. Equations employed for computation of the alpha activity of radon and its decay products as well as the description of program functions are given. (author). 2 refs, 4 figs.

  5. Manual cross check of computed dose times for motorised wedged fields

    International Nuclear Information System (INIS)

    Porte, J.

    2001-01-01

    If a mass of tissue equivalent material is exposed in turn to wedged and open radiation fields of the same size, for equal times, it is incorrect to assume that the resultant isodose pattern will be effectively that of a wedge having half the angle of the wedged field. Computer programs have been written to address the problem of creating an intermediate wedge field, commonly known as a motorized wedge. The total exposure time is apportioned between the open and wedged fields, to produce a beam modification equivalent to that of a wedged field of a given wedge angle. (author)

  6. Total vaginectomy and urethral lengthening at time of neourethral prelamination in transgender men.

    Science.gov (United States)

    Medina, Carlos A; Fein, Lydia A; Salgado, Christopher J

    2017-11-29

    For transgender men (TGM), gender-affirmation surgery (GAS) is often the final stage of their gender transition. GAS involves creating a neophallus, typically using tissue remote from the genital region, such as radial forearm free-flap phalloplasty. Essential to this process is vaginectomy. Complexity of vaginal fascial attachments, atrophy due to testosterone use, and need to preserve integrity of the vaginal epithelium for tissue rearrangement add to the intricacy of the procedure during GAS. We designed the technique presented here to minimize complications and contribute to overall success of the phalloplasty procedure. After obtaining approval from the Institutional Review Board, our transgender (TG) database at the University of Miami Hospital was reviewed to identify cases with vaginectomy and urethral elongation performed at the time of radial forearm free-flap phalloplasty prelamination. Surgical technique for posterior vaginectomy and anterior vaginal wall-flap harvest with subsequent urethral lengthening is detailed. Six patients underwent total vaginectomy and urethral elongation at the time of radial forearm free-flap phalloplasty prelamination. Mean estimated blood loss (EBL) was 290 ± 199.4 ml for the vaginectomy and urethral elongation, and no one required transfusion. There were no intraoperative complications (cystotomy, ureteral obstruction, enterotomy, proctotomy, or neurological injury). One patient had a urologic complication (urethral stricture) in the neobulbar urethra. Total vaginectomy and urethral lengthening procedures at the time of GAS are relatively safe procedures, and using the described technique provides excellent tissue for urethral prelamination and a low complication rate in both the short and long term.

  7. Timing of urinary catheter removal after uncomplicated total abdominal hysterectomy: a prospective randomized trial.

    Science.gov (United States)

    Ahmed, Magdy R; Sayed Ahmed, Waleed A; Atwa, Khaled A; Metwally, Lobna

    2014-05-01

    To assess whether immediate (0h), intermediate (after 6h) or delayed (after 24h) removal of an indwelling urinary catheter after uncomplicated abdominal hysterectomy can affect the rate of re-catheterization due to urinary retention, rate of urinary tract infection, ambulation time and length of hospital stay. Prospective randomized controlled trial conducted at Suez Canal University Hospital, Egypt. Two hundred and twenty-one women underwent total abdominal hysterectomy for benign gynecological diseases and were randomly allocated into three groups. Women in group A (73 patients) had their urinary catheter removed immediately after surgery. Group B (81 patients) had the catheter removed 6h post-operatively while in group C (67 patients) the catheter was removed after 24h. The main outcome measures were the frequency of urinary retention, urinary tract infections, ambulation time and length of hospital stay. There was a significantly higher number of urinary retention episodes requiring re-catheterization in the immediate removal group compared to the intermediate and delayed removal groups (16.4% versus 2.5% and 0% respectively). Delayed urinary catheter removal was associated with a higher incidence of urinary tract infections (15%), delayed ambulation time (10.3h) and longer hospital stay (5.6 days) compared to the early (1.4%, 4.1h and 3.2 days respectively) and intermediate (3.7%, 6.8h and 3.4 days respectively) removal groups. Removal of the urinary catheter 6h postoperatively appears to be more advantageous than early or late removal in cases of uncomplicated total abdominal hysterectomy. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.

  8. Aligator: A computational tool for optimizing total chemical synthesis of large proteins.

    Science.gov (United States)

    Jacobsen, Michael T; Erickson, Patrick W; Kay, Michael S

    2017-09-15

    The scope of chemical protein synthesis (CPS) continues to expand, driven primarily by advances in chemical ligation tools (e.g., reversible solubilizing groups and novel ligation chemistries). However, the design of an optimal synthesis route can be an arduous and fickle task due to the large number of theoretically possible, and in many cases problematic, synthetic strategies. In this perspective, we highlight recent CPS tool advances and then introduce a new and easy-to-use program, Aligator (Automated Ligator), for analyzing and designing the most efficient strategies for constructing large targets using CPS. As a model set, we selected the E. coli ribosomal proteins and associated factors for computational analysis. Aligator systematically scores and ranks all feasible synthetic strategies for a particular CPS target. The Aligator script methodically evaluates potential peptide segments for a target using a scoring function that includes solubility, ligation site quality, segment lengths, and number of ligations to provide a ranked list of potential synthetic strategies. We demonstrate the utility of Aligator by analyzing three recent CPS projects from our lab: TNFα (157 aa), GroES (97 aa), and DapA (312 aa). As the limits of CPS are extended, we expect that computational tools will play an increasingly important role in the efficient execution of ambitious CPS projects such as production of a mirror-image ribosome. Copyright © 2017 Elsevier Ltd. All rights reserved.

  9. Saddlepoint approximation to the distribution of the total distance of the continuous time random walk

    Science.gov (United States)

    Gatto, Riccardo

    2017-12-01

    This article considers the random walk over Rp, with p ≥ 2, where a given particle starts at the origin and moves stepwise with uniformly distributed step directions and step lengths following a common distribution. Step directions and step lengths are independent. The case where the number of steps of the particle is fixed and the more general case where it follows an independent continuous time inhomogeneous counting process are considered. Saddlepoint approximations to the distribution of the distance from the position of the particle to the origin are provided. Despite the p-dimensional nature of the random walk, the computations of the saddlepoint approximations are one-dimensional and thus simple. Explicit formulae are derived with dimension p = 3: for uniformly and exponentially distributed step lengths, for fixed and for Poisson distributed number of steps. In these situations, the high accuracy of the saddlepoint approximations is illustrated by numerical comparisons with Monte Carlo simulation. Contribution to the "Topical Issue: Continuous Time Random Walk Still Trendy: Fifty-year History, Current State and Outlook", edited by Ryszard Kutner and Jaume Masoliver.

  10. Observed and simulated time evolution of HCl, ClONO2, and HF total column abundances

    Directory of Open Access Journals (Sweden)

    B.-M. Sinnhuber

    2012-04-01

    Full Text Available Time series of total column abundances of hydrogen chloride (HCl, chlorine nitrate (ClONO2, and hydrogen fluoride (HF were determined from ground-based Fourier transform infrared (FTIR spectra recorded at 17 sites belonging to the Network for the Detection of Atmospheric Composition Change (NDACC and located between 80.05° N and 77.82° S. By providing such a near-global overview on ground-based measurements of the two major stratospheric chlorine reservoir species, HCl and ClONO2, the present study is able to confirm the decrease of the atmospheric inorganic chlorine abundance during the last few years. This decrease is expected following the 1987 Montreal Protocol and its amendments and adjustments, where restrictions and a subsequent phase-out of the prominent anthropogenic chlorine source gases (solvents, chlorofluorocarbons were agreed upon to enable a stabilisation and recovery of the stratospheric ozone layer. The atmospheric fluorine content is expected to be influenced by the Montreal Protocol, too, because most of the banned anthropogenic gases also represent important fluorine sources. But many of the substitutes to the banned gases also contain fluorine so that the HF total column abundance is expected to have continued to increase during the last few years. The measurements are compared with calculations from five different models: the two-dimensional Bremen model, the two chemistry-transport models KASIMA and SLIMCAT, and the two chemistry-climate models EMAC and SOCOL. Thereby, the ability of the models to reproduce the absolute total column amounts, the seasonal cycles, and the temporal evolution found in the FTIR measurements is investigated and inter-compared. This is especially interesting because the models have different architectures. The overall agreement between the measurements and models for the total column abundances and the seasonal cycles is good. Linear trends of HCl, ClONO2, and HF are calculated from both

  11. The influence of tourniquet use and operative time on the incidence of deep vein thrombosis in total knee arthroplasty.

    Science.gov (United States)

    Hernandez, Arnaldo José; Almeida, Adriano Marques de; Fávaro, Edmar; Sguizzato, Guilherme Turola

    2012-09-01

    To evaluate the association between tourniquet and total operative time during total knee arthroplasty and the occurrence of deep vein thrombosis. Seventy-eight consecutive patients from our institution underwent cemented total knee arthroplasty for degenerative knee disorders. The pneumatic tourniquet time and total operative time were recorded in minutes. Four categories were established for total tourniquet time: 120 minutes. Three categories were defined for operative time: 150 minutes. Between 7 and 12 days after surgery, the patients underwent ascending venography to evaluate the presence of distal or proximal deep vein thrombosis. We evaluated the association between the tourniquet time and total operative time and the occurrence of deep vein thrombosis after total knee arthroplasty. In total, 33 cases (42.3%) were positive for deep vein thrombosis; 13 (16.7%) cases involved the proximal type. We found no statistically significant difference in tourniquet time or operative time between patients with or without deep vein thrombosis. We did observe a higher frequency of proximal deep vein thrombosis in patients who underwent surgery lasting longer than 120 minutes. The mean total operative time was also higher in patients with proximal deep vein thrombosis. The tourniquet time did not significantly differ in these patients. We concluded that surgery lasting longer than 120 minutes increases the risk of proximal deep vein thrombosis.

  12. Wait time management strategies for total joint replacement surgery: sustainability and unintended consequences.

    Science.gov (United States)

    Pomey, Marie-Pascale; Clavel, Nathalie; Amar, Claudia; Sabogale-Olarte, Juan Carlos; Sanmartin, Claudia; De Coster, Carolyn; Noseworthy, Tom

    2017-09-07

    In Canada, long waiting times for core specialized services have consistently been identified as a key barrier to access. Governments and organizations have responded with strategies for better access management, notably for total joint replacement (TJR) of the hip and knee. While wait time management strategies (WTMS) are promising, the factors which influence their sustainable implementation at the organizational level are understudied. Consequently, this study examined organizational and systemic factors that made it possible to sustain waiting times for TJR within federally established limits and for at least 18 months or more. The research design is a multiple case study of WTMS implementation. Five cases were selected across five Canadian provinces. Three success levels were pre-defined: 1) the WTMS maintained compliance with requirements for more than 18 months; 2) the WTMS met requirements for 18 months but could not sustain the level thereafter; 3) the WTMS never met requirements. For each case, we collected documents and interviewed key informants. We analyzed systemic and organizational factors, with particular attention to governance and leadership, culture, resources, methods, and tools. We found that successful organizations had specific characteristics: 1) management of the whole care continuum, 2) strong clinical leadership; 3) dedicated committees to coordinate and sustain strategy; 4) a culture based on trust and innovation. All strategies led to relatively similar unintended consequences. The main negative consequence was an initial increase in waiting times for TJR and the main positive consequence was operational enhancement of other areas of specialization based on the TJR model. This study highlights important differences in factors which help to achieve and sustain waiting times. To be sustainable, a WTMS needs to generate greater synergies between contextual-level strategy (provincial or regional) and organizational objectives and

  13. The position of a standard optical computer mouse affects cardiorespiratory responses during the operation of a computer under time constraints

    Directory of Open Access Journals (Sweden)

    Shunji Sako

    2014-08-01

    Full Text Available Objectives: This study investigated the association between task-induced stress and fatigue by examining the cardiovascular responses of subjects using different mouse positions while operating a computer under time constraints. Material and Methods: The study was participated by 16 young, healthy men and examined the use of optical mouse devices affixed to laptop computers. Two mouse positions were investigated: (1 the distal position (DP, in which the subjects place their forearms on the desk accompanied by the abduction and flexion of their shoulder joints, and (2 the proximal position (PP, in which the subjects place only their wrists on the desk without using an armrest. The subjects continued each task for 16 min. We assessed differences in several characteristics according to mouse position, including expired gas values, autonomic nerve activities (based on cardiorespiratory responses, operating efficiencies (based on word counts, and fatigue levels (based on the visual analog scale – VAS. Results: Oxygen consumption (VO2, the ratio of inspiration time to respiration time (Ti/Ttotal, respiratory rate (RR, minute ventilation (VE, and the ratio of expiration to inspiration (Te/Ti were significantly lower when the participants were performing the task in the DP than those obtained in the PP. Tidal volume (VT, carbon dioxide output rates (VCO2/VE, and oxygen extraction fractions (VO2/VE were significantly higher for the DP than they were for the PP. No significant difference in VAS was observed between the positions; however, as the task progressed, autonomic nerve activities were lower and operating efficiencies were significantly higher for the DP than they were for the PP. Conclusions: Our results suggest that the DP has fewer effects on cardiorespiratory functions, causes lower levels of sympathetic nerve activity and mental stress, and produces a higher total workload than the PP. This suggests that the DP is preferable to the PP when

  14. Evaluation of a semi-automated computer algorithm for measuring total fat and visceral fat content in lambs undergoing in vivo whole body computed tomography.

    Science.gov (United States)

    Rosenblatt, Alana J; Scrivani, Peter V; Boisclair, Yves R; Reeves, Anthony P; Ramos-Nieves, Jose M; Xie, Yiting; Erb, Hollis N

    2017-10-01

    Computed tomography (CT) is a suitable tool for measuring body fat, since it is non-destructive and can be used to differentiate metabolically active visceral fat from total body fat. Whole body analysis of body fat is likely to be more accurate than single CT slice estimates of body fat. The aim of this study was to assess the agreement between semi-automated computer analysis of whole body volumetric CT data and conventional proximate (chemical) analysis of body fat in lambs. Data were collected prospectively from 12 lambs that underwent duplicate whole body CT, followed by slaughter and carcass analysis by dissection and chemical analysis. Agreement between methods for quantification of total and visceral fat was assessed by Bland-Altman plot analysis. The repeatability of CT was assessed for these measures using the mean difference of duplicated measures. When compared to chemical analysis, CT systematically underestimated total and visceral fat contents by more than 10% of the mean fat weight. Therefore, carcass analysis and semi-automated CT computer measurements were not interchangeable for quantifying body fat content without the use of a correction factor. CT acquisition was repeatable, with a mean difference of repeated measures being close to zero. Therefore, uncorrected whole body CT might have an application for assessment of relative changes in fat content, especially in growing lambs. Copyright © 2017 Elsevier Ltd. All rights reserved.

  15. A rapid infusion protocol is safe for total dose iron polymaltose: time for change.

    Science.gov (United States)

    Garg, M; Morrison, G; Friedman, A; Lau, A; Lau, D; Gibson, P R

    2011-07-01

    Intravenous correction of iron deficiency by total dose iron polymaltose is inexpensive and safe, but current protocols entail prolonged administration over more than 4 h. This results in reduced patient acceptance, and hospital resource strain. We aimed to assess prospectively the safety of a rapid intravenous protocol and compare this with historical controls. Consecutive patients in whom intravenous iron replacement was indicated were invited to have up to 1.5 g iron polymaltose by a 58-min infusion protocol after an initial 15-min test dose without pre-medication. Infusion-related adverse events (AE) and delayed AE over the ensuing 5 days were also prospectively documented and graded as mild, moderate or severe. One hundred patients, 63 female, mean age 54 (range 18-85) years were studied. Thirty-four infusion-related AE to iron polymaltose occurred in a total of 24 patients--25 mild, 8 moderate and 1 severe; higher than previously reported for a slow protocol iron infusion. Thirty-one delayed AE occurred in 26 patients--26 mild, 3 moderate and 2 severe; similar to previously reported. All but five patients reported they would prefer iron replacement through the rapid protocol again. The presence of inflammatory bowel disease (IBD) predicted infusion-related reactions (54% vs 14% without IBD, P cost, resource utilization and time benefits for the patient and hospital system. © 2011 The Authors. Internal Medicine Journal © 2011 Royal Australasian College of Physicians.

  16. Feasibility and safety of augmented-reality glass for computed tomography-assisted percutaneous revascularization of coronary chronic total occlusion: A single center prospective pilot study.

    Science.gov (United States)

    Opolski, Maksymilian P; Debski, Artur; Borucki, Bartosz A; Staruch, Adam D; Kepka, Cezary; Rokicki, Jakub K; Sieradzki, Bartosz; Witkowski, Adam

    2017-11-01

    Percutaneous coronary intervention (PCI) of chronic total occlusion (CTO) may be facilitated by projection of coronary computed tomography angiography (CTA) datasets in the catheterization laboratory. There is no data on the feasibility and safety outcomes of CTA-assisted CTO PCI using a wearable augmented-reality glass. A total of 15 patients scheduled for elective antegrade CTO intervention were prospectively enrolled and underwent preprocedural coronary CTA. Three-dimensional and curved multiplanar CT reconstructions were transmitted to a head-mounted hands-free computer worn by interventional cardiologists during CTO PCI to provide additional information on CTO tortuosity and calcification. The results of CTO PCI using a wearable computer were compared with a time-matched prospective angiographic registry of 59 patients undergoing antegrade CTO PCI without a wearable computer. Operators' satisfaction was assessed by a 5-point Likert scale. Mean age was 64 ± 8 years and the mean J-CTO score was 2.1 ± 0.9 in the CTA-assisted group. The voice-activated co-registration and review of CTA images in a wearable computer during CTO PCI were feasible and highly rated by PCI operators (4.7/5 points). There were no major adverse cardiovascular events. Compared with standard CTO PCI, CTA-assisted recanalization of CTO using a wearable computer showed more frequent selection of the first-choice stiff wire (0% vs 40%, p augmented-reality glass is feasible and safe, and might reduce the resources required for the interventional treatment of CTO. Copyright © 2017 Society of Cardiovascular Computed Tomography. Published by Elsevier Inc. All rights reserved.

  17. Optimum filters with time width constraints for liquid argon total-absorption detectors

    International Nuclear Information System (INIS)

    Gatti, E.; Radeka, V.

    1977-10-01

    Optimum filter responses are found for triangular current input pulses occurring in liquid argon ionization chambers used as total absorption detectors. The filters considered are subject to the following constraints: finite width of the output pulse having a prescribed ratio to the width of the triangular input current pulse and zero area of a bipolar antisymmetrical pulse or of a three lobe pulse, as required for high event rates. The feasibility of pulse shaping giving an output equal to, or shorter than, the input one is demonstrated. It is shown that the signal-to-noise ratio remains constant for the chamber interelectrode gap which gives an input pulse width (i.e., electron drift time) greater than one third of the required output pulse width

  18. Real-time fusion of coronary CT angiography with X-ray fluoroscopy during chronic total occlusion PCI

    Energy Technology Data Exchange (ETDEWEB)

    Ghoshhajra, Brian B.; Takx, Richard A.P. [Harvard Medical School, Cardiac MR PET CT Program, Massachusetts General Hospital, Department of Radiology and Division of Cardiology, Boston, MA (United States); Stone, Luke L.; Yeh, Robert W.; Jaffer, Farouc A. [Harvard Medical School, Cardiac Cathetrization Laboratory, Cardiology Division, Massachusetts General Hospital, Boston, MA (United States); Girard, Erin E. [Siemens Healthcare, Princeton, NJ (United States); Brilakis, Emmanouil S. [Cardiology Division, Dallas VA Medical Center and UT Southwestern Medical Center, Dallas, TX (United States); Lombardi, William L. [University of Washington, Cardiology Division, Seattle, WA (United States)

    2017-06-15

    The purpose of this study was to demonstrate the feasibility of real-time fusion of coronary computed tomography angiography (CTA) centreline and arterial wall calcification with X-ray fluoroscopy during chronic total occlusion (CTO) percutaneous coronary intervention (PCI). Patients undergoing CTO PCI were prospectively enrolled. Pre-procedural CT scans were integrated with conventional coronary fluoroscopy using prototype software. We enrolled 24 patients who underwent CTO PCI using the prototype CT fusion software, and 24 consecutive CTO PCI patients without CT guidance served as a control group. Mean age was 66 ± 11 years, and 43/48 patients were men. Real-time CTA fusion during CTO PCI provided additional information regarding coronary arterial calcification and tortuosity that generated new insights into antegrade wiring, antegrade dissection/reentry, and retrograde wiring during CTO PCI. Overall CTO success rates and procedural outcomes remained similar between the two groups, despite a trend toward higher complexity in the fusion CTA group. This study demonstrates that real-time automated co-registration of coronary CTA centreline and calcification onto live fluoroscopic images is feasible and provides new insights into CTO PCI, and in particular, antegrade dissection reentry-based CTO PCI. (orig.)

  19. Estimating time-based instantaneous total mortality rate based on the age-structured abundance index

    Science.gov (United States)

    Wang, Yingbin; Jiao, Yan

    2015-05-01

    The instantaneous total mortality rate ( Z) of a fish population is one of the important parameters in fisheries stock assessment. The estimation of Z is crucial to fish population dynamics analysis, abundance and catch forecast, and fisheries management. A catch curve-based method for estimating time-based Z and its change trend from catch per unit effort (CPUE) data of multiple cohorts is developed. Unlike the traditional catch-curve method, the method developed here does not need the assumption of constant Z throughout the time, but the Z values in n continuous years are assumed constant, and then the Z values in different n continuous years are estimated using the age-based CPUE data within these years. The results of the simulation analyses show that the trends of the estimated time-based Z are consistent with the trends of the true Z, and the estimated rates of change from this approach are close to the true change rates (the relative differences between the change rates of the estimated Z and the true Z are smaller than 10%). Variations of both Z and recruitment can affect the estimates of Z value and the trend of Z. The most appropriate value of n can be different given the effects of different factors. Therefore, the appropriate value of n for different fisheries should be determined through a simulation analysis as we demonstrated in this study. Further analyses suggested that selectivity and age estimation are also two factors that can affect the estimated Z values if there is error in either of them, but the estimated change rates of Z are still close to the true change rates. We also applied this approach to the Atlantic cod ( Gadus morhua) fishery of eastern Newfoundland and Labrador from 1983 to 1997, and obtained reasonable estimates of time-based Z.

  20. Total and domain-specific sitting time among employees in desk-based work settings in Australia.

    Science.gov (United States)

    Bennie, Jason A; Pedisic, Zeljko; Timperio, Anna; Crawford, David; Dunstan, David; Bauman, Adrian; van Uffelen, Jannique; Salmon, Jo

    2015-06-01

    To describe the total and domain-specific daily sitting time among a sample of Australian office-based employees. In April 2010, paper-based surveys were provided to desk-based employees (n=801) in Victoria, Australia. Total daily and domain-specific (work, leisure-time and transport-related) sitting time (minutes/day) were assessed by validated questionnaires. Differences in sitting time were examined across socio-demographic (age, sex, occupational status) and lifestyle characteristics (physical activity levels, body mass index [BMI]) using multiple linear regression analyses. The median (95% confidence interval [CI]) of total daily sitting time was 540 (531-557) minutes/day. Insufficiently active adults (median=578 minutes/day, [95%CI: 564-602]), younger adults aged 18-29 years (median=561 minutes/day, [95%CI: 540-577]) reported the highest total daily sitting times. Occupational sitting time accounted for almost 60% of total daily sitting time. In multivariate analyses, total daily sitting time was negatively associated with age (unstandardised regression coefficient [B]=-1.58, pphysical activity (minutes/week) (B=-0.03, pemployees reported that more than half of their total daily sitting time was accrued in the work setting. Given the high contribution of occupational sitting to total daily sitting time among desk-based employees, interventions should focus on the work setting. © 2014 Public Health Association of Australia.

  1. Time-Domain Terahertz Computed Axial Tomography NDE System

    Science.gov (United States)

    Zimdars, David

    2012-01-01

    NASA has identified the need for advanced non-destructive evaluation (NDE) methods to characterize aging and durability in aircraft materials to improve the safety of the nation's airline fleet. 3D THz tomography can play a major role in detection and characterization of flaws and degradation in aircraft materials, including Kevlar-based composites and Kevlar and Zylon fabric covers for soft-shell fan containment where aging and durability issues are critical. A prototype computed tomography (CT) time-domain (TD) THz imaging system has been used to generate 3D images of several test objects including a TUFI tile (a thermal protection system tile used on the Space Shuttle and possibly the Orion or similar capsules). This TUFI tile had simulated impact damage that was located and the depth of damage determined. The CT motion control gan try was designed and constructed, and then integrated with a T-Ray 4000 control unit and motion controller to create a complete CT TD-THz imaging system prototype. A data collection software script was developed that takes multiple z-axis slices in sequence and saves the data for batch processing. The data collection software was integrated with the ability to batch process the slice data with the CT TD-THz image reconstruction software. The time required to take a single CT slice was decreased from six minutes to approximately one minute by replacing the 320 ps, 100-Hz waveform acquisition system with an 80 ps, 1,000-Hz waveform acquisition system. The TD-THZ computed tomography system was built from pre-existing commercial off-the-shelf subsystems. A CT motion control gantry was constructed from COTS components that can handle larger samples. The motion control gantry allows inspection of sample sizes of up to approximately one cubic foot (.0.03 cubic meters). The system reduced to practice a CT-TDTHz system incorporating a COTS 80- ps/l-kHz waveform scanner. The incorporation of this scanner in the system allows acquisition of 3D

  2. Computed tomography for preoperative planning in minimal-invasive total hip arthroplasty: Radiation exposure and cost analysis

    Energy Technology Data Exchange (ETDEWEB)

    Huppertz, Alexander, E-mail: Alexander.Huppertz@charite.de [Imaging Science Institute Charite Berlin, Robert-Koch-Platz 7, D-10115 Berlin (Germany); Department of Radiology, Medical Physics, Charite-University Hospitals of Berlin, Chariteplatz 1, D-10117 Berlin (Germany); Radmer, Sebastian, E-mail: s.radmer@immanuel.de [Department of Orthopedic Surgery and Rheumatology, Immanuel-Krankenhaus, Koenigstr. 63, D-14109, Berlin (Germany); Asbach, Patrick, E-mail: Patrick.Asbach@charite.de [Department of Radiology, Medical Physics, Charite-University Hospitals of Berlin, Chariteplatz 1, D-10117 Berlin (Germany); Juran, Ralf, E-mail: ralf.juran@charite.de [Department of Radiology, Medical Physics, Charite-University Hospitals of Berlin, Chariteplatz 1, D-10117 Berlin (Germany); Schwenke, Carsten, E-mail: carsten.schwenke@scossis.de [Biostatistician, Scossis Statistical Consulting, Zeltinger Str. 58G, D-13465 Berlin (Germany); Diederichs, Gerd, E-mail: gerd.diederichs@charite.de [Department of Radiology, Medical Physics, Charite-University Hospitals of Berlin, Chariteplatz 1, D-10117 Berlin (Germany); Hamm, Bernd, E-mail: Bernd.Hamm@charite.de [Department of Radiology, Medical Physics, Charite-University Hospitals of Berlin, Chariteplatz 1, D-10117 Berlin (Germany); Sparmann, Martin, E-mail: m.sparmann@immanuel.de [Department of Orthopedic Surgery and Rheumatology, Immanuel-Krankenhaus, Koenigstr. 63, D-14109, Berlin (Germany)

    2011-06-15

    Computed tomography (CT) was used for preoperative planning of minimal-invasive total hip arthroplasty (THA). 92 patients (50 males, 42 females, mean age 59.5 years) with a mean body-mass-index (BMI) of 26.5 kg/m{sup 2} underwent 64-slice CT to depict the pelvis, the knee and the ankle in three independent acquisitions using combined x-, y-, and z-axis tube current modulation. Arthroplasty planning was performed using 3D-Hip Plan (Symbios, Switzerland) and patient radiation dose exposure was determined. The effects of BMI, gender, and contralateral THA on the effective dose were evaluated by an analysis-of-variance. A process-cost-analysis from the hospital perspective was done. All CT examinations were of sufficient image quality for 3D-THA planning. A mean effective dose of 4.0 mSv (SD 0.9 mSv) modeled by the BMI (p < 0.0001) was calculated. The presence of a contralateral THA (9/92 patients; p = 0.15) and the difference between males and females were not significant (p = 0.08). Personnel involved were the radiologist (4 min), the surgeon (16 min), the radiographer (12 min), and administrative personnel (4 min). A CT operation time of 11 min and direct per-patient costs of 52.80 Euro were recorded. Preoperative CT for THA was associated with a slight and justifiable increase of radiation exposure in comparison to conventional radiographs and low per-patient costs.

  3. The Syncardia™ total artificial heart: in vivo, in vitro, and computational modeling studies

    Science.gov (United States)

    Slepian, Marvin J.; Alemu, Yared; Soares, João Silva; Smith, Richard G.; Einav, Shmuel; Bluestein, Danny

    2014-01-01

    The SynCardia™ total artificial heart (TAH) is the only FDA-approved TAH in the world. The SynCardia™ TAH is a pneumatically driven, pulsatile system capable of flows of >9 L/min. The TAH is indicated for use as a bridge to transplantation (BTT) in patients at imminent risk of death from non-reversible bi-ventricular failure. In the Pivotal US approval trial the TAH achieved a BTT rate of >79%. Recently a multi-center, post-market approval study similarly demonstrated a comparable BTT rate. A major milestone was recently achieved for the TAH, with over 1100 TAHs having been implanted to date, with the bulk of implantation occurring at an ever increasing rate in the past few years. The TAH is most commonly utilized to save the lives of patients dying from end-stage bi-ventricular heart failure associated with ischemic or non-ischemic dilated cardiomyopathy. Beyond progressive chronic heart failure, the TAH has demonstrated great efficacy in supporting patients with acute irreversible heart failure associated with massive acute myocardial infarction. In recent years several diverse clinical scenarios have also proven to be well served by the TAH including severe heart failure associated with advanced congenital heart disease. failed or burned-out transplants, infiltrative and restrictive cardiomyopathies and failed ventricular assist devices. Looking to the future a major unmet need remains in providing total heart support for children and small adults. As such, the present TAH design must be scaled to fit the smaller patient, while providing equivalent, if not superior flow characteristics, shear profiles and overall device thrombogenicity. To aid in the development of a new “pediatric,” TAH an engineering methodology known as “Device Thrombogenicity Emulation (DTE)”, that we have recently developed and described, is being employed. Recently, to further our engineering understanding of the TAH, as steps towards next generation designs we have: (1

  4. Image reconstruction in circular cone-beam computed tomography by constrained, total-variation minimization

    International Nuclear Information System (INIS)

    Sidky, Emil Y; Pan Xiaochuan

    2008-01-01

    An iterative algorithm, based on recent work in compressive sensing, is developed for volume image reconstruction from a circular cone-beam scan. The algorithm minimizes the total variation (TV) of the image subject to the constraint that the estimated projection data is within a specified tolerance of the available data and that the values of the volume image are non-negative. The constraints are enforced by the use of projection onto convex sets (POCS) and the TV objective is minimized by steepest descent with an adaptive step-size. The algorithm is referred to as adaptive-steepest-descent-POCS (ASD-POCS). It appears to be robust against cone-beam artifacts, and may be particularly useful when the angular range is limited or when the angular sampling rate is low. The ASD-POCS algorithm is tested with the Defrise disk and jaw computerized phantoms. Some comparisons are performed with the POCS and expectation-maximization (EM) algorithms. Although the algorithm is presented in the context of circular cone-beam image reconstruction, it can also be applied to scanning geometries involving other x-ray source trajectories

  5. Estimating spatio-temporal dynamics of stream total phosphate concentration by soft computing techniques.

    Science.gov (United States)

    Chang, Fi-John; Chen, Pin-An; Chang, Li-Chiu; Tsai, Yu-Hsuan

    2016-08-15

    This study attempts to model the spatio-temporal dynamics of total phosphate (TP) concentrations along a river for effective hydro-environmental management. We propose a systematical modeling scheme (SMS), which is an ingenious modeling process equipped with a dynamic neural network and three refined statistical methods, for reliably predicting the TP concentrations along a river simultaneously. Two different types of artificial neural network (BPNN-static neural network; NARX network-dynamic neural network) are constructed in modeling the dynamic system. The Dahan River in Taiwan is used as a study case, where ten-year seasonal water quality data collected at seven monitoring stations along the river are used for model training and validation. Results demonstrate that the NARX network can suitably capture the important dynamic features and remarkably outperforms the BPNN model, and the SMS can effectively identify key input factors, suitably overcome data scarcity, significantly increase model reliability, satisfactorily estimate site-specific TP concentration at seven monitoring stations simultaneously, and adequately reconstruct seasonal TP data into a monthly scale. The proposed SMS can reliably model the dynamic spatio-temporal water pollution variation in a river system for missing, hazardous or costly data of interest. Copyright © 2016 Elsevier B.V. All rights reserved.

  6. Television Viewing, Computer Use, Time Driving and All‐Cause Mortality: The SUN Cohort

    Science.gov (United States)

    Basterra‐Gortari, Francisco Javier; Bes‐Rastrollo, Maira; Gea, Alfredo; Núñez‐Córdoba, Jorge María; Toledo, Estefanía; Martínez‐González, Miguel Ángel

    2014-01-01

    Background Sedentary behaviors have been directly associated with all‐cause mortality. However, little is known about different types of sedentary behaviors in relation to overall mortality. Our objective was to assess the association between different sedentary behaviors and all‐cause mortality. Methods and Results In this prospective, dynamic cohort study (the SUN Project) 13 284 Spanish university graduates with a mean age of 37 years were followed‐up for a median of 8.2 years. Television, computer, and driving time were assessed at baseline. Poisson regression models were fitted to examine the association between each sedentary behavior and total mortality. All‐cause mortality incidence rate ratios (IRRs) per 2 hours per day were 1.40 (95% confidence interval (CI): 1.06 to 1.84) for television viewing, 0.96 (95% CI: 0.79 to 1.18) for computer use, and 1.14 (95% CI: 0.90 to 1.44) for driving, after adjustment for age, sex, smoking status, total energy intake, Mediterranean diet adherence, body mass index, and physical activity. The risk of mortality was twofold higher for participants reporting ≥3 h/day of television viewing than for those reporting Television viewing was directly associated with all‐cause mortality. However, computer use and time spent driving were not significantly associated with higher mortality. Further cohort studies and trials designed to assess whether reductions in television viewing are able to reduce mortality are warranted. The lack of association between computer use or time spent driving and mortality needs further confirmation. PMID:24965030

  7. Time-of-Flight Sensors in Computer Graphics

    DEFF Research Database (Denmark)

    Kolb, Andreas; Barth, Erhardt; Koch, Reinhard

    2009-01-01

    , including Computer Graphics, Computer Vision and Man Machine Interaction (MMI). These technologies are starting to have an impact on research and commercial applications. The upcoming generation of ToF sensors, however, will be even more powerful and will have the potential to become “ubiquitous real...

  8. Cluster Computing For Real Time Seismic Array Analysis.

    Science.gov (United States)

    Martini, M.; Giudicepietro, F.

    A seismic array is an instrument composed by a dense distribution of seismic sen- sors that allow to measure the directional properties of the wavefield (slowness or wavenumber vector) radiated by a seismic source. Over the last years arrays have been widely used in different fields of seismological researches. In particular they are applied in the investigation of seismic sources on volcanoes where they can be suc- cessfully used for studying the volcanic microtremor and long period events which are critical for getting information on the volcanic systems evolution. For this reason arrays could be usefully employed for the volcanoes monitoring, however the huge amount of data produced by this type of instruments and the processing techniques which are quite time consuming limited their potentiality for this application. In order to favor a direct application of arrays techniques to continuous volcano monitoring we designed and built a small PC cluster able to near real time computing the kinematics properties of the wavefield (slowness or wavenumber vector) produced by local seis- mic source. The cluster is composed of 8 Intel Pentium-III bi-processors PC working at 550 MHz, and has 4 Gigabytes of RAM memory. It runs under Linux operating system. The developed analysis software package is based on the Multiple SIgnal Classification (MUSIC) algorithm and is written in Fortran. The message-passing part is based upon the LAM programming environment package, an open-source imple- mentation of the Message Passing Interface (MPI). The developed software system includes modules devote to receiving date by internet and graphical applications for the continuous displaying of the processing results. The system has been tested with a data set collected during a seismic experiment conducted on Etna in 1999 when two dense seismic arrays have been deployed on the northeast and the southeast flanks of this volcano. A real time continuous acquisition system has been simulated by

  9. Determination of time-dependent uncertainty of the total solar irradiance records from 1978 to present

    Directory of Open Access Journals (Sweden)

    Fröhlich Claus

    2016-01-01

    Full Text Available Aims. The existing records of total solar irradiance (TSI since 1978 differ not only in absolute values, but also show different trends. For the study of TSI variability these records need to be combined and three composites have been devised; however, the results depend on the choice of the records and the way they are combined. A new composite should be based on all existing records with an individual qualification. It is proposed to use a time-dependent uncertainty for weighting of the individual records. Methods. The determination of the time-dependent deviation of the TSI records is performed by comparison with the square root of the sunspot number (SSN. However, this correlation is only valid for timescales of the order of a year or more because TSI and SSN react quite differently to solar activity changes on shorter timescales. Hence the results concern only periods longer than the one-year-low-pass filter used in the analysis. Results. Besides the main objective to determine an investigator-independent uncertainty, the comparison of TSI with √SSN turns out to be a powerful tool for the study of the TSI long-term changes. The correlation of √SSN with TSI replicates very well the TSI minima, especially the very low value of the recent minimum. The results of the uncertainty determination confirm not only the need for adequate corrections for degradation, but also show that a rather detailed analysis is needed. The daily average of all TSI values available on that day, weighted with the correspondingly determined uncertainty, is used to construct a “new” composite, which, overall, compares well with the Physikalisch-Meteorologisches Observatorium Davos (PMOD composite. Finally, the TSI − √SSN comparison proves to be an important diagnostic tool not only for estimating uncertainties of observations, but also for a better understanding of the long-term variability of TSI.

  10. Enhanced responses to tumor immunization following total body irradiation are time-dependent.

    Directory of Open Access Journals (Sweden)

    Adi Diab

    Full Text Available The development of successful cancer vaccines is contingent on the ability to induce effective and persistent anti-tumor immunity against self-antigens that do not typically elicit immune responses. In this study, we examine the effects of a non-myeloablative dose of total body irradiation on the ability of tumor-naïve mice to respond to DNA vaccines against melanoma. We demonstrate that irradiation followed by lymphocyte infusion results in a dramatic increase in responsiveness to tumor vaccination, with augmentation of T cell responses to tumor antigens and tumor eradication. In irradiated mice, infused CD8(+ T cells expand in an environment that is relatively depleted in regulatory T cells, and this correlates with improved CD8(+ T cell functionality. We also observe an increase in the frequency of dendritic cells displaying an activated phenotype within lymphoid organs in the first 24 hours after irradiation. Intriguingly, both the relative decrease in regulatory T cells and increase in activated dendritic cells correspond with a brief window of augmented responsiveness to immunization. After this 24 hour window, the numbers of dendritic cells decline, as does the ability of mice to respond to immunizations. When immunizations are initiated within the period of augmented dendritic cell activation, mice develop anti-tumor responses that show increased durability as well as magnitude, and this approach leads to improved survival in experiments with mice bearing established tumors as well as in a spontaneous melanoma model. We conclude that irradiation can produce potent immune adjuvant effects independent of its ability to induce tumor ablation, and that the timing of immunization and lymphocyte infusion in the irradiated host are crucial for generating optimal anti-tumor immunity. Clinical strategies using these approaches must therefore optimize such parameters, as the correct timing of infusion and vaccination may mean the difference

  11. Brake response time before and after total knee arthroplasty: a prospective cohort study

    Directory of Open Access Journals (Sweden)

    Niederseer David

    2010-11-01

    Full Text Available Abstract Background Although the numbers of total knee arthroplasty (TKA are increasing, there is only a small number of studies investigating driving safety after TKA. The parameter 'Brake Response Time (BRT' is one of the most important criteria for driving safety and was therefore chosen for investigation. The present study was conducted to test the hypotheses that patients with right- or left-sided TKA show a significant increase in BRT from pre-operative (pre-op, 1 day before surgery to post-operative (post-op, 2 weeks post surgery, and a significant decrease in BRT from post-op to the follow-up investigation (FU, 8 weeks post surgery. Additionally, it was hypothesized that the BRT of patients after TKA is significantly higher than that of healthy controls. Methods 31 of 70 consecutive patients (mean age 65.7 +/- 10.2 years receiving TKA were tested for their BRT pre-op, post-op and at FU. BRT was assessed using a custom-made driving simulator. We used normative BRT data from 31 healthy controls for comparison. Results There were no significant increases between pre-op and post-op BRT values for patients who had undergone left- or right-sided TKA. Even the proportion of patients above a BRT threshold of 700 ms was not significantly increased postop. Controls had a BRT which was significantly better than the BRT of patients with right- or left-sided TKA at all three time points. Conclusion The present study showed a small and insignificant postoperative increase in the BRT of patients who had undergone right- or left-sided TKA. Therefore, we believe it is not justified to impair the patient's quality of social and occupational life post-surgery by imposing restrictions on driving motor vehicles beyond an interval of two weeks after surgery.

  12. Modulation of Total Sleep Time by Transcranial Direct Current Stimulation (tDCS).

    Science.gov (United States)

    Frase, Lukas; Piosczyk, Hannah; Zittel, Sulamith; Jahn, Friederike; Selhausen, Peter; Krone, Lukas; Feige, Bernd; Mainberger, Florian; Maier, Jonathan G; Kuhn, Marion; Klöppel, Stefan; Normann, Claus; Sterr, Annette; Spiegelhalder, Kai; Riemann, Dieter; Nitsche, Michael A; Nissen, Christoph

    2016-09-01

    Arousal and sleep are fundamental physiological processes, and their modulation is of high clinical significance. This study tested the hypothesis that total sleep time (TST) in humans can be modulated by the non-invasive brain stimulation technique transcranial direct current stimulation (tDCS) targeting a 'top-down' cortico-thalamic pathway of sleep-wake regulation. Nineteen healthy participants underwent a within-subject, repeated-measures protocol across five nights in the sleep laboratory with polysomnographic monitoring (adaptation, baseline, three experimental nights). tDCS was delivered via bi-frontal target electrodes and bi-parietal return electrodes before sleep (anodal 'activation', cathodal 'deactivation', and sham stimulation). Bi-frontal anodal stimulation significantly decreased TST, compared with cathodal and sham stimulation. This effect was location specific. Bi-frontal cathodal stimulation did not significantly increase TST, potentially due to ceiling effects in good sleepers. Exploratory resting-state EEG analyses before and after the tDCS protocols were consistent with the notion of increased cortical arousal after anodal stimulation and decreased cortical arousal after cathodal stimulation. The study provides proof-of-concept that TST can be decreased by non-invasive bi-frontal anodal tDCS in healthy humans. Further elucidating the 'top-down' pathway of sleep-wake regulation is expected to increase knowledge on the fundamentals of sleep-wake regulation and to contribute to the development of novel treatments for clinical conditions of disturbed arousal and sleep.

  13. On the Laws of Total Local Times for -Paths and Bridges of Symmetric Lévy Processes

    Directory of Open Access Journals (Sweden)

    Masafumi Hayashi

    2013-01-01

    Full Text Available The joint law of the total local times at two levels for -paths of symmetric Lévy processes is shown to admit an explicit representation in terms of the laws of the squared Bessel processes of dimensions two and zero. The law of the total local time at a single level for bridges is also discussed.

  14. Control bandwidth improvements in GRAVITY fringe tracker by switching to a synchronous real time computer architecture

    Science.gov (United States)

    Abuter, Roberto; Dembet, Roderick; Lacour, Sylvestre; di Lieto, Nicola; Woillez, Julien; Eisenhauer, Frank; Fedou, Pierre; Phan Duc, Than

    2016-08-01

    The new VLTI (Very Large Telescope Interferometer) 1 instrument GRAVITY5, 22, 23 is equipped with a fringe tracker16 able to stabilize the K-band fringes on six baselines at the same time. It has been designed to achieve a performance for average seeing conditions of a residual OPD (Optical Path Difference) lower than 300 nm with objects brighter than K = 10. The control loop implementing the tracking is composed of a four stage real time computer system compromising: a sensor where the detector pixels are read in and the OPD and GD (Group Delay) are calculated; a controller receiving the computed sensor quantities and producing commands for the piezo actuators; a concentrator which combines both the OPD commands with the real time tip/tilt corrections offloading them to the piezo actuator; and finally a Kalman15 parameter estimator. This last stage is used to monitor current measurements over a window of few seconds and estimate new values for the main Kalman15 control loop parameters. The hardware and software implementation of this design runs asynchronously and communicates the four computers for data transfer via the Reflective Memory Network3. With the purpose of improving the performance of the GRAVITY5, 23 fringe tracking16, 22 control loop, a deviation from the standard asynchronous communication mechanism has been proposed and implemented. This new scheme operates the four independent real time computers involved in the tracking loop synchronously using the Reflective Memory Interrupts2 as the coordination signal. This synchronous mechanism had the effect of reducing the total pure delay of the loop from 3.5 [ms] to 2.0 [ms] which then translates on a better stabilization of the fringes as the bandwidth of the system is substantially improved. This paper will explain in detail the real time architecture of the fringe tracker in both is synchronous and synchronous implementation. The achieved improvements on reducing the delay via this mechanism will be

  15. Surgical time and complications of total transvaginal (total-NOTES, single-port laparoscopic-assisted and conventional ovariohysterectomy in bitches

    Directory of Open Access Journals (Sweden)

    M.A.M. Silva

    2015-06-01

    Full Text Available The recently developed minimally invasive techniques of ovariohysterectomy (OVH have been studied in dogs in order to optimize their benefits and decrease risks to the patients. The purpose of this study was to compare surgical time, complications and technical difficulties of transvaginal total-NOTES, single-port laparoscopic-assisted and conventional OVH in bitches. Twelve bitches were submitted to total-NOTES (NOTES group, while 13 underwent single-port laparoscopic-assisted (SPLA group and 15 were submitted to conventional OVH (OPEN group. Intra-operative period was divided into 7 stages: (1 access to abdominal cavity; (2 pneumoperitoneum; approach to the right (3 and left (4 ovarian pedicle and uterine body (5; (6 abdominal or vaginal synthesis, performed in 6 out of 12 patients of NOTES; (7 inoperative time. Overall and stages operative times, intra and postoperative complications and technical difficulties were compared among groups. Mean overall surgical time in NOTES (25.7±6.8 minutes and SPLA (23.1±4.0 minutes groups were shorter than in the OPEN group (34.0±6.4 minutes (P<0.05. The intraoperative stage that required the longest time was the approach to the uterine body in the NOTES group and abdominal and cutaneous sutures in the OPEN group. There was no difference regarding the rates of complications. Major complications included postoperative bleeding requiring reoperation in a bitch in the OPEN group, while minor complications included mild vaginal discharge in four patients in the NOTES group and seroma in three bitches in the SPLA group. In conclusion, total-NOTES and SPLA OVH were less time-consuming then conventional OVH in bitches. All techniques presented complications, which were properly managed.

  16. Joint association of physical activity in leisure and total sitting time with metabolic syndrome amongst 15,235 Danish adults

    DEFF Research Database (Denmark)

    Petersen, Christina Bjørk; Nielsen, Asser Jon; Bauman, Adrian

    2014-01-01

    and total daily sitting time were assessed by self-report in 15,235 men and women in the Danish Health Examination Survey 2007-2008. Associations between leisure time physical activity, total sitting time and metabolic syndrome were investigated in logistic regression analysis. RESULTS: Adjusted odds ratios......BACKGROUND: Recent studies suggest that physical inactivity as well as sitting time are associated with metabolic syndrome. Our aim was to examine joint associations of leisure time physical activity and total daily sitting time with metabolic syndrome. METHODS: Leisure time physical activity...... (OR) for metabolic syndrome were 2.14 (95% CI: 1.88-2.43) amongst participants who were inactive in leisure time compared to the most active, and 1.42 (95% CI: 1.26-1.61) amongst those who sat for ≥10h/day compared to physical activity, sitting time...

  17. A comprehensive combined experimental and computational framework for pre-clinical wear simulation of total knee replacements.

    Science.gov (United States)

    Abdelgaied, A; Fisher, J; Jennings, L M

    2018-02-01

    A more robust pre-clinical wear simulation framework is required in order to simulate wider and higher ranges of activities, observed in different patient populations such as younger more active patients. Such a framework will help to understand and address the reported higher failure rates for younger and more active patients (National_Joint_Registry, 2016). The current study has developed and validated a comprehensive combined experimental and computational framework for pre-clinical wear simulation of total knee replacements (TKR). The input mechanical (elastic modulus and Poisson's ratio) and wear parameters of the moderately cross-linked ultra-high molecular weight polyethylene (UHMWPE) bearing material were independently measured from experimental studies under realistic test conditions, similar to the loading conditions found in the total knee replacements. The wear predictions from the computational wear simulation were validated against the direct experimental wear measurements for size 3 Sigma curved total knee replacements (DePuy, UK) in an independent experimental wear simulation study under three different daily activities; walking, deep squat, and stairs ascending kinematic conditions. The measured compressive mechanical properties of the moderately cross-linked UHMWPE material were more than 20% lower than that reported in the literature under tensile test conditions. The pin-on-plate wear coefficient of moderately cross-linked UHMWPE was significantly dependant of the contact stress and the degree of cross-shear at the articulating surfaces. The computational wear predictions for the TKR from the current framework were consistent and in a good agreement with the independent full TKR experimental wear simulation measurements, with 0.94 coefficient of determination of the framework. In addition, the comprehensive combined experimental and computational framework was able to explain the complex experimental wear trends from the three different daily

  18. Storm-time total electron content and its response to penetration electric fields over South America

    Directory of Open Access Journals (Sweden)

    P. M. de Siqueira

    2011-10-01

    Full Text Available In this work the response of the ionosphere due to the severe magnetic storm of 7–10 November 2004 is investigated by analyzing GPS Total Electron Content (TEC maps constructed for the South America sector. In order to verify the disturbed zonal electric fields in South America during the superstorm, ionospheric vertical drift data obtained from modeling results are used in the analysis. The vertical drifts were inferred from ΔH magnetometer data (Jicamarca-Piura following the methodology presented by Anderson et al. (2004. Also used were vertical drifts measured by the Jicamarca ISR. Data from a digisonde located at São Luís, Brazil (2.33° S, 44.2° W, dip latitude 0.25° are presented to complement the Jicamarca equatorial data. Penetration electric fields were observed by the comparison between the equatorial vertical drifts and the Interplanetary Electric Field (IEF. The TEC maps obtained from GPS data reflect the ionospheric response over the South America low-latitude and equatorial region. They reveal unexpected plasma distributions and TEC levels during the main phase of the superstorm on 7 November, which is coincident with the local post-sunset hours. At this time an increase in the pre-reversal enhancement was expected to develop the Equatorial Ionization Anomaly (EIA but we observed the absence of EIA. The results also reveal well known characteristics of the plasma distributions on 8, 9, and 10 November. The emphasized features are the expansion and intensification of EIA due to prompt penetration electric fields on 9 November and the inhibition of EIA during post-sunset hours on 7, 8, and 10 November. One important result is that the TEC maps provided a bi-dimensional view of the ionospheric changes offering a spatial description of the electrodynamics involved, which is an advantage over TEC measured by isolated GPS receivers.

  19. When is it safe to resume driving after total hip and total knee arthroplasty? a meta-analysis of literature on post-operative brake reaction times.

    Science.gov (United States)

    van der Velden, C A; Tolk, J J; Janssen, R P A; Reijman, M

    2017-05-01

    The aim of this study was to assess the current available evidence about when patients might resume driving after elective, primary total hip (THA) or total knee arthroplasty (TKA) undertaken for osteoarthritis (OA). In February 2016, EMBASE, MEDLINE, Web of Science, Scopus, Cochrane, PubMed Publisher, CINAHL, EBSCO and Google Scholar were searched for clinical studies reporting on 'THA', 'TKA', 'car driving', 'reaction time' and 'brake response time'. Two researchers (CAV and JJT) independently screened the titles and abstracts for eligibility and assessed the risk of bias. Both fixed and random effects were used to pool data and calculate mean differences (MD) and 95% confidence intervals (CI) between pre- and post-operative total brake response time (TBRT). A total of 19 studies were included. The assessment of the risk of bias showed that one study was at high risk, six studies at moderate risk and 12 studies at low risk. Meta-analysis of TBRT showed a MD decrease of 25.54 ms (95% CI -32.02 to 83.09) two weeks after right-sided THA, and of 18.19 ms (95% CI -6.13 to 42.50) four weeks after a right-sided TKA, when compared with the pre-operative value. The TBRT returned to baseline two weeks after a right-sided THA and four weeks after a right-sided TKA. These results may serve as guidelines for orthopaedic surgeons when advising patients when to resume driving. However, the advice should be individualised. Cite this article: Bone Joint J 2017;99-B:566-76. ©2017 The British Editorial Society of Bone & Joint Surgery.

  20. 12 CFR 516.10 - How does OTS compute time periods under this part?

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 5 2010-01-01 2010-01-01 false How does OTS compute time periods under this part? 516.10 Section 516.10 Banks and Banking OFFICE OF THRIFT SUPERVISION, DEPARTMENT OF THE TREASURY APPLICATION PROCESSING PROCEDURES § 516.10 How does OTS compute time periods under this part? In computing...

  1. How Many Times Should One Run a Computational Simulation?

    DEFF Research Database (Denmark)

    Seri, Raffaello; Secchi, Davide

    2017-01-01

    This chapter is an attempt to answer the question “how many runs of a computational simulation should one do,” and it gives an answer by means of statistical analysis. After defining the nature of the problem and which types of simulation are mostly affected by it, the article introduces statisti......This chapter is an attempt to answer the question “how many runs of a computational simulation should one do,” and it gives an answer by means of statistical analysis. After defining the nature of the problem and which types of simulation are mostly affected by it, the article introduces...

  2. A general algorithm for computing distance transforms in linear time

    NARCIS (Netherlands)

    Meijster, A.; Roerdink, J.B.T.M.; Hesselink, W.H.; Goutsias, J; Vincent, L; Bloomberg, DS

    2000-01-01

    A new general algorithm fur computing distance transforms of digital images is presented. The algorithm consists of two phases. Both phases consist of two scans, a forward and a backward scan. The first phase scans the image column-wise, while the second phase scans the image row-wise. Since the

  3. Constrained Total Generalized p-Variation Minimization for Few-View X-Ray Computed Tomography Image Reconstruction.

    Science.gov (United States)

    Zhang, Hanming; Wang, Linyuan; Yan, Bin; Li, Lei; Cai, Ailong; Hu, Guoen

    2016-01-01

    Total generalized variation (TGV)-based computed tomography (CT) image reconstruction, which utilizes high-order image derivatives, is superior to total variation-based methods in terms of the preservation of edge information and the suppression of unfavorable staircase effects. However, conventional TGV regularization employs l1-based form, which is not the most direct method for maximizing sparsity prior. In this study, we propose a total generalized p-variation (TGpV) regularization model to improve the sparsity exploitation of TGV and offer efficient solutions to few-view CT image reconstruction problems. To solve the nonconvex optimization problem of the TGpV minimization model, we then present an efficient iterative algorithm based on the alternating minimization of augmented Lagrangian function. All of the resulting subproblems decoupled by variable splitting admit explicit solutions by applying alternating minimization method and generalized p-shrinkage mapping. In addition, approximate solutions that can be easily performed and quickly calculated through fast Fourier transform are derived using the proximal point method to reduce the cost of inner subproblems. The accuracy and efficiency of the simulated and real data are qualitatively and quantitatively evaluated to validate the efficiency and feasibility of the proposed method. Overall, the proposed method exhibits reasonable performance and outperforms the original TGV-based method when applied to few-view problems.

  4. Increased epicardial fat volume quantified by 64-multidetector computed tomography is associated with coronary atherosclerosis and totally occlusive lesions

    International Nuclear Information System (INIS)

    Ueno, Koji; Anzai, Toshihisa; Jinzaki, Masahiro

    2009-01-01

    The relationship between the epicardial fat volume measured by 64-slice multidetector computed tomography (MDCT) and the extension and severity of coronary atherosclerosis was investigated. Both MDCT and conventional coronary angiography (CAG) were performed in 71 consecutive patients who presented with effort angina. The volume of epicardial adipose tissue (EAT) was measured by MDCT. The severity of coronary atherosclerosis was assessed by evaluating the extension of coronary plaques in 790 segments using MDCT data, and the percentage diameter stenosis in 995 segments using CAG data. The estimated volume of EAT indexed by body surface area was defined as VEAT. Increased VEAT was associated with advanced age, male sex, degree of metabolic alterations, a history of acute coronary syndrome (ACS) and the presence of total occlusions, and showed positive correlation with the stenosis score r=0.28, P=0.02) and the atheromatosis score (r=0.67, P 3 /m 2 ) to be the strongest independent determinant of the presence of total occlusions odds ratio 4.64. P=0.02). VEAT correlates with the degree of metabolic alterations and coronary atheromatosis. Excessive accumulation of EAT might contribute to the development of ACS and coronary total occlusions. (author)

  5. Quo vadis? : persuasive computing using real time queue information

    NARCIS (Netherlands)

    Meys, Wouter; Groen, Maarten

    2014-01-01

    By presenting tourists with real-time information an increase in efficiency and satisfaction of their day planning can be achieved. At the same time, real-time information services can offer the municipality the opportunity to spread the tourists throughout the city centre. An important factor for

  6. Event Based Simulator for Parallel Computing over the Wide Area Network for Real Time Visualization

    Science.gov (United States)

    Sundararajan, Elankovan; Harwood, Aaron; Kotagiri, Ramamohanarao; Satria Prabuwono, Anton

    As the computational requirement of applications in computational science continues to grow tremendously, the use of computational resources distributed across the Wide Area Network (WAN) becomes advantageous. However, not all applications can be executed over the WAN due to communication overhead that can drastically slowdown the computation. In this paper, we introduce an event based simulator to investigate the performance of parallel algorithms executed over the WAN. The event based simulator known as SIMPAR (SIMulator for PARallel computation), simulates the actual computations and communications involved in parallel computation over the WAN using time stamps. Visualization of real time applications require steady stream of processed data flow for visualization purposes. Hence, SIMPAR may prove to be a valuable tool to investigate types of applications and computing resource requirements to provide uninterrupted flow of processed data for real time visualization purposes. The results obtained from the simulation show concurrence with the expected performance using the L-BSP model.

  7. A Computer Navigation System Analysis of the Accuracy of the Extramedullary (Tibial Alignment Technique in Total Knee Arthroplasty (TKA

    Directory of Open Access Journals (Sweden)

    EK Chee

    2010-07-01

    Full Text Available In total knee arthroplasty, mechanical alignment guides have improved the accuracy of implant alignment, but errors are not uncommon. In the present study, an image-free computer-assisted navigation system was used to analyse the accuracy of an extramedullary (tibial alignment system, which is based on predetermined, fixed anatomical landmarks. Comparisons were made between two surgeons, with different levels of competency in order to determine if experience affected the accuracy of extramedullary tibial jig placement, in either the coronal and sagittal planes or both planes. The results showed that the accuracy of the extramedullary tibial alignment system, in the coronal plane (in up to 80-87% of cases was much better than for posterior slope, and sagittal plane. Surgeon experience was not a significant factor.

  8. Quantum computing without wavefunctions: time-dependent density functional theory for universal quantum computation.

    Science.gov (United States)

    Tempel, David G; Aspuru-Guzik, Alán

    2012-01-01

    We prove that the theorems of TDDFT can be extended to a class of qubit Hamiltonians that are universal for quantum computation. The theorems of TDDFT applied to universal Hamiltonians imply that single-qubit expectation values can be used as the basic variables in quantum computation and information theory, rather than wavefunctions. From a practical standpoint this opens the possibility of approximating observables of interest in quantum computations directly in terms of single-qubit quantities (i.e. as density functionals). Additionally, we also demonstrate that TDDFT provides an exact prescription for simulating universal Hamiltonians with other universal Hamiltonians that have different, and possibly easier-to-realize two-qubit interactions. This establishes the foundations of TDDFT for quantum computation and opens the possibility of developing density functionals for use in quantum algorithms.

  9. The importance of bony impingement in restricting flexion after total knee arthroplasty: computer simulation model with clinical correlation.

    Science.gov (United States)

    Mizu-Uchi, Hideki; Colwell, Clifford W; Fukagawa, Shingo; Matsuda, Shuichi; Iwamoto, Yukihide; D'Lima, Darryl D

    2012-10-01

    We constructed patient-specific models from computed tomography data after total knee arthroplasty to predict knee flexion based on implant-bone impingement. The maximum flexion before impingement between the femur and the tibial insert was computed using a musculoskeletal modeling program (KneeSIM; LifeModeler, Inc, San Clemente, California) during a weight-bearing deep knee bend. Postoperative flexion was measured in a clinical cohort of 21 knees (low-flex group: 6 knees with 125° of flexion at 2 years). Average predicted flexion angles were within 2° of clinical measurements for the high-flex group. In the low-flex group, 4 cases had impingement involving the bone cut at the posterior condyle, and the average predicted knee flexion was 102° compared with 93° measured clinically. These results indicate that the level of the distal femoral resection should be carefully planned and that exposed bone proximal to the tips of the posterior condyles of the femoral component should be removed if there is risk of impingement. Copyright © 2012 Elsevier Inc. All rights reserved.

  10. Adaptive-weighted total variation minimization for sparse data toward low-dose x-ray computed tomography image reconstruction.

    Science.gov (United States)

    Liu, Yan; Ma, Jianhua; Fan, Yi; Liang, Zhengrong

    2012-12-07

    Previous studies have shown that by minimizing the total variation (TV) of the to-be-estimated image with some data and other constraints, piecewise-smooth x-ray computed tomography (CT) can be reconstructed from sparse-view projection data without introducing notable artifacts. However, due to the piecewise constant assumption for the image, a conventional TV minimization algorithm often suffers from over-smoothness on the edges of the resulting image. To mitigate this drawback, we present an adaptive-weighted TV (AwTV) minimization algorithm in this paper. The presented AwTV model is derived by considering the anisotropic edge property among neighboring image voxels, where the associated weights are expressed as an exponential function and can be adaptively adjusted by the local image-intensity gradient for the purpose of preserving the edge details. Inspired by the previously reported TV-POCS (projection onto convex sets) implementation, a similar AwTV-POCS implementation was developed to minimize the AwTV subject to data and other constraints for the purpose of sparse-view low-dose CT image reconstruction. To evaluate the presented AwTV-POCS algorithm, both qualitative and quantitative studies were performed by computer simulations and phantom experiments. The results show that the presented AwTV-POCS algorithm can yield images with several notable gains, in terms of noise-resolution tradeoff plots and full-width at half-maximum values, as compared to the corresponding conventional TV-POCS algorithm.

  11. Central axis dose verification in patients treated with total body irradiation of photons using a Computed Radiography system

    International Nuclear Information System (INIS)

    Rubio Rivero, A.; Caballero Pinelo, R.; Gonzalez Perez, Y.

    2015-01-01

    To propose and evaluate a method for the central axis dose verification in patients treated with total body irradiation (TBI) of photons using images obtained through a Computed Radiography (CR) system. It was used the Computed Radiography (Fuji) portal imaging cassette readings and correlate with measured of absorbed dose in water using 10 x 10 irradiation fields with ionization chamber in the 60 Co equipment. The analytical and graphic expression is obtained through software 'Origin8', the TBI patient portal verification images were processed using software ImageJ, to obtain the patient dose. To validate the results, the absorbed dose in RW3 models was measured with ionization chamber with different thickness, simulating TBI real conditions. Finally it was performed a retrospective study over the last 4 years obtaining the patients absorbed dose based on the reading in the image and comparing with the planned dose. The analytical equation obtained permits estimate the absorbed dose using image pixel value and the dose measured with ionization chamber and correlated with patient clinical records. Those results are compared with reported evidence obtaining a difference less than 02%, the 3 methods were compared and the results are within 10%. (Author)

  12. 5 CFR 831.703 - Computation of annuities for part-time service.

    Science.gov (United States)

    2010-01-01

    ... 5 Administrative Personnel 2 2010-01-01 2010-01-01 false Computation of annuities for part-time... part-time service. (a) Purpose. The computational method in this section shall be used to determine the annuity for an employee who has part-time service on or after April 7, 1986. (b) Definitions. In this...

  13. Storm blueprints patterns for distributed real-time computation

    CERN Document Server

    Goetz, P Taylor

    2014-01-01

    A blueprints book with 10 different projects built in 10 different chapters which demonstrate the various use cases of storm for both beginner and intermediate users, grounded in real-world example applications.Although the book focuses primarily on Java development with Storm, the patterns are more broadly applicable and the tips, techniques, and approaches described in the book apply to architects, developers, and operations.Additionally, the book should provoke and inspire applications of distributed computing to other industries and domains. Hadoop enthusiasts will also find this book a go

  14. Macroprocessing is the computing design principle for the times

    CERN Multimedia

    2001-01-01

    In a keynote speech, Intel Corporation CEO Craig Barrett emphasized that "macroprocessing" provides innovative and cost effective solutions to companies that they can customize and scale to match their own data needs. Barrett showcased examples of macroprocessing implementations from business, government and the scientific community, which use the power of Intel Architecture and Oracle9i Real Application Clusters to build large complex and scalable database solutions. A testimonial from CERN explained how the need for high performance computing to perform scientific research on sub-atomic particles was accomplished by using clusters of Xeon processor-based servers.

  15. Bulk metal concentrations versus total suspended solids in rivers: Time-invariant & catchment-specific relationships.

    Science.gov (United States)

    Nasrabadi, Touraj; Ruegner, Hermann; Schwientek, Marc; Bennett, Jeremy; Fazel Valipour, Shahin; Grathwohl, Peter

    2018-01-01

    Suspended particles in rivers can act as carriers of potentially bioavailable metal species and are thus an emerging area of interest in river system monitoring. The delineation of bulk metals concentrations in river water into dissolved and particulate components is also important for risk assessment. Linear relationships between bulk metal concentrations in water (CW,tot) and total suspended solids (TSS) in water can be used to easily evaluate dissolved (CW, intercept) and particle-bound metal fluxes (CSUS, slope) in streams (CW,tot = CW + CSUS TSS). In this study, we apply this principle to catchments in Iran (Haraz) and Germany (Ammer, Goldersbach, and Steinlach) that show differences in geology, geochemistry, land use and hydrological characteristics. For each catchment, particle-bound and dissolved concentrations for a suite of metals in water were calculated based on linear regressions of total suspended solids and total metal concentrations. Results were replicable across sampling campaigns in different years and seasons (between 2013 and 2016) and could be reproduced in a laboratory sedimentation experiment. CSUS values generally showed little variability in different catchments and agree well with soil background values for some metals (e.g. lead and nickel) while other metals (e.g. copper) indicate anthropogenic influences. CW was elevated in the Haraz (Iran) catchment, indicating higher bioavailability and potential human and ecological health concerns (where higher values of CSUS/CW are considered as a risk indicator).

  16. Total factor productivity (TFP) growth agriculture in pakistan: trends in different time horizons

    International Nuclear Information System (INIS)

    Ali, A.; Mushtaq, K.; Ashfaq, M.

    2008-01-01

    The present study estimated total factor productivity (TFP) growth of agriculture sector of Pakistan for the period 1971-2006 by employing Tornqvist-Theil (T-T) index number methodology. Most of the conventional inputs were used in constructing the input index. The output index includes major crops, minor crops, important fruits and vegetables and four categories of livestock products. The study estimated TFP growth rates for different decades. The results showed that TFP growth rate was lowest during the decade of 70s (0.96 percent) and highest during the last six years of the study period (2.86 percent). The decade of 80s and 90s registered TFP growth rate of 2.24 percent and 2.46 percent, respectively. The results also explained that TFP growth contributed about 33 percent to total agricultural output growth during the decade of 70s and this contribution increased up to 83 percent during the last six years of the study period. The contribution of TFP growth to total agricultural output growth was 53 and 81 percent during the decades of 80s and 90s, respectively. The study observed that macro level government policies, institutional factors and weather conditions are the major key factors that influenced TFP growth. (author)

  17. Critical Care Admissions following Total Laryngectomy: Is It Time to Change Our Practice?

    Science.gov (United States)

    Walijee, Hussein; Morgan, Alexandria; Gibson, Bethan; Berry, Sandeep; Jaffery, Ali

    2016-01-01

    Critical Care Unit (CCU) beds are a limited resource and in increasing demand. Studies have shown that complex head and neck patients can be safely managed on a ward setting given the appropriate staffing and support. This retrospective case series aims to quantify the CCU care received by patients following total laryngectomy (TL) at a District General Hospital (DGH) and compare patient outcomes in an attempt to inform current practice. Data relating to TL were collected over a 5-year period from 1st January 2010 to 31st December 2015. A total of 22 patients were included. All patients were admitted to CCU postoperatively for an average length of stay of 25.5 hours. 95% of these patients were admitted to CCU for the purpose of close monitoring only, not requiring any active treatment prior to discharge to the ward. 73% of total complications were encountered after the first 24 hours postoperatively at which point patients had been stepped down to ward care. Avoiding the use of CCU beds and instead providing the appropriate level of care on the ward would result in a potential cost saving of approximately £8,000 with no influence on patient morbidity and mortality.

  18. Real-time exposure fusion on a mobile computer

    CSIR Research Space (South Africa)

    Bachoo, AK

    2009-12-01

    Full Text Available information in these scenarios. An image captured using a short exposure time will not saturate bright image re- gions while an image captured with a long exposure time will show more detail in the dark regions. The pixel depth provided by most camera.... The auto exposure also creates strong blown-out highlights in the foreground (the grass patch). The short shutter time (Exposure 1) correctly exposes the grass while the long shutter time (Exposure 3) is able to correctly expose the camouflaged dummy...

  19. Correlates of occupational, leisure and total sitting time in working adults: results from the Singapore multi-ethnic cohort.

    Science.gov (United States)

    Uijtdewilligen, Léonie; Yin, Jason Dean-Chen; van der Ploeg, Hidde P; Müller-Riemenschneider, Falk

    2017-12-13

    Evidence on the health risks of sitting is accumulating. However, research identifying factors influencing sitting time in adults is limited, especially in Asian populations. This study aimed to identify socio-demographic and lifestyle correlates of occupational, leisure and total sitting time in a sample of Singapore working adults. Data were collected between 2004 and 2010 from participants of the Singapore Multi Ethnic Cohort (MEC). Medical exclusion criteria for cohort participation were cancer, heart disease, stroke, renal failure and serious mental illness. Participants who were not working over the past 12 months and without data on sitting time were excluded from the analyses. Multivariable regression analyses were used to examine cross-sectional associations of self-reported age, gender, ethnicity, marital status, education, smoking, caloric intake and moderate-to-vigorous leisure time physical activity (LTPA) with self-reported occupational, leisure and total sitting time. Correlates were also studied separately for Chinese, Malays and Indians. The final sample comprised 9384 participants (54.8% male): 50.5% were Chinese, 24.0% Malay, and 25.5% Indian. For the total sample, mean occupational sitting time was 2.71 h/day, mean leisure sitting time was 2.77 h/day and mean total sitting time was 5.48 h/day. Sitting time in all domains was highest among Chinese. Age, gender, education, and caloric intake were associated with higher occupational sitting time, while ethnicity, marital status and smoking were associated with lower occupational sitting time. Marital status, smoking, caloric intake and LTPA were associated with higher leisure sitting time, while age, gender and ethnicity were associated with lower leisure sitting time. Gender, marital status, education, caloric intake and LTPA were associated with higher total sitting time, while ethnicity was associated with lower total sitting time. Stratified analyses revealed different associations within

  20. Total donor ischemic time: relationship to early hemodynamics and intensive care morbidity in pediatric cardiac transplant recipients.

    Science.gov (United States)

    Rodrigues, Warren; Carr, Michelle; Ridout, Deborah; Carter, Katherine; Hulme, Sara Louise; Simmonds, Jacob; Elliott, Martin; Hoskote, Aparna; Burch, Michael; Brown, Kate L

    2011-11-01

    Single-center studies have failed to link modest increases in total donor ischemic time to mortality after pediatric orthotopic heart transplant. We aimed to investigate whether prolonged total donor ischemic time is linked to pediatric intensive care morbidity after orthotopic heart transplant. Retrospective cohort review. Tertiary pediatric transplant center in the United Kingdom. Ninety-three pediatric orthotopic heart transplants between 2002 and 2006. Total donor ischemic time was investigated for association with early post-orthotopic heart transplant hemodynamics and intensive care unit morbidities. Of 43 males and 50 females with median age 7.2 (interquartile range 2.2, 13.0) yrs, 62 (68%) had dilated cardiomyopathy, 20 (22%) had congenital heart disease, and nine (10%) had restrictive cardiomyopathy. The mean total donor ischemic time was 225.9 (sd 65.6) mins. In the first 24 hrs after orthotopic heart transplant, age-adjusted mean arterial blood pressure increased (p total donor ischemic time was significantly associated with lower mean arterial blood pressure (p care unit (p = .004), and longer post-orthotopic heart transplant stay in hospital (p = .02). Total donor ischemic time was not related to levels of mean pulmonary arterial pressure (p = .62), left atrial pressure (p = .38), or central venous pressure (p = .76) early after orthotopic heart transplant. Prolonged total donor ischemic time has an adverse effect on the donor organ, contributing to lower mean arterial blood pressure, as well as more prolonged ventilation and intensive care unit and hospital stays post-orthotopic heart transplant, reflecting increased morbidity.

  1. Project Energise: Using participatory approaches and real time computer prompts to reduce occupational sitting and increase work time physical activity in office workers.

    Science.gov (United States)

    Gilson, Nicholas D; Ng, Norman; Pavey, Toby G; Ryde, Gemma C; Straker, Leon; Brown, Wendy J

    2016-11-01

    This efficacy study assessed the added impact real time computer prompts had on a participatory approach to reduce occupational sedentary exposure and increase physical activity. Quasi-experimental. 57 Australian office workers (mean [SD]; age=47 [11] years; BMI=28 [5]kg/m 2 ; 46 men) generated a menu of 20 occupational 'sit less and move more' strategies through participatory workshops, and were then tasked with implementing strategies for five months (July-November 2014). During implementation, a sub-sample of workers (n=24) used a chair sensor/software package (Sitting Pad) that gave real time prompts to interrupt desk sitting. Baseline and intervention sedentary behaviour and physical activity (GENEActiv accelerometer; mean work time percentages), and minutes spent sitting at desks (Sitting Pad; mean total time and longest bout) were compared between non-prompt and prompt workers using a two-way ANOVA. Workers spent close to three quarters of their work time sedentary, mostly sitting at desks (mean [SD]; total desk sitting time=371 [71]min/day; longest bout spent desk sitting=104 [43]min/day). Intervention effects were four times greater in workers who used real time computer prompts (8% decrease in work time sedentary behaviour and increase in light intensity physical activity; pcomputer prompts facilitated the impact of a participatory approach on reductions in occupational sedentary exposure, and increases in physical activity. Copyright © 2016 Sports Medicine Australia. Published by Elsevier Ltd. All rights reserved.

  2. Imageless navigation total hip arthroplasty – an evaluation of operative time

    Directory of Open Access Journals (Sweden)

    Valsamis Epaminondas Markos

    2018-01-01

    Discussion: This is the first study that demonstrates no added operative time when using imageless navigation in THA, achieved with an improved workflow. The results also demonstrate a very reasonable learning curve.

  3. Computer-controlled neutron time-of-flight spectrometer. Part II

    International Nuclear Information System (INIS)

    Merriman, S.H.

    1979-12-01

    A time-of-flight spectrometer for neutron inelastic scattering research has been interfaced to a PDP-15/30 computer. The computer is used for experimental data acquisition and analysis and for apparatus control. This report was prepared to summarize the functions of the computer and to act as a users' guide to the software system

  4. 20 Years of Total and Tropical Ozone Time Series Based on European Satellite Observations

    Science.gov (United States)

    Loyola, D. G.; Heue, K. P.; Coldewey-Egbers, M.

    2016-12-01

    Ozone is an important trace gas in the atmosphere, while the stratospheric ozone layer protects the earth surface from the incident UV radiation, the tropospheric ozone acts as green house gas and causes health damages as well as crop loss. The total ozone column is dominated by the stratospheric column, the tropospheric columns only contributes about 10% to the total column.The ozone column data from the European satellite instruments GOME, SCIAMACHY, OMI, GOME-2A and GOME-2B are available within the ESA Climate Change Initiative project with a high degree of inter-sensor consistency. The tropospheric ozone columns are based on the convective cloud differential algorithm. The datasets encompass a period of more than 20 years between 1995 and 2015, for the trend analysis the data sets were harmonized relative to one of the instruments. For the tropics we found an increase in the tropospheric ozone column of 0.75 ± 0.12 DU decade^{-1} with local variations between 1.8 and -0.8. The largest trends were observed over southern Africa and the Atlantic Ocean. A seasonal trend analysis led to the assumption that the increase is caused by additional forest fires.The trend for the total column was not that certain, based on model predicted trend data and the measurement uncertainty we estimated that another 10 to 15 years of observations will be required to observe a statistical significant trend. In the mid latitudes the trends are currently hidden in the large variability and for the tropics the modelled trends are low. Also the possibility of diverging trends at different altitudes must be considered; an increase in the tropospheric ozone might be accompanied by decreasing stratospheric ozone.The European satellite data record will be extended over the next two decades with the atmospheric satellite missions Sentinel 5 Precursor (launch end of 2016), Sentinel 4 and Sentinel 5.

  5. Computation and evaluation of scheduled waiting time for railway networks

    DEFF Research Database (Denmark)

    Landex, Alex

    2010-01-01

    Timetables are affected by scheduled waiting time (SWT) that prolongs the travel times for trains and thereby passengers. SWT occurs when a train hinders another train to run with the wanted speed. The SWT affects both the trains and the passengers in the trains. The passengers may be further...... affected due to longer transfer times to other trains. SWT can be estimated analytically for a given timetable or by simulation of timetables and/or plans of operation. The simulation of SWT has the benefit that it is possible to examine the entire network. This makes it possible to improve the future...

  6. Patient-specific instrumentation for total knee arthroplasty does not match the pre-operative plan as assessed by intra-operative computer-assisted navigation.

    Science.gov (United States)

    Scholes, Corey; Sahni, Varun; Lustig, Sebastien; Parker, David A; Coolican, Myles R J

    2014-03-01

    The introduction of patient-specific instruments (PSI) for guiding bone cuts could increase the incidence of malalignment in primary total knee arthroplasty. The purpose of this study was to assess the agreement between one type of patient-specific instrumentation (Zimmer PSI) and the pre-operative plan with respect to bone cuts and component alignment during TKR using imageless computer navigation. A consecutive series of 30 femoral and tibial guides were assessed in-theatre by the same surgeon using computer navigation. Following surgical exposure, the PSI cutting guides were placed on the joint surface and alignment assessed using the navigation tracker. The difference between in-theatre data and the pre-operative plan was recorded and analysed. The error between in-theatre measurements and pre-operative plan for the femoral and tibial components exceeded 3° for 3 and 17% of the sample, respectively, while the error for total coronal alignment exceeded 3° for 27% of the sample. The present results indicate that alignment with Zimmer PSI cutting blocks, assessed by imageless navigation, does not match the pre-operative plan in a proportion of cases. To prevent unnecessary increases in the incidence of malalignment in primary TKR, it is recommended that these devices should not be used without objective verification of alignment, either in real-time or with post-operative imaging. Further work is required to identify the source of discrepancies and validate these devices prior to routine use. II.

  7. A computer program for the estimation of time of death

    DEFF Research Database (Denmark)

    Lynnerup, N

    1993-01-01

    In the 1960s Marshall and Hoare presented a "Standard Cooling Curve" based on their mathematical analyses on the postmortem cooling of bodies. Although fairly accurate under standard conditions, the "curve" or formula is based on the assumption that the ambience temperature is constant and that t......In the 1960s Marshall and Hoare presented a "Standard Cooling Curve" based on their mathematical analyses on the postmortem cooling of bodies. Although fairly accurate under standard conditions, the "curve" or formula is based on the assumption that the ambience temperature is constant...... cooling of bodies is presented. It is proposed that by having a computer program that solves the equation, giving the length of the cooling period in response to a certain rectal temperature, and which allows easy comparison of multiple solutions, the uncertainties related to ambience temperature...

  8. 42 CFR 93.509 - Computation of time.

    Science.gov (United States)

    2010-10-01

    ... holiday observed by the Federal government, in which case it includes the next business day. (b) When the... required or authorized under the rules in this part to be filed for good cause shown. When time permits...

  9. Computational complexity of time-dependent density functional theory

    International Nuclear Information System (INIS)

    Whitfield, J D; Yung, M-H; Tempel, D G; Aspuru-Guzik, A; Boixo, S

    2014-01-01

    Time-dependent density functional theory (TDDFT) is rapidly emerging as a premier method for solving dynamical many-body problems in physics and chemistry. The mathematical foundations of TDDFT are established through the formal existence of a fictitious non-interacting system (known as the Kohn–Sham system), which can reproduce the one-electron reduced probability density of the actual system. We build upon these works and show that on the interior of the domain of existence, the Kohn–Sham system can be efficiently obtained given the time-dependent density. We introduce a V-representability parameter which diverges at the boundary of the existence domain and serves to quantify the numerical difficulty of constructing the Kohn-Sham potential. For bounded values of V-representability, we present a polynomial time quantum algorithm to generate the time-dependent Kohn–Sham potential with controllable error bounds. (paper)

  10. On some methods for improving time of reachability sets computation for the dynamic system control problem

    Science.gov (United States)

    Zimovets, Artem; Matviychuk, Alexander; Ushakov, Vladimir

    2016-12-01

    The paper presents two different approaches to reduce the time of computer calculation of reachability sets. First of these two approaches use different data structures for storing the reachability sets in the computer memory for calculation in single-threaded mode. Second approach is based on using parallel algorithms with reference to the data structures from the first approach. Within the framework of this paper parallel algorithm of approximate reachability set calculation on computer with SMP-architecture is proposed. The results of numerical modelling are presented in the form of tables which demonstrate high efficiency of parallel computing technology and also show how computing time depends on the used data structure.

  11. Total sitting time, leisure time physical activity and risk of hospitalization due to low back pain: The Danish Health Examination Survey cohort 2007-2008.

    Science.gov (United States)

    Balling, Mie; Holmberg, Teresa; Petersen, Christina B; Aadahl, Mette; Meyrowitsch, Dan W; Tolstrup, Janne S

    2018-02-01

    This study aimed to test the hypotheses that a high total sitting time and vigorous physical activity in leisure time increase the risk of low back pain and herniated lumbar disc disease. A total of 76,438 adults answered questions regarding their total sitting time and physical activity during leisure time in the Danish Health Examination Survey 2007-2008. Information on low back pain diagnoses up to 10 September 2015 was obtained from The National Patient Register. The mean follow-up time was 7.4 years. Data were analysed using Cox regression analysis with adjustment for potential confounders. Multiple imputations were performed for missing values. During the follow-up period, 1796 individuals were diagnosed with low back pain, of whom 479 were diagnosed with herniated lumbar disc disease. Total sitting time was not associated with low back pain or herniated lumbar disc disease. However, moderate or vigorous physical activity, as compared to light physical activity, was associated with increased risk of low back pain (HR = 1.16, 95% CI: 1.03-1.30 and HR = 1.45, 95% CI: 1.15-1.83). Moderate, but not vigorous physical activity was associated with increased risk of herniated lumbar disc disease. The results suggest that total sitting time is not associated with low back pain, but moderate and vigorous physical activity is associated with increased risk of low back pain compared with light physical activity.

  12. A strategy for reducing turnaround time in design optimization using a distributed computer system

    Science.gov (United States)

    Young, Katherine C.; Padula, Sharon L.; Rogers, James L.

    1988-01-01

    There is a need to explore methods for reducing lengthly computer turnaround or clock time associated with engineering design problems. Different strategies can be employed to reduce this turnaround time. One strategy is to run validated analysis software on a network of existing smaller computers so that portions of the computation can be done in parallel. This paper focuses on the implementation of this method using two types of problems. The first type is a traditional structural design optimization problem, which is characterized by a simple data flow and a complicated analysis. The second type of problem uses an existing computer program designed to study multilevel optimization techniques. This problem is characterized by complicated data flow and a simple analysis. The paper shows that distributed computing can be a viable means for reducing computational turnaround time for engineering design problems that lend themselves to decomposition. Parallel computing can be accomplished with a minimal cost in terms of hardware and software.

  13. Contrast timing in computed tomography: Effect of different contrast media concentrations on bolus geometry

    International Nuclear Information System (INIS)

    Mahnken, Andreas H.; Jost, Gregor; Seidensticker, Peter; Kuhl, Christiane; Pietsch, Hubertus

    2012-01-01

    Objective: To assess the effect of low-osmolar, monomeric contrast media with different iodine concentrations on bolus shape in aortic CT angiography. Materials and methods: Repeated sequential computed tomography scanning of the descending aorta of eight beagle dogs (5 male, 12.7 ± 3.1 kg) was performed without table movement with a standardized CT scan protocol. Iopromide 300 (300 mg I/mL), iopromide 370 (370 mg I/mL) and iomeprol 400 (400 mg I/mL) were administered via a foreleg vein with an identical iodine delivery rate of 1.2 g I/s and a total iodine dose of 300 mg I/kg body weight. Time-enhancement curves were computed and analyzed. Results: Iopromide 300 showed the highest peak enhancement (445.2 ± 89.1 HU), steepest up-slope (104.2 ± 17.5 HU/s) and smallest full width at half maximum (FWHM; 5.8 ± 1.0 s). Peak enhancement, duration of FWHM, enhancement at FWHM and up-slope differed significantly between iopromide 300 and iomeprol 400 (p 0.05). Conclusions: Low viscous iopromide 300 results in a better defined bolus with a significantly higher peak enhancement, steeper up-slope and smaller FWHM when compared to iomeprol 400. These characteristics potentially affect contrast timing.

  14. Assessment of Tandem Measurements of pH and Total Gut Transit Time in Healthy Volunteers

    OpenAIRE

    Mikolajczyk, Adam E; Watson, Sydeaka; Surma, Bonnie L; Rubin, David T

    2015-01-01

    Objectives: The variation of luminal pH and transit time in an individual is unknown, yet is necessary to interpret single measurements. This study aimed to assess the intrasubject variability of gut pH and transit time in healthy volunteers using SmartPill devices (Covidien, Minneapolis, MN). Methods: Each subject (n=10) ingested two SmartPill devices separated by 24?h. Mean pH values were calculated for 30?min after gastric emptying (AGE), before the ileocecal (BIC) valve, after the ileocec...

  15. Newmark local time stepping on high-performance computing architectures

    KAUST Repository

    Rietmann, Max

    2016-11-25

    In multi-scale complex media, finite element meshes often require areas of local refinement, creating small elements that can dramatically reduce the global time-step for wave-propagation problems due to the CFL condition. Local time stepping (LTS) algorithms allow an explicit time-stepping scheme to adapt the time-step to the element size, allowing near-optimal time-steps everywhere in the mesh. We develop an efficient multilevel LTS-Newmark scheme and implement it in a widely used continuous finite element seismic wave-propagation package. In particular, we extend the standard LTS formulation with adaptations to continuous finite element methods that can be implemented very efficiently with very strong element-size contrasts (more than 100×). Capable of running on large CPU and GPU clusters, we present both synthetic validation examples and large scale, realistic application examples to demonstrate the performance and applicability of the method and implementation on thousands of CPU cores and hundreds of GPUs.

  16. Newmark local time stepping on high-performance computing architectures

    Energy Technology Data Exchange (ETDEWEB)

    Rietmann, Max, E-mail: max.rietmann@erdw.ethz.ch [Institute for Computational Science, Università della Svizzera italiana, Lugano (Switzerland); Institute of Geophysics, ETH Zurich (Switzerland); Grote, Marcus, E-mail: marcus.grote@unibas.ch [Department of Mathematics and Computer Science, University of Basel (Switzerland); Peter, Daniel, E-mail: daniel.peter@kaust.edu.sa [Institute for Computational Science, Università della Svizzera italiana, Lugano (Switzerland); Institute of Geophysics, ETH Zurich (Switzerland); Schenk, Olaf, E-mail: olaf.schenk@usi.ch [Institute for Computational Science, Università della Svizzera italiana, Lugano (Switzerland)

    2017-04-01

    In multi-scale complex media, finite element meshes often require areas of local refinement, creating small elements that can dramatically reduce the global time-step for wave-propagation problems due to the CFL condition. Local time stepping (LTS) algorithms allow an explicit time-stepping scheme to adapt the time-step to the element size, allowing near-optimal time-steps everywhere in the mesh. We develop an efficient multilevel LTS-Newmark scheme and implement it in a widely used continuous finite element seismic wave-propagation package. In particular, we extend the standard LTS formulation with adaptations to continuous finite element methods that can be implemented very efficiently with very strong element-size contrasts (more than 100x). Capable of running on large CPU and GPU clusters, we present both synthetic validation examples and large scale, realistic application examples to demonstrate the performance and applicability of the method and implementation on thousands of CPU cores and hundreds of GPUs.

  17. Newmark local time stepping on high-performance computing architectures

    KAUST Repository

    Rietmann, Max; Grote, Marcus; Peter, Daniel; Schenk, Olaf

    2016-01-01

    In multi-scale complex media, finite element meshes often require areas of local refinement, creating small elements that can dramatically reduce the global time-step for wave-propagation problems due to the CFL condition. Local time stepping (LTS) algorithms allow an explicit time-stepping scheme to adapt the time-step to the element size, allowing near-optimal time-steps everywhere in the mesh. We develop an efficient multilevel LTS-Newmark scheme and implement it in a widely used continuous finite element seismic wave-propagation package. In particular, we extend the standard LTS formulation with adaptations to continuous finite element methods that can be implemented very efficiently with very strong element-size contrasts (more than 100×). Capable of running on large CPU and GPU clusters, we present both synthetic validation examples and large scale, realistic application examples to demonstrate the performance and applicability of the method and implementation on thousands of CPU cores and hundreds of GPUs.

  18. Determinantal Representation of the Time-Dependent Stationary Correlation Function for the Totally Asymmetric Simple Exclusion Model

    Directory of Open Access Journals (Sweden)

    Nikolay M. Bogoliubov

    2009-04-01

    Full Text Available The basic model of the non-equilibrium low dimensional physics the so-called totally asymmetric exclusion process is related to the 'crystalline limit' (q → ∞ of the SU_q(2 quantum algebra. Using the quantum inverse scattering method we obtain the exact expression for the time-dependent stationary correlation function of the totally asymmetric simple exclusion process on a one dimensional lattice with the periodic boundary conditions.

  19. Invariant set computation for constrained uncertain discrete-time systems

    NARCIS (Netherlands)

    Athanasopoulos, N.; Bitsoris, G.

    2010-01-01

    In this article a novel approach to the determination of polytopic invariant sets for constrained discrete-time linear uncertain systems is presented. First, the problem of stabilizing a prespecified initial condition set in the presence of input and state constraints is addressed. Second, the

  20. 10 CFR 13.27 - Computation of time.

    Science.gov (United States)

    2010-01-01

    ...; and (2) By 11:59 p.m. Eastern Time for a document served by the E-Filing system. [72 FR 49153, Aug. 28... the calculation of additional days when a participant is not entitled to receive an entire filing... same filing and service method, the number of days for service will be determined by the presiding...

  1. 10 CFR 2.306 - Computation of time.

    Science.gov (United States)

    2010-01-01

    ...:59 p.m. Eastern Time for a document served by the E-Filing system. [72 FR 49151, Aug. 28, 2007] ... the calculation of additional days when a participant is not entitled to receive an entire filing... filing and service method, the number of days for service will be determined by the presiding officer...

  2. Real time operating system for a nuclear power plant computer

    International Nuclear Information System (INIS)

    Alger, L.S.; Lala, J.H.

    1986-01-01

    A quadruply redundant synchronous fault tolerant processor (FTP) is now under fabrication at the C.S. Draper Laboratory to be used initially as a trip monitor for the Experimental Breeder Reactor EBR-II operated by the Argonne National Laboratory in Idaho Falls, Idaho. The real time operating system for this processor is described

  3. Total time on test processes and applications to failure data analysis

    International Nuclear Information System (INIS)

    Barlow, R.E.; Campo, R.

    1975-01-01

    This paper describes a new method for analyzing data. The method applies to non-negative observations such as times to failure of devices and survival times of biological organisms and involves a plot of the data. These plots are useful in choosing a probabilistic model to represent the failure behavior of the data. They also furnish information about the failure rate function and aid in its estimation. An important feature of these data plots is that incomplete data can be analyzed. The underlying random variables are, however, assumed to be independent and identically distributed. The plots have a theoretical basis, and converge to a transform of the underlying probability distribution as the sample size increases

  4. Effect of temperature, time, and milling process on yield, flavonoid, and total phenolic content of Zingiber officinale water extract

    Science.gov (United States)

    Andriyani, R.; Kosasih, W.; Ningrum, D. R.; Pudjiraharti, S.

    2017-03-01

    Several parameters such as temperature, time of extraction, and size of simplicia play significant role in medicinal herb extraction. This study aimed to investigate the effect of those parameters on yield extract, flavonoid, and total phenolic content in water extract of Zingiber officinale. The temperatures used were 50, 70 and 90°C and the extraction times were 30, 60 and 90 min. Z. officinale in the form of powder and chips were used to study the effect of milling treatment. The correlation among those variables was analysed using ANOVA two-way factors without replication. The result showed that time and temperature did not influence the yield of extract of Powder simplicia. However, time of extraction influenced the extract of simplicia treated without milling process. On the other hand, flavonoid and total phenolic content were not influenced by temperature, time, and milling treatment.

  5. The reliable solution and computation time of variable parameters Logistic model

    OpenAIRE

    Pengfei, Wang; Xinnong, Pan

    2016-01-01

    The reliable computation time (RCT, marked as Tc) when applying a double precision computation of a variable parameters logistic map (VPLM) is studied. First, using the method proposed, the reliable solutions for the logistic map are obtained. Second, for a time-dependent non-stationary parameters VPLM, 10000 samples of reliable experiments are constructed, and the mean Tc is then computed. The results indicate that for each different initial value, the Tcs of the VPLM are generally different...

  6. Time- and dose-dependent effects of total-body ionizing radiation on muscle stem cells

    Science.gov (United States)

    Masuda, Shinya; Hisamatsu, Tsubasa; Seko, Daiki; Urata, Yoshishige; Goto, Shinji; Li, Tao-Sheng; Ono, Yusuke

    2015-01-01

    Exposure to high levels of genotoxic stress, such as high-dose ionizing radiation, increases both cancer and noncancer risks. However, it remains debatable whether low-dose ionizing radiation reduces cellular function, or rather induces hormetic health benefits. Here, we investigated the effects of total-body γ-ray radiation on muscle stem cells, called satellite cells. Adult C57BL/6 mice were exposed to γ-radiation at low- to high-dose rates (low, 2 or 10 mGy/day; moderate, 50 mGy/day; high, 250 mGy/day) for 30 days. No hormetic responses in proliferation, differentiation, or self-renewal of satellite cells were observed in low-dose radiation-exposed mice at the acute phase. However, at the chronic phase, population expansion of satellite cell-derived progeny was slightly decreased in mice exposed to low-dose radiation. Taken together, low-dose ionizing irradiation may suppress satellite cell function, rather than induce hormetic health benefits, in skeletal muscle in adult mice. PMID:25869487

  7. Fall Risk Score at the Time of Discharge Predicts Readmission Following Total Joint Arthroplasty.

    Science.gov (United States)

    Ravi, Bheeshma; Nan, Zhang; Schwartz, Adam J; Clarke, Henry D

    2017-07-01

    Readmission among Medicare recipients is a leading driver of healthcare expenditure. To date, most predictive tools are too coarse for direct clinical application. Our objective in this study is to determine if a pre-existing tool to identify patients at increased risk for inpatient falls, the Hendrich Fall Risk Score, could be used to accurately identify Medicare patients at increased risk for readmission following arthroplasty, regardless of whether the readmission was due to a fall. This study is a retrospective cohort study. We identified 2437 Medicare patients who underwent a primary elective total joint arthroplasty (TJA) of the hip or knee for osteoarthritis between 2011 and 2014. The Hendrich Fall Risk score was recorded for each patient preoperatively and postoperatively. Our main outcome measure was hospital readmission within 30 days of discharge. Of 2437 eligible TJA recipients, there were 226 (9.3%) patients who had a score ≥6. These patients were more likely to have an unplanned readmission (unadjusted odds ratio 2.84, 95% confidence interval 1.70-4.76, P 3 days (49.6% vs 36.6%, P = .0001), and were less likely to be sent home after discharge (20.8% vs 35.8%, P fall risk score after TJA is strongly associated with unplanned readmission. Application of this tool will allow hospitals to identify these patients and plan their discharge. Copyright © 2017 Elsevier Inc. All rights reserved.

  8. 22 CFR 1429.21 - Computation of time for filing papers.

    Science.gov (United States)

    2010-04-01

    ... 22 Foreign Relations 2 2010-04-01 2010-04-01 true Computation of time for filing papers. 1429.21... MISCELLANEOUS AND GENERAL REQUIREMENTS General Requirements § 1429.21 Computation of time for filing papers. In... subchapter requires the filing of any paper, such document must be received by the Board or the officer or...

  9. Time-ordered product expansions for computational stochastic system biology

    International Nuclear Information System (INIS)

    Mjolsness, Eric

    2013-01-01

    The time-ordered product framework of quantum field theory can also be used to understand salient phenomena in stochastic biochemical networks. It is used here to derive Gillespie’s stochastic simulation algorithm (SSA) for chemical reaction networks; consequently, the SSA can be interpreted in terms of Feynman diagrams. It is also used here to derive other, more general simulation and parameter-learning algorithms including simulation algorithms for networks of stochastic reaction-like processes operating on parameterized objects, and also hybrid stochastic reaction/differential equation models in which systems of ordinary differential equations evolve the parameters of objects that can also undergo stochastic reactions. Thus, the time-ordered product expansion can be used systematically to derive simulation and parameter-fitting algorithms for stochastic systems. (paper)

  10. Wake force computation in the time domain for long structures

    International Nuclear Information System (INIS)

    Bane, K.; Weiland, T.

    1983-07-01

    One is often interested in calculating the wake potentials for short bunches in long structures using TBCI. For ultra-relativistic particles it is sufficient to solve for the fields only over a window containing the bunch and moving along with it. This technique reduces both the memory and the running time required by a factor that equals the ratio of the structure length to the window length. For example, for a bunch with sigma/sub z/ of one picosecond traversing a single SLAC cell this improvement factor is 15. It is thus possible to solve for the wakefields in very long structures: for a given problem, increasing the structure length will not change the memory required while only adding linearly to the CPU time needed

  11. Real-time FPGA architectures for computer vision

    Science.gov (United States)

    Arias-Estrada, Miguel; Torres-Huitzil, Cesar

    2000-03-01

    This paper presents an architecture for real-time generic convolution of a mask and an image. The architecture is intended for fast low level image processing. The FPGA-based architecture takes advantage of the availability of registers in FPGAs to implement an efficient and compact module to process the convolutions. The architecture is designed to minimize the number of accesses to the image memory and is based on parallel modules with internal pipeline operation in order to improve its performance. The architecture is prototyped in a FPGA, but it can be implemented on a dedicated VLSI to reach higher clock frequencies. Complexity issues, FPGA resources utilization, FPGA limitations, and real time performance are discussed. Some results are presented and discussed.

  12. HOPE: Just-in-time Python compiler for astrophysical computations

    Science.gov (United States)

    Akeret, Joel; Gamper, Lukas; Amara, Adam; Refregier, Alexandre

    2014-11-01

    HOPE is a specialized Python just-in-time (JIT) compiler designed for numerical astrophysical applications. HOPE focuses on a subset of the language and is able to translate Python code into C++ while performing numerical optimization on mathematical expressions at runtime. To enable the JIT compilation, the user only needs to add a decorator to the function definition. By using HOPE, the user benefits from being able to write common numerical code in Python while getting the performance of compiled implementation.

  13. Computational micromagnetics: prediction of time dependent and thermal properties

    International Nuclear Information System (INIS)

    Schrefl, T.; Scholz, W.; Suess, Dieter; Fidler, J.

    2001-01-01

    Finite element modeling treats magnetization processes on a length scale of several nanometers and thus gives a quantitative correlation between the microstructure and the magnetic properties of ferromagnetic materials. This work presents a novel finite element/boundary element micro-magnetics solver that combines a wavelet-based matrix compression technique for magnetostatic field calculations with a BDF/GMRES method for the time integration of the Gilbert equation of motion. The simulations show that metastable energy minima and nonuniform magnetic states within the grains are important factors in the reversal dynamics at finite temperature. The numerical solution of the Gilbert equation shows how reversed domains nucleate and expand. The switching time of submicron magnetic elements depends on the shape of the elements. Elements with slanted ends decrease the overall reversal time, as a transverse demagnetizing field suppresses oscillations of the magnetization. Thermal activated processes can be included adding a random thermal field to the effective magnetic field. Thermally assisted reversal was studied for CoCrPtTa thin-film media

  14. Investigating the influence of eating habits, body weight and television programme preferences on television viewing time and domestic computer usage.

    Science.gov (United States)

    Raptou, Elena; Papastefanou, Georgios; Mattas, Konstadinos

    2017-01-01

    The present study explored the influence of eating habits, body weight and television programme preference on television viewing time and domestic computer usage, after adjusting for sociodemographic characteristics and home media environment indicators. In addition, potential substitution or complementarity in screen time was investigated. Individual level data were collected via questionnaires that were administered to a random sample of 2,946 Germans. The econometric analysis employed a seemingly unrelated bivariate ordered probit model to conjointly estimate television viewing time and time engaged in domestic computer usage. Television viewing and domestic computer usage represent two independent behaviours in both genders and across all age groups. Dietary habits have a significant impact on television watching with less healthy food choices associated with increasing television viewing time. Body weight is found to be positively correlated with television screen time in both men and women, and overweight individuals have a higher propensity for heavy television viewing. Similar results were obtained for age groups where an increasing body mass index (BMI) in adults over 24 years old is more likely to be positively associated with a higher duration of television watching. With respect to dietary habits of domestic computer users, participants aged over 24 years of both genders seem to adopt more healthy dietary patterns. A downward trend in the BMI of domestic computer users was observed in women and adults aged 25-60 years. On the contrary, young domestic computer users 18-24 years old have a higher body weight than non-users. Television programme preferences also affect television screen time with clear differences to be observed between genders and across different age groups. In order to reduce total screen time, health interventions should target different types of screen viewing audiences separately.

  15. Total body height estimation using sacrum height in Anatolian Caucasians: multidetector computed tomography-based virtual anthropometry

    Energy Technology Data Exchange (ETDEWEB)

    Karakas, Hakki Muammer [Inonu University Medical Faculty, Turgut Ozal Medical Center, Department of Radiology, Malatya (Turkey); Celbis, Osman [Inonu University Medical Faculty Turgut Ozal Medical Center, Department of Forensic Medicine, Malatya (Turkey); Harma, Ahmet [Inonu University Medical Faculty Turgut Ozal Medical Center, Department of Orthopaedics and Traumatology, Malatya (Turkey); Alicioglu, Banu [Trakya University Medical Faculty, Department of Radiology, Edirne (Turkey); Trakya University Health Sciences Institute, Department of Anatomy, Edirne (Turkey)

    2011-05-15

    Estimation of total body height is a major step when a subject has to be identified from his/her skeletal structures. In the presence of decomposed skeletons and missing bones, estimation is usually based on regression equation for intact long bones. If these bones are fragmented or missing, alternative structures must be used. In this study, the value of sacrum height (SH) in total body height (TBH) estimation was investigated in a contemporary population of adult Anatolian Caucasians. Sixty-six men (41.6 {+-} 14.9 years) and 43 women (41.1 {+-} 14.2 years) were scanned with 64-row multidetector computed tomography (MDCT) to obtain high-resolution anthropometric data. SH of midsagittal sections was electronically measured. The technique and methodology were validated on a standard skeletal model. Sacrum height was 111.2 {+-} 12.6 mm (77-138 mm) in men and 104.7 {+-} 8.2 (89-125 mm) in women. The difference between the two sexes regarding SH was significant (p < 0.0001). SH did not significantly correlate with age in men, whereas the correlation was significant in women (p < 0.03). The correlation between SH and the stature was significant in men (r = 0.427, p < 0.0001) and was insignificant in women. For men the regression equation was [Stature = (0.306 x SH)+137.9] (r = 0.54, SEE = 56.9, p < 0.0001). Sacrum height is not susceptible to sex, or to age in men. In the presence of incomplete male skeletons, SH helps to determine the stature. This study is also one of the initial applications of MDCT in virtual anthropometric research. (orig.)

  16. Total body height estimation using sacrum height in Anatolian Caucasians: multidetector computed tomography-based virtual anthropometry

    International Nuclear Information System (INIS)

    Karakas, Hakki Muammer; Celbis, Osman; Harma, Ahmet; Alicioglu, Banu

    2011-01-01

    Estimation of total body height is a major step when a subject has to be identified from his/her skeletal structures. In the presence of decomposed skeletons and missing bones, estimation is usually based on regression equation for intact long bones. If these bones are fragmented or missing, alternative structures must be used. In this study, the value of sacrum height (SH) in total body height (TBH) estimation was investigated in a contemporary population of adult Anatolian Caucasians. Sixty-six men (41.6 ± 14.9 years) and 43 women (41.1 ± 14.2 years) were scanned with 64-row multidetector computed tomography (MDCT) to obtain high-resolution anthropometric data. SH of midsagittal sections was electronically measured. The technique and methodology were validated on a standard skeletal model. Sacrum height was 111.2 ± 12.6 mm (77-138 mm) in men and 104.7 ± 8.2 (89-125 mm) in women. The difference between the two sexes regarding SH was significant (p < 0.0001). SH did not significantly correlate with age in men, whereas the correlation was significant in women (p < 0.03). The correlation between SH and the stature was significant in men (r = 0.427, p < 0.0001) and was insignificant in women. For men the regression equation was [Stature = (0.306 x SH)+137.9] (r = 0.54, SEE = 56.9, p < 0.0001). Sacrum height is not susceptible to sex, or to age in men. In the presence of incomplete male skeletons, SH helps to determine the stature. This study is also one of the initial applications of MDCT in virtual anthropometric research. (orig.)

  17. Surgical planning of total hip arthroplasty: accuracy of computer-assisted EndoMap software in predicting component size

    International Nuclear Information System (INIS)

    Davila, Jesse A.; Kransdorf, Mark J.; Duffy, Gavan P.

    2006-01-01

    The purpose of our study was to assess the accuracy of a computer-assisted templating in the surgical planning of patients undergoing total hip arthroplasty utilizing EndoMap software (Siemans AG, Medical Solutions, Erlangen, Germany). Endomap Software is an electronic program that uses DICOM images to analyze standard anteroposterior radiographs for determination of optimal prosthesis component size. We retrospectively reviewed the preoperative radiographs of 36 patients undergoing uncomplicated primary total hip arthroplasty, utilizing EndoMap software, Version VA20. DICOM anteroposterior radiographs were analyzed using standard manufacturer supplied electronic templates to determine acetabular and femoral component sizes. No additional clinical information was reviewed. Acetabular and femoral component sizes were assessed by an orthopedic surgeon and two radiologists. Mean and estimated component size was compared with component size as documented in operative reports. The mean estimated acetabular component size was 53 mm (range 48-60 mm), 1 mm larger than the mean implanted size of 52 mm (range 48-62 mm). Thirty-one of 36 acetabular component sizes (86%) were accurate within one size. The mean calculated femoral component size was 4 (range 2-7), 1 size smaller than the actual mean component size of 5 (range 2-9). Twenty-six of 36 femoral component sizes (72%) were accurate within one size, and accurate within two sizes in all but four cases (94%). EndoMap Software predicted femoral component size well, with 72% within one component size of that used, and 94% within two sizes. Acetabular component size was predicted slightly better with 86% within one component size and 94% within two component sizes. (orig.)

  18. Assessing total nitrogen in surface-water samples--precision and bias of analytical and computational methods

    Science.gov (United States)

    Rus, David L.; Patton, Charles J.; Mueller, David K.; Crawford, Charles G.

    2013-01-01

    The characterization of total-nitrogen (TN) concentrations is an important component of many surface-water-quality programs. However, three widely used methods for the determination of total nitrogen—(1) derived from the alkaline-persulfate digestion of whole-water samples (TN-A); (2) calculated as the sum of total Kjeldahl nitrogen and dissolved nitrate plus nitrite (TN-K); and (3) calculated as the sum of dissolved nitrogen and particulate nitrogen (TN-C)—all include inherent limitations. A digestion process is intended to convert multiple species of nitrogen that are present in the sample into one measureable species, but this process may introduce bias. TN-A results can be negatively biased in the presence of suspended sediment, and TN-K data can be positively biased in the presence of elevated nitrate because some nitrate is reduced to ammonia and is therefore counted twice in the computation of total nitrogen. Furthermore, TN-C may not be subject to bias but is comparatively imprecise. In this study, the effects of suspended-sediment and nitrate concentrations on the performance of these TN methods were assessed using synthetic samples developed in a laboratory as well as a series of stream samples. A 2007 laboratory experiment measured TN-A and TN-K in nutrient-fortified solutions that had been mixed with varying amounts of sediment-reference materials. This experiment identified a connection between suspended sediment and negative bias in TN-A and detected positive bias in TN-K in the presence of elevated nitrate. A 2009–10 synoptic-field study used samples from 77 stream-sampling sites to confirm that these biases were present in the field samples and evaluated the precision and bias of TN methods. The precision of TN-C and TN-K depended on the precision and relative amounts of the TN-component species used in their respective TN computations. Particulate nitrogen had an average variability (as determined by the relative standard deviation) of 13

  19. An Enhanced Discrete Artificial Bee Colony Algorithm to Minimize the Total Flow Time in Permutation Flow Shop Scheduling with Limited Buffers

    Directory of Open Access Journals (Sweden)

    Guanlong Deng

    2016-01-01

    Full Text Available This paper presents an enhanced discrete artificial bee colony algorithm for minimizing the total flow time in the flow shop scheduling problem with buffer capacity. First, the solution in the algorithm is represented as discrete job permutation to directly convert to active schedule. Then, we present a simple and effective scheme called best insertion for the employed bee and onlooker bee and introduce a combined local search exploring both insertion and swap neighborhood. To validate the performance of the presented algorithm, a computational campaign is carried out on the Taillard benchmark instances, and computations and comparisons show that the proposed algorithm is not only capable of solving the benchmark set better than the existing discrete differential evolution algorithm and iterated greedy algorithm, but also capable of performing better than two recently proposed discrete artificial bee colony algorithms.

  20. A close-form solution to predict the total melting time of an ablating slab in contact with a plasma

    International Nuclear Information System (INIS)

    Yeh, F.-B.

    2007-01-01

    An exact melt-through time is derived for a one-dimensional heated slab in contact with a plasma when the melted material is immediately removed. The plasma is composed of a collisionless presheath and sheath on a slab, which partially reflects and secondarily emits ions and electrons. The energy transport from plasma to the surface accounting for the presheath and sheath is determined from the kinetic analysis. This work proposes a semi-analytical model to calculate the total melting time of a slab based on a direct integration of the unsteady heat conduction equation, and provides quantitative results applicable to control the total melting time of the slab. The total melting time as a function of plasma parameters and thermophysical properties of the slab are obtained. The predicted energy transmission factor as a function of dimensionless wall potential agrees well with the experimental data. The effects of reflectivities of the ions and electrons on the wall, electron-to-ion source temperature ratio at the presheath edge, charge number, ion-to-electron mass ratio, ionization energy, plasma flow work-to-heat conduction ratios, Stefan number, melting temperature, Biot number and bias voltage on the total melting time of the slab are quantitatively provided in this work

  1. Does the brake response time of the right leg change after left total knee arthroplasty? A prospective study.

    Science.gov (United States)

    Marques, Carlos J; Barreiros, João; Cabri, Jan; Carita, Ana I; Friesecke, Christian; Loehr, Jochen F

    2008-08-01

    Patients undergoing total knee arthroplasty often ask when they can safely resume car driving. There is little evidence available on which physicians can rely when advising patients on this issue. In a prospective study we assessed the brake response time of 24 patients admitted to the clinic for left total knee arthroplasty preoperatively and then 10 days after surgery. On each measurement day the patients performed two tasks, a simple and a complex brake response time task in a car simulator. Ten days after left TKA the brake response time for the simple task had decreased by 3.6% (p=0.24), the reaction time by 3.1% (p=0.34) and the movement time by 6.6% (p=0.07). However, the performance improvement was not statistically significant. Task complexity increased brake response time at both time points. A 5.8% increase was significant (p=0.01) at 10 days after surgery. Based on our results, we suggest that patients who have undergone left total knee arthroplasty may resume car driving 10 days after surgery as long as they drive a car with automatic transmission.

  2. A Comparison of Plain Radiography with Computer Tomography in Determining Coronal and Sagittal Alignments following Total Knee Arthroplasty

    Directory of Open Access Journals (Sweden)

    Solayar GN

    2017-07-01

    Full Text Available INTRODUCTION: Optimal coronal and sagittal component positioning is important in achieving a successful outcome following total knee arthroplasty (TKA. Modalities to determine post-operative alignment include plain radiography and computer tomography (CT imaging. This study aims to determine the accuracy and reliability of plain radiographs in measuring coronal and sagittal alignment following TKA. MATERIALS AND METHODS: A prospective, consecutive study of 58 patients undergoing TKA was performed comparing alignment data from plain radiographs and CT imaging. Hip- knee-angle (HKA, sagittal femoral angle (SFA and sagittal tibial angle (STA measurements were taken by two observers from plain radiographs and compared with CT alignment. Intra- and inter-observer correlation was calculated for each measurement. RESULTS: Intra-observer correlation was excellent for HKA (r>0.89 with a mean difference of 0.95 and STA (r>0.8 compared to SFA (r=0.5. When comparing modalities (radiographs vs CT, HKA estimations for both observers showed the least maximum and mean differences while SFA observations were the least accurate. CONCLUSION: Radiographic estimation of HKA showed excellent intra- and inter-observer correlation and corresponds well with CT imaging. However, radiographic estimation of sagittal plane alignment was less reliably measured and correlated less with CT imaging. Plain radiography was found to be inferior to CT for estimation of biplanar prosthetic alignment following TKA.

  3. Computer-assisted surgery simulations and directed practice of total knee arthroplasty: educational benefits to the trainee.

    Science.gov (United States)

    Myden, C A; Anglin, C; Kopp, G D; Hutchison, C R

    2012-01-01

    Orthopaedic residents typically learn to perform total knee arthroplasty (TKA) through an apprenticeship-type model, which is a necessarily slow process. Surgical skills courses, using artificial bones, have been shown to improve technical and cognitive skills significantly within a couple of days. The addition of computer-assisted surgery (CAS) simulations challenges the participants to consider the same task in a different context, promoting cognitive flexibility. We designed a hands-on educational intervention for junior residents with a conventional tibiofemoral TKA station, two different tibiofemoral CAS stations, and a CAS and conventional patellar resection station, including both qualitative and quantitative analyses. Qualitatively, structured interviews before and after the course were analyzed for recurring themes. Quantitatively, subjects were evaluated on their technical skills before and after the course, and on a multiple-choice knowledge test and error detection test after the course, in comparison to senior residents who performed only the testing. Four themes emerged: confidence, awareness, deepening knowledge and changed perspectives. The residents' attitudes to CAS changed from negative before the course to neutral or positive afterwards. The junior resident group completed 23% of tasks in the pre-course skills test and 75% of tasks on the post-test (peducational interventions, promoting cognitive flexibility, would benefit trainees, attending surgeons, the healthcare system and patients.

  4. Case-related factors affecting cutting errors of the proximal tibia in total knee arthroplasty assessed by computer navigation.

    Science.gov (United States)

    Tsukeoka, Tadashi; Tsuneizumi, Yoshikazu; Yoshino, Kensuke; Suzuki, Mashiko

    2018-05-01

    The aim of this study was to determine factors that contribute to bone cutting errors of conventional instrumentation for tibial resection in total knee arthroplasty (TKA) as assessed by an image-free navigation system. The hypothesis is that preoperative varus alignment is a significant contributory factor to tibial bone cutting errors. This was a prospective study of a consecutive series of 72 TKAs. The amount of the tibial first-cut errors with reference to the planned cutting plane in both coronal and sagittal planes was measured by an image-free computer navigation system. Multiple regression models were developed with the amount of tibial cutting error in the coronal and sagittal planes as dependent variables and sex, age, disease, height, body mass index, preoperative alignment, patellar height (Insall-Salvati ratio) and preoperative flexion angle as independent variables. Multiple regression analysis showed that sex (male gender) (R = 0.25 p = 0.047) and preoperative varus alignment (R = 0.42, p = 0.001) were positively associated with varus tibial cutting errors in the coronal plane. In the sagittal plane, none of the independent variables was significant. When performing TKA in varus deformity, careful confirmation of the bone cutting surface should be performed to avoid varus alignment. The results of this study suggest technical considerations that can help a surgeon achieve more accurate component placement. IV.

  5. Applications of parallel computer architectures to the real-time simulation of nuclear power systems

    International Nuclear Information System (INIS)

    Doster, J.M.; Sills, E.D.

    1988-01-01

    In this paper the authors report on efforts to utilize parallel computer architectures for the thermal-hydraulic simulation of nuclear power systems and current research efforts toward the development of advanced reactor operator aids and control systems based on this new technology. Many aspects of reactor thermal-hydraulic calculations are inherently parallel, and the computationally intensive portions of these calculations can be effectively implemented on modern computers. Timing studies indicate faster-than-real-time, high-fidelity physics models can be developed when the computational algorithms are designed to take advantage of the computer's architecture. These capabilities allow for the development of novel control systems and advanced reactor operator aids. Coupled with an integral real-time data acquisition system, evolving parallel computer architectures can provide operators and control room designers improved control and protection capabilities. Current research efforts are currently under way in this area

  6. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction CMS distributed computing system performed well during the 2011 start-up. The events in 2011 have more pile-up and are more complex than last year; this results in longer reconstruction times and harder events to simulate. Significant increases in computing capacity were delivered in April for all computing tiers, and the utilisation and load is close to the planning predictions. All computing centre tiers performed their expected functionalities. Heavy-Ion Programme The CMS Heavy-Ion Programme had a very strong showing at the Quark Matter conference. A large number of analyses were shown. The dedicated heavy-ion reconstruction facility at the Vanderbilt Tier-2 is still involved in some commissioning activities, but is available for processing and analysis. Facilities and Infrastructure Operations Facility and Infrastructure operations have been active with operations and several important deployment tasks. Facilities participated in the testing and deployment of WMAgent and WorkQueue+Request...

  7. [Total quality management in times of crisis. The case of Argentina].

    Science.gov (United States)

    Larroca, Norberto

    2003-01-01

    Healthcare organizations were faced with so great a challenge following the financial slump that they were forced to 'sharpen their wits' in order to survive. Integrated Quality Management (in Spanish GIC), proved the ideal instrument. GIC has four foundational elements, the application of which allow for successful management of crisis situations. They are as follows: TRAINING of human resources EVALUATION of healthcare institutions SELF-EVALUATION by institutions QUALITY ACCREDITATION of institutions All our organizations have the appropriate tools to carry out these activities which form the basis of our project: CAES (Argentinean Chamber of Healthcare Institutions) -training-, CIDCAM (Inter-institutional Committee for Quality Development in Medical Care) -evaluation and self-evaluation-, CENAS (Specialist Centre for the Standardization and Accreditation in Health Care)-accreditation-. In times of crisis, we play an active part, that is, instead of withdrawing our efforts, we do our best to achieve the best and most adequate objective in order to meet the needs of the population through Integrated Quality Management. Eventually, when the results are examined, medical care that meets the best quality standards is found to be, after all, the most economical (that is best results, better satisfaction of healthcare users and providers as well as less mistakes).

  8. Total testosterone levels are often more than three times elevated in patients with androgen-secreting tumours

    DEFF Research Database (Denmark)

    Glintborg, Dorte; Lambaa Altinok, Magda; Petersen, Kresten Rubeck

    2015-01-01

    surgery. Terminal hair growth on lip and chin gradually increases after menopause, which complicates distinction from normal physiological variation. Precise testosterone assays have just recently become available in the daily clinic. We present three women diagnosed with testosterone-producing tumours...... when total testosterone levels are above three times the upper reference limit....

  9. Balancing Exploration, Uncertainty Representation and Computational Time in Many-Objective Reservoir Policy Optimization

    Science.gov (United States)

    Zatarain-Salazar, J.; Reed, P. M.; Quinn, J.; Giuliani, M.; Castelletti, A.

    2016-12-01

    As we confront the challenges of managing river basin systems with a large number of reservoirs and increasingly uncertain tradeoffs impacting their operations (due to, e.g. climate change, changing energy markets, population pressures, ecosystem services, etc.), evolutionary many-objective direct policy search (EMODPS) solution strategies will need to address the computational demands associated with simulating more uncertainties and therefore optimizing over increasingly noisy objective evaluations. Diagnostic assessments of state-of-the-art many-objective evolutionary algorithms (MOEAs) to support EMODPS have highlighted that search time (or number of function evaluations) and auto-adaptive search are key features for successful optimization. Furthermore, auto-adaptive MOEA search operators are themselves sensitive to having a sufficient number of function evaluations to learn successful strategies for exploring complex spaces and for escaping from local optima when stagnation is detected. Fortunately, recent parallel developments allow coordinated runs that enhance auto-adaptive algorithmic learning and can handle scalable and reliable search with limited wall-clock time, but at the expense of the total number of function evaluations. In this study, we analyze this tradeoff between parallel coordination and depth of search using different parallelization schemes of the Multi-Master Borg on a many-objective stochastic control problem. We also consider the tradeoff between better representing uncertainty in the stochastic optimization, and simplifying this representation to shorten the function evaluation time and allow for greater search. Our analysis focuses on the Lower Susquehanna River Basin (LSRB) system where multiple competing objectives for hydropower production, urban water supply, recreation and environmental flows need to be balanced. Our results provide guidance for balancing exploration, uncertainty, and computational demands when using the EMODPS

  10. Computation of a long-time evolution in a Schroedinger system

    International Nuclear Information System (INIS)

    Girard, R.; Kroeger, H.; Labelle, P.; Bajzer, Z.

    1988-01-01

    We compare different techniques for the computation of a long-time evolution and the S matrix in a Schroedinger system. As an application we consider a two-nucleon system interacting via the Yamaguchi potential. We suggest computation of the time evolution for a very short time using Pade approximants, the long-time evolution being obtained by iterative squaring. Within the technique of strong approximation of Moller wave operators (SAM) we compare our calculation with computation of the time evolution in the eigenrepresentation of the Hamiltonian and with the standard Lippmann-Schwinger solution for the S matrix. We find numerical agreement between these alternative methods for time-evolution computation up to half the number of digits of internal machine precision, and fairly rapid convergence of both techniques towards the Lippmann-Schwinger solution

  11. Independent and combined associations of total sedentary time and television viewing time with food intake patterns of 9- to 11-year-old Canadian children.

    Science.gov (United States)

    Borghese, Michael M; Tremblay, Mark S; Leduc, Genevieve; Boyer, Charles; Bélanger, Priscilla; LeBlanc, Allana G; Francis, Claire; Chaput, Jean-Philippe

    2014-08-01

    The relationships among sedentary time, television viewing time, and dietary patterns in children are not fully understood. The aim of this paper was to determine which of self-reported television viewing time or objectively measured sedentary time is a better correlate of the frequency of consumption of healthy and unhealthy foods. A cross-sectional study was conducted of 9- to 11-year-old children (n = 523; 57.1% female) from Ottawa, Ontario, Canada. Accelerometers were used to determine total sedentary time, and questionnaires were used to determine the number of hours of television watching and the frequency of consumption of foods per week. Television viewing was negatively associated with the frequency of consumption of fruits, vegetables, and green vegetables, and positively associated with the frequency of consumption of sweets, soft drinks, diet soft drinks, pastries, potato chips, French fries, fruit juices, ice cream, fried foods, and fast food. Except for diet soft drinks and fruit juices, these associations were independent of covariates, including sedentary time. Total sedentary time was negatively associated with the frequency of consumption of sports drinks, independent of covariates, including television viewing. In combined sedentary time and television viewing analyses, children watching >2 h of television per day consumed several unhealthy food items more frequently than did children watching ≤2 h of television, regardless of sedentary time. In conclusion, this paper provides evidence to suggest that television viewing time is more strongly associated with unhealthy dietary patterns than is total sedentary time. Future research should focus on reducing television viewing time, as a means of improving dietary patterns and potentially reducing childhood obesity.

  12. The relationship between TV/computer time and adolescents' health-promoting behavior: a secondary data analysis.

    Science.gov (United States)

    Chen, Mei-Yen; Liou, Yiing-Mei; Wu, Jen-Yee

    2008-03-01

    Television and computers provide significant benefits for learning about the world. Some studies have linked excessive television (TV) watching or computer game playing to disadvantage of health status or some unhealthy behavior among adolescents. However, the relationships between watching TV/playing computer games and adolescents adopting health promoting behavior were limited. This study aimed to discover the relationship between time spent on watching TV and on leisure use of computers and adolescents' health promoting behavior, and associated factors. This paper used secondary data analysis from part of a health promotion project in Taoyuan County, Taiwan. A cross-sectional design was used and purposive sampling was conducted among adolescents in the original project. A total of 660 participants answered the questions appropriately for this work between January and June 2004. Findings showed the mean age of the respondents was 15.0 +/- 1.7 years. The mean numbers of TV watching hours were 2.28 and 4.07 on weekdays and weekends respectively. The mean hours of leisure (non-academic) computer use were 1.64 and 3.38 on weekdays and weekends respectively. Results indicated that adolescents spent significant time watching TV and using the computer, which was negatively associated with adopting health-promoting behaviors such as life appreciation, health responsibility, social support and exercise behavior. Moreover, being boys, being overweight, living in a rural area, and being middle-school students were significantly associated with spending long periods watching TV and using the computer. Therefore, primary health care providers should record the TV and non-academic computer time of youths when conducting health promotion programs, and educate parents on how to become good and healthy electronic media users.

  13. LHC Computing Grid Project Launches intAction with International Support. A thousand times more computing power by 2006

    CERN Multimedia

    2001-01-01

    The first phase of the LHC Computing Grid project was approved at an extraordinary meeting of the Council on 20 September 2001. CERN is preparing for the unprecedented avalanche of data that will be produced by the Large Hadron Collider experiments. A thousand times more computer power will be needed by 2006! CERN's need for a dramatic advance in computing capacity is urgent. As from 2006, the four giant detectors observing trillions of elementary particle collisions at the LHC will accumulate over ten million Gigabytes of data, equivalent to the contents of about 20 million CD-ROMs, each year of its operation. A thousand times more computing power will be needed than is available to CERN today. The strategy the collabortations have adopted to analyse and store this unprecedented amount of data is the coordinated deployment of Grid technologies at hundreds of institutes which will be able to search out and analyse information from an interconnected worldwide grid of tens of thousands of computers and storag...

  14. Influence of different maceration time and temperatures on total phenols, colour and sensory properties of Cabernet Sauvignon wines.

    Science.gov (United States)

    Şener, Hasan; Yildirim, Hatice Kalkan

    2013-12-01

    Maceration and fermentation time and temperatures are important factors affecting wine quality. In this study different maceration times (3 and 6 days) and temperatures (15  and 25 ) during production of red wine (Vitis vinifera L. Cabernet Sauvignon) were investigated. In all wines standard wine chemical parameters and some specific parameters as total phenols, tartaric esters, total flavonols and colour parameters (CD, CI, T, dA%, %Y, %R, %B, CIELAB values) were determined. Sensory evaluation was performed by descriptive sensory analysis. The results demonstrated not only the importance of skin contact time and temperature during maceration but also the effects of transition temperatures (different maceration and fermentation temperatures) on wine quality as a whole. The results of sensory descriptive analyses revealed that the temperature significantly affected the aroma and flavour attributes of wines. The highest scores for 'cassis', 'clove', 'fresh fruity' and 'rose' characters were obtained in wines produced at low temperature (15 ) of maceration (6 days) and fermentation.

  15. Brake response time is significantly impaired after total knee arthroplasty: investigation of performing an emergency stop while driving a car.

    Science.gov (United States)

    Jordan, Maurice; Hofmann, Ulf-Krister; Rondak, Ina; Götze, Marco; Kluba, Torsten; Ipach, Ingmar

    2015-09-01

    The objective of this study was to investigate whether total knee arthroplasty (TKA) impairs the ability to perform an emergency stop. An automatic transmission brake simulator was developed to evaluate total brake response time. A prospective repeated-measures design was used. Forty patients (20 left/20 right) were measured 8 days and 6, 12, and 52 wks after surgery. Eight days postoperative total brake response time increased significantly by 30% in right TKA and insignificantly by 2% in left TKA. Brake force significantly decreased by 35% in right TKA and by 25% in left TKA during this period. Baseline values were reached at week 12 in right TKA; the impairment of outcome measures, however, was no longer significant at week 6 compared with preoperative values. Total brake response time and brake force in left TKA fell below baseline values at weeks 6 and 12. Brake force in left TKA was the only outcome measure significantly impaired 8 days postoperatively. This study highlights that categorical statements cannot be provided. This study's findings on automatic transmission driving suggest that right TKA patients may resume driving 6 wks postoperatively. Fitness to drive in left TKA is not fully recovered 8 days postoperatively. If testing is not available, patients should refrain from driving until they return from rehabilitation.

  16. Multiscale Space-Time Computational Methods for Fluid-Structure Interactions

    Science.gov (United States)

    2015-09-13

    thermo-fluid analysis of a ground vehicle and its tires ST-SI Computational Analysis of a Vertical - Axis Wind Turbine We have successfully...of a vertical - axis wind turbine . Multiscale Compressible-Flow Computation with Particle Tracking We have successfully tested the multiscale...Tezduyar, Spenser McIntyre, Nikolay Kostov, Ryan Kolesar, Casey Habluetzel. Space–time VMS computation of wind - turbine rotor and tower aerodynamics

  17. A Non-Linear Digital Computer Model Requiring Short Computation Time for Studies Concerning the Hydrodynamics of the BWR

    Energy Technology Data Exchange (ETDEWEB)

    Reisch, F; Vayssier, G

    1969-05-15

    This non-linear model serves as one of the blocks in a series of codes to study the transient behaviour of BWR or PWR type reactors. This program is intended to be the hydrodynamic part of the BWR core representation or the hydrodynamic part of the PWR heat exchanger secondary side representation. The equations have been prepared for the CSMP digital simulation language. By using the most suitable integration routine available, the ratio of simulation time to real time is about one on an IBM 360/75 digital computer. Use of the slightly different language DSL/40 on an IBM 7044 computer takes about four times longer. The code has been tested against the Eindhoven loop with satisfactory agreement.

  18. A Modular Environment for Geophysical Inversion and Run-time Autotuning using Heterogeneous Computing Systems

    Science.gov (United States)

    Myre, Joseph M.

    Heterogeneous computing systems have recently come to the forefront of the High-Performance Computing (HPC) community's interest. HPC computer systems that incorporate special purpose accelerators, such as Graphics Processing Units (GPUs), are said to be heterogeneous. Large scale heterogeneous computing systems have consistently ranked highly on the Top500 list since the beginning of the heterogeneous computing trend. By using heterogeneous computing systems that consist of both general purpose processors and special- purpose accelerators, the speed and problem size of many simulations could be dramatically increased. Ultimately this results in enhanced simulation capabilities that allows, in some cases for the first time, the execution of parameter space and uncertainty analyses, model optimizations, and other inverse modeling techniques that are critical for scientific discovery and engineering analysis. However, simplifying the usage and optimization of codes for heterogeneous computing systems remains a challenge. This is particularly true for scientists and engineers for whom understanding HPC architectures and undertaking performance analysis may not be primary research objectives. To enable scientists and engineers to remain focused on their primary research objectives, a modular environment for geophysical inversion and run-time autotuning on heterogeneous computing systems is presented. This environment is composed of three major components: 1) CUSH---a framework for reducing the complexity of programming heterogeneous computer systems, 2) geophysical inversion routines which can be used to characterize physical systems, and 3) run-time autotuning routines designed to determine configurations of heterogeneous computing systems in an attempt to maximize the performance of scientific and engineering codes. Using three case studies, a lattice-Boltzmann method, a non-negative least squares inversion, and a finite-difference fluid flow method, it is shown that

  19. Real Time Animation of Trees Based on BBSC in Computer Games

    Directory of Open Access Journals (Sweden)

    Xuefeng Ao

    2009-01-01

    Full Text Available That researchers in the field of computer games usually find it is difficult to simulate the motion of actual 3D model trees lies in the fact that the tree model itself has very complicated structure, and many sophisticated factors need to be considered during the simulation. Though there are some works on simulating 3D tree and its motion, few of them are used in computer games due to the high demand for real-time in computer games. In this paper, an approach of animating trees in computer games based on a novel tree model representation—Ball B-Spline Curves (BBSCs are proposed. By taking advantage of the good features of the BBSC-based model, physical simulation of the motion of leafless trees with wind blowing becomes easier and more efficient. The method can generate realistic 3D tree animation in real-time, which meets the high requirement for real time in computer games.

  20. Ubiquitous computing technology for just-in-time motivation of behavior change.

    Science.gov (United States)

    Intille, Stephen S

    2004-01-01

    This paper describes a vision of health care where "just-in-time" user interfaces are used to transform people from passive to active consumers of health care. Systems that use computational pattern recognition to detect points of decision, behavior, or consequences automatically can present motivational messages to encourage healthy behavior at just the right time. Further, new ubiquitous computing and mobile computing devices permit information to be conveyed to users at just the right place. In combination, computer systems that present messages at the right time and place can be developed to motivate physical activity and healthy eating. Computational sensing technologies can also be used to measure the impact of the motivational technology on behavior.

  1. Online Operation Guidance of Computer System Used in Real-Time Distance Education Environment

    Science.gov (United States)

    He, Aiguo

    2011-01-01

    Computer system is useful for improving real time and interactive distance education activities. Especially in the case that a large number of students participate in one distance lecture together and every student uses their own computer to share teaching materials or control discussions over the virtual classrooms. The problem is that within…

  2. 78 FR 38949 - Computer Security Incident Coordination (CSIC): Providing Timely Cyber Incident Response

    Science.gov (United States)

    2013-06-28

    ... exposed to various forms of cyber attack. In some cases, attacks can be thwarted through the use of...-3383-01] Computer Security Incident Coordination (CSIC): Providing Timely Cyber Incident Response... systems will be successfully attacked. When a successful attack occurs, the job of a Computer Security...

  3. Hydrologic Response to Climate Change: Missing Precipitation Data Matters for Computed Timing Trends

    Science.gov (United States)

    Daniels, B.

    2016-12-01

    This work demonstrates the derivation of climate timing statistics and applying them to determine resulting hydroclimate impacts. Long-term daily precipitation observations from 50 California stations were used to compute climate trends of precipitation event Intensity, event Duration and Pause between events. Each precipitation event trend was then applied as input to a PRMS hydrology model which showed hydrology changes to recharge, baseflow, streamflow, etc. An important concern was precipitation uncertainty induced by missing observation values and causing errors in quantification of precipitation trends. Many standard statistical techniques such as ARIMA and simple endogenous or even exogenous imputation were applied but failed to help resolve these uncertainties. What helped resolve these uncertainties was use of multiple imputation techniques. This involved fitting of Weibull probability distributions to multiple imputed values for the three precipitation trends.Permutation resampling techniques using Monte Carlo processing were then applied to the multiple imputation values to derive significance p-values for each trend. Significance at the 95% level for Intensity was found for 11 of the 50 stations, Duration from 16 of the 50, and Pause from 19, of which 12 were 99% significant. The significance weighted trends for California are Intensity -4.61% per decade, Duration +3.49% per decade, and Pause +3.58% per decade. Two California basins with PRMS hydrologic models were studied: Feather River in the northern Sierra Nevada mountains and the central coast Soquel-Aptos. Each local trend was changed without changing the other trends or the total precipitation. Feather River Basin's critical supply to Lake Oroville and the State Water Project benefited from a total streamflow increase of 1.5%. The Soquel-Aptos Basin water supply was impacted by a total groundwater recharge decrease of -7.5% and streamflow decrease of -3.2%.

  4. Terahertz time-domain attenuated total reflection spectroscopy applied to the rapid discrimination of the botanical origin of honeys

    Science.gov (United States)

    Liu, Wen; Zhang, Yuying; Yang, Si; Han, Donghai

    2018-05-01

    A new technique to identify the floral resources of honeys is demanded. Terahertz time-domain attenuated total reflection spectroscopy combined with chemometrics methods was applied to discriminate different categorizes (Medlar honey, Vitex honey, and Acacia honey). Principal component analysis (PCA), cluster analysis (CA) and partial least squares-discriminant analysis (PLS-DA) have been used to find information of the botanical origins of honeys. Spectral range also was discussed to increase the precision of PLS-DA model. The accuracy of 88.46% for validation set was obtained, using PLS-DA model in 0.5-1.5 THz. This work indicated terahertz time-domain attenuated total reflection spectroscopy was an available approach to evaluate the quality of honey rapidly.

  5. Computer tomography urography assisted real-time ultrasound-guided percutaneous nephrolithotomy on renal calculus.

    Science.gov (United States)

    Fang, You-Qiang; Wu, Jie-Ying; Li, Teng-Cheng; Zheng, Hao-Feng; Liang, Guan-Can; Chen, Yan-Xiong; Hong, Xiao-Bin; Cai, Wei-Zhong; Zang, Zhi-Jun; Di, Jin-Ming

    2017-06-01

    This study aimed to assess the role of pre-designed route on computer tomography urography (CTU) in the ultrasound-guided percutaneous nephrolithotomy (PCNL) for renal calculus.From August 2013 to May 2016, a total of 100 patients diagnosed with complex renal calculus in our hospital were randomly divided into CTU group and control group (without CTU assistance). CTU was used to design a rational route for puncturing in CTU group. Ultrasound was used in both groups to establish a working trace in the operation areas. Patients' perioperative parameters and postoperative complications were recorded.All operations were successfully performed, without transferring to open surgery. Time of channel establishment in CTU group (6.5 ± 4.3 minutes) was shorter than the control group (10.0 ± 6.7 minutes) (P = .002). In addition, there was shorter operation time, lower rates of blood transfusion, secondary operation, and less establishing channels. The incidence of postoperative complications including residual stones, sepsis, severe hemorrhage, and perirenal hematoma was lower in CTU group than in control group.Pre-designing puncture route on CTU images would improve the puncturing accuracy, lessen establishing channels as well as improve the security in the ultrasound-guided PCNL for complex renal calculus, but at the cost of increased radiation exposure.

  6. Application verification research of cloud computing technology in the field of real time aerospace experiment

    Science.gov (United States)

    Wan, Junwei; Chen, Hongyan; Zhao, Jing

    2017-08-01

    According to the requirements of real-time, reliability and safety for aerospace experiment, the single center cloud computing technology application verification platform is constructed. At the IAAS level, the feasibility of the cloud computing technology be applied to the field of aerospace experiment is tested and verified. Based on the analysis of the test results, a preliminary conclusion is obtained: Cloud computing platform can be applied to the aerospace experiment computing intensive business. For I/O intensive business, it is recommended to use the traditional physical machine.

  7. COMPUTING

    CERN Multimedia

    I. Fisk

    2013-01-01

    Computing activity had ramped down after the completion of the reprocessing of the 2012 data and parked data, but is increasing with new simulation samples for analysis and upgrade studies. Much of the Computing effort is currently involved in activities to improve the computing system in preparation for 2015. Operations Office Since the beginning of 2013, the Computing Operations team successfully re-processed the 2012 data in record time, not only by using opportunistic resources like the San Diego Supercomputer Center which was accessible, to re-process the primary datasets HTMHT and MultiJet in Run2012D much earlier than planned. The Heavy-Ion data-taking period was successfully concluded in February collecting almost 500 T. Figure 3: Number of events per month (data) In LS1, our emphasis is to increase efficiency and flexibility of the infrastructure and operation. Computing Operations is working on separating disk and tape at the Tier-1 sites and the full implementation of the xrootd federation ...

  8. Total and segmental colon transit time in constipated children assessed by scintigraphy with 111In-DTPA given orally.

    Science.gov (United States)

    Vattimo, A; Burroni, L; Bertelli, P; Messina, M; Meucci, D; Tota, G

    1993-12-01

    Serial colon scintigraphy using 111In-DTPA (2 MBq) given orally was performed in 39 children referred for constipation, and the total and segmental colon transit times were measured. The bowel movements during the study were recorded and the intervals between defecations (ID) were calculated. This method proved able to identify children with normal colon morphology (no. = 32) and those with dolichocolon (no. = 7). Normal children were not included for ethical reasons and we used the normal range determined by others using x-ray methods (29 +/- 4 hours). Total and segmental colon transit times were found to be prolonged in all children with dolichocolon (TC: 113.55 +/- 41.20 hours; RC: 39.85 +/- 26.39 hours; LC: 43.05 +/- 18.30 hours; RS: 30.66 +/- 26.89 hours). In the group of children with a normal colon shape, 13 presented total and segmental colon transit times within the referred normal value (TC: 27.79 +/- 4.10 hours; RC: 9.11 +/- 2.53 hours; LC: 9.80 +/- 3.50 hours; RS: 8.88 +/- 4.09 hours) and normal bowel function (ID: 23.37 +/- 5.93 hours). In the remaining children, 5 presented prolonged retention in the rectum (RS: 53.36 +/- 29.66 hours), and 14 a prolonged transit time in all segments. A good correlation was found between the transit time and bowel function. From the point of view of radiation dosimetry, the most heavily irradiated organs were the lower large intestine and the ovaries, and the level of radiation burden depended on the colon transit time. We can conclude that the described method results safe, accurate and fully diagnostic.

  9. Time Domain Terahertz Axial Computed Tomography Non Destructive Evaluation, Phase I

    Data.gov (United States)

    National Aeronautics and Space Administration — We propose to demonstrate key elements of feasibility for a high speed automated time domain terahertz computed axial tomography (TD-THz CT) non destructive...

  10. Time Domain Terahertz Axial Computed Tomography Non Destructive Evaluation, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — In this Phase 2 project, we propose to develop, construct, and deliver to NASA a computed axial tomography time-domain terahertz (CT TD-THz) non destructive...

  11. Objectively measured physical environmental neighbourhood factors are not associated with accelerometer-determined total sedentary time in adults

    OpenAIRE

    Compernolle, Sofie; De Cocker, Katrien; Mackenbach, Joreintje D.; Van Nassau, Femke; Lakerveld, Jeroen; Cardon, Greet; De Bourdeaudhuij, Ilse

    2017-01-01

    Background: The physical neighbourhood environment may influence adults' sedentary behaviour. Yet, most studies examining the association between the physical neighbourhood environment and sedentary behaviour rely on self-reported data of either the physical neighbourhood environment and/or sedentary behaviour. The aim of this study was to investigate the associations between objectively measured physical environmental neighbourhood factors and accelerometer-determined total sedentary time in...

  12. Decreasing Transition Times in Elementary School Classrooms: Using Computer-Assisted Instruction to Automate Intervention Components

    Science.gov (United States)

    Hine, Jeffrey F.; Ardoin, Scott P.; Foster, Tori E.

    2015-01-01

    Research suggests that students spend a substantial amount of time transitioning between classroom activities, which may reduce time spent academically engaged. This study used an ABAB design to evaluate the effects of a computer-assisted intervention that automated intervention components previously shown to decrease transition times. We examined…

  13. Clinical responses after total body irradiation by over permissible dose of γ-rays in one time

    International Nuclear Information System (INIS)

    Jiang Benrong; Wang Guilin; Liu Huilan; Tang Xingsheng; Ai Huisheng

    1990-01-01

    The clinical responses of patients after total body over permissilbe dose γ-ray irradiation were observed and analysed. The results showed: when the dose was above 5 cGy, there was some immunological depression, but no significant change in hematopoietic functions. 5 cases showed some transient changes of ECG, perhaps due to vagotonia caused by psychological imbalance, One case vomitted 3-4 times after 28 cGy irradiation, this suggested that a few times of vomitting had no significance in the estimation of the irradiated dose and the whole clinical manifestations must be concretely analysed

  14. COMPUTING

    CERN Multimedia

    I. Fisk

    2010-01-01

    Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion An activity that is still in progress is computing for the heavy-ion program. The heavy-ion events are collected without zero suppression, so the event size is much large at roughly 11 MB per event of RAW. The central collisions are more complex and...

  15. CROSAT: A digital computer program for statistical-spectral analysis of two discrete time series

    International Nuclear Information System (INIS)

    Antonopoulos Domis, M.

    1978-03-01

    The program CROSAT computes directly from two discrete time series auto- and cross-spectra, transfer and coherence functions, using a Fast Fourier Transform subroutine. Statistical analysis of the time series is optional. While of general use the program is constructed to be immediately compatible with the ICL 4-70 and H316 computers at AEE Winfrith, and perhaps with minor modifications, with any other hardware system. (author)

  16. Hard Real-Time Task Scheduling in Cloud Computing Using an Adaptive Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Amjad Mahmood

    2017-04-01

    Full Text Available In the Infrastructure-as-a-Service cloud computing model, virtualized computing resources in the form of virtual machines are provided over the Internet. A user can rent an arbitrary number of computing resources to meet their requirements, making cloud computing an attractive choice for executing real-time tasks. Economical task allocation and scheduling on a set of leased virtual machines is an important problem in the cloud computing environment. This paper proposes a greedy and a genetic algorithm with an adaptive selection of suitable crossover and mutation operations (named as AGA to allocate and schedule real-time tasks with precedence constraint on heterogamous virtual machines. A comprehensive simulation study has been done to evaluate the performance of the proposed algorithms in terms of their solution quality and efficiency. The simulation results show that AGA outperforms the greedy algorithm and non-adaptive genetic algorithm in terms of solution quality.

  17. First-in-Man Computed Tomography-Guided Percutaneous Revascularization of Coronary Chronic Total Occlusion Using a Wearable Computer: Proof of Concept.

    Science.gov (United States)

    Opolski, Maksymilian P; Debski, Artur; Borucki, Bartosz A; Szpak, Marcin; Staruch, Adam D; Kepka, Cezary; Witkowski, Adam

    2016-06-01

    We report a case of successful computed tomography-guided percutaneous revascularization of a chronically occluded right coronary artery using a wearable, hands-free computer with a head-mounted display worn by interventional cardiologists in the catheterization laboratory. The projection of 3-dimensional computed tomographic reconstructions onto the screen of virtual reality glass allowed the operators to clearly visualize the distal coronary vessel, and verify the direction of the guide wire advancement relative to the course of the occluded vessel segment. This case provides proof of concept that wearable computers can improve operator comfort and procedure efficiency in interventional cardiology. Copyright © 2016 Canadian Cardiovascular Society. Published by Elsevier Inc. All rights reserved.

  18. Time expenditure in computer aided time studies implemented for highly mechanized forest equipment

    Directory of Open Access Journals (Sweden)

    Elena Camelia Mușat

    2016-06-01

    Full Text Available Time studies represent important tools that are used in forest operations research to produce empirical models or to comparatively assess the performance of two or more operational alternatives with the general aim to predict the performance of operational behavior, choose the most adequate equipment or eliminate the useless time. There is a long tradition in collecting the needed data in a traditional fashion, but this approach has its limitations, and it is likely that in the future the use of professional software would be extended is such preoccupations as this kind of tools have been already implemented. However, little to no information is available in what concerns the performance of data analyzing tasks when using purpose-built professional time studying software in such research preoccupations, while the resources needed to conduct time studies, including here the time may be quite intensive. Our study aimed to model the relations between the variation of time needed to analyze the video-recorded time study data and the variation of some measured independent variables for a complex organization of a work cycle. The results of our study indicate that the number of work elements which were separated within a work cycle as well as the delay-free cycle time and the software functionalities that were used during data analysis, significantly affected the time expenditure needed to analyze the data (α=0.01, p<0.01. Under the conditions of this study, where the average duration of a work cycle was of about 48 seconds and the number of separated work elements was of about 14, the speed that was usedto replay the video files significantly affected the mean time expenditure which averaged about 273 seconds for half of the real speed and about 192 seconds for an analyzing speed that equaled the real speed. We argue that different study designs as well as the parameters used within the software are likely to produce

  19. VNAP2: a computer program for computation of two-dimensional, time-dependent, compressible, turbulent flow

    Energy Technology Data Exchange (ETDEWEB)

    Cline, M.C.

    1981-08-01

    VNAP2 is a computer program for calculating turbulent (as well as laminar and inviscid), steady, and unsteady flow. VNAP2 solves the two-dimensional, time-dependent, compressible Navier-Stokes equations. The turbulence is modeled with either an algebraic mixing-length model, a one-equation model, or the Jones-Launder two-equation model. The geometry may be a single- or a dual-flowing stream. The interior grid points are computed using the unsplit MacCormack scheme. Two options to speed up the calculations for high Reynolds number flows are included. The boundary grid points are computed using a reference-plane-characteristic scheme with the viscous terms treated as source functions. An explicit artificial viscosity is included for shock computations. The fluid is assumed to be a perfect gas. The flow boundaries may be arbitrary curved solid walls, inflow/outflow boundaries, or free-jet envelopes. Typical problems that can be solved concern nozzles, inlets, jet-powered afterbodies, airfoils, and free-jet expansions. The accuracy and efficiency of the program are shown by calculations of several inviscid and turbulent flows. The program and its use are described completely, and six sample cases and a code listing are included.

  20. COMPUTING

    CERN Multimedia

    M. Kasemann P. McBride Edited by M-C. Sawley with contributions from: P. Kreuzer D. Bonacorsi S. Belforte F. Wuerthwein L. Bauerdick K. Lassila-Perini M-C. Sawley

    Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the comput...

  1. Optimization of scan time in MRI for total hip prostheses. SEMAC tailoring for prosthetic implants containing different types of metals

    Energy Technology Data Exchange (ETDEWEB)

    Deligianni, X. [University of Basel Hospital, Basel (Switzerland). Div. of Radiological Physics; Merian Iselin Klinik, Basel (Switzerland). Inst. of Radiology; Bieri, O. [University of Basel Hospital, Basel (Switzerland). Div. of Radiological Physics; Elke, R. [Orthomerian, Basel (Switzerland); Wischer, T.; Egelhof, T. [Merian Iselin Klinik, Basel (Switzerland). Inst. of Radiology

    2015-12-15

    Magnetic resonance imaging (MRI) of soft tissues after total hip arthroplasty is of clinical interest for the diagnosis of various pathologies that are usually invisible with other imaging modalities. As a result, considerable effort has been put into the development of metal artifact reduction MRI strategies, such as slice encoding for metal artifact correction (SEMAC). Generally, the degree of metal artifact reduction with SEMAC directly relates to the overall time spent for acquisition, but there is no specific consensus about the most efficient sequence setup depending on the implant material. The aim of this article is to suggest material-tailored SEMAC protocol settings. Five of the most common total hip prostheses (1. Revision prosthesis (S-Rom), 2. Titanium alloy, 3. Mueller type (CoNiCRMo alloy), 4. Old Charnley prosthesis (Exeter/Stryker), 5. MS-30 stem (stainless-steel)) were scanned on a 1.5 T MRI clinical scanner with a SEMAC sequence with a range of artifact-resolving slice encoding steps (SES: 2 - 23) along the slice direction (yielding a total variable scan time ranging from 1 to 10 min). The reduction of the artifact volume in comparison with maximal artifact suppression was evaluated both quantitatively and qualitatively in order to establish a recommended number of steps for each case. The number of SES that reduced the artifact volume below approximately 300 mm{sup 3} ranged from 3 to 13, depending on the material. Our results showed that although 3 SES steps can be sufficient for artifact reduction for titanium prostheses, at least 11 SES should be used for prostheses made of materials such as certain alloys of stainless steel. Tailoring SES to the implant material and to the desired degree of metal artifact reduction represents a simple tool for workflow optimization of SEMAC imaging near total hip arthroplasty in a clinical setting.

  2. Optimization of scan time in MRI for total hip prostheses. SEMAC tailoring for prosthetic implants containing different types of metals

    International Nuclear Information System (INIS)

    Deligianni, X.; Wischer, T.; Egelhof, T.

    2015-01-01

    Magnetic resonance imaging (MRI) of soft tissues after total hip arthroplasty is of clinical interest for the diagnosis of various pathologies that are usually invisible with other imaging modalities. As a result, considerable effort has been put into the development of metal artifact reduction MRI strategies, such as slice encoding for metal artifact correction (SEMAC). Generally, the degree of metal artifact reduction with SEMAC directly relates to the overall time spent for acquisition, but there is no specific consensus about the most efficient sequence setup depending on the implant material. The aim of this article is to suggest material-tailored SEMAC protocol settings. Five of the most common total hip prostheses (1. Revision prosthesis (S-Rom), 2. Titanium alloy, 3. Mueller type (CoNiCRMo alloy), 4. Old Charnley prosthesis (Exeter/Stryker), 5. MS-30 stem (stainless-steel)) were scanned on a 1.5 T MRI clinical scanner with a SEMAC sequence with a range of artifact-resolving slice encoding steps (SES: 2 - 23) along the slice direction (yielding a total variable scan time ranging from 1 to 10 min). The reduction of the artifact volume in comparison with maximal artifact suppression was evaluated both quantitatively and qualitatively in order to establish a recommended number of steps for each case. The number of SES that reduced the artifact volume below approximately 300 mm 3 ranged from 3 to 13, depending on the material. Our results showed that although 3 SES steps can be sufficient for artifact reduction for titanium prostheses, at least 11 SES should be used for prostheses made of materials such as certain alloys of stainless steel. Tailoring SES to the implant material and to the desired degree of metal artifact reduction represents a simple tool for workflow optimization of SEMAC imaging near total hip arthroplasty in a clinical setting.

  3. Quantitative analysis of orthopedic metal artefact reduction in 64-slice computed tomography scans in large head metal-on-metal total hip replacement, a phantom study

    NARCIS (Netherlands)

    Boomsma, Martijn F.; Warringa, Niek; Edens, Mireille A.; Mueller, Dirk; Ettema, Harmen B.; Verheyen, Cees C. P. M.; Maas, Mario

    2016-01-01

    Purpose: Quantification of the effect of O-MAR on decreasing metal artefacts caused by large head metal on metal total hip arthroplasty (MoM THA) in a dedicated phantom setup of the hip. Background: Pathological reactions of the hip capsule on Computed tomography (CT) can be difficult to diagnose

  4. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion The Tier 0 infrastructure was able to repack and promptly reconstruct heavy-ion collision data. Two copies were made of the data at CERN using a large CASTOR disk pool, and the core physics sample was replicated ...

  5. Real-time computing platform for spiking neurons (RT-spike).

    Science.gov (United States)

    Ros, Eduardo; Ortigosa, Eva M; Agís, Rodrigo; Carrillo, Richard; Arnold, Michael

    2006-07-01

    A computing platform is described for simulating arbitrary networks of spiking neurons in real time. A hybrid computing scheme is adopted that uses both software and hardware components to manage the tradeoff between flexibility and computational power; the neuron model is implemented in hardware and the network model and the learning are implemented in software. The incremental transition of the software components into hardware is supported. We focus on a spike response model (SRM) for a neuron where the synapses are modeled as input-driven conductances. The temporal dynamics of the synaptic integration process are modeled with a synaptic time constant that results in a gradual injection of charge. This type of model is computationally expensive and is not easily amenable to existing software-based event-driven approaches. As an alternative we have designed an efficient time-based computing architecture in hardware, where the different stages of the neuron model are processed in parallel. Further improvements occur by computing multiple neurons in parallel using multiple processing units. This design is tested using reconfigurable hardware and its scalability and performance evaluated. Our overall goal is to investigate biologically realistic models for the real-time control of robots operating within closed action-perception loops, and so we evaluate the performance of the system on simulating a model of the cerebellum where the emulation of the temporal dynamics of the synaptic integration process is important.

  6. Variation in computer time with geometry prescription in monte carlo code KENO-IV

    International Nuclear Information System (INIS)

    Gopalakrishnan, C.R.

    1988-01-01

    In most studies, the Monte Carlo criticality code KENO-IV has been compared with other Monte Carlo codes, but evaluation of its performance with different box descriptions has not been done so far. In Monte Carlo computations, any fractional savings of computing time is highly desirable. Variation in computation time with box description in KENO for two different fast reactor fuel subassemblies of FBTR and PFBR is studied. The K eff of an infinite array of fuel subassemblies is calculated by modelling the subassemblies in two different ways (i) multi-region, (ii) multi-box. In addition to these two cases, excess reactivity calculations of FBTR are also performed in two ways to study this effect in a complex geometry. It is observed that the K eff values calculated by multi-region and multi-box models agree very well. However the increase in computation time from the multi-box to the multi-region is considerable, while the difference in computer storage requirements for the two models is negligible. This variation in computing time arises from the way the neutron is tracked in the two cases. (author)

  7. Impact of operative time on early joint infection and deep vein thrombosis in primary total hip arthroplasty.

    Science.gov (United States)

    Wills, B W; Sheppard, E D; Smith, W R; Staggers, J R; Li, P; Shah, A; Lee, S R; Naranje, S M

    2018-03-22

    Infections and deep vein thrombosis (DVT) after total hip arthroplasty (THA) are challenging problems for both the patient and surgeon. Previous studies have identified numerous risk factors for infections and DVT after THA but have often been limited by sample size. We aimed to evaluate the effect of operative time on early postoperative infection as well as DVT rates following THA. We hypothesized that an increase in operative time would result in increased odds of acquiring an infection as well as a DVT. We conducted a retrospective analysis of prospectively collected data using the American College of Surgeons National Surgical Quality Improvement Program (NSQIP) database from 2006 to 2015 for all patients undergoing primary THA. Associations between operative time and infection or DVT were evaluated with multivariable logistic regressions controlling for demographics and several known risks factors for infection. Three different types of infections were evaluated: (1) superficial surgical site infection (SSI), an infection involving the skin or subcutaneous tissue, (2) deep SSI, an infection involving the muscle or fascial layers beneath the subcutaneous tissue, and (3) organ/space infection, an infection involving any part of the anatomy manipulated during surgery other than the incisional components. In total, 103,044 patients who underwent THA were included in our study. Our results suggested a significant association between superficial SSIs and operative time. Specifically, the adjusted odds of suffering a superficial SSI increased by 6% (CI=1.04-1.08, ptime. When using dichotomized operative time (90minutes), the adjusted odds of suffering a superficial SSI was 56% higher for patients with prolonged operative time (CI=1.05-2.32, p=0.0277). The adjusted odds of suffering a deep SSI increased by 7% for every 10-minute increase in operative time (CI=1.01-1.14, p=0.0335). No significant associations were detected between organ/space infection, wound

  8. Computing the Maximum Detour of a Plane Graph in Subquadratic Time

    DEFF Research Database (Denmark)

    Wulff-Nilsen, Christian

    Let G be a plane graph where each edge is a line segment. We consider the problem of computing the maximum detour of G, defined as the maximum over all pairs of distinct points p and q of G of the ratio between the distance between p and q in G and the distance |pq|. The fastest known algorithm f...... for this problem has O(n^2) running time. We show how to obtain O(n^{3/2}*(log n)^3) expected running time. We also show that if G has bounded treewidth, its maximum detour can be computed in O(n*(log n)^3) expected time....

  9. Continuous-variable quantum computing in optical time-frequency modes using quantum memories.

    Science.gov (United States)

    Humphreys, Peter C; Kolthammer, W Steven; Nunn, Joshua; Barbieri, Marco; Datta, Animesh; Walmsley, Ian A

    2014-09-26

    We develop a scheme for time-frequency encoded continuous-variable cluster-state quantum computing using quantum memories. In particular, we propose a method to produce, manipulate, and measure two-dimensional cluster states in a single spatial mode by exploiting the intrinsic time-frequency selectivity of Raman quantum memories. Time-frequency encoding enables the scheme to be extremely compact, requiring a number of memories that are a linear function of only the number of different frequencies in which the computational state is encoded, independent of its temporal duration. We therefore show that quantum memories can be a powerful component for scalable photonic quantum information processing architectures.

  10. Computation of transit times using the milestoning method with applications to polymer translocation

    Science.gov (United States)

    Hawk, Alexander T.; Konda, Sai Sriharsha M.; Makarov, Dmitrii E.

    2013-08-01

    Milestoning is an efficient approximation for computing long-time kinetics and thermodynamics of large molecular systems, which are inaccessible to brute-force molecular dynamics simulations. A common use of milestoning is to compute the mean first passage time (MFPT) for a conformational transition of interest. However, the MFPT is not always the experimentally observed timescale. In particular, the duration of the transition path, or the mean transit time, can be measured in single-molecule experiments, such as studies of polymers translocating through pores and fluorescence resonance energy transfer studies of protein folding. Here we show how to use milestoning to compute transit times and illustrate our approach by applying it to the translocation of a polymer through a narrow pore.

  11. A combined time-of-flight and depth-of-interaction detector for total-body positron emission tomography

    Energy Technology Data Exchange (ETDEWEB)

    Berg, Eric, E-mail: eberg@ucdavis.edu; Roncali, Emilie; Du, Junwei; Cherry, Simon R. [Department of Biomedical Engineering, University of California, Davis, One Shields Avenue, Davis, California 95616 (United States); Kapusta, Maciej [Molecular Imaging, Siemens Healthcare, Knoxville, Tennessee 37932 (United States)

    2016-02-15

    Purpose: In support of a project to build a total-body PET scanner with an axial field-of-view of 2 m, the authors are developing simple, cost-effective block detectors with combined time-of-flight (TOF) and depth-of-interaction (DOI) capabilities. Methods: This work focuses on investigating the potential of phosphor-coated crystals with conventional PMT-based block detector readout to provide DOI information while preserving timing resolution. The authors explored a variety of phosphor-coating configurations with single crystals and crystal arrays. Several pulse shape discrimination techniques were investigated, including decay time, delayed charge integration (DCI), and average signal shapes. Results: Pulse shape discrimination based on DCI provided the lowest DOI positioning error: 2 mm DOI positioning error was obtained with single phosphor-coated crystals while 3–3.5 mm DOI error was measured with the block detector module. Minimal timing resolution degradation was observed with single phosphor-coated crystals compared to uncoated crystals, and a timing resolution of 442 ps was obtained with phosphor-coated crystals in the block detector compared to 404 ps without phosphor coating. Flood maps showed a slight degradation in crystal resolvability with phosphor-coated crystals; however, all crystals could be resolved. Energy resolution was degraded by 3%–7% with phosphor-coated crystals compared to uncoated crystals. Conclusions: These results demonstrate the feasibility of obtaining TOF–DOI capabilities with simple block detector readout using phosphor-coated crystals.

  12. A combined time-of-flight and depth-of-interaction detector for total-body positron emission tomography

    International Nuclear Information System (INIS)

    Berg, Eric; Roncali, Emilie; Du, Junwei; Cherry, Simon R.; Kapusta, Maciej

    2016-01-01

    Purpose: In support of a project to build a total-body PET scanner with an axial field-of-view of 2 m, the authors are developing simple, cost-effective block detectors with combined time-of-flight (TOF) and depth-of-interaction (DOI) capabilities. Methods: This work focuses on investigating the potential of phosphor-coated crystals with conventional PMT-based block detector readout to provide DOI information while preserving timing resolution. The authors explored a variety of phosphor-coating configurations with single crystals and crystal arrays. Several pulse shape discrimination techniques were investigated, including decay time, delayed charge integration (DCI), and average signal shapes. Results: Pulse shape discrimination based on DCI provided the lowest DOI positioning error: 2 mm DOI positioning error was obtained with single phosphor-coated crystals while 3–3.5 mm DOI error was measured with the block detector module. Minimal timing resolution degradation was observed with single phosphor-coated crystals compared to uncoated crystals, and a timing resolution of 442 ps was obtained with phosphor-coated crystals in the block detector compared to 404 ps without phosphor coating. Flood maps showed a slight degradation in crystal resolvability with phosphor-coated crystals; however, all crystals could be resolved. Energy resolution was degraded by 3%–7% with phosphor-coated crystals compared to uncoated crystals. Conclusions: These results demonstrate the feasibility of obtaining TOF–DOI capabilities with simple block detector readout using phosphor-coated crystals

  13. An atomic orbital based real-time time-dependent density functional theory for computing electronic circular dichroism band spectra

    Energy Technology Data Exchange (ETDEWEB)

    Goings, Joshua J.; Li, Xiaosong, E-mail: xsli@uw.edu [Department of Chemistry, University of Washington, Seattle, Washington 98195 (United States)

    2016-06-21

    One of the challenges of interpreting electronic circular dichroism (ECD) band spectra is that different states may have different rotatory strength signs, determined by their absolute configuration. If the states are closely spaced and opposite in sign, observed transitions may be washed out by nearby states, unlike absorption spectra where transitions are always positive additive. To accurately compute ECD bands, it is necessary to compute a large number of excited states, which may be prohibitively costly if one uses the linear-response time-dependent density functional theory (TDDFT) framework. Here we implement a real-time, atomic-orbital based TDDFT method for computing the entire ECD spectrum simultaneously. The method is advantageous for large systems with a high density of states. In contrast to previous implementations based on real-space grids, the method is variational, independent of nuclear orientation, and does not rely on pseudopotential approximations, making it suitable for computation of chiroptical properties well into the X-ray regime.

  14. Real-time data acquisition and feedback control using Linux Intel computers

    International Nuclear Information System (INIS)

    Penaflor, B.G.; Ferron, J.R.; Piglowski, D.A.; Johnson, R.D.; Walker, M.L.

    2006-01-01

    This paper describes the experiences of the DIII-D programming staff in adapting Linux based Intel computing hardware for use in real-time data acquisition and feedback control systems. Due to the highly dynamic and unstable nature of magnetically confined plasmas in tokamak fusion experiments, real-time data acquisition and feedback control systems are in routine use with all major tokamaks. At DIII-D, plasmas are created and sustained using a real-time application known as the digital plasma control system (PCS). During each experiment, the PCS periodically samples data from hundreds of diagnostic signals and provides these data to control algorithms implemented in software. These algorithms compute the necessary commands to send to various actuators that affect plasma performance. The PCS consists of a group of rack mounted Intel Xeon computer systems running an in-house customized version of the Linux operating system tailored specifically to meet the real-time performance needs of the plasma experiments. This paper provides a more detailed description of the real-time computing hardware and custom developed software, including recent work to utilize dual Intel Xeon equipped computers within the PCS

  15. Kajian dan Implementasi Real TIME Operating System pada Single Board Computer Berbasis Arm

    OpenAIRE

    A, Wiedjaja; M, Handi; L, Jonathan; Christian, Benyamin; Kristofel, Luis

    2014-01-01

    Operating System is an important software in computer system. For personal and office use the operating system is sufficient. However, to critical mission applications such as nuclear power plants and braking system on the car (auto braking system) which need a high level of reliability, it requires operating system which operates in real time. The study aims to assess the implementation of the Linux-based operating system on a Single Board Computer (SBC) ARM-based, namely Pandaboard ES with ...

  16. Y2K issues for real time computer systems for fast breeder test reactor

    International Nuclear Information System (INIS)

    Swaminathan, P.

    1999-01-01

    Presentation shows the classification of real time systems related to operation, control and monitoring of the fast breeder test reactor. Software life cycle includes software requirement specification, software design description, coding, commissioning, operation and management. A software scheme in supervisory computer of fast breeder test rector is described with the twenty years of experience in design, development, installation, commissioning, operation and maintenance of computer based supervision control system for nuclear installation with a particular emphasis on solving the Y2K problem

  17. The Time Course of Knee Swelling Post Total Knee Arthroplasty and Its Associations with Quadriceps Strength and Gait Speed.

    Science.gov (United States)

    Pua, Yong-Hao

    2015-07-01

    This study examines the time course of knee swelling post total knee arthroplasty (TKA) and its associations with quadriceps strength and gait speed. Eighty-five patients with unilateral TKA participated. Preoperatively and on post-operative days (PODs) 1, 4, 14, and 90, knee swelling was measured using bioimpedance spectrometry. Preoperatively and on PODs 14 and 90, quadriceps strength was measured using isokinetic dynamometry while fast gait speed was measured using the timed 10-meter walk. On POD1, knee swelling increased ~35% from preoperative levels after which, knee swelling reduced but remained at ~11% above preoperative levels on POD90. In longitudinal, multivariable analyses, knee swelling was associated with quadriceps weakness (P<0.01) and slower gait speed (P=0.03). Interventions to reduce post-TKA knee swelling may be indicated to improve quadriceps strength and gait speed. Copyright © 2015 Elsevier Inc. All rights reserved.

  18. Near real-time digital holographic microscope based on GPU parallel computing

    Science.gov (United States)

    Zhu, Gang; Zhao, Zhixiong; Wang, Huarui; Yang, Yan

    2018-01-01

    A transmission near real-time digital holographic microscope with in-line and off-axis light path is presented, in which the parallel computing technology based on compute unified device architecture (CUDA) and digital holographic microscopy are combined. Compared to other holographic microscopes, which have to implement reconstruction in multiple focal planes and are time-consuming the reconstruction speed of the near real-time digital holographic microscope can be greatly improved with the parallel computing technology based on CUDA, so it is especially suitable for measurements of particle field in micrometer and nanometer scale. Simulations and experiments show that the proposed transmission digital holographic microscope can accurately measure and display the velocity of particle field in micrometer scale, and the average velocity error is lower than 10%.With the graphic processing units(GPU), the computing time of the 100 reconstruction planes(512×512 grids) is lower than 120ms, while it is 4.9s using traditional reconstruction method by CPU. The reconstruction speed has been raised by 40 times. In other words, it can handle holograms at 8.3 frames per second and the near real-time measurement and display of particle velocity field are realized. The real-time three-dimensional reconstruction of particle velocity field is expected to achieve by further optimization of software and hardware. Keywords: digital holographic microscope,

  19. Challenges in reducing the computational time of QSTS simulations for distribution system analysis.

    Energy Technology Data Exchange (ETDEWEB)

    Deboever, Jeremiah [Georgia Inst. of Technology, Atlanta, GA (United States); Zhang, Xiaochen [Georgia Inst. of Technology, Atlanta, GA (United States); Reno, Matthew J. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Broderick, Robert Joseph [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Grijalva, Santiago [Georgia Inst. of Technology, Atlanta, GA (United States); Therrien, Francis [CME International T& D, St. Bruno, QC (Canada)

    2017-06-01

    The rapid increase in penetration of distributed energy resources on the electric power distribution system has created a need for more comprehensive interconnection modelling and impact analysis. Unlike conventional scenario - based studies , quasi - static time - series (QSTS) simulation s can realistically model time - dependent voltage controllers and the diversity of potential impacts that can occur at different times of year . However, to accurately model a distribution system with all its controllable devices, a yearlong simulation at 1 - second resolution is often required , which could take conventional computers a computational time of 10 to 120 hours when an actual unbalanced distribution feeder is modeled . This computational burden is a clear l imitation to the adoption of QSTS simulation s in interconnection studies and for determining optimal control solutions for utility operations . Our ongoing research to improve the speed of QSTS simulation has revealed many unique aspects of distribution system modelling and sequential power flow analysis that make fast QSTS a very difficult problem to solve. In this report , the most relevant challenges in reducing the computational time of QSTS simulations are presented: number of power flows to solve, circuit complexity, time dependence between time steps, multiple valid power flow solutions, controllable element interactions, and extensive accurate simulation analysis.

  20. The total length of myocytes and capillaries, and total number of myocyte nuclei in the rat heart are time-dependently increased by growth hormone

    DEFF Research Database (Denmark)

    Brüel, Annemarie; Oxlund, Hans; Nyengaard, Jens Randel

    2005-01-01

    /kg/day) or vehicle for 5, 10, 20, 40, or 80 days. From the left ventricle (LV) histological sections were made and stereological methods applied. Linear regression showed that GH time-dependently increased: LV volume (r=0.96, P

  1. IMU-based Real-time Pose Measurement system for Anterior Pelvic Plane in Total Hip Replacement Surgeries.

    Science.gov (United States)

    Zhe Cao; Shaojie Su; Hao Tang; Yixin Zhou; Zhihua Wang; Hong Chen

    2017-07-01

    With the aging of population, the number of Total Hip Replacement Surgeries (THR) increased year by year. In THR, inaccurate position of the implanted prosthesis may lead to the failure of the operation. In order to reduce the failure rate and acquire the real-time pose of Anterior Pelvic Plane (APP), we propose a measurement system in this paper. The measurement system includes two parts: Initial Pose Measurement Instrument (IPMI) and Real-time Pose Measurement Instrument (RPMI). IPMI is used to acquire the initial pose of the APP, and RPMI is used to estimate the real-time pose of the APP. Both are composed of an Inertial Measurement Unit (IMU) and magnetometer sensors. To estimate the attitude of the measurement system, the Extended Kalman Filter (EKF) is adopted in this paper. The real-time pose of the APP could be acquired together with the algorithm designed in the paper. The experiment results show that the Root Mean Square Error (RMSE) is within 1.6 degrees, which meets the requirement of THR operations.

  2. Reconstruction of MODIS total suspended matter time series maps by DINEOF and validation with autonomous platform data

    Science.gov (United States)

    Nechad, Bouchra; Alvera-Azcaràte, Aida; Ruddick, Kevin; Greenwood, Naomi

    2011-08-01

    In situ measurements of total suspended matter (TSM) over the period 2003-2006, collected with two autonomous platforms from the Centre for Environment, Fisheries and Aquatic Sciences (Cefas) measuring the optical backscatter (OBS) in the southern North Sea, are used to assess the accuracy of TSM time series extracted from satellite data. Since there are gaps in the remote sensing (RS) data, due mainly to cloud cover, the Data Interpolating Empirical Orthogonal Functions (DINEOF) is used to fill in the TSM time series and build a continuous daily "recoloured" dataset. The RS datasets consist of TSM maps derived from MODIS imagery using the bio-optical model of Nechad et al. (Rem Sens Environ 114: 854-866, 2010). In this study, the DINEOF time series are compared to the in situ OBS measured in moderately to very turbid waters respectively in West Gabbard and Warp Anchorage, in the southern North Sea. The discrepancies between instantaneous RS, DINEOF-filled RS data and Cefas data are analysed in terms of TSM algorithm uncertainties, space-time variability and DINEOF reconstruction uncertainty.

  3. Explicit time marching methods for the time-dependent Euler computations

    International Nuclear Information System (INIS)

    Tai, C.H.; Chiang, D.C.; Su, Y.P.

    1997-01-01

    Four explicit type time marching methods, including one proposed by the authors, are examined. The TVD conditions of this method are analyzed with the linear conservation law as the model equation. Performance of these methods when applied to the Euler equations are numerically tested. Seven examples are tested, the main concern is the performance of the methods when discontinuities with different strengths are encountered. When the discontinuity is getting stronger, spurious oscillation shows up for three existing methods, while the method proposed by the authors always gives the results with satisfaction. The effect of the limiter is also investigated. To put these methods in the same basis for the comparison the same spatial discretization is used. Roe's solver is used to evaluate the fluxes at the cell interface; spatially second-order accuracy is achieved by the MUSCL reconstruction. 19 refs., 8 figs

  4. COMPUTING

    CERN Multimedia

    P. McBride

    It has been a very active year for the computing project with strong contributions from members of the global community. The project has focused on site preparation and Monte Carlo production. The operations group has begun processing data from P5 as part of the global data commissioning. Improvements in transfer rates and site availability have been seen as computing sites across the globe prepare for large scale production and analysis as part of CSA07. Preparations for the upcoming Computing Software and Analysis Challenge CSA07 are progressing. Ian Fisk and Neil Geddes have been appointed as coordinators for the challenge. CSA07 will include production tests of the Tier-0 production system, reprocessing at the Tier-1 sites and Monte Carlo production at the Tier-2 sites. At the same time there will be a large analysis exercise at the Tier-2 centres. Pre-production simulation of the Monte Carlo events for the challenge is beginning. Scale tests of the Tier-0 will begin in mid-July and the challenge it...

  5. COMPUTING

    CERN Multimedia

    M. Kasemann

    Introduction During the past six months, Computing participated in the STEP09 exercise, had a major involvement in the October exercise and has been working with CMS sites on improving open issues relevant for data taking. At the same time operations for MC production, real data reconstruction and re-reconstructions and data transfers at large scales were performed. STEP09 was successfully conducted in June as a joint exercise with ATLAS and the other experiments. It gave good indication about the readiness of the WLCG infrastructure with the two major LHC experiments stressing the reading, writing and processing of physics data. The October Exercise, in contrast, was conducted as an all-CMS exercise, where Physics, Computing and Offline worked on a common plan to exercise all steps to efficiently access and analyze data. As one of the major results, the CMS Tier-2s demonstrated to be fully capable for performing data analysis. In recent weeks, efforts were devoted to CMS Computing readiness. All th...

  6. COMPUTING

    CERN Multimedia

    I. Fisk

    2012-01-01

    Introduction Computing continued with a high level of activity over the winter in preparation for conferences and the start of the 2012 run. 2012 brings new challenges with a new energy, more complex events, and the need to make the best use of the available time before the Long Shutdown. We expect to be resource constrained on all tiers of the computing system in 2012 and are working to ensure the high-priority goals of CMS are not impacted. Heavy ions After a successful 2011 heavy-ion run, the programme is moving to analysis. During the run, the CAF resources were well used for prompt analysis. Since then in 2012 on average 200 job slots have been used continuously at Vanderbilt for analysis workflows. Operations Office As of 2012, the Computing Project emphasis has moved from commissioning to operation of the various systems. This is reflected in the new organisation structure where the Facilities and Data Operations tasks have been merged into a common Operations Office, which now covers everything ...

  7. COMPUTING

    CERN Multimedia

    M. Kasemann

    CCRC’08 challenges and CSA08 During the February campaign of the Common Computing readiness challenges (CCRC’08), the CMS computing team had achieved very good results. The link between the detector site and the Tier0 was tested by gradually increasing the number of parallel transfer streams well beyond the target. Tests covered the global robustness at the Tier0, processing a massive number of very large files and with a high writing speed to tapes.  Other tests covered the links between the different Tiers of the distributed infrastructure and the pre-staging and reprocessing capacity of the Tier1’s: response time, data transfer rate and success rate for Tape to Buffer staging of files kept exclusively on Tape were measured. In all cases, coordination with the sites was efficient and no serious problem was found. These successful preparations prepared the ground for the second phase of the CCRC’08 campaign, in May. The Computing Software and Analysis challen...

  8. COMPUTING

    CERN Multimedia

    M. Kasemann

    Introduction More than seventy CMS collaborators attended the Computing and Offline Workshop in San Diego, California, April 20-24th to discuss the state of readiness of software and computing for collisions. Focus and priority were given to preparations for data taking and providing room for ample dialog between groups involved in Commissioning, Data Operations, Analysis and MC Production. Throughout the workshop, aspects of software, operating procedures and issues addressing all parts of the computing model were discussed. Plans for the CMS participation in STEP’09, the combined scale testing for all four experiments due in June 2009, were refined. The article in CMS Times by Frank Wuerthwein gave a good recap of the highly collaborative atmosphere of the workshop. Many thanks to UCSD and to the organizers for taking care of this workshop, which resulted in a long list of action items and was definitely a success. A considerable amount of effort and care is invested in the estimate of the co...

  9. COMPUTING

    CERN Multimedia

    Contributions from I. Fisk

    2012-01-01

    Introduction The start of the 2012 run has been busy for Computing. We have reconstructed, archived, and served a larger sample of new data than in 2011, and we are in the process of producing an even larger new sample of simulations at 8 TeV. The running conditions and system performance are largely what was anticipated in the plan, thanks to the hard work and preparation of many people. Heavy ions Heavy Ions has been actively analysing data and preparing for conferences.  Operations Office Figure 6: Transfers from all sites in the last 90 days For ICHEP and the Upgrade efforts, we needed to produce and process record amounts of MC samples while supporting the very successful data-taking. This was a large burden, especially on the team members. Nevertheless the last three months were very successful and the total output was phenomenal, thanks to our dedicated site admins who keep the sites operational and the computing project members who spend countless hours nursing the...

  10. Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure

    International Nuclear Information System (INIS)

    Wang, Henry; Ma Yunzhi; Pratx, Guillem; Xing Lei

    2011-01-01

    Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47x speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed. (note)

  11. Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Henry [Department of Electrical Engineering, Stanford University, Stanford, CA 94305 (United States); Ma Yunzhi; Pratx, Guillem; Xing Lei, E-mail: hwang41@stanford.edu [Department of Radiation Oncology, Stanford University School of Medicine, Stanford, CA 94305-5847 (United States)

    2011-09-07

    Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47x speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed. (note)

  12. Toward real-time Monte Carlo simulation using a commercial cloud computing infrastructure.

    Science.gov (United States)

    Wang, Henry; Ma, Yunzhi; Pratx, Guillem; Xing, Lei

    2011-09-07

    Monte Carlo (MC) methods are the gold standard for modeling photon and electron transport in a heterogeneous medium; however, their computational cost prohibits their routine use in the clinic. Cloud computing, wherein computing resources are allocated on-demand from a third party, is a new approach for high performance computing and is implemented to perform ultra-fast MC calculation in radiation therapy. We deployed the EGS5 MC package in a commercial cloud environment. Launched from a single local computer with Internet access, a Python script allocates a remote virtual cluster. A handshaking protocol designates master and worker nodes. The EGS5 binaries and the simulation data are initially loaded onto the master node. The simulation is then distributed among independent worker nodes via the message passing interface, and the results aggregated on the local computer for display and data analysis. The described approach is evaluated for pencil beams and broad beams of high-energy electrons and photons. The output of cloud-based MC simulation is identical to that produced by single-threaded implementation. For 1 million electrons, a simulation that takes 2.58 h on a local computer can be executed in 3.3 min on the cloud with 100 nodes, a 47× speed-up. Simulation time scales inversely with the number of parallel nodes. The parallelization overhead is also negligible for large simulations. Cloud computing represents one of the most important recent advances in supercomputing technology and provides a promising platform for substantially improved MC simulation. In addition to the significant speed up, cloud computing builds a layer of abstraction for high performance parallel computing, which may change the way dose calculations are performed and radiation treatment plans are completed.

  13. A computationally simple and robust method to detect determinism in a time series

    DEFF Research Database (Denmark)

    Lu, Sheng; Ju, Ki Hwan; Kanters, Jørgen K.

    2006-01-01

    We present a new, simple, and fast computational technique, termed the incremental slope (IS), that can accurately distinguish between deterministic from stochastic systems even when the variance of noise is as large or greater than the signal, and remains robust for time-varying signals. The IS ......We present a new, simple, and fast computational technique, termed the incremental slope (IS), that can accurately distinguish between deterministic from stochastic systems even when the variance of noise is as large or greater than the signal, and remains robust for time-varying signals...

  14. SLMRACE: a noise-free RACE implementation with reduced computational time

    Science.gov (United States)

    Chauvin, Juliet; Provenzi, Edoardo

    2017-05-01

    We present a faster and noise-free implementation of the RACE algorithm. RACE has mixed characteristics between the famous Retinex model of Land and McCann and the automatic color equalization (ACE) color-correction algorithm. The original random spray-based RACE implementation suffers from two main problems: its computational time and the presence of noise. Here, we will show that it is possible to adapt two techniques recently proposed by Banić et al. to the RACE framework in order to drastically decrease the computational time and noise generation. The implementation will be called smart-light-memory-RACE (SLMRACE).

  15. Ultrasonic divergent-beam scanner for time-of-flight tomography with computer evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Glover, G H

    1978-03-02

    The rotatable ultrasonic divergent-beam scanner is designed for time-of-flight tomography with computer evaluation. With it there can be measured parameters that are of importance for the structure of soft tissues, e.g. time as a function of the velocity distribution along a certain path of flight(the method is analogous to the transaxial X-ray tomography). Moreover it permits to perform the quantitative measurement of two-dimensional velocity distributions and may therefore be applied to serial examinations for detecting cancer of the breast. As computers digital memories as well as analog-digital-hybrid systems are suitable.

  16. A Hybrid Metaheuristic Approach for Minimizing the Total Flow Time in A Flow Shop Sequence Dependent Group Scheduling Problem

    Directory of Open Access Journals (Sweden)

    Antonio Costa

    2014-07-01

    Full Text Available Production processes in Cellular Manufacturing Systems (CMS often involve groups of parts sharing the same technological requirements in terms of tooling and setup. The issue of scheduling such parts through a flow-shop production layout is known as the Flow-Shop Group Scheduling (FSGS problem or, whether setup times are sequence-dependent, the Flow-Shop Sequence-Dependent Group Scheduling (FSDGS problem. This paper addresses the FSDGS issue, proposing a hybrid metaheuristic procedure integrating features from Genetic Algorithms (GAs and Biased Random Sampling (BRS search techniques with the aim of minimizing the total flow time, i.e., the sum of completion times of all jobs. A well-known benchmark of test cases, entailing problems with two, three, and six machines, is employed for both tuning the relevant parameters of the developed procedure and assessing its performances against two metaheuristic algorithms recently presented by literature. The obtained results and a properly arranged ANOVA analysis highlight the superiority of the proposed approach in tackling the scheduling problem under investigation.

  17. A Flexible Job Shop Scheduling Problem with Controllable Processing Times to Optimize Total Cost of Delay and Processing

    Directory of Open Access Journals (Sweden)

    Hadi Mokhtari

    2015-11-01

    Full Text Available In this paper, the flexible job shop scheduling problem with machine flexibility and controllable process times is studied. The main idea is that the processing times of operations may be controlled by consumptions of additional resources. The purpose of this paper to find the best trade-off between processing cost and delay cost in order to minimize the total costs. The proposed model, flexible job shop scheduling with controllable processing times (FJCPT, is formulated as an integer non-linear programming (INLP model and then it is converted into an integer linear programming (ILP model. Due to NP-hardness of FJCPT, conventional analytic optimization methods are not efficient. Hence, in order to solve the problem, a Scatter Search (SS, as an efficient metaheuristic method, is developed. To show the effectiveness of the proposed method, numerical experiments are conducted. The efficiency of the proposed algorithm is compared with that of a genetic algorithm (GA available in the literature for solving FJSP problem. The results showed that the proposed SS provide better solutions than the existing GA.

  18. Computer-games for gravitational wave science outreach: Black Hole Pong and Space Time Quest

    International Nuclear Information System (INIS)

    Carbone, L; Bond, C; Brown, D; Brückner, F; Grover, K; Lodhia, D; Mingarelli, C M F; Fulda, P; Smith, R J E; Unwin, R; Vecchio, A; Wang, M; Whalley, L; Freise, A

    2012-01-01

    We have established a program aimed at developing computer applications and web applets to be used for educational purposes as well as gravitational wave outreach activities. These applications and applets teach gravitational wave physics and technology. The computer programs are generated in collaboration with undergraduates and summer students as part of our teaching activities, and are freely distributed on a dedicated website. As part of this program, we have developed two computer-games related to gravitational wave science: 'Black Hole Pong' and 'Space Time Quest'. In this article we present an overview of our computer related outreach activities and discuss the games and their educational aspects, and report on some positive feedback received.

  19. Computational Procedures for a Class of GI/D/k Systems in Discrete Time

    Directory of Open Access Journals (Sweden)

    Md. Mostafizur Rahman

    2009-01-01

    Full Text Available A class of discrete time GI/D/k systems is considered for which the interarrival times have finite support and customers are served in first-in first-out (FIFO order. The system is formulated as a single server queue with new general independent interarrival times and constant service duration by assuming cyclic assignment of customers to the identical servers. Then the queue length is set up as a quasi-birth-death (QBD type Markov chain. It is shown that this transformed GI/D/1 system has special structures which make the computation of the matrix R simple and efficient, thereby reducing the number of multiplications in each iteration significantly. As a result we were able to keep the computation time very low. Moreover, use of the resulting structural properties makes the computation of the distribution of queue length of the transformed system efficient. The computation of the distribution of waiting time is also shown to be simple by exploiting the special structures.

  20. Experimental technique for study on three-particle reactions in kinematically total experiments with usage of the two-processor complex on the M-400 computer basis

    International Nuclear Information System (INIS)

    Berezin, F.N.; Kisurin, V.A.; Nemets, O.F.; Ofengenden, R.G.; Pugach, V.M.; Pavlenko, Yu.N.; Patlan', Yu.V.; Savrasov, S.S.

    1981-01-01

    Experimental technique for investigation of three-particle nuclear reactions in kinematically total experiments is described. The technique provides the storage of one-dimensional and two- dimensional energy spectra from several detectors. A block diagram of the measuring system, using this technique, is presented. The measuring system consists of analog equipment for rapid-slow coincidences and of a two-processor complex on the base of the M-400 computer with a general bus. Application of a two-processor complex, each computer of which has a possibility of direct access to memory of another computer, permits to separate functions of data collection and data operational presentation and to perform necessary physical calculations. Software of the measuring complex which includes programs written using the ASSEMBLER language for the first computer and functional programs written using the BASIC language for the second computer, is considered. Software of the first computer includes the DISPETCHER dialog control program, driver package for control of external devices, of applied program package and system modules. The technique, described, is tested in experiment on investigation of d+ 10 B→α+α+α three- particle reaction at deutron energy of 13.6 MeV. The two-dimensional energy spectrum reaction obtained with the help of the technique described is presented [ru

  1. Dissociated time course between peak torque and total work recovery following bench press training in resistance trained men.

    Science.gov (United States)

    Ferreira, Diogo V; Gentil, Paulo; Ferreira-Junior, João B; Soares, Saulo R S; Brown, Lee E; Bottaro, Martim

    2017-10-01

    To evaluate the time course of peak torque and total work recovery after a resistance training session involving the bench press exercise. Repeated measures with a within subject design. Twenty-six resistance-trained men (age: 23.7±3.7years; height: 176.0±5.7cm; mass: 79.65±7.61kg) performed one session involving eight sets of the bench press exercise performed to momentary muscle failure with 2-min rest between sets. Shoulder horizontal adductors peak torque (PT), total work (TW), delayed onset muscle soreness (DOMS) and subjective physical fitness were measured pre, immediately post, 24, 48, 72 and 96h following exercise. The exercise protocol resulted in significant pectoralis major DOMS that lasted for 72h. Immediately after exercise, the reduction in shoulder horizontal adductors TW (25%) was greater than PT (17%). TW, as a percentage of baseline values, was also less than PT at 24, 48 and 96h after exercise. Additionally, PT returned to baseline at 96h, while TW did not. Resistance trained men presented dissimilar PT and TW recovery following free weight bench press exercise. This indicates that recovery of maximal voluntary contraction does not reflect the capability to perform multiple contractions. Strength and conditioning professionals should be cautious when evaluating muscle recovery by peak torque, since it can lead to the repetition of a training session sooner than recommended. Copyright © 2017. Published by Elsevier Inc.

  2. Total neutron-counting plutonium inventory measurement systems (PIMS) and their potential application to near real time materials accountancy (NRTMA)

    International Nuclear Information System (INIS)

    Driscall, I.; Fox, G.H.; Orr, C.H.; Whitehouse, K.R.

    1988-01-01

    A radiometric method of determining the inventory of an operating plutonium plant is described. An array of total neutron counters distributed across the plant is used to estimate hold-up at each plant item. Corrections for the sensitivity of detectors to plutonium in adjacent plant items are achieved through a matrix approach. This paper describes our experience in design, calibration and operation of a Plutonium Inventory Measurement System (PIMS) on an oxalate precipitation plutonium finishing line. Data from a recent trial of Near-Real-Time Materials Accounting (NRTMA) using the PIMS are presented and used to illustrate its present performance and problem areas. The reader is asked to consider what role PIMS might have in future accountancy systems

  3. Iterated greedy algorithms to minimize the total family flow time for job-shop scheduling with job families and sequence-dependent set-ups

    Science.gov (United States)

    Kim, Ji-Su; Park, Jung-Hyeon; Lee, Dong-Ho

    2017-10-01

    This study addresses a variant of job-shop scheduling in which jobs are grouped into job families, but they are processed individually. The problem can be found in various industrial systems, especially in reprocessing shops of remanufacturing systems. If the reprocessing shop is a job-shop type and has the component-matching requirements, it can be regarded as a job shop with job families since the components of a product constitute a job family. In particular, sequence-dependent set-ups in which set-up time depends on the job just completed and the next job to be processed are also considered. The objective is to minimize the total family flow time, i.e. the maximum among the completion times of the jobs within a job family. A mixed-integer programming model is developed and two iterated greedy algorithms with different local search methods are proposed. Computational experiments were conducted on modified benchmark instances and the results are reported.

  4. The thermodynamic approach to boron chemical vapour deposition based on a computer minimization of the total Gibbs free energy

    International Nuclear Information System (INIS)

    Naslain, R.; Thebault, J.; Hagenmuller, P.; Bernard, C.

    1979-01-01

    A thermodynamic approach based on the minimization of the total Gibbs free energy of the system is used to study the chemical vapour deposition (CVD) of boron from BCl 3 -H 2 or BBr 3 -H 2 mixtures on various types of substrates (at 1000 < T< 1900 K and 1 atm). In this approach it is assumed that states close to equilibrium are reached in the boron CVD apparatus. (Auth.)

  5. Extending total parenteral nutrition hang time in the neonatal intensive care unit: is it safe and cost effective?

    Science.gov (United States)

    Balegar V, Kiran Kumar; Azeem, Mohammad Irfan; Spence, Kaye; Badawi, Nadia

    2013-01-01

    To investigate the effects of prolonging hang time of total parenteral nutrition (TPN) fluid on central line-associated blood stream infection (CLABSI), TPN-related cost and nursing workload. A before-after observational study comparing the practice of hanging TPN bags for 48 h (6 February 2009-5 February 2010) versus 24 h (6 February 2008-5 February 2009) in a tertiary neonatal intensive care unit was conducted. The main outcome measures were CLABSI, TPN-related expenses and nursing workload. One hundred thirty-six infants received 24-h TPN bags and 124 received 48-h TPN bags. Median (inter-quartile range) gestation (37 weeks (33,39) vs. 36 weeks (33,39)), mean (±standard deviation) admission weight of 2442 g (±101) versus 2476 g (±104) and TPN duration (9.7 days (±12.7) vs. 9.9 days (±13.4)) were similar (P > 0.05) between the 24- and 48-h TPN groups. There was no increase in CLABSI with longer hang time (0.8 vs. 0.4 per 1000 line days in the 24-h vs. 48-h group; P < 0.05). Annual cost saving using 48-h TPN was AUD 97,603.00. By using 48-h TPN, 68.3% of nurses indicated that their workload decreased and 80.5% indicated that time spent changing TPN reduced. Extending TPN hang time from 24 to 48 h did not alter CLABSI rate and was associated with a reduced TPN-related cost and perceived nursing workload. Larger randomised controlled trials are needed to more clearly delineate these effects. © 2012 The Authors. Journal of Paediatrics and Child Health © 2012 Paediatrics and Child Health Division (Royal Australasian College of Physicians).

  6. Efficient Buffer Capacity and Scheduler Setting Computation for Soft Real-Time Stream Processing Applications

    NARCIS (Netherlands)

    Bekooij, Marco; Bekooij, Marco Jan Gerrit; Wiggers, M.H.; van Meerbergen, Jef

    2007-01-01

    Soft real-time applications that process data streams can often be intuitively described as dataflow process networks. In this paper we present a novel analysis technique to compute conservative estimates of the required buffer capacities in such process networks. With the same analysis technique

  7. Computational time analysis of the numerical solution of 3D electrostatic Poisson's equation

    Science.gov (United States)

    Kamboh, Shakeel Ahmed; Labadin, Jane; Rigit, Andrew Ragai Henri; Ling, Tech Chaw; Amur, Khuda Bux; Chaudhary, Muhammad Tayyab

    2015-05-01

    3D Poisson's equation is solved numerically to simulate the electric potential in a prototype design of electrohydrodynamic (EHD) ion-drag micropump. Finite difference method (FDM) is employed to discretize the governing equation. The system of linear equations resulting from FDM is solved iteratively by using the sequential Jacobi (SJ) and sequential Gauss-Seidel (SGS) methods, simulation results are also compared to examine the difference between the results. The main objective was to analyze the computational time required by both the methods with respect to different grid sizes and parallelize the Jacobi method to reduce the computational time. In common, the SGS method is faster than the SJ method but the data parallelism of Jacobi method may produce good speedup over SGS method. In this study, the feasibility of using parallel Jacobi (PJ) method is attempted in relation to SGS method. MATLAB Parallel/Distributed computing environment is used and a parallel code for SJ method is implemented. It was found that for small grid size the SGS method remains dominant over SJ method and PJ method while for large grid size both the sequential methods may take nearly too much processing time to converge. Yet, the PJ method reduces computational time to some extent for large grid sizes.

  8. Math modeling and computer mechanization for real time simulation of rotary-wing aircraft

    Science.gov (United States)

    Howe, R. M.

    1979-01-01

    Mathematical modeling and computer mechanization for real time simulation of rotary wing aircraft is discussed. Error analysis in the digital simulation of dynamic systems, such as rotary wing aircraft is described. The method for digital simulation of nonlinearities with discontinuities, such as exist in typical flight control systems and rotor blade hinges, is discussed.

  9. Efficient Geo-Computational Algorithms for Constructing Space-Time Prisms in Road Networks

    Directory of Open Access Journals (Sweden)

    Hui-Ping Chen

    2016-11-01

    Full Text Available The Space-time prism (STP is a key concept in time geography for analyzing human activity-travel behavior under various Space-time constraints. Most existing time-geographic studies use a straightforward algorithm to construct STPs in road networks by using two one-to-all shortest path searches. However, this straightforward algorithm can introduce considerable computational overhead, given the fact that accessible links in a STP are generally a small portion of the whole network. To address this issue, an efficient geo-computational algorithm, called NTP-A*, is proposed. The proposed NTP-A* algorithm employs the A* and branch-and-bound techniques to discard inaccessible links during two shortest path searches, and thereby improves the STP construction performance. Comprehensive computational experiments are carried out to demonstrate the computational advantage of the proposed algorithm. Several implementation techniques, including the label-correcting technique and the hybrid link-node labeling technique, are discussed and analyzed. Experimental results show that the proposed NTP-A* algorithm can significantly improve STP construction performance in large-scale road networks by a factor of 100, compared with existing algorithms.

  10. Computer-generated versus nurse-determined strategy for incubator humidity and time to regain birthweight

    NARCIS (Netherlands)

    Helder, Onno K.; Mulder, Paul G. H.; van Goudoever, Johannes B.

    2008-01-01

    To compare effects on premature infants' weight gain of a computer-generated and a nurse-determined incubator humidity strategy. An optimal humidity protocol is thought to reduce time to regain birthweight. Prospective randomized controlled design. Level IIIC neonatal intensive care unit in the

  11. Computing Camps for Girls : A First-Time Experience at the University of Limerick

    NARCIS (Netherlands)

    McInerney, Clare; Lamprecht, A.L.; Margaria, Tiziana

    2018-01-01

    Increasing the number of females in ICT-related university courses has been a major concern for several years. In 2015, we offered a girls-only computing summer camp for the first time, as a new component in our education and outreach activities to foster students’ interest in our discipline. In

  12. A Real-Time Plagiarism Detection Tool for Computer-Based Assessments

    Science.gov (United States)

    Jeske, Heimo J.; Lall, Manoj; Kogeda, Okuthe P.

    2018-01-01

    Aim/Purpose: The aim of this article is to develop a tool to detect plagiarism in real time amongst students being evaluated for learning in a computer-based assessment setting. Background: Cheating or copying all or part of source code of a program is a serious concern to academic institutions. Many academic institutions apply a combination of…

  13. Modeling of requirement specification for safety critical real time computer system using formal mathematical specifications

    International Nuclear Information System (INIS)

    Sankar, Bindu; Sasidhar Rao, B.; Ilango Sambasivam, S.; Swaminathan, P.

    2002-01-01

    Full text: Real time computer systems are increasingly used for safety critical supervision and control of nuclear reactors. Typical application areas are supervision of reactor core against coolant flow blockage, supervision of clad hot spot, supervision of undesirable power excursion, power control and control logic for fuel handling systems. The most frequent cause of fault in safety critical real time computer system is traced to fuzziness in requirement specification. To ensure the specified safety, it is necessary to model the requirement specification of safety critical real time computer systems using formal mathematical methods. Modeling eliminates the fuzziness in the requirement specification and also helps to prepare the verification and validation schemes. Test data can be easily designed from the model of the requirement specification. Z and B are the popular languages used for modeling the requirement specification. A typical safety critical real time computer system for supervising the reactor core of prototype fast breeder reactor (PFBR) against flow blockage is taken as case study. Modeling techniques and the actual model are explained in detail. The advantages of modeling for ensuring the safety are summarized

  14. Green computing: power optimisation of VFI-based real-time multiprocessor dataflow applications (extended version)

    NARCIS (Netherlands)

    Ahmad, W.; Holzenspies, P.K.F.; Stoelinga, Mariëlle Ida Antoinette; van de Pol, Jan Cornelis

    2015-01-01

    Execution time is no longer the only performance metric for computer systems. In fact, a trend is emerging to trade raw performance for energy savings. Techniques like Dynamic Power Management (DPM, switching to low power state) and Dynamic Voltage and Frequency Scaling (DVFS, throttling processor

  15. Green computing: power optimisation of vfi-based real-time multiprocessor dataflow applications

    NARCIS (Netherlands)

    Ahmad, W.; Holzenspies, P.K.F.; Stoelinga, Mariëlle Ida Antoinette; van de Pol, Jan Cornelis

    2015-01-01

    Execution time is no longer the only performance metric for computer systems. In fact, a trend is emerging to trade raw performance for energy savings. Techniques like Dynamic Power Management (DPM, switching to low power state) and Dynamic Voltage and Frequency Scaling (DVFS, throttling processor

  16. Effect of exposure time reduction towards sensitivity and SNR for computed radiography (CR) application in NDT

    International Nuclear Information System (INIS)

    Sapizah Rahim; Khairul Anuar Mohd Salleh; Noorhazleena Azaman; Shaharudin Sayuti; Siti Madiha Muhammad Amir; Arshad Yassin; Abdul Razak Hamzah

    2010-01-01

    Signal-to-noise ratio (SNR) and sensitivity study of Computed Radiography (CR) system with reduction of exposure time is presented. The purposes of this research are to determine the behavior of SNR toward three different thicknesses (step wedge; 5, 10 and 15 mm) and the ability of CR system to recognize hole type penetrameter when the exposure time decreased up to 80 % according to the exposure chart (D7; ISOVOLT Titan E). It is shown that the SNR is decreased with decreasing of exposure time percentage but the high quality image is achieved until 80 % reduction of exposure time. (author)

  17. Application of queueing models to multiprogrammed computer systems operating in a time-critical environment

    Science.gov (United States)

    Eckhardt, D. E., Jr.

    1979-01-01

    A model of a central processor (CPU) which services background applications in the presence of time critical activity is presented. The CPU is viewed as an M/M/1 queueing system subject to periodic interrupts by deterministic, time critical process. The Laplace transform of the distribution of service times for the background applications is developed. The use of state of the art queueing models for studying the background processing capability of time critical computer systems is discussed and the results of a model validation study which support this application of queueing models are presented.

  18. A computer-based time study system for timber harvesting operations

    Science.gov (United States)

    Jingxin Wang; Joe McNeel; John Baumgras

    2003-01-01

    A computer-based time study system was developed for timber harvesting operations. Object-oriented techniques were used to model and design the system. The front-end of the time study system resides on the MS Windows CE and the back-end is supported by MS Access. The system consists of three major components: a handheld system, data transfer interface, and data storage...

  19. Latitude-Time Total Electron Content Anomalies as Precursors to Japan's Large Earthquakes Associated with Principal Component Analysis

    Directory of Open Access Journals (Sweden)

    Jyh-Woei Lin

    2011-01-01

    Full Text Available The goal of this study is to determine whether principal component analysis (PCA can be used to process latitude-time ionospheric TEC data on a monthly basis to identify earthquake associated TEC anomalies. PCA is applied to latitude-time (mean-of-a-month ionospheric total electron content (TEC records collected from the Japan GEONET network to detect TEC anomalies associated with 18 earthquakes in Japan (M≥6.0 from 2000 to 2005. According to the results, PCA was able to discriminate clear TEC anomalies in the months when all 18 earthquakes occurred. After reviewing months when no M≥6.0 earthquakes occurred but geomagnetic storm activity was present, it is possible that the maximal principal eigenvalues PCA returned for these 18 earthquakes indicate earthquake associated TEC anomalies. Previously PCA has been used to discriminate earthquake-associated TEC anomalies recognized by other researchers, who found that statistical association between large earthquakes and TEC anomalies could be established in the 5 days before earthquake nucleation; however, since PCA uses the characteristics of principal eigenvalues to determine earthquake related TEC anomalies, it is possible to show that such anomalies existed earlier than this 5-day statistical window.

  20. Time-series MODIS image-based retrieval and distribution analysis of total suspended matter concentrations in Lake Taihu (China).

    Science.gov (United States)

    Zhang, Yuchao; Lin, Shan; Liu, Jianping; Qian, Xin; Ge, Yi

    2010-09-01

    Although there has been considerable effort to use remotely sensed images to provide synoptic maps of total suspended matter (TSM), there are limited studies on universal TSM retrieval models. In this paper, we have developed a TSM retrieval model for Lake Taihu using TSM concentrations measured in situ and a time series of quasi-synchronous MODIS 250 m images from 2005. After simple geometric and atmospheric correction, we found a significant relationship (R = 0.8736, N = 166) between in situ measured TSM concentrations and MODIS band normalization difference of band 3 and band 1. From this, we retrieved TSM concentrations in eight regions of Lake Taihu in 2007 and analyzed the characteristic distribution and variation of TSM. Synoptic maps of model-estimated TSM of 2007 showed clear geographical and seasonal variations. TSM in Central Lake and Southern Lakeshore were consistently higher than in other regions, while TSM in East Taihu was generally the lowest among the regions throughout the year. Furthermore, a wide range of TSM concentrations appeared from winter to summer. TSM in winter could be several times that in summer.

  1. A Novel Ant Colony Algorithm for the Single-Machine Total Weighted Tardiness Problem with Sequence Dependent Setup Times

    Directory of Open Access Journals (Sweden)

    Fardin Ahmadizar

    2011-08-01

    Full Text Available This paper deals with the NP-hard single-machine total weighted tardiness problem with sequence dependent setup times. Incorporating fuzzy sets and genetic operators, a novel ant colony optimization algorithm is developed for the problem. In the proposed algorithm, artificial ants construct solutions as orders of jobs based on the heuristic information as well as pheromone trails. To calculate the heuristic information, three well-known priority rules are adopted as fuzzy sets and then aggregated. When all artificial ants have terminated their constructions, genetic operators such as crossover and mutation are applied to generate new regions of the solution space. A local search is then performed to improve the performance quality of some of the solutions found. Moreover, at run-time the pheromone trails are locally as well as globally updated, and limited between lower and upper bounds. The proposed algorithm is experimented on a set of benchmark problems from the literature and compared with other metaheuristics.

  2. Low Computational Signal Acquisition for GNSS Receivers Using a Resampling Strategy and Variable Circular Correlation Time

    Directory of Open Access Journals (Sweden)

    Yeqing Zhang

    2018-02-01

    Full Text Available For the objective of essentially decreasing computational complexity and time consumption of signal acquisition, this paper explores a resampling strategy and variable circular correlation time strategy specific to broadband multi-frequency GNSS receivers. In broadband GNSS receivers, the resampling strategy is established to work on conventional acquisition algorithms by resampling the main lobe of received broadband signals with a much lower frequency. Variable circular correlation time is designed to adapt to different signal strength conditions and thereby increase the operation flexibility of GNSS signal acquisition. The acquisition threshold is defined as the ratio of the highest and second highest correlation results in the search space of carrier frequency and code phase. Moreover, computational complexity of signal acquisition is formulated by amounts of multiplication and summation operations in the acquisition process. Comparative experiments and performance analysis are conducted on four sets of real GPS L2C signals with different sampling frequencies. The results indicate that the resampling strategy can effectively decrease computation and time cost by nearly 90–94% with just slight loss of acquisition sensitivity. With circular correlation time varying from 10 ms to 20 ms, the time cost of signal acquisition has increased by about 2.7–5.6% per millisecond, with most satellites acquired successfully.

  3. Low Computational Signal Acquisition for GNSS Receivers Using a Resampling Strategy and Variable Circular Correlation Time

    Science.gov (United States)

    Zhang, Yeqing; Wang, Meiling; Li, Yafeng

    2018-01-01

    For the objective of essentially decreasing computational complexity and time consumption of signal acquisition, this paper explores a resampling strategy and variable circular correlation time strategy specific to broadband multi-frequency GNSS receivers. In broadband GNSS receivers, the resampling strategy is established to work on conventional acquisition algorithms by resampling the main lobe of received broadband signals with a much lower frequency. Variable circular correlation time is designed to adapt to different signal strength conditions and thereby increase the operation flexibility of GNSS signal acquisition. The acquisition threshold is defined as the ratio of the highest and second highest correlation results in the search space of carrier frequency and code phase. Moreover, computational complexity of signal acquisition is formulated by amounts of multiplication and summation operations in the acquisition process. Comparative experiments and performance analysis are conducted on four sets of real GPS L2C signals with different sampling frequencies. The results indicate that the resampling strategy can effectively decrease computation and time cost by nearly 90–94% with just slight loss of acquisition sensitivity. With circular correlation time varying from 10 ms to 20 ms, the time cost of signal acquisition has increased by about 2.7–5.6% per millisecond, with most satellites acquired successfully. PMID:29495301

  4. Toward a web-based real-time radiation treatment planning system in a cloud computing environment.

    Science.gov (United States)

    Na, Yong Hum; Suh, Tae-Suk; Kapp, Daniel S; Xing, Lei

    2013-09-21

    To exploit the potential dosimetric advantages of intensity modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT), an in-depth approach is required to provide efficient computing methods. This needs to incorporate clinically related organ specific constraints, Monte Carlo (MC) dose calculations, and large-scale plan optimization. This paper describes our first steps toward a web-based real-time radiation treatment planning system in a cloud computing environment (CCE). The Amazon Elastic Compute Cloud (EC2) with a master node (named m2.xlarge containing 17.1 GB of memory, two virtual cores with 3.25 EC2 Compute Units each, 420 GB of instance storage, 64-bit platform) is used as the backbone of cloud computing for dose calculation and plan optimization. The master node is able to scale the workers on an 'on-demand' basis. MC dose calculation is employed to generate accurate beamlet dose kernels by parallel tasks. The intensity modulation optimization uses total-variation regularization (TVR) and generates piecewise constant fluence maps for each initial beam direction in a distributed manner over the CCE. The optimized fluence maps are segmented into deliverable apertures. The shape of each aperture is iteratively rectified to be a sequence of arcs using the manufacture's constraints. The output plan file from the EC2 is sent to the simple storage service. Three de-identified clinical cancer treatment plans have been studied for evaluating the performance of the new planning platform with 6 MV flattening filter free beams (40 × 40 cm(2)) from the Varian TrueBeam(TM) STx linear accelerator. A CCE leads to speed-ups of up to 14-fold for both dose kernel calculations and plan optimizations in the head and neck, lung, and prostate cancer cases considered in this study. The proposed system relies on a CCE that is able to provide an infrastructure for parallel and distributed computing. The resultant plans from the cloud computing are

  5. Toward a web-based real-time radiation treatment planning system in a cloud computing environment

    International Nuclear Information System (INIS)

    Na, Yong Hum; Kapp, Daniel S; Xing, Lei; Suh, Tae-Suk

    2013-01-01

    To exploit the potential dosimetric advantages of intensity modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT), an in-depth approach is required to provide efficient computing methods. This needs to incorporate clinically related organ specific constraints, Monte Carlo (MC) dose calculations, and large-scale plan optimization. This paper describes our first steps toward a web-based real-time radiation treatment planning system in a cloud computing environment (CCE). The Amazon Elastic Compute Cloud (EC2) with a master node (named m2.xlarge containing 17.1 GB of memory, two virtual cores with 3.25 EC2 Compute Units each, 420 GB of instance storage, 64-bit platform) is used as the backbone of cloud computing for dose calculation and plan optimization. The master node is able to scale the workers on an ‘on-demand’ basis. MC dose calculation is employed to generate accurate beamlet dose kernels by parallel tasks. The intensity modulation optimization uses total-variation regularization (TVR) and generates piecewise constant fluence maps for each initial beam direction in a distributed manner over the CCE. The optimized fluence maps are segmented into deliverable apertures. The shape of each aperture is iteratively rectified to be a sequence of arcs using the manufacture’s constraints. The output plan file from the EC2 is sent to the simple storage service. Three de-identified clinical cancer treatment plans have been studied for evaluating the performance of the new planning platform with 6 MV flattening filter free beams (40 × 40 cm 2 ) from the Varian TrueBeam TM STx linear accelerator. A CCE leads to speed-ups of up to 14-fold for both dose kernel calculations and plan optimizations in the head and neck, lung, and prostate cancer cases considered in this study. The proposed system relies on a CCE that is able to provide an infrastructure for parallel and distributed computing. The resultant plans from the cloud computing are

  6. Computationally determining the salience of decision points for real-time wayfinding support

    Directory of Open Access Journals (Sweden)

    Makoto Takemiya

    2012-06-01

    Full Text Available This study introduces the concept of computational salience to explain the discriminatory efficacy of decision points, which in turn may have applications to providing real-time assistance to users of navigational aids. This research compared algorithms for calculating the computational salience of decision points and validated the results via three methods: high-salience decision points were used to classify wayfinders; salience scores were used to weight a conditional probabilistic scoring function for real-time wayfinder performance classification; and salience scores were correlated with wayfinding-performance metrics. As an exploratory step to linking computational and cognitive salience, a photograph-recognition experiment was conducted. Results reveal a distinction between algorithms useful for determining computational and cognitive saliences. For computational salience, information about the structural integration of decision points is effective, while information about the probability of decision-point traversal shows promise for determining cognitive salience. Limitations from only using structural information and motivations for future work that include non-structural information are elicited.

  7. In-Network Computation is a Dumb Idea Whose Time Has Come

    KAUST Repository

    Sapio, Amedeo; Abdelaziz, Ibrahim; Aldilaijan, Abdulla; Canini, Marco; Kalnis, Panos

    2017-01-01

    Programmable data plane hardware creates new opportunities for infusing intelligence into the network. This raises a fundamental question: what kinds of computation should be delegated to the network? In this paper, we discuss the opportunities and challenges for co-designing data center distributed systems with their network layer. We believe that the time has finally come for offloading part of their computation to execute in-network. However, in-network computation tasks must be judiciously crafted to match the limitations of the network machine architecture of programmable devices. With the help of our experiments on machine learning and graph analytics workloads, we identify that aggregation functions raise opportunities to exploit the limited computation power of networking hardware to lessen network congestion and improve the overall application performance. Moreover, as a proof-of-concept, we propose DAIET, a system that performs in-network data aggregation. Experimental results with an initial prototype show a large data reduction ratio (86.9%-89.3%) and a similar decrease in the workers' computation time.

  8. In-Network Computation is a Dumb Idea Whose Time Has Come

    KAUST Repository

    Sapio, Amedeo

    2017-11-27

    Programmable data plane hardware creates new opportunities for infusing intelligence into the network. This raises a fundamental question: what kinds of computation should be delegated to the network? In this paper, we discuss the opportunities and challenges for co-designing data center distributed systems with their network layer. We believe that the time has finally come for offloading part of their computation to execute in-network. However, in-network computation tasks must be judiciously crafted to match the limitations of the network machine architecture of programmable devices. With the help of our experiments on machine learning and graph analytics workloads, we identify that aggregation functions raise opportunities to exploit the limited computation power of networking hardware to lessen network congestion and improve the overall application performance. Moreover, as a proof-of-concept, we propose DAIET, a system that performs in-network data aggregation. Experimental results with an initial prototype show a large data reduction ratio (86.9%-89.3%) and a similar decrease in the workers\\' computation time.

  9. Real-time computer treatment of THz passive device images with the high image quality

    Science.gov (United States)

    Trofimov, Vyacheslav A.; Trofimov, Vladislav V.

    2012-06-01

    We demonstrate real-time computer code improving significantly the quality of images captured by the passive THz imaging system. The code is not only designed for a THz passive device: it can be applied to any kind of such devices and active THz imaging systems as well. We applied our code for computer processing of images captured by four passive THz imaging devices manufactured by different companies. It should be stressed that computer processing of images produced by different companies requires using the different spatial filters usually. The performance of current version of the computer code is greater than one image per second for a THz image having more than 5000 pixels and 24 bit number representation. Processing of THz single image produces about 20 images simultaneously corresponding to various spatial filters. The computer code allows increasing the number of pixels for processed images without noticeable reduction of image quality. The performance of the computer code can be increased many times using parallel algorithms for processing the image. We develop original spatial filters which allow one to see objects with sizes less than 2 cm. The imagery is produced by passive THz imaging devices which captured the images of objects hidden under opaque clothes. For images with high noise we develop an approach which results in suppression of the noise after using the computer processing and we obtain the good quality image. With the aim of illustrating the efficiency of the developed approach we demonstrate the detection of the liquid explosive, ordinary explosive, knife, pistol, metal plate, CD, ceramics, chocolate and other objects hidden under opaque clothes. The results demonstrate the high efficiency of our approach for the detection of hidden objects and they are a very promising solution for the security problem.

  10. COMPUTING

    CERN Multimedia

    2010-01-01

    Introduction Just two months after the “LHC First Physics” event of 30th March, the analysis of the O(200) million 7 TeV collision events in CMS accumulated during the first 60 days is well under way. The consistency of the CMS computing model has been confirmed during these first weeks of data taking. This model is based on a hierarchy of use-cases deployed between the different tiers and, in particular, the distribution of RECO data to T1s, who then serve data on request to T2s, along a topology known as “fat tree”. Indeed, during this period this model was further extended by almost full “mesh” commissioning, meaning that RECO data were shipped to T2s whenever possible, enabling additional physics analyses compared with the “fat tree” model. Computing activities at the CMS Analysis Facility (CAF) have been marked by a good time response for a load almost evenly shared between ALCA (Alignment and Calibration tasks - highest p...

  11. COMPUTING

    CERN Multimedia

    I. Fisk

    2012-01-01

      Introduction Computing activity has been running at a sustained, high rate as we collect data at high luminosity, process simulation, and begin to process the parked data. The system is functional, though a number of improvements are planned during LS1. Many of the changes will impact users, we hope only in positive ways. We are trying to improve the distributed analysis tools as well as the ability to access more data samples more transparently.  Operations Office Figure 2: Number of events per month, for 2012 Since the June CMS Week, Computing Operations teams successfully completed data re-reconstruction passes and finished the CMSSW_53X MC campaign with over three billion events available in AOD format. Recorded data was successfully processed in parallel, exceeding 1.2 billion raw physics events per month for the first time in October 2012 due to the increase in data-parking rate. In parallel, large efforts were dedicated to WMAgent development and integrati...

  12. Quantification of Artifact Reduction With Real-Time Cine Four-Dimensional Computed Tomography Acquisition Methods

    International Nuclear Information System (INIS)

    Langner, Ulrich W.; Keall, Paul J.

    2010-01-01

    Purpose: To quantify the magnitude and frequency of artifacts in simulated four-dimensional computed tomography (4D CT) images using three real-time acquisition methods- direction-dependent displacement acquisition, simultaneous displacement and phase acquisition, and simultaneous displacement and velocity acquisition- and to compare these methods with commonly used retrospective phase sorting. Methods and Materials: Image acquisition for the four 4D CT methods was simulated with different displacement and velocity tolerances for spheres with radii of 0.5 cm, 1.5 cm, and 2.5 cm, using 58 patient-measured tumors and respiratory motion traces. The magnitude and frequency of artifacts, CT doses, and acquisition times were computed for each method. Results: The mean artifact magnitude was 50% smaller for the three real-time methods than for retrospective phase sorting. The dose was ∼50% lower, but the acquisition time was 20% to 100% longer for the real-time methods than for retrospective phase sorting. Conclusions: Real-time acquisition methods can reduce the frequency and magnitude of artifacts in 4D CT images, as well as the imaging dose, but they increase the image acquisition time. The results suggest that direction-dependent displacement acquisition is the preferred real-time 4D CT acquisition method, because on average, the lowest dose is delivered to the patient and the acquisition time is the shortest for the resulting number and magnitude of artifacts.

  13. An Efficient Integer Coding and Computing Method for Multiscale Time Segment

    Directory of Open Access Journals (Sweden)

    TONG Xiaochong

    2016-12-01

    Full Text Available This article focus on the exist problem and status of current time segment coding, proposed a new set of approach about time segment coding: multi-scale time segment integer coding (MTSIC. This approach utilized the tree structure and the sort by size formed among integer, it reflected the relationship among the multi-scale time segments: order, include/contained, intersection, etc., and finally achieved an unity integer coding processing for multi-scale time. On this foundation, this research also studied the computing method for calculating the time relationships of MTSIC, to support an efficient calculation and query based on the time segment, and preliminary discussed the application method and prospect of MTSIC. The test indicated that, the implement of MTSIC is convenient and reliable, and the transformation between it and the traditional method is convenient, it has the very high efficiency in query and calculating.

  14. COMPUTING

    CERN Multimedia

    I. Fisk

    2011-01-01

    Introduction The Computing Team successfully completed the storage, initial processing, and distribution for analysis of proton-proton data in 2011. There are still a variety of activities ongoing to support winter conference activities and preparations for 2012. Heavy ions The heavy-ion run for 2011 started in early November and has already demonstrated good machine performance and success of some of the more advanced workflows planned for 2011. Data collection will continue until early December. Facilities and Infrastructure Operations Operational and deployment support for WMAgent and WorkQueue+Request Manager components, routinely used in production by Data Operations, are provided. The GlideInWMS and components installation are now deployed at CERN, which is added to the GlideInWMS factory placed in the US. There has been new operational collaboration between the CERN team and the UCSD GlideIn factory operators, covering each others time zones by monitoring/debugging pilot jobs sent from the facto...

  15. An assessment of the real-time application capabilities of the SIFT computer system

    Science.gov (United States)

    Butler, R. W.

    1982-01-01

    The real-time capabilities of the SIFT computer system, a highly reliable multicomputer architecture developed to support the flight controls of a relaxed static stability aircraft, are discussed. The SIFT computer system was designed to meet extremely high reliability requirements and to facilitate a formal proof of its correctness. Although SIFT represents a significant achievement in fault-tolerant system research it presents an unusual and restrictive interface to its users. The characteristics of the user interface and its impact on application system design are assessed.

  16. Theory and computation of disturbance invariant sets for discrete-time linear systems

    Directory of Open Access Journals (Sweden)

    Kolmanovsky Ilya

    1998-01-01

    Full Text Available This paper considers the characterization and computation of invariant sets for discrete-time, time-invariant, linear systems with disturbance inputs whose values are confined to a specified compact set but are otherwise unknown. The emphasis is on determining maximal disturbance-invariant sets X that belong to a specified subset Γ of the state space. Such d-invariant sets have important applications in control problems where there are pointwise-in-time state constraints of the form χ ( t ∈ Γ . One purpose of the paper is to unite and extend in a rigorous way disparate results from the prior literature. In addition there are entirely new results. Specific contributions include: exploitation of the Pontryagin set difference to clarify conceptual matters and simplify mathematical developments, special properties of maximal invariant sets and conditions for their finite determination, algorithms for generating concrete representations of maximal invariant sets, practical computational questions, extension of the main results to general Lyapunov stable systems, applications of the computational techniques to the bounding of state and output response. Results on Lyapunov stable systems are applied to the implementation of a logic-based, nonlinear multimode regulator. For plants with disturbance inputs and state-control constraints it enlarges the constraint-admissible domain of attraction. Numerical examples illustrate the various theoretical and computational results.

  17. Evaluation of the level of skill required of operators of a computer-assisted radiologic total lung capacity measurement system

    International Nuclear Information System (INIS)

    Mazzeo, J.

    1985-01-01

    This research was conducted to obtain information regarding the feasibility of using non-medical personnel to obtain measurements of radiologic total lung capacity (TLC). Operators from each of four groups (general undergraduates, nursing students, medical students, radiologists) differing in the amount of medical training and/or experience reading x-rays, performed each of two tasks. The first task was the measurement of radiologic TLC for a set of twenty x-rays. The second task consisted of tracing the outline of the anatomical structures that must be identified in the execution of the radiologic TLC measurement task. Data from the radiologic TLC measurement task were used to identify possible group differences in the reliability and validity of the measures. The reliability analyses were performed within the framework of Generalizability Theory. While the results are not conclusive, due to small sizes, the analyses suggest that group differences in reliability of the measures, if they exist, are small

  18. An additional reference axis improves femoral rotation alignment in image-free computer navigation assisted total knee arthroplasty.

    Science.gov (United States)

    Inui, Hiroshi; Taketomi, Shuji; Nakamura, Kensuke; Sanada, Takaki; Tanaka, Sakae; Nakagawa, Takumi

    2013-05-01

    Few studies have demonstrated improvement in accuracy of rotational alignment using image-free navigation systems mainly due to the inconsistent registration of anatomical landmarks. We have used an image-free navigation for total knee arthroplasty, which adopts the average algorithm between two reference axes (transepicondylar axis and axis perpendicular to the Whiteside axis) for femoral component rotation control. We hypothesized that addition of another axis (condylar twisting axis measured on a preoperative radiograph) would improve the accuracy. One group using the average algorithm (double-axis group) was compared with the other group using another axis to confirm the accuracy of the average algorithm (triple-axis group). Femoral components were more accurately implanted for rotational alignment in the triple-axis group (ideal: triple-axis group 100%, double-axis group 82%, P<0.05). Copyright © 2013 Elsevier Inc. All rights reserved.

  19. Three-dimensional computer graphics-based ankle morphometry with computerized tomography for total ankle replacement design and positioning.

    Science.gov (United States)

    Kuo, Chien-Chung; Lu, Hsuan-Lun; Leardini, Alberto; Lu, Tung-Wu; Kuo, Mei-Ying; Hsu, Horng-Chaung

    2014-05-01

    Morphometry of the bones of the ankle joint is important for the design of joint replacements and their surgical implantations. However, very little three-dimensional (3D) data are available and not a single study has addressed the Chinese population. Fifty-eight fresh frozen Chinese cadaveric ankle specimens, 26 females, and 32 males, were CT-scanned in the neutral position and their 3D computer graphics-based models were reconstructed. The 3D morphology of the distal tibia/fibula segment and the full talus was analyzed by measuring 31 parameters, defining the relevant dimensions, areas, and volumes from the models. The measurements were compared statistically between sexes and with previously reported data from Caucasian subjects. The results showed that, within a general similarity of ankle morphology between the current Chinese and previous Caucasian subjects groups, there were significant differences in 9 out of the 31 parameters analyzed. From a quantitative comparison with available prostheses designed for the Caucasian population, few of these designs have both tibial and talar components suitable in dimension for the Chinese population. The current data will be helpful for the sizing, design, and surgical positioning of ankle replacements and for surgical instruments, especially for the Chinese population. Copyright © 2013 Wiley Periodicals, Inc.

  20. First Zenith Total Delay and Integrated Water Vapour Estimates from the Near Real-Time GNSS Data Processing Systems at the University of Luxembourg

    Science.gov (United States)

    Ahmed, F.; Teferle, F. N.; Bingley, R. M.

    2012-04-01

    Since September 2011 the University of Luxembourg in collaboration with the University of Nottingham has been setting up two near real-time processing systems for ground-based GNSS data for the provision of zenith total delay (ZTD) and integrated water vapour (IWV) estimates. Both systems are based on Bernese v5.0, use the double-differenced network processing strategy and operate with a 1-hour (NRT1h) and 15-minutes (NRT15m) update cycle. Furthermore, the systems follow the approach of the E-GVAP METO and IES2 systems in that the normal equations for the latest data are combined with those from the previous four updates during the estimation of the ZTDs. NRT1h currently takes the hourly data from over 130 GNSS stations in Europe whereas NRT15m is primarily using the real-time streams of EUREF-IP. Both networks include additional GNSS stations in Luxembourg, Belgium and France. The a priori station coordinates for all of these stem from a moving average computed over the last 20 to 50 days and are based on the precise point positioning processing strategy. In this study we present the first ZTD and IWV estimates obtained from the NRT1h and NRT15m systems in development at the University of Luxembourg. In a preliminary evaluation we compare their performance to the IES2 system at the University of Nottingham and find the IWV estimates to agree at the sub-millimetre level.

  1. The utility of the Philips SRI-100 real time portal imaging device in a case of postoperative irradiation for prevention of heterotopic bone formation following total hip replacement

    International Nuclear Information System (INIS)

    Kiffer, J.D.; Quong, G.; Lawlor, M.; Schumer, W.; Aitken, L.; Wallace, A.

    1994-01-01

    The new Radiation Oncology Department at the Heidelberg Repatriation Hospital in Melbourne, Australia commenced operation in June 1992. As part of quality control the Philips SL-15 linear accelerator was fitted with the Philips SRI-100 Real Time Portal Imaging Device (RTPID), the first such apparatus in Australia. One of its major advantages over older systems is its ability to provide a permanent hard copy of the image of the field treated. The computer image can be immediately manipulated and enhanced on the screen (with respect to such qualities as brightness and contrast) prior to the printing of the hard copy. This is a significant improvement over the more cumbersome older port films that required developing time, without any pre-assessment of the image quality. The utility of the Philips SRI-100 RTPID is demonstrated in the case of a patient irradiated soon after total hip replacement, as prophylaxis against heterotopic bone formation (HBF). The rapidity and quality of image production is a major advantage in these patients where post operative pain may result in positional change between film exposure and image production. Extremely accurate shielding block position is essential to shield the prosthesis(and allow bone ingrowth for fixation) whilst avoiding inadvertent shielding of the areas at risk for HBF. A review of the literature on this topic is provided. 14 refs., 4 figs

  2. Hardware architecture design of image restoration based on time-frequency domain computation

    Science.gov (United States)

    Wen, Bo; Zhang, Jing; Jiao, Zipeng

    2013-10-01

    The image restoration algorithms based on time-frequency domain computation is high maturity and applied widely in engineering. To solve the high-speed implementation of these algorithms, the TFDC hardware architecture is proposed. Firstly, the main module is designed, by analyzing the common processing and numerical calculation. Then, to improve the commonality, the iteration control module is planed for iterative algorithms. In addition, to reduce the computational cost and memory requirements, the necessary optimizations are suggested for the time-consuming module, which include two-dimensional FFT/IFFT and the plural calculation. Eventually, the TFDC hardware architecture is adopted for hardware design of real-time image restoration system. The result proves that, the TFDC hardware architecture and its optimizations can be applied to image restoration algorithms based on TFDC, with good algorithm commonality, hardware realizability and high efficiency.

  3. Joint Time-Frequency-Space Classification of EEG in a Brain-Computer Interface Application

    Directory of Open Access Journals (Sweden)

    Molina Gary N Garcia

    2003-01-01

    Full Text Available Brain-computer interface is a growing field of interest in human-computer interaction with diverse applications ranging from medicine to entertainment. In this paper, we present a system which allows for classification of mental tasks based on a joint time-frequency-space decorrelation, in which mental tasks are measured via electroencephalogram (EEG signals. The efficiency of this approach was evaluated by means of real-time experimentations on two subjects performing three different mental tasks. To do so, a number of protocols for visualization, as well as training with and without feedback, were also developed. Obtained results show that it is possible to obtain good classification of simple mental tasks, in view of command and control, after a relatively small amount of training, with accuracies around 80%, and in real time.

  4. Flexible structure control experiments using a real-time workstation for computer-aided control engineering

    Science.gov (United States)

    Stieber, Michael E.

    1989-01-01

    A Real-Time Workstation for Computer-Aided Control Engineering has been developed jointly by the Communications Research Centre (CRC) and Ruhr-Universitaet Bochum (RUB), West Germany. The system is presently used for the development and experimental verification of control techniques for large space systems with significant structural flexibility. The Real-Time Workstation essentially is an implementation of RUB's extensive Computer-Aided Control Engineering package KEDDC on an INTEL micro-computer running under the RMS real-time operating system. The portable system supports system identification, analysis, control design and simulation, as well as the immediate implementation and test of control systems. The Real-Time Workstation is currently being used by CRC to study control/structure interaction on a ground-based structure called DAISY, whose design was inspired by a reflector antenna. DAISY emulates the dynamics of a large flexible spacecraft with the following characteristics: rigid body modes, many clustered vibration modes with low frequencies and extremely low damping. The Real-Time Workstation was found to be a very powerful tool for experimental studies, supporting control design and simulation, and conducting and evaluating tests withn one integrated environment.

  5. Optimisation of high-quality total ribonucleic acid isolation from cartilaginous tissues for real-time polymerase chain reaction analysis.

    Science.gov (United States)

    Peeters, M; Huang, C L; Vonk, L A; Lu, Z F; Bank, R A; Helder, M N; Doulabi, B Zandieh

    2016-11-01

    Studies which consider the molecular mechanisms of degeneration and regeneration of cartilaginous tissues are seriously hampered by problematic ribonucleic acid (RNA) isolations due to low cell density and the dense, proteoglycan-rich extracellular matrix of cartilage. Proteoglycans tend to co-purify with RNA, they can absorb the full spectrum of UV light and they are potent inhibitors of polymerase chain reaction (PCR). Therefore, the objective of the present study is to compare and optimise different homogenisation methods and RNA isolation kits for an array of cartilaginous tissues. Tissue samples such as the nucleus pulposus (NP), annulus fibrosus (AF), articular cartilage (AC) and meniscus, were collected from goats and homogenised by either the MagNA Lyser or Freezer Mill. RNA of duplicate samples was subsequently isolated by either TRIzol (benchmark), or the RNeasy Lipid Tissue, RNeasy Fibrous Tissue, or Aurum Total RNA Fatty and Fibrous Tissue kits. RNA yield, purity, and integrity were determined and gene expression levels of type II collagen and aggrecan were measured by real-time PCR. No differences between the two homogenisation methods were found. RNA isolation using the RNeasy Fibrous and Lipid kits resulted in the purest RNA (A260/A280 ratio), whereas TRIzol isolations resulted in RNA that is not as pure, and show a larger difference in gene expression of duplicate samples compared with both RNeasy kits. The Aurum kit showed low reproducibility. For the extraction of high-quality RNA from cartilaginous structures, we suggest homogenisation of the samples by the MagNA Lyser. For AC, NP and AF we recommend the RNeasy Fibrous kit, whereas for the meniscus the RNeasy Lipid kit is advised.Cite this article: M. Peeters, C. L. Huang, L. A. Vonk, Z. F. Lu, R. A. Bank, M. N. Helder, B. Zandieh Doulabi. Optimisation of high-quality total ribonucleic acid isolation from cartilaginous tissues for real-time polymerase chain reaction analysis. Bone Joint Res 2016

  6. Using a Cloud Computing System to Reduce Door-to-Balloon Time in Acute ST-Elevation Myocardial Infarction Patients Transferred for Percutaneous Coronary Intervention.

    Science.gov (United States)

    Ho, Chi-Kung; Chen, Fu-Cheng; Chen, Yung-Lung; Wang, Hui-Ting; Lee, Chien-Ho; Chung, Wen-Jung; Lin, Cheng-Jui; Hsueh, Shu-Kai; Hung, Shin-Chiang; Wu, Kuan-Han; Liu, Chu-Feng; Kung, Chia-Te; Cheng, Cheng-I

    2017-01-01

    This study evaluated the impact on clinical outcomes using a cloud computing system to reduce percutaneous coronary intervention hospital door-to-balloon (DTB) time for ST segment elevation myocardial infarction (STEMI). A total of 369 patients before and after implementation of the transfer protocol were enrolled. Of these patients, 262 were transferred through protocol while the other 107 patients were transferred through the traditional referral process. There were no significant differences in DTB time, pain to door of STEMI receiving center arrival time, and pain to balloon time between the two groups. Pain to electrocardiography time in patients with Killip I/II and catheterization laboratory to balloon time in patients with Killip III/IV were significantly reduced in transferred through protocol group compared to in traditional referral process group (both p cloud computing system in our present protocol did not reduce DTB time.

  7. Real time computer control of a nonlinear Multivariable System via Linearization and Stability Analysis

    International Nuclear Information System (INIS)

    Raza, K.S.M.

    2004-01-01

    This paper demonstrates that if a complicated nonlinear, non-square, state-coupled multi variable system is smartly linearized and subjected to a thorough stability analysis then we can achieve our design objectives via a controller which will be quite simple (in term of resource usage and execution time) and very efficient (in terms of robustness). Further the aim is to implement this controller via computer in a real time environment. Therefore first a nonlinear mathematical model of the system is achieved. An intelligent work is done to decouple the multivariable system. Linearization and stability analysis techniques are employed for the development of a linearized and mathematically sound control law. Nonlinearities like the saturation in actuators are also been catered. The controller is then discretized using Runge-Kutta integration. Finally the discretized control law is programmed in a computer in a real time environment. The programme is done in RT -Linux using GNU C for the real time realization of the control scheme. The real time processes, like sampling and controlled actuation, and the non real time processes, like graphical user interface and display, are programmed as different tasks. The issue of inter process communication, between real time and non real time task is addressed quite carefully. The results of this research pursuit are presented graphically. (author)

  8. Five times sit-to-stand test in subjects with total knee replacement: Reliability and relationship with functional mobility tests.

    Science.gov (United States)

    Medina-Mirapeix, Francesc; Vivo-Fernández, Iván; López-Cañizares, Juan; García-Vidal, José A; Benítez-Martínez, Josep Carles; Del Baño-Aledo, María Elena

    2018-01-01

    The objective was to determine the inter-observer and test/retest reliability of the "Five-repetition sit-to-stand" (5STS) test in patients with total knee replacement (TKR). To explore correlation between 5STS and two mobility tests. A reliability study was conducted among 24 (mean age 72.13, S.D. 10.67; 50% were women) outpatients with TKR. They were recruited from a traumatology unit of a public hospital via convenience sampling. A physiotherapist and trauma physician assessed each patient at the same time. The same physiotherapist realized a 5STS second measurement 45-60min after the first one. Reliability was assessed with intraclass correlation coefficients (ICCs) and Bland-Altman plots. Pearson coefficient was calculated to assess the correlation between 5STS, time up to go test (TUG) and four meters gait speed (4MGS). ICC for inter-observer and test-retest reliability of the 5STS were 0.998 (95% confidence interval [CI], 0.995-0.999) and 0.982 (95% CI, 0.959-0.992). Bland-Altman plot inter-observer showed limits between -0.82 and 1.06 with a mean of 0.11 and no heteroscedasticity within the data. Bland-Altman plot for test-retest showed the limits between 1.76 and 4.16, a mean of 1.20 and heteroscedasticity within the data. Pearson correlation coefficient revealed significant correlation between 5STS and TUG (r=0.7, ptest-retest reliability when it is used in people with TKR, and also significant correlation with other functional mobility tests. These findings support the use of 5STS as outcome measure in TKR population. Copyright © 2017 Elsevier B.V. All rights reserved.

  9. Performance evaluation of linear time-series ionospheric Total Electron Content model over low latitude Indian GPS stations

    Science.gov (United States)

    Dabbakuti, J. R. K. Kumar; Venkata Ratnam, D.

    2017-10-01

    Precise modeling of the ionospheric Total Electron Content (TEC) is a critical aspect of Positioning, Navigation, and Timing (PNT) services intended for the Global Navigation Satellite Systems (GNSS) applications as well as Earth Observation System (EOS), satellite communication, and space weather forecasting applications. In this paper, linear time series modeling has been carried out on ionospheric TEC at two different locations at Koneru Lakshmaiah University (KLU), Guntur (geographic 16.44° N, 80.62° E; geomagnetic 7.55° N) and Bangalore (geographic 12.97° N, 77.59° E; geomagnetic 4.53° N) at the northern low-latitude region, for the year 2013 in the 24th solar cycle. The impact of the solar and geomagnetic activity on periodic oscillations of TEC has been investigated. Results confirm that the correlation coefficient of the estimated TEC from the linear model TEC and the observed GPS-TEC is around 93%. Solar activity is the key component that influences ionospheric daily averaged TEC while periodic component reveals the seasonal dependency of TEC. Furthermore, it is observed that the influence of geomagnetic activity component on TEC is different at both the latitudes. The accuracy of the model has been assessed by comparing the International Reference Ionosphere (IRI) 2012 model TEC and TEC measurements. Moreover, the absence of winter anomaly is remarkable, as determined by the Root Mean Square Error (RMSE) between the linear model TEC and GPS-TEC. On the contrary, the IRI2012 model TEC evidently failed to predict the absence of winter anomaly in the Equatorial Ionization Anomaly (EIA) crest region. The outcome of this work will be useful for improving the ionospheric now-casting models under various geophysical conditions.

  10. Characterizing the hydration state of L-threonine in solution using terahertz time-domain attenuated total reflection spectroscopy

    Science.gov (United States)

    Huang, Huachuan; Liu, Qiao; Zhu, Liguo; Li, Zeren

    2018-01-01

    The hydration of biomolecules is closely related to the dynamic process of their functional expression, therefore, characterizing hydration phenomena is a subject of keen interest. However, direct measurements on the global hydration state of biomolecules couldn't have been acquired using traditional techniques such as thermodynamics, ultrasound, microwave spectroscopy or viscosity, etc. In order to realize global hydration characterization of amino acid such as L-threonine, terahertz time-domain attenuated total reflectance spectroscopy (THz-TDS-ATR) was adopted in this paper. By measuring the complex permittivity of L-threonine solutions with various concentrations in the THz region, the hydration state and its concentration dependence were obtained, indicating that the number of hydrous water decreased with the increase of concentration. The hydration number was evaluated to be 17.8 when the molar concentration of L-threonine was 0.34 mol/L, and dropped to 13.2 when the molar concentration increased to 0.84 mol/L, when global hydration was taken into account. According to the proposed direct measurements, it is believed that the THz-TDS-ATR technique is a powerful tool for studying the picosecond molecular dynamics of amino acid solutions.

  11. Copyright and Computer Generated Materials – Is it Time to Reboot the Discussion About Authorship?

    Directory of Open Access Journals (Sweden)

    Anne Fitzgerald

    2013-12-01

    Full Text Available Computer generated materials are ubiquitous and we encounter them on a daily basis, even though most people are unaware that this is the case. Blockbuster movies, television weather reports and telephone directories all include material that is produced by utilising computer technologies. Copyright protection for materials generated by a programmed computer was considered by the Federal Court and Full Court of the Federal Court in Telstra Corporation Limited v Phone Directories Company Pty Ltd.  The court held that the White and Yellow pages telephone directories produced by Telstra and its subsidiary, Sensis, were not protected by copyright because they were computer-generated works which lacked the requisite human authorship.The Copyright Act 1968 (Cth does not contain specific provisions on the subsistence of copyright in computer-generated materials. Although the issue of copyright protection for computer-generated materials has been examined in Australia on two separate occasions by independently-constituted Copyright Law Review Committees over a period of 10 years (1988 to 1998, the Committees’ recommendations for legislative clarification by the enactment of specific amendments to the Copyright Act have not yet been implemented and the legal position remains unclear. In the light of the decision of the Full Federal Court in Telstra v Phone Directories it is timely to consider whether specific provisions should be enacted to clarify the position of computer-generated works under copyright law and, in particular, whether the requirement of human authorship for original works protected under Part III of the Copyright Act should now be reconceptualised to align with the realities of how copyright materials are created in the digital era.

  12. Bound on quantum computation time: Quantum error correction in a critical environment

    International Nuclear Information System (INIS)

    Novais, E.; Mucciolo, Eduardo R.; Baranger, Harold U.

    2010-01-01

    We obtain an upper bound on the time available for quantum computation for a given quantum computer and decohering environment with quantum error correction implemented. First, we derive an explicit quantum evolution operator for the logical qubits and show that it has the same form as that for the physical qubits but with a reduced coupling strength to the environment. Using this evolution operator, we find the trace distance between the real and ideal states of the logical qubits in two cases. For a super-Ohmic bath, the trace distance saturates, while for Ohmic or sub-Ohmic baths, there is a finite time before the trace distance exceeds a value set by the user.

  13. Real-time computing in environmental monitoring of a nuclear power plant

    International Nuclear Information System (INIS)

    Deme, S.; Lang, E.; Nagy, Gy.

    1987-06-01

    A real-time computing method is described for calculating the environmental radiation exposure due to a nuclear power plant both at normal operation and at accident. The effects of the Gaussian plume are recalculated in every ten minutes based on meteorological parameters measured at a height of 20 and 120 m as well as on emission data. At normal operation the quantity of radioactive materials released through the stacks is measured and registered while, at an accident, the source strength is unknown and the calculated relative data are normalized to the values measured at the eight environmental monitoring stations. The doses due to noble gases and to dry and wet deposition as well as the time integral of 131 I concentration are calculated and stored by a professional personal computer for 720 points of the environment of 11 km radius. (author)

  14. The Educator´s Approach to Media Training and Computer Games within Leisure Time of School-children

    OpenAIRE

    MORAVCOVÁ, Dagmar

    2009-01-01

    The paper describes possible ways of approaching computer games playing as part of leisure time of school-children and deals with the significance of media training in leisure time. At first it specifies the concept of leisure time and its functions, then shows some positive and negative effects of the media. It further describes classical computer games, the problem of excess computer game playing and means of prevention. The paper deals with the educator's personality and the importance of ...

  15. Computation of the Short-Time Linear Canonical Transform with Dual Window

    Directory of Open Access Journals (Sweden)

    Lei Huang

    2017-01-01

    Full Text Available The short-time linear canonical transform (STLCT, which maps the time domain signal into the joint time and frequency domain, has recently attracted some attention in the area of signal processing. However, its applications are still limited due to the fact that selection of coefficients of the short-time linear canonical series (STLCS is not unique, because time and frequency elementary functions (together known as basis function of STLCS do not constitute an orthogonal basis. To solve this problem, this paper investigates a dual window solution. First, the nonorthogonal problem that suffered from original window is fulfilled by orthogonal condition with dual window. Then based on the obtained condition, a dual window computation approach of the GT is extended to the STLCS. In addition, simulations verify the validity of the proposed condition and solutions. Furthermore, some possible applied directions are discussed.

  16. Memristive Computational Architecture of an Echo State Network for Real-Time Speech Emotion Recognition

    Science.gov (United States)

    2015-05-28

    recognition is simpler and requires less computational resources compared to other inputs such as facial expressions . The Berlin database of Emotional ...Processing Magazine, IEEE, vol. 18, no. 1, pp. 32– 80, 2001. [15] K. R. Scherer, T. Johnstone, and G. Klasmeyer, “Vocal expression of emotion ...Network for Real-Time Speech- Emotion Recognition 5a. CONTRACT NUMBER IN-HOUSE 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 62788F 6. AUTHOR(S) Q

  17. Winter-time size distribution and source apportionment of total suspended particulate matter and associated metals in Delhi

    Science.gov (United States)

    Srivastava, Arun; Gupta, Sandeep; Jain, V. K.

    2009-03-01

    A study of the winter time size distribution and source apportionment of total suspended particulate matter (TSPM) and associated heavy metal concentrations have been carried out for the city of Delhi. This study is important from the point of view of implementation of compressed natural gas (CNG) as alternate of diesel fuel in the public transport system in 2001 to reduce the pollution level. TSPM were collected using a five-stage cascade impactor at six sites in the winters of 2005-06. The results of size distribution indicate that a major portion (~ 40%) of TSPM concentration is in the form of PM0.7 (heavy metals associated with various size fractions of TSPM. A very good correlation between coarse and fine size fraction of TSPM was observed. It was also observed that the metals associated with coarse particles have more chances of correlation with other metals; rather they are associated with fine particles. Source apportionment was carried out separately in coarse and fine size modes of TSPM by Chemical Mass Balance Receptor Model (CMB8) as well as by Principle Component Analysis (PCA) of SPSS. Source apportionment by PCA reveals that there are two major sources (possibly vehicular and crustal re-suspension) in both coarse and fine size fractions. Results obtained by CMB8 show the dominance of vehicular pollutants and crustal dust in fine and coarse size mode respectively. Noticeably the dominance of vehicular pollutants are now confined to fine size only whilst during pre CNG era it dominated both coarse and fine size mode. An increase of 42.5, 44.4, 48.2, 38.6 and 38.9% in the concentrations of TSPM, PM10.9, coarse particles, fine particles and lead respectively was observed during pre (2001) to post CNG (2005-06) period.

  18. Computational model for real-time determination of tritium inventory in a detritiation installation

    International Nuclear Information System (INIS)

    Bornea, Anisia; Stefanescu, Ioan; Zamfirache, Marius; Stefan, Iuliana; Sofalca, Nicolae; Bidica, Nicolae

    2008-01-01

    Full text: At ICIT Rm.Valcea an experimental pilot plant was built having as main objective the development of a technology for detritiation of heavy water processed in the CANDU-type reactors of the nuclear power plant at Cernavoda, Romania. The aspects related to safeguards and safety for such a detritiation installation being of great importance, a complex computational model has been developed. The model allows real-time calculation of tritium inventory in a working installation. The applied detritiation technology is catalyzed isotopic exchange coupled with cryogenic distillation. Computational models for non-steady working conditions have been developed for each process of isotopic exchange. By coupling these processes tritium inventory can be determined in real-time. The computational model was developed based on the experience gained on the pilot installation. The model uses a set of parameters specific to isotopic exchange processes. These parameters were experimentally determined in the pilot installation. The model is included in the monitoring system and uses as input data the parameters acquired in real-time from automation system of the pilot installation. A friendly interface has been created to visualize the final results as data or graphs. (authors)

  19. A real-time computational model for estimating kinematics of ankle ligaments.

    Science.gov (United States)

    Zhang, Mingming; Davies, T Claire; Zhang, Yanxin; Xie, Sheng Quan

    2016-01-01

    An accurate assessment of ankle ligament kinematics is crucial in understanding the injury mechanisms and can help to improve the treatment of an injured ankle, especially when used in conjunction with robot-assisted therapy. A number of computational models have been developed and validated for assessing the kinematics of ankle ligaments. However, few of them can do real-time assessment to allow for an input into robotic rehabilitation programs. An ankle computational model was proposed and validated to quantify the kinematics of ankle ligaments as the foot moves in real-time. This model consists of three bone segments with three rotational degrees of freedom (DOFs) and 12 ankle ligaments. This model uses inputs for three position variables that can be measured from sensors in many ankle robotic devices that detect postures within the foot-ankle environment and outputs the kinematics of ankle ligaments. Validation of this model in terms of ligament length and strain was conducted by comparing it with published data on cadaver anatomy and magnetic resonance imaging. The model based on ligament lengths and strains is in concurrence with those from the published studies but is sensitive to ligament attachment positions. This ankle computational model has the potential to be used in robot-assisted therapy for real-time assessment of ligament kinematics. The results provide information regarding the quantification of kinematics associated with ankle ligaments related to the disability level and can be used for optimizing the robotic training trajectory.

  20. Polynomial-time computability of the edge-reliability of graphs using Gilbert's formula

    Directory of Open Access Journals (Sweden)

    Marlowe Thomas J.

    1998-01-01

    Full Text Available Reliability is an important consideration in analyzing computer and other communication networks, but current techniques are extremely limited in the classes of graphs which can be analyzed efficiently. While Gilbert's formula establishes a theoretically elegant recursive relationship between the edge reliability of a graph and the reliability of its subgraphs, naive evaluation requires consideration of all sequences of deletions of individual vertices, and for many graphs has time complexity essentially Θ (N!. We discuss a general approach which significantly reduces complexity, encoding subgraph isomorphism in a finer partition by invariants, and recursing through the set of invariants. We illustrate this approach using threshhold graphs, and show that any computation of reliability using Gilbert's formula will be polynomial-time if and only if the number of invariants considered is polynomial; we then show families of graphs with polynomial-time, and non-polynomial reliability computation, and show that these encompass most previously known results. We then codify our approach to indicate how it can be used for other classes of graphs, and suggest several classes to which the technique can be applied.

  1. Extending the length and time scales of Gram–Schmidt Lyapunov vector computations

    Energy Technology Data Exchange (ETDEWEB)

    Costa, Anthony B., E-mail: acosta@northwestern.edu [Department of Chemistry, Northwestern University, Evanston, IL 60208 (United States); Green, Jason R., E-mail: jason.green@umb.edu [Department of Chemistry, Northwestern University, Evanston, IL 60208 (United States); Department of Chemistry, University of Massachusetts Boston, Boston, MA 02125 (United States)

    2013-08-01

    Lyapunov vectors have found growing interest recently due to their ability to characterize systems out of thermodynamic equilibrium. The computation of orthogonal Gram–Schmidt vectors requires multiplication and QR decomposition of large matrices, which grow as N{sup 2} (with the particle count). This expense has limited such calculations to relatively small systems and short time scales. Here, we detail two implementations of an algorithm for computing Gram–Schmidt vectors. The first is a distributed-memory message-passing method using Scalapack. The second uses the newly-released MAGMA library for GPUs. We compare the performance of both codes for Lennard–Jones fluids from N=100 to 1300 between Intel Nahalem/Infiniband DDR and NVIDIA C2050 architectures. To our best knowledge, these are the largest systems for which the Gram–Schmidt Lyapunov vectors have been computed, and the first time their calculation has been GPU-accelerated. We conclude that Lyapunov vector calculations can be significantly extended in length and time by leveraging the power of GPU-accelerated linear algebra.

  2. Extending the length and time scales of Gram–Schmidt Lyapunov vector computations

    International Nuclear Information System (INIS)

    Costa, Anthony B.; Green, Jason R.

    2013-01-01

    Lyapunov vectors have found growing interest recently due to their ability to characterize systems out of thermodynamic equilibrium. The computation of orthogonal Gram–Schmidt vectors requires multiplication and QR decomposition of large matrices, which grow as N 2 (with the particle count). This expense has limited such calculations to relatively small systems and short time scales. Here, we detail two implementations of an algorithm for computing Gram–Schmidt vectors. The first is a distributed-memory message-passing method using Scalapack. The second uses the newly-released MAGMA library for GPUs. We compare the performance of both codes for Lennard–Jones fluids from N=100 to 1300 between Intel Nahalem/Infiniband DDR and NVIDIA C2050 architectures. To our best knowledge, these are the largest systems for which the Gram–Schmidt Lyapunov vectors have been computed, and the first time their calculation has been GPU-accelerated. We conclude that Lyapunov vector calculations can be significantly extended in length and time by leveraging the power of GPU-accelerated linear algebra

  3. A neuro-fuzzy computing technique for modeling hydrological time series

    Science.gov (United States)

    Nayak, P. C.; Sudheer, K. P.; Rangan, D. M.; Ramasastri, K. S.

    2004-05-01

    Intelligent computing tools such as artificial neural network (ANN) and fuzzy logic approaches are proven to be efficient when applied individually to a variety of problems. Recently there has been a growing interest in combining both these approaches, and as a result, neuro-fuzzy computing techniques have evolved. This approach has been tested and evaluated in the field of signal processing and related areas, but researchers have only begun evaluating the potential of this neuro-fuzzy hybrid approach in hydrologic modeling studies. This paper presents the application of an adaptive neuro fuzzy inference system (ANFIS) to hydrologic time series modeling, and is illustrated by an application to model the river flow of Baitarani River in Orissa state, India. An introduction to the ANFIS modeling approach is also presented. The advantage of the method is that it does not require the model structure to be known a priori, in contrast to most of the time series modeling techniques. The results showed that the ANFIS forecasted flow series preserves the statistical properties of the original flow series. The model showed good performance in terms of various statistical indices. The results are highly promising, and a comparative analysis suggests that the proposed modeling approach outperforms ANNs and other traditional time series models in terms of computational speed, forecast errors, efficiency, peak flow estimation etc. It was observed that the ANFIS model preserves the potential of the ANN approach fully, and eases the model building process.

  4. iFLOOD: A Real Time Flood Forecast System for Total Water Modeling in the National Capital Region

    Science.gov (United States)

    Sumi, S. J.; Ferreira, C.

    2017-12-01

    Extreme flood events are the costliest natural hazards impacting the US and frequently cause extensive damages to infrastructure, disruption to economy and loss of lives. In 2016, Hurricane Matthew brought severe damage to South Carolina and demonstrated the importance of accurate flood hazard predictions that requires the integration of riverine and coastal model forecasts for total water prediction in coastal and tidal areas. The National Weather Service (NWS) and the National Ocean Service (NOS) provide flood forecasts for almost the entire US, still there are service-gap areas in tidal regions where no official flood forecast is available. The National capital region is vulnerable to multi-flood hazards including high flows from annual inland precipitation events and surge driven coastal inundation along the tidal Potomac River. Predicting flood levels on such tidal areas in river-estuarine zone is extremely challenging. The main objective of this study is to develop the next generation of flood forecast systems capable of providing accurate and timely information to support emergency management and response in areas impacted by multi-flood hazards. This forecast system is capable of simulating flood levels in the Potomac and Anacostia River incorporating the effects of riverine flooding from the upstream basins, urban storm water and tidal oscillations from the Chesapeake Bay. Flood forecast models developed so far have been using riverine data to simulate water levels for Potomac River. Therefore, the idea is to use forecasted storm surge data from a coastal model as boundary condition of this system. Final output of this validated model will capture the water behavior in river-estuary transition zone far better than the one with riverine data only. The challenge for this iFLOOD forecast system is to understand the complex dynamics of multi-flood hazards caused by storm surges, riverine flow, tidal oscillation and urban storm water. Automated system

  5. Minimally invasive computer-navigated total hip arthroplasty, following the concept of femur first and combined anteversion: design of a blinded randomized controlled trial

    Directory of Open Access Journals (Sweden)

    Woerner Michael

    2011-08-01

    Full Text Available Abstract Background Impingement can be a serious complication after total hip arthroplasty (THA, and is one of the major causes of postoperative pain, dislocation, aseptic loosening, and implant breakage. Minimally invasive THA and computer-navigated surgery were introduced several years ago. We have developed a novel, computer-assisted operation method for THA following the concept of "femur first"/"combined anteversion", which incorporates various aspects of performing a functional optimization of the cup position, and comprehensively addresses range of motion (ROM as well as cup containment and alignment parameters. Hence, the purpose of this study is to assess whether the artificial joint's ROM can be improved by this computer-assisted operation method. Second, the clinical and radiological outcome will be evaluated. Methods/Design A registered patient- and observer-blinded randomized controlled trial will be conducted. Patients between the ages of 50 and 75 admitted for primary unilateral THA will be included. Patients will be randomly allocated to either receive minimally invasive computer-navigated "femur first" THA or the conventional minimally invasive THA procedure. Self-reported functional status and health-related quality of life (questionnaires will be assessed both preoperatively and postoperatively. Perioperative complications will be registered. Radiographic evaluation will take place up to 6 weeks postoperatively with a computed tomography (CT scan. Component position will be evaluated by an independent external institute on a 3D reconstruction of the femur/pelvis using image-processing software. Postoperative ROM will be calculated by an algorithm which automatically determines bony and prosthetic impingements. Discussion In the past, computer navigation has improved the accuracy of component positioning. So far, there are only few objective data quantifying the risks and benefits of computer navigated THA. Therefore, this

  6. Reliability of real-time computing with radiation data feedback at accidental release

    International Nuclear Information System (INIS)

    Deme, S.; Feher, I.; Lang, E.

    1990-01-01

    At the first workshop in 1985 we reported on the real-time dose computing method used at the Paks Nuclear Power Plant and on the telemetric system developed for the normalization of the computed data. At present, the computing method normalized for the telemetric data represents the primary information for deciding on any necessary counter measures in case of a nuclear reactor accident. In this connection we analyzed the reliability of the results obtained in this manner. The points of the analysis were: how the results are influenced by the choice of certain parameters that cannot be determined by direct methods and how the improperly chosen diffusion parameters would distort the determination of environmental radiation parameters normalized on the basis of the measurements ( 131 I activity concentration, gamma dose rate) at points lying at a given distance from the measuring stations. A further source of errors may be that, when determining the level of gamma radiation, the radionuclide doses in the cloud and on the ground surface are measured together by the environmental monitoring stations, whereas these doses appear separately in the computations. At the Paks NPP it is the time integral of the aiborne activity concentration of vapour form 131 I which is determined. This quantity includes neither the other physical and chemical forms of 131 I nor the other isotopes of radioiodine. We gave numerical examples for the uncertainties due to the above factors. As a result, we arrived at the conclusions that there is a need to decide on accident-related measures based on the computing method that the dose uncertainties may reach one order of magnitude for points lying far from the monitoring stations. Different measures are discussed to make the uncertainties significantly lower

  7. Reaction time for processing visual stimulus in a computer-assisted rehabilitation environment.

    Science.gov (United States)

    Sanchez, Yerly; Pinzon, David; Zheng, Bin

    2017-10-01

    To examine the reaction time when human subjects process information presented in the visual channel under both a direct vision and a virtual rehabilitation environment when walking was performed. Visual stimulus included eight math problems displayed on the peripheral vision to seven healthy human subjects in a virtual rehabilitation training (computer-assisted rehabilitation environment (CAREN)) and a direct vision environment. Subjects were required to verbally report the results of these math calculations in a short period of time. Reaction time measured by Tobii Eye tracker and calculation accuracy were recorded and compared between the direct vision and virtual rehabilitation environment. Performance outcomes measured for both groups included reaction time, reading time, answering time and the verbal answer score. A significant difference between the groups was only found for the reaction time (p = .004). Participants had more difficulty recognizing the first equation of the virtual environment. Participants reaction time was faster in the direct vision environment. This reaction time delay should be kept in mind when designing skill training scenarios in virtual environments. This was a pilot project to a series of studies assessing cognition ability of stroke patients who are undertaking a rehabilitation program with a virtual training environment. Implications for rehabilitation Eye tracking is a reliable tool that can be employed in rehabilitation virtual environments. Reaction time changes between direct vision and virtual environment.

  8. Computer simulation of the time evolution of a quenched model alloy in the nucleation region

    International Nuclear Information System (INIS)

    Marro, J.; Lebowitz, J.L.; Kalos, M.H.

    1979-01-01

    The time evolution of the structure function and of the cluster (or grain) distribution following quenching in a model binary alloy with a small concentration of minority atoms is obtained from computer simulations. The structure function S-bar (k,t) obeys a simple scaling relation, S-bar (k,t) = K -3 F (k/K) with K (t) proportional t/sup -a/, a approx. = 0.25, during the latter and larger part of the evolution. During the same period, the mean cluster size grows approximately linearly with time

  9. The Impact of Total Ischemic Time, Donor Age and the Pathway of Donor Death on Graft Outcomes After Deceased Donor Kidney Transplantation.

    Science.gov (United States)

    Wong, Germaine; Teixeira-Pinto, Armando; Chapman, Jeremy R; Craig, Jonathan C; Pleass, Henry; McDonald, Stephen; Lim, Wai H

    2017-06-01

    Prolonged ischemia is a known risk factor for delayed graft function (DGF) and its interaction with donor characteristics, the pathways of donor death, and graft outcomes may have important implications for allocation policies. Using data from the Australian and New Zealand Dialysis and Transplant registry (1994-2013), we examined the relationship between total ischemic time with graft outcomes among recipients who received their first deceased donor kidney transplants. Total ischemic time (in hours) was defined as the time of the donor renal artery interruption or aortic clamp, until the time of release of the clamp on the renal artery in the recipient. A total of 7542 recipients were followed up over a median follow-up time of 5.3 years (interquartile range of 8.2 years). Of these, 1823 (24.6%) experienced DGF and 2553 (33.9%) experienced allograft loss. Recipients with total ischemic time of 14 hours or longer experienced an increased odd of DGF compared with those with total ischemic time less than 14 hours. This effect was most marked among those with older donors (P value for interaction = 0.01). There was a significant interaction between total ischemic time, donor age, and graft loss (P value for interaction = 0.03). There was on average, a 9% increase in the overall risk of graft loss per hour increase in the total ischemic time (adjusted hazard ratio, 1.09; 95% confidence interval, 1.01-1.18; P = 0.02) in recipients with older donation after circulatory death grafts. There is a clinically important interaction between donor age, the pathway of donor death, and total ischemic time on graft outcomes, such that the duration of ischemic time has the greatest impact on graft survival in recipients with older donation after circulatory death kidneys.

  10. COMPUTING

    CERN Multimedia

    M. Kasemann

    CMS relies on a well functioning, distributed computing infrastructure. The Site Availability Monitoring (SAM) and the Job Robot submission have been very instrumental for site commissioning in order to increase availability of more sites such that they are available to participate in CSA07 and are ready to be used for analysis. The commissioning process has been further developed, including "lessons learned" documentation via the CMS twiki. Recently the visualization, presentation and summarizing of SAM tests for sites has been redesigned, it is now developed by the central ARDA project of WLCG. Work to test the new gLite Workload Management System was performed; a 4 times increase in throughput with respect to LCG Resource Broker is observed. CMS has designed and launched a new-generation traffic load generator called "LoadTest" to commission and to keep exercised all data transfer routes in the CMS PhE-DEx topology. Since mid-February, a transfer volume of about 12 P...

  11. Virtual photons in imaginary time: Computing exact Casimir forces via standard numerical electromagnetism techniques

    International Nuclear Information System (INIS)

    Rodriguez, Alejandro; Ibanescu, Mihai; Joannopoulos, J. D.; Johnson, Steven G.; Iannuzzi, Davide

    2007-01-01

    We describe a numerical method to compute Casimir forces in arbitrary geometries, for arbitrary dielectric and metallic materials, with arbitrary accuracy (given sufficient computational resources). Our approach, based on well-established integration of the mean stress tensor evaluated via the fluctuation-dissipation theorem, is designed to directly exploit fast methods developed for classical computational electromagnetism, since it only involves repeated evaluation of the Green's function for imaginary frequencies (equivalently, real frequencies in imaginary time). We develop the approach by systematically examining various formulations of Casimir forces from the previous decades and evaluating them according to their suitability for numerical computation. We illustrate our approach with a simple finite-difference frequency-domain implementation, test it for known geometries such as a cylinder and a plate, and apply it to new geometries. In particular, we show that a pistonlike geometry of two squares sliding between metal walls, in both two and three dimensions with both perfect and realistic metallic materials, exhibits a surprising nonmonotonic ''lateral'' force from the walls

  12. Computational derivation of quantum relativist electromagnetic systems with forward-backward space-time shifts

    International Nuclear Information System (INIS)

    Dubois, Daniel M.

    2000-01-01

    This paper is a continuation of our preceding paper dealing with computational derivation of the Klein-Gordon quantum relativist equation and the Schroedinger quantum equation with forward and backward space-time shifts. The first part introduces forward and backward derivatives for discrete and continuous systems. Generalized complex discrete and continuous derivatives are deduced. The second part deduces the Klein-Gordon equation from the space-time complex continuous derivatives. These derivatives take into account forward-backward space-time shifts related to an internal phase velocity u. The internal group velocity v is related to the speed of light u.v=c 2 and to the external group and phase velocities u.v=v g .v p . Without time shift, the Schroedinger equation is deduced, with a supplementary term, which could represent a reference potential. The third part deduces the Quantum Relativist Klein-Gordon equation for a particle in an electromagnetic field

  13. Evaluation of computed tomography coronary angiography in patients with a high heart rate using 16-slice spiral computed tomography with 0.37-s gantry rotation time

    International Nuclear Information System (INIS)

    Zhang, Shi-Zheng; Hu, Xiu-Hua; Zhang, Qiao-Wei; Huang, Wen-Xin

    2005-01-01

    The aim of our study is to evaluate computed tomography (CT) coronary angiography in patients with a high heart rate using 16-slice spiral CT with 0.37-s gantry rotation time. We compare the image quality of patients whose heart rates were over 70 beats per minute (bpm) with that of patients whose heart rates were 70 bpm or less. Sixty patients with various heart rates underwent retrospectively ECG-gated multislice spiral CT (MSCT) coronary angiography. Two experienced observers who were blind to the heart rates of the patients evaluated all the MSCT coronary angiographic images and calculated the assessable segments. A total of 620 out of 891 (69.6%) segments were satisfactorily visualized. On average, 10.3 coronary artery segments per patient could be evaluated. In 36 patients whose heart rates were below 70 bpm [mean 62.2 bpm±5.32 (standard deviation, SD)], the number of assessable segments was 10.72±2.02 (SD). In the other 24 patients whose heart rates were above 70 bpm [mean 78.6 bpm±8.24 (SD)], the corresponding number was 9.75±1.74 (SD). No statistically significant difference was found in these two subgroups' t test, P>0.05. The new generation of 16-slice spiral CT with 0.37-s rotation time can satisfactorily evaluate the coronary arteries of patients with high heart rates (above 70 bpm, up to 102 bpm). (orig.)

  14. Resolving time of scintillation camera-computer system and methods of correction for counting loss, 2

    International Nuclear Information System (INIS)

    Iinuma, Takeshi; Fukuhisa, Kenjiro; Matsumoto, Toru

    1975-01-01

    Following the previous work, counting-rate performance of camera-computer systems was investigated for two modes of data acquisition. The first was the ''LIST'' mode in which image data and timing signals were sequentially stored on magnetic disk or tape via a buffer memory. The second was the ''HISTOGRAM'' mode in which image data were stored in a core memory as digital images and then the images were transfered to magnetic disk or tape by the signal of frame timing. Firstly, the counting-rates stored in the buffer memory was measured as a function of display event-rates of the scintillation camera for the two modes. For both modes, stored counting-rated (M) were expressed by the following formula: M=N(1-Ntau) where N was the display event-rates of the camera and tau was the resolving time including analog-to-digital conversion time and memory cycle time. The resolving time for each mode may have been different, but it was about 10 μsec for both modes in our computer system (TOSBAC 3400 model 31). Secondly, the date transfer speed from the buffer memory to the external memory such as magnetic disk or tape was considered for the two modes. For the ''LIST'' mode, the maximum value of stored counting-rates from the camera was expressed in terms of size of the buffer memory, access time and data transfer-rate of the external memory. For the ''HISTOGRAM'' mode, the minimum time of the frame was determined by size of the buffer memory, access time and transfer rate of the external memory. In our system, the maximum value of stored counting-rates were about 17,000 counts/sec. with the buffer size of 2,000 words, and minimum frame time was about 130 msec. with the buffer size of 1024 words. These values agree well with the calculated ones. From the author's present analysis, design of the camera-computer system becomes possible for quantitative dynamic imaging and future improvements are suggested. (author)

  15. Computation and Communication Evaluation of an Authentication Mechanism for Time-Triggered Networked Control Systems

    Science.gov (United States)

    Martins, Goncalo; Moondra, Arul; Dubey, Abhishek; Bhattacharjee, Anirban; Koutsoukos, Xenofon D.

    2016-01-01

    In modern networked control applications, confidentiality and integrity are important features to address in order to prevent against attacks. Moreover, network control systems are a fundamental part of the communication components of current cyber-physical systems (e.g., automotive communications). Many networked control systems employ Time-Triggered (TT) architectures that provide mechanisms enabling the exchange of precise and synchronous messages. TT systems have computation and communication constraints, and with the aim to enable secure communications in the network, it is important to evaluate the computational and communication overhead of implementing secure communication mechanisms. This paper presents a comprehensive analysis and evaluation of the effects of adding a Hash-based Message Authentication (HMAC) to TT networked control systems. The contributions of the paper include (1) the analysis and experimental validation of the communication overhead, as well as a scalability analysis that utilizes the experimental result for both wired and wireless platforms and (2) an experimental evaluation of the computational overhead of HMAC based on a kernel-level Linux implementation. An automotive application is used as an example, and the results show that it is feasible to implement a secure communication mechanism without interfering with the existing automotive controller execution times. The methods and results of the paper can be used for evaluating the performance impact of security mechanisms and, thus, for the design of secure wired and wireless TT networked control systems. PMID:27463718

  16. Cloud computing platform for real-time measurement and verification of energy performance

    International Nuclear Information System (INIS)

    Ke, Ming-Tsun; Yeh, Chia-Hung; Su, Cheng-Jie

    2017-01-01

    Highlights: • Application of PSO algorithm can improve the accuracy of the baseline model. • M&V cloud platform automatically calculates energy performance. • M&V cloud platform can be applied in all energy conservation measures. • Real-time operational performance can be monitored through the proposed platform. • M&V cloud platform facilitates the development of EE programs and ESCO industries. - Abstract: Nations worldwide are vigorously promoting policies to improve energy efficiency. The use of measurement and verification (M&V) procedures to quantify energy performance is an essential topic in this field. Currently, energy performance M&V is accomplished via a combination of short-term on-site measurements and engineering calculations. This requires extensive amounts of time and labor and can result in a discrepancy between actual energy savings and calculated results. In addition, the M&V period typically lasts for periods as long as several months or up to a year, the failure to immediately detect abnormal energy performance not only decreases energy performance, results in the inability to make timely correction, and misses the best opportunity to adjust or repair equipment and systems. In this study, a cloud computing platform for the real-time M&V of energy performance is developed. On this platform, particle swarm optimization and multivariate regression analysis are used to construct accurate baseline models. Instantaneous and automatic calculations of the energy performance and access to long-term, cumulative information about the energy performance are provided via a feature that allows direct uploads of the energy consumption data. Finally, the feasibility of this real-time M&V cloud platform is tested for a case study involving improvements to a cold storage system in a hypermarket. Cloud computing platform for real-time energy performance M&V is applicable to any industry and energy conservation measure. With the M&V cloud platform, real-time

  17. Automated selection of brain regions for real-time fMRI brain-computer interfaces

    Science.gov (United States)

    Lührs, Michael; Sorger, Bettina; Goebel, Rainer; Esposito, Fabrizio

    2017-02-01

    Objective. Brain-computer interfaces (BCIs) implemented with real-time functional magnetic resonance imaging (rt-fMRI) use fMRI time-courses from predefined regions of interest (ROIs). To reach best performances, localizer experiments and on-site expert supervision are required for ROI definition. To automate this step, we developed two unsupervised computational techniques based on the general linear model (GLM) and independent component analysis (ICA) of rt-fMRI data, and compared their performances on a communication BCI. Approach. 3 T fMRI data of six volunteers were re-analyzed in simulated real-time. During a localizer run, participants performed three mental tasks following visual cues. During two communication runs, a letter-spelling display guided the subjects to freely encode letters by performing one of the mental tasks with a specific timing. GLM- and ICA-based procedures were used to decode each letter, respectively using compact ROIs and whole-brain distributed spatio-temporal patterns of fMRI activity, automatically defined from subject-specific or group-level maps. Main results. Letter-decoding performances were comparable to supervised methods. In combination with a similarity-based criterion, GLM- and ICA-based approaches successfully decoded more than 80% (average) of the letters. Subject-specific maps yielded optimal performances. Significance. Automated solutions for ROI selection may help accelerating the translation of rt-fMRI BCIs from research to clinical applications.

  18. A sub-cubic time algorithm for computing the quartet distance between two general trees

    DEFF Research Database (Denmark)

    Nielsen, Jesper; Kristensen, Anders Kabell; Mailund, Thomas

    2011-01-01

    Background When inferring phylogenetic trees different algorithms may give different trees. To study such effects a measure for the distance between two trees is useful. Quartet distance is one such measure, and is the number of quartet topologies that differ between two trees. Results We have...... derived a new algorithm for computing the quartet distance between a pair of general trees, i.e. trees where inner nodes can have any degree ≥ 3. The time and space complexity of our algorithm is sub-cubic in the number of leaves and does not depend on the degree of the inner nodes. This makes...... it the fastest algorithm so far for computing the quartet distance between general trees independent of the degree of the inner nodes. Conclusions We have implemented our algorithm and two of the best competitors. Our new algorithm is significantly faster than the competition and seems to run in close...

  19. Computer vision system in real-time for color determination on flat surface food

    Directory of Open Access Journals (Sweden)

    Erick Saldaña

    2013-03-01

    Full Text Available Artificial vision systems also known as computer vision are potent quality inspection tools, which can be applied in pattern recognition for fruits and vegetables analysis. The aim of this research was to design, implement and calibrate a new computer vision system (CVS in real-time for the color measurement on flat surface food. For this purpose was designed and implemented a device capable of performing this task (software and hardware, which consisted of two phases: a image acquisition and b image processing and analysis. Both the algorithm and the graphical interface (GUI were developed in Matlab. The CVS calibration was performed using a conventional colorimeter (Model CIEL* a* b*, where were estimated the errors of the color parameters: eL* = 5.001%, and ea* = 2.287%, and eb* = 4.314 % which ensure adequate and efficient automation application in industrial processes in the quality control in the food industry sector.

  20. Computer vision system in real-time for color determination on flat surface food

    Directory of Open Access Journals (Sweden)

    Erick Saldaña

    2013-01-01

    Full Text Available Artificial vision systems also known as computer vision are potent quality inspection tools, which can be applied in pattern recognition for fruits and vegetables analysis. The aim of this research was to design, implement and calibrate a new computer vision system (CVS in real - time f or the color measurement on flat surface food. For this purpose was designed and implemented a device capable of performing this task (software and hardware, which consisted of two phases: a image acquisition and b image processing and analysis. Both th e algorithm and the graphical interface (GUI were developed in Matlab. The CVS calibration was performed using a conventional colorimeter (Model CIEL* a* b*, where were estimated the errors of the color parameters: e L* = 5.001%, and e a* = 2.287%, and e b* = 4.314 % which ensure adequate and efficient automation application in industrial processes in the quality control in the food industry sector.

  1. A stable computational scheme for stiff time-dependent constitutive equations

    International Nuclear Information System (INIS)

    Shih, C.F.; Delorenzi, H.G.; Miller, A.K.

    1977-01-01

    Viscoplasticity and creep type constitutive equations are increasingly being employed in finite element codes for evaluating the deformation of high temperature structural members. These constitutive equations frequently exhibit stiff regimes which makes an analytical assessment of the structure very costly. A computational scheme for handling deformation in stiff regimes is proposed in this paper. By the finite element discretization, the governing partial differential equations in the spatial (x) and time (t) variables are reduced to a system of nonlinear ordinary differential equations in the independent variable t. The constitutive equations are expanded in a Taylor's series about selected values of t. The resulting system of differential equations are then integrated by an implicit scheme which employs a predictor technique to initiate the Newton-Raphson procedure. To examine the stability and accuracy of the computational scheme, a series of calculations were carried out for uniaxial specimens and thick wall tubes subjected to mechanical and thermal loading. (Auth.)

  2. Real-time dynamics of lattice gauge theories with a few-qubit quantum computer

    Science.gov (United States)

    Martinez, Esteban A.; Muschik, Christine A.; Schindler, Philipp; Nigg, Daniel; Erhard, Alexander; Heyl, Markus; Hauke, Philipp; Dalmonte, Marcello; Monz, Thomas; Zoller, Peter; Blatt, Rainer

    2016-06-01

    Gauge theories are fundamental to our understanding of interactions between the elementary constituents of matter as mediated by gauge bosons. However, computing the real-time dynamics in gauge theories is a notorious challenge for classical computational methods. This has recently stimulated theoretical effort, using Feynman’s idea of a quantum simulator, to devise schemes for simulating such theories on engineered quantum-mechanical devices, with the difficulty that gauge invariance and the associated local conservation laws (Gauss laws) need to be implemented. Here we report the experimental demonstration of a digital quantum simulation of a lattice gauge theory, by realizing (1 + 1)-dimensional quantum electrodynamics (the Schwinger model) on a few-qubit trapped-ion quantum computer. We are interested in the real-time evolution of the Schwinger mechanism, describing the instability of the bare vacuum due to quantum fluctuations, which manifests itself in the spontaneous creation of electron-positron pairs. To make efficient use of our quantum resources, we map the original problem to a spin model by eliminating the gauge fields in favour of exotic long-range interactions, which can be directly and efficiently implemented on an ion trap architecture. We explore the Schwinger mechanism of particle-antiparticle generation by monitoring the mass production and the vacuum persistence amplitude. Moreover, we track the real-time evolution of entanglement in the system, which illustrates how particle creation and entanglement generation are directly related. Our work represents a first step towards quantum simulation of high-energy theories using atomic physics experiments—the long-term intention is to extend this approach to real-time quantum simulations of non-Abelian lattice gauge theories.

  3. Fault tolerant distributed real time computer systems for I and C of prototype fast breeder reactor

    Energy Technology Data Exchange (ETDEWEB)

    Manimaran, M., E-mail: maran@igcar.gov.in; Shanmugam, A.; Parimalam, P.; Murali, N.; Satya Murty, S.A.V.

    2014-03-15

    Highlights: • Architecture of distributed real time computer system (DRTCS) used in I and C of PFBR is explained. • Fault tolerant (hot standby) architecture, fault detection and switch over are detailed. • Scaled down model was used to study functional and performance requirements of DRTCS. • Quality of service parameters for scaled down model was critically studied. - Abstract: Prototype fast breeder reactor (PFBR) is in the advanced stage of construction at Kalpakkam, India. Three-tier architecture is adopted for instrumentation and control (I and C) of PFBR wherein bottom tier consists of real time computer (RTC) systems, middle tier consists of process computers and top tier constitutes of display stations. These RTC systems are geographically distributed and networked together with process computers and display stations. Hot standby architecture comprising of dual redundant RTC systems with switch over logic system is deployed in order to achieve fault tolerance. Fault tolerant dual redundant network connectivity is provided in each RTC system and TCP/IP protocol is selected for network communication. In order to assess the performance of distributed RTC systems, scaled down model was developed with 9 representative systems and nearly 15% of I and C signals of PFBR were connected and monitored. Functional and performance testing were carried out for each RTC system and the fault tolerant characteristics were studied by creating various faults into the system and observed the performance. Various quality of service parameters like connection establishment delay, priority parameter, transit delay, throughput, residual error ratio, etc., are critically studied for the network.

  4. 21 CFR 10.20 - Submission of documents to Division of Dockets Management; computation of time; availability for...

    Science.gov (United States)

    2010-04-01

    ... Management; computation of time; availability for public disclosure. 10.20 Section 10.20 Food and Drugs FOOD... Management; computation of time; availability for public disclosure. (a) A submission to the Division of Dockets Management of a petition, comment, objection, notice, compilation of information, or any other...

  5. COMPUTING

    CERN Multimedia

    M. Kasemann

    Overview In autumn the main focus was to process and handle CRAFT data and to perform the Summer08 MC production. The operational aspects were well covered by regular Computing Shifts, experts on duty and Computing Run Coordination. At the Computing Resource Board (CRB) in October a model to account for service work at Tier 2s was approved. The computing resources for 2009 were reviewed for presentation at the C-RRB. The quarterly resource monitoring is continuing. Facilities/Infrastructure operations Operations during CRAFT data taking ran fine. This proved to be a very valuable experience for T0 workflows and operations. The transfers of custodial data to most T1s went smoothly. A first round of reprocessing started at the Tier-1 centers end of November; it will take about two weeks. The Computing Shifts procedure was tested full scale during this period and proved to be very efficient: 30 Computing Shifts Persons (CSP) and 10 Computing Resources Coordinators (CRC). The shift program for the shut down w...

  6. Computer modelling of structures with account of the construction stages and the time dependent material properties

    Directory of Open Access Journals (Sweden)

    Traykov Alexander

    2015-01-01

    Full Text Available Numerical studies are performed on computer models taking into account the stages of construction and time dependent material properties defined in two forms. A 2D model of three storey two spans frame is created. The first form deals with material defined in the usual design practice way - without taking into account the time dependent properties of the concrete. The second form creep and shrinkage of the concrete are taken into account. Displacements and internal forces in specific elements and sections are reported. The influence of the time dependent material properties on the displacement and the internal forces in the main structural elements is tracked down. The results corresponding to the two forms of material definition are compared together as well as with the results obtained by the usual design calculations. Conclusions on the influence of the concrete creep and shrinkage during the construction towards structural behaviour are made.

  7. Computational imaging with multi-camera time-of-flight systems

    KAUST Repository

    Shrestha, Shikhar

    2016-07-11

    Depth cameras are a ubiquitous technology used in a wide range of applications, including robotic and machine vision, human computer interaction, autonomous vehicles as well as augmented and virtual reality. In this paper, we explore the design and applications of phased multi-camera time-of-flight (ToF) systems. We develop a reproducible hardware system that allows for the exposure times and waveforms of up to three cameras to be synchronized. Using this system, we analyze waveform interference between multiple light sources in ToF applications and propose simple solutions to this problem. Building on the concept of orthogonal frequency design, we demonstrate state-of-the-art results for instantaneous radial velocity capture via Doppler time-of-flight imaging and we explore new directions for optically probing global illumination, for example by de-scattering dynamic scenes and by non-line-of-sight motion detection via frequency gating. © 2016 ACM.

  8. hctsa: A Computational Framework for Automated Time-Series Phenotyping Using Massive Feature Extraction.

    Science.gov (United States)

    Fulcher, Ben D; Jones, Nick S

    2017-11-22

    Phenotype measurements frequently take the form of time series, but we currently lack a systematic method for relating these complex data streams to scientifically meaningful outcomes, such as relating the movement dynamics of organisms to their genotype or measurements of brain dynamics of a patient to their disease diagnosis. Previous work addressed this problem by comparing implementations of thousands of diverse scientific time-series analysis methods in an approach termed highly comparative time-series analysis. Here, we introduce hctsa, a software tool for applying this methodological approach to data. hctsa includes an architecture for computing over 7,700 time-series features and a suite of analysis and visualization algorithms to automatically select useful and interpretable time-series features for a given application. Using exemplar applications to high-throughput phenotyping experiments, we show how hctsa allows researchers to leverage decades of time-series research to quantify and understand informative structure in time-series data. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  9. Preliminary Investigation of Time Remaining Display on the Computer-based Emergency Operating Procedure

    Science.gov (United States)

    Suryono, T. J.; Gofuku, A.

    2018-02-01

    One of the important thing in the mitigation of accidents in nuclear power plant accidents is time management. The accidents should be resolved as soon as possible in order to prevent the core melting and the release of radioactive material to the environment. In this case, operators should follow the emergency operating procedure related with the accident, in step by step order and in allowable time. Nowadays, the advanced main control rooms are equipped with computer-based procedures (CBPs) which is make it easier for operators to do their tasks of monitoring and controlling the reactor. However, most of the CBPs do not include the time remaining display feature which informs operators of time available for them to execute procedure steps and warns them if the they reach the time limit. Furthermore, the feature will increase the awareness of operators about their current situation in the procedure. This paper investigates this issue. The simplified of emergency operating procedure (EOP) of steam generator tube rupture (SGTR) accident of PWR plant is applied. In addition, the sequence of actions on each step of the procedure is modelled using multilevel flow modelling (MFM) and influenced propagation rule. The prediction of action time on each step is acquired based on similar case accidents and the Support Vector Regression. The derived time will be processed and then displayed on a CBP user interface.

  10. Early assessment of the likely cost-effectiveness of a new technology: A Markov model with probabilistic sensitivity analysis of computer-assisted total knee replacement.

    Science.gov (United States)

    Dong, Hengjin; Buxton, Martin

    2006-01-01

    The objective of this study is to apply a Markov model to compare cost-effectiveness of total knee replacement (TKR) using computer-assisted surgery (CAS) with that of TKR using a conventional manual method in the absence of formal clinical trial evidence. A structured search was carried out to identify evidence relating to the clinical outcome, cost, and effectiveness of TKR. Nine Markov states were identified based on the progress of the disease after TKR. Effectiveness was expressed by quality-adjusted life years (QALYs). The simulation was carried out initially for 120 cycles of a month each, starting with 1,000 TKRs. A discount rate of 3.5 percent was used for both cost and effectiveness in the incremental cost-effectiveness analysis. Then, a probabilistic sensitivity analysis was carried out using a Monte Carlo approach with 10,000 iterations. Computer-assisted TKR was a long-term cost-effective technology, but the QALYs gained were small. After the first 2 years, the incremental cost per QALY of computer-assisted TKR was dominant because of cheaper and more QALYs. The incremental cost-effectiveness ratio (ICER) was sensitive to the "effect of CAS," to the CAS extra cost, and to the utility of the state "Normal health after primary TKR," but it was not sensitive to utilities of other Markov states. Both probabilistic and deterministic analyses produced similar cumulative serious or minor complication rates and complex or simple revision rates. They also produced similar ICERs. Compared with conventional TKR, computer-assisted TKR is a cost-saving technology in the long-term and may offer small additional QALYs. The "effect of CAS" is to reduce revision rates and complications through more accurate and precise alignment, and although the conclusions from the model, even when allowing for a full probabilistic analysis of uncertainty, are clear, the "effect of CAS" on the rate of revisions awaits long-term clinical evidence.

  11. A users manual for a computer program which calculates time optical geocentric transfers using solar or nuclear electric and high thrust propulsion

    Science.gov (United States)

    Sackett, L. L.; Edelbaum, T. N.; Malchow, H. L.

    1974-01-01

    This manual is a guide for using a computer program which calculates time optimal trajectories for high-and low-thrust geocentric transfers. Either SEP or NEP may be assumed and a one or two impulse, fixed total delta V, initial high thrust phase may be included. Also a single impulse of specified delta V may be included after the low thrust state. The low thrust phase utilizes equinoctial orbital elements to avoid the classical singularities and Kryloff-Boguliuboff averaging to help insure more rapid computation time. The program is written in FORTRAN 4 in double precision for use on an IBM 360 computer. The manual includes a description of the problem treated, input/output information, examples of runs, and source code listings.

  12. The reliable solution and computation time of variable parameters logistic model

    Science.gov (United States)

    Wang, Pengfei; Pan, Xinnong

    2018-05-01

    The study investigates the reliable computation time (RCT, termed as T c) by applying a double-precision computation of a variable parameters logistic map (VPLM). Firstly, by using the proposed method, we obtain the reliable solutions for the logistic map. Secondly, we construct 10,000 samples of reliable experiments from a time-dependent non-stationary parameters VPLM and then calculate the mean T c. The results indicate that, for each different initial value, the T cs of the VPLM are generally different. However, the mean T c trends to a constant value when the sample number is large enough. The maximum, minimum, and probable distribution functions of T c are also obtained, which can help us to identify the robustness of applying a nonlinear time series theory to forecasting by using the VPLM output. In addition, the T c of the fixed parameter experiments of the logistic map is obtained, and the results suggest that this T c matches the theoretical formula-predicted value.

  13. Computing moment to moment BOLD activation for real-time neurofeedback

    Science.gov (United States)

    Hinds, Oliver; Ghosh, Satrajit; Thompson, Todd W.; Yoo, Julie J.; Whitfield-Gabrieli, Susan; Triantafyllou, Christina; Gabrieli, John D.E.

    2013-01-01

    Estimating moment to moment changes in blood oxygenation level dependent (BOLD) activation levels from functional magnetic resonance imaging (fMRI) data has applications for learned regulation of regional activation, brain state monitoring, and brain-machine interfaces. In each of these contexts, accurate estimation of the BOLD signal in as little time as possible is desired. This is a challenging problem due to the low signal-to-noise ratio of fMRI data. Previous methods for real-time fMRI analysis have either sacrificed the ability to compute moment to moment activation changes by averaging several acquisitions into a single activation estimate or have sacrificed accuracy by failing to account for prominent sources of noise in the fMRI signal. Here we present a new method for computing the amount of activation present in a single fMRI acquisition that separates moment to moment changes in the fMRI signal intensity attributable to neural sources from those due to noise, resulting in a feedback signal more reflective of neural activation. This method computes an incremental general linear model fit to the fMRI timeseries, which is used to calculate the expected signal intensity at each new acquisition. The difference between the measured intensity and the expected intensity is scaled by the variance of the estimator in order to transform this residual difference into a statistic. Both synthetic and real data were used to validate this method and compare it to the only other published real-time fMRI method. PMID:20682350

  14. Region-oriented CT image representation for reducing computing time of Monte Carlo simulations

    International Nuclear Information System (INIS)

    Sarrut, David; Guigues, Laurent

    2008-01-01

    Purpose. We propose a new method for efficient particle transportation in voxelized geometry for Monte Carlo simulations. We describe its use for calculating dose distribution in CT images for radiation therapy. Material and methods. The proposed approach, based on an implicit volume representation named segmented volume, coupled with an adapted segmentation procedure and a distance map, allows us to minimize the number of boundary crossings, which slows down simulation. The method was implemented with the GEANT4 toolkit and compared to four other methods: One box per voxel, parameterized volumes, octree-based volumes, and nested parameterized volumes. For each representation, we compared dose distribution, time, and memory consumption. Results. The proposed method allows us to decrease computational time by up to a factor of 15, while keeping memory consumption low, and without any modification of the transportation engine. Speeding up is related to the geometry complexity and the number of different materials used. We obtained an optimal number of steps with removal of all unnecessary steps between adjacent voxels sharing a similar material. However, the cost of each step is increased. When the number of steps cannot be decreased enough, due for example, to the large number of material boundaries, such a method is not considered suitable. Conclusion. This feasibility study shows that optimizing the representation of an image in memory potentially increases computing efficiency. We used the GEANT4 toolkit, but we could potentially use other Monte Carlo simulation codes. The method introduces a tradeoff between speed and geometry accuracy, allowing computational time gain. However, simulations with GEANT4 remain slow and further work is needed to speed up the procedure while preserving the desired accuracy

  15. Real-time computation of parameter fitting and image reconstruction using graphical processing units

    Science.gov (United States)

    Locans, Uldis; Adelmann, Andreas; Suter, Andreas; Fischer, Jannis; Lustermann, Werner; Dissertori, Günther; Wang, Qiulin

    2017-06-01

    In recent years graphical processing units (GPUs) have become a powerful tool in scientific computing. Their potential to speed up highly parallel applications brings the power of high performance computing to a wider range of users. However, programming these devices and integrating their use in existing applications is still a challenging task. In this paper we examined the potential of GPUs for two different applications. The first application, created at Paul Scherrer Institut (PSI), is used for parameter fitting during data analysis of μSR (muon spin rotation, relaxation and resonance) experiments. The second application, developed at ETH, is used for PET (Positron Emission Tomography) image reconstruction and analysis. Applications currently in use were examined to identify parts of the algorithms in need of optimization. Efficient GPU kernels were created in order to allow applications to use a GPU, to speed up the previously identified parts. Benchmarking tests were performed in order to measure the achieved speedup. During this work, we focused on single GPU systems to show that real time data analysis of these problems can be achieved without the need for large computing clusters. The results show that the currently used application for parameter fitting, which uses OpenMP to parallelize calculations over multiple CPU cores, can be accelerated around 40 times through the use of a GPU. The speedup may vary depending on the size and complexity of the problem. For PET image analysis, the obtained speedups of the GPU version were more than × 40 larger compared to a single core CPU implementation. The achieved results show that it is possible to improve the execution time by orders of magnitude.

  16. 38 CFR 3.22 - DIC benefits for survivors of certain veterans rated totally disabled at time of death.

    Science.gov (United States)

    2010-07-01

    ... benefits under paragraph (a) of this section receives any money or property pursuant to a judicial... amount of money received and the fair market value of the property received. The provisions of this... veteran. The amount to be reported is the total of the amount of money received and the fair market value...

  17. A real-time PCR method for quantification of the total and major variant strains of the deformed wing virus.

    Directory of Open Access Journals (Sweden)

    Emma L Bradford

    Full Text Available European honey bees (Apis mellifera are critically important to global food production by virtue of their pollination services but are severely threatened by deformed wing virus (DWV especially in the presence of the external parasite Varroa destructor. DWV exists as many viral strains with the two major variants (DWV-A and DWV-B varying in virulence. A single plasmid standard was constructed containing three sections for the specific determination of DWV-A (VP2 capsid region, DWV-B (IRES and a conserved region suitable for total DWV (helicase region. The assays were confirmed as specific and discriminatory with limits of detections of 25, 25 and 50 genome equivalents for DWV-A, DWV-B and total-DWV, respectively. The methods were successfully tested on Apis mellifera and V. destructor samples with varying DWV profiles. The new method determined a more accurate total DWV titre in samples with substantial DWV-B than the method currently described in the COLOSS Beebook. The proposed assays could be utilized for the screening of large quantities of bee material for both a total DWV load overview along with more detailed investigations into DWV-A and DWV-B profiles.

  18. Total internal reflection fluorescence (TIRF) microscopy for real-time imaging of nanoparticle-cell plasma membrane interaction

    DEFF Research Database (Denmark)

    Parhamifar, Ladan; Moghimi, Seyed Moien

    2012-01-01

    Nanoparticulate systems are widely used for site-specific drug and gene delivery as well as for medical imaging. The mode of nanoparticle-cell interaction may have a significant effect on the pathway of nanoparticle internalization and subsequent intracellular trafficking. Total internal reflection...

  19. COMPUTING

    CERN Multimedia

    P. McBride

    The Computing Project is preparing for a busy year where the primary emphasis of the project moves towards steady operations. Following the very successful completion of Computing Software and Analysis challenge, CSA06, last fall, we have reorganized and established four groups in computing area: Commissioning, User Support, Facility/Infrastructure Operations and Data Operations. These groups work closely together with groups from the Offline Project in planning for data processing and operations. Monte Carlo production has continued since CSA06, with about 30M events produced each month to be used for HLT studies and physics validation. Monte Carlo production will continue throughout the year in the preparation of large samples for physics and detector studies ramping to 50 M events/month for CSA07. Commissioning of the full CMS computing system is a major goal for 2007. Site monitoring is an important commissioning component and work is ongoing to devise CMS specific tests to be included in Service Availa...

  20. COMPUTING

    CERN Multimedia

    M. Kasemann

    Overview During the past three months activities were focused on data operations, testing and re-enforcing shift and operational procedures for data production and transfer, MC production and on user support. Planning of the computing resources in view of the new LHC calendar in ongoing. Two new task forces were created for supporting the integration work: Site Commissioning, which develops tools helping distributed sites to monitor job and data workflows, and Analysis Support, collecting the user experience and feedback during analysis activities and developing tools to increase efficiency. The development plan for DMWM for 2009/2011 was developed at the beginning of the year, based on the requirements from the Physics, Computing and Offline groups (see Offline section). The Computing management meeting at FermiLab on February 19th and 20th was an excellent opportunity discussing the impact and for addressing issues and solutions to the main challenges facing CMS computing. The lack of manpower is particul...

  1. Present and future aspects of PROSA - A computer program for near real time accountancy

    International Nuclear Information System (INIS)

    Beedgen, R.

    1987-01-01

    The methods of near real time accountancy (NRTA) for safeguarding nuclear material received a lot of attention in the last years. They developed PROSA 1.0 as a computer program to evaluate a sequence of material balance data based on three statistical tests for a selected false alarm probability. A new NRTA test procedure will be included and an option for the calculation of detection probabilities of hypothetical loss patterns will be made available in future releases of PROSA. Under a non-loss assumption, PROSA may also be used for the analysis of facility measurement models

  2. Polynomial-time computability of the edge-reliability of graphs using Gilbert's formula

    Directory of Open Access Journals (Sweden)

    Thomas J. Marlowe

    1998-01-01

    Full Text Available Reliability is an important consideration in analyzing computer and other communication networks, but current techniques are extremely limited in the classes of graphs which can be analyzed efficiently. While Gilbert's formula establishes a theoretically elegant recursive relationship between the edge reliability of a graph and the reliability of its subgraphs, naive evaluation requires consideration of all sequences of deletions of individual vertices, and for many graphs has time complexity essentially Θ (N!. We discuss a general approach which significantly reduces complexity, encoding subgraph isomorphism in a finer partition by invariants, and recursing through the set of invariants.

  3. Using real-time fMRI brain-computer interfacing to treat eating disorders.

    Science.gov (United States)

    Sokunbi, Moses O

    2018-05-15

    Real-time functional magnetic resonance imaging based brain-computer interfacing (fMRI neurofeedback) has shown encouraging outcomes in the treatment of psychiatric and behavioural disorders. However, its use in the treatment of eating disorders is very limited. Here, we give a brief overview of how to design and implement fMRI neurofeedback intervention for the treatment of eating disorders, considering the basic and essential components. We also attempt to develop potential adaptations of fMRI neurofeedback intervention for the treatment of anorexia nervosa, bulimia nervosa and binge eating disorder. Copyright © 2018 Elsevier B.V. All rights reserved.

  4. Evolution of perturbed dynamical systems: analytical computation with time independent accuracy

    Energy Technology Data Exchange (ETDEWEB)

    Gurzadyan, A.V. [Russian-Armenian (Slavonic) University, Department of Mathematics and Mathematical Modelling, Yerevan (Armenia); Kocharyan, A.A. [Monash University, School of Physics and Astronomy, Clayton (Australia)

    2016-12-15

    An analytical method for investigation of the evolution of dynamical systems with independent on time accuracy is developed for perturbed Hamiltonian systems. The error-free estimation using of computer algebra enables the application of the method to complex multi-dimensional Hamiltonian and dissipative systems. It also opens principal opportunities for the qualitative study of chaotic trajectories. The performance of the method is demonstrated on perturbed two-oscillator systems. It can be applied to various non-linear physical and astrophysical systems, e.g. to long-term planetary dynamics. (orig.)

  5. Self-Motion Perception: Assessment by Real-Time Computer Generated Animations

    Science.gov (United States)

    Parker, Donald E.

    1999-01-01

    Our overall goal is to develop materials and procedures for assessing vestibular contributions to spatial cognition. The specific objective of the research described in this paper is to evaluate computer-generated animations as potential tools for studying self-orientation and self-motion perception. Specific questions addressed in this study included the following. First, does a non- verbal perceptual reporting procedure using real-time animations improve assessment of spatial orientation? Are reports reliable? Second, do reports confirm expectations based on stimuli to vestibular apparatus? Third, can reliable reports be obtained when self-motion description vocabulary training is omitted?

  6. Towards OpenVL: Improving Real-Time Performance of Computer Vision Applications

    Science.gov (United States)

    Shen, Changsong; Little, James J.; Fels, Sidney

    Meeting constraints for real-time performance is a main issue for computer vision, especially for embedded computer vision systems. This chapter presents our progress on our open vision library (OpenVL), a novel software architecture to address efficiency through facilitating hardware acceleration, reusability, and scalability for computer vision systems. A logical image understanding pipeline is introduced to allow parallel processing. We also discuss progress on our middleware—vision library utility toolkit (VLUT)—that enables applications to operate transparently over a heterogeneous collection of hardware implementations. OpenVL works as a state machine,with an event-driven mechanismto provide users with application-level interaction. Various explicit or implicit synchronization and communication methods are supported among distributed processes in the logical pipelines. The intent of OpenVL is to allow users to quickly and easily recover useful information from multiple scenes, in a cross-platform, cross-language manner across various software environments and hardware platforms. To validate the critical underlying concepts of OpenVL, a human tracking system and a local positioning system are implemented and described. The novel architecture separates the specification of algorithmic details from the underlying implementation, allowing for different components to be implemented on an embedded system without recompiling code.

  7. Real time analysis for atmospheric dispersions for Fukushima nuclear accident: Mobile phone based cloud computing assessment

    International Nuclear Information System (INIS)

    Woo, Tae Ho

    2014-01-01

    Highlights: • Possible nuclear accident is simulated for the atmospheric contaminations. • The simulations results give the relative importance of the fallouts. • The cloud computing of IT is performed successfully. • One can prepare for the possible damages of such a NPP accident. • Some other variables can be considered in the modeling. - Abstract: The radioactive material dispersion is investigated by the system dynamics (SD) method. The non-linear complex algorithm could give the information about the hazardous material behavior in the case of nuclear accident. The prevailing westerlies region is modeled for the dynamical consequences of the Fukushima nuclear accident. The event sequence shows the scenario from earthquake to dispersion of the radionuclides. Then, the dispersion reaches two cities in Korea. The importance of the radioactive dispersion is related to the fast and reliable data processing, which could be accomplished by cloud computing concept. The values of multiplications for the wind, plume concentrations, and cloud computing factor are obtained. The highest value is 94.13 in the 206th day for Seoul. In Pusan, the highest value is 15.48 in the 219th day. The source is obtained as dispersion of radionuclide multiplied by 100. The real time safety assessment is accomplished by mobile phone

  8. A computationally efficient electricity price forecasting model for real time energy markets

    International Nuclear Information System (INIS)

    Feijoo, Felipe; Silva, Walter; Das, Tapas K.

    2016-01-01

    Highlights: • A fast hybrid forecast model for electricity prices. • Accurate forecast model that combines K-means and machine learning techniques. • Low computational effort by elimination of feature selection techniques. • New benchmark results by using market data for year 2012 and 2015. - Abstract: Increased significance of demand response and proliferation of distributed energy resources will continue to demand faster and more accurate models for forecasting locational marginal prices. This paper presents such a model (named K-SVR). While yielding prediction accuracy comparable with the best known models in the literature, K-SVR requires a significantly reduced computational time. The computational reduction is attained by eliminating the use of a feature selection process, which is commonly used by the existing models in the literature. K-SVR is a hybrid model that combines clustering algorithms, support vector machine, and support vector regression. K-SVR is tested using Pennsylvania–New Jersey–Maryland market data from the periods 2005–6, 2011–12, and 2014–15. Market data from 2006 has been used to measure performance of many of the existing models. Authors chose these models to compare performance and demonstrate strengths of K-SVR. Results obtained from K-SVR using the market data from 2012 and 2015 are new, and will serve as benchmark for future models.

  9. A cascadic monotonic time-discretized algorithm for finite-level quantum control computation

    Science.gov (United States)

    Ditz, P.; Borzi`, A.

    2008-03-01

    A computer package (CNMS) is presented aimed at the solution of finite-level quantum optimal control problems. This package is based on a recently developed computational strategy known as monotonic schemes. Quantum optimal control problems arise in particular in quantum optics where the optimization of a control representing laser pulses is required. The purpose of the external control field is to channel the system's wavefunction between given states in its most efficient way. Physically motivated constraints, such as limited laser resources, are accommodated through appropriately chosen cost functionals. Program summaryProgram title: CNMS Catalogue identifier: ADEB_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADEB_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 770 No. of bytes in distributed program, including test data, etc.: 7098 Distribution format: tar.gz Programming language: MATLAB 6 Computer: AMD Athlon 64 × 2 Dual, 2:21 GHz, 1:5 GB RAM Operating system: Microsoft Windows XP Word size: 32 Classification: 4.9 Nature of problem: Quantum control Solution method: Iterative Running time: 60-600 sec

  10. Computer-aided detection system for chest radiography: reducing report turnaround times of examinations with abnormalities.

    Science.gov (United States)

    Kao, E-Fong; Liu, Gin-Chung; Lee, Lo-Yeh; Tsai, Huei-Yi; Jaw, Twei-Shiun

    2015-06-01

    The ability to give high priority to examinations with pathological findings could be very useful to radiologists with large work lists who wish to first evaluate the most critical studies. A computer-aided detection (CAD) system for identifying chest examinations with abnormalities has therefore been developed. To evaluate the effectiveness of a CAD system on report turnaround times of chest examinations with abnormalities. The CAD system was designed to automatically mark chest examinations with possible abnormalities in the work list of radiologists interpreting chest examinations. The system evaluation was performed in two phases: two radiologists interpreted the chest examinations without CAD in phase 1 and with CAD in phase 2. The time information recorded by the radiology information system was then used to calculate the turnaround times. All chest examinations were reviewed by two other radiologists and were divided into normal and abnormal groups. The turnaround times for the examinations with pathological findings with and without the CAD system assistance were compared. The sensitivity and specificity of the CAD for chest abnormalities were 0.790 and 0.697, respectively, and use of the CAD system decreased the turnaround time for chest examinations with abnormalities by 44%. The turnaround times required for radiologists to identify chest examinations with abnormalities could be reduced by using the CAD system. This system could be useful for radiologists with large work lists who wish to first evaluate the most critical studies. © The Foundation Acta Radiologica 2014 Reprints and permissions: sagepub.co.uk/journalsPermissions.nav.

  11. Computer-assisted assessment of ultrasound real-time elastography: initial experience in 145 breast lesions.

    Science.gov (United States)

    Zhang, Xue; Xiao, Yang; Zeng, Jie; Qiu, Weibao; Qian, Ming; Wang, Congzhi; Zheng, Rongqin; Zheng, Hairong

    2014-01-01

    To develop and evaluate a computer-assisted method of quantifying five-point elasticity scoring system based on ultrasound real-time elastography (RTE), for classifying benign and malignant breast lesions, with pathologic results as the reference standard. Conventional ultrasonography (US) and RTE images of 145 breast lesions (67 malignant, 78 benign) were performed in this study. Each lesion was automatically contoured on the B-mode image by the level set method and mapped on the RTE image. The relative elasticity value of each pixel was reconstructed and classified into hard or soft by the fuzzy c-means clustering method. According to the hardness degree inside lesion and its surrounding tissue, the elasticity score of the RTE image was computed in an automatic way. Visual assessments of the radiologists were used for comparing the diagnostic performance. Histopathologic examination was used as the reference standard. The Student's t test and receiver operating characteristic (ROC) curve analysis were performed for statistical analysis. Considering score 4 or higher as test positive for malignancy, the diagnostic accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) were 93.8% (136/145), 92.5% (62/67), 94.9% (74/78), 93.9% (62/66), and 93.7% (74/79) for the computer-assisted scheme, and 89.7% (130/145), 85.1% (57/67), 93.6% (73/78), 92.0% (57/62), and 88.0% (73/83) for manual assessment. Area under ROC curve (Az value) for the proposed method was higher than the Az value for visual assessment (0.96 vs. 0.93). Computer-assisted quantification of classical five-point scoring system can significantly eliminate the interobserver variability and thereby improve the diagnostic confidence of classifying the breast lesions to avoid unnecessary biopsy. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.

  12. Computer-assisted assessment of ultrasound real-time elastography: Initial experience in 145 breast lesions

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Xue; Xiao, Yang [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China); Zeng, Jie [Department of Medical Ultrasonics, Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou (China); Qiu, Weibao; Qian, Ming; Wang, Congzhi [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China); Zheng, Rongqin, E-mail: zhengronggin@hotmail.com [Department of Medical Ultrasonics, Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou (China); Zheng, Hairong, E-mail: hr.zheng@siat.ac.cn [Shenzhen Key Lab for Molecular Imaging, Paul C. Lauterbur Research Center for Biomedical Imaging, Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen (China)

    2014-01-15

    Purpose: To develop and evaluate a computer-assisted method of quantifying five-point elasticity scoring system based on ultrasound real-time elastography (RTE), for classifying benign and malignant breast lesions, with pathologic results as the reference standard. Materials and methods: Conventional ultrasonography (US) and RTE images of 145 breast lesions (67 malignant, 78 benign) were performed in this study. Each lesion was automatically contoured on the B-mode image by the level set method and mapped on the RTE image. The relative elasticity value of each pixel was reconstructed and classified into hard or soft by the fuzzy c-means clustering method. According to the hardness degree inside lesion and its surrounding tissue, the elasticity score of the RTE image was computed in an automatic way. Visual assessments of the radiologists were used for comparing the diagnostic performance. Histopathologic examination was used as the reference standard. The Student's t test and receiver operating characteristic (ROC) curve analysis were performed for statistical analysis. Results: Considering score 4 or higher as test positive for malignancy, the diagnostic accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) were 93.8% (136/145), 92.5% (62/67), 94.9% (74/78), 93.9% (62/66), and 93.7% (74/79) for the computer-assisted scheme, and 89.7% (130/145), 85.1% (57/67), 93.6% (73/78), 92.0% (57/62), and 88.0% (73/83) for manual assessment. Area under ROC curve (A{sub z} value) for the proposed method was higher than the A{sub z} value for visual assessment (0.96 vs. 0.93). Conclusion: Computer-assisted quantification of classical five-point scoring system can significantly eliminate the interobserver variability and thereby improve the diagnostic confidence of classifying the breast lesions to avoid unnecessary biopsy.

  13. Time complexity analysis for distributed memory computers: implementation of parallel conjugate gradient method

    NARCIS (Netherlands)

    Hoekstra, A.G.; Sloot, P.M.A.; Haan, M.J.; Hertzberger, L.O.; van Leeuwen, J.

    1991-01-01

    New developments in Computer Science, both hardware and software, offer researchers, such as physicists, unprecedented possibilities to solve their computational intensive problems.However, full exploitation of e.g. new massively parallel computers, parallel languages or runtime environments

  14. Effect of time lapse on the diagnostic accuracy of cone beam computed tomography for detection of vertical root fractures

    Energy Technology Data Exchange (ETDEWEB)

    Eskandarloo, Amir; Shokri, Abbas, E-mail: Dr.a.shokri@gmail.com [Dental Research Center, Department of Oral and Maxillofacial Radiology, Hamadan University of Medical Sciences, Hamadan (Iran, Islamic Republic of); Asl, Amin Mahdavi [Department of Oral and Maxillofacial Radiology, Golestan University of Medical Sciences, Gorgan (Iran, Islamic Republic of); Jalalzadeh, Mohsen [Department of Endodontics, Hamadan University of Medical Sciences, Hamadan (Iran, Islamic Republic of); Tayari, Maryam [Department of Pedodontics, Golestan University of Medical Sciences, Gorgan (Iran, Islamic Republic of); Hosseinipanah, Mohammad [Department of Anatomy, School of Medicine, Hamadan University of Medical Sciences, Hamadan (Iran, Islamic Republic of); Fardmal, Javad [Research Center for Health Sciences and Department of Epidemiology and Biostatistics, School of Public Health, Hamadan University of Medical Sciences, Hamadan (Iran, Islamic Republic of)

    2016-01-15

    Accurate and early diagnosis of vertical root fractures (VRFs) is imperative to prevent extensive bone loss and unnecessary endodontic and prosthodontic treatments. The aim of this study was to assess the effect of time lapse on the diagnostic accuracy of cone beam computed tomography (CBCT) for VRFs in endodontically treated dog’s teeth. Forty eight incisors and premolars of three adult male dogs underwent root canal therapy. The teeth were assigned to two groups: VRFs were artificially induced in the first group (n=24) while the teeth in the second group remained intact (n=24). The CBCT scans were obtained by NewTom 3G unit immediately after inducing VRFs and after one, two, three, four, eight, 12 and 16 weeks. Three oral and maxillofacial radiologists blinded to the date of radiographs assessed the presence/absence of VRFs on CBCT scans. The sensitivity, specificity and accuracy values were calculated and data were analyzed using SPSS v.16 software and ANOVA. The total accuracy of detection of VRFs immediately after surgery, one, two, three, four, eight, 12 and 16 weeks was 67.3%, 68.7%, 66.6%, 64.6%, 64.5%, 69.4%, 68.7%, 68% respectively. The effect of time lapse on detection of VRFs was not significant (p>0.05). Overall sensitivity, specificity and accuracy of CBCT for detection of VRFs were 74.3%, 62.2%, 67.2% respectively. Cone beam computed tomography is a valuable tool for detection of VRFs. Time lapse (four months) had no effect on detection of VRFs on CBCT scans. (author)

  15. Effect of time lapse on the diagnostic accuracy of cone beam computed tomography for detection of vertical root fractures

    International Nuclear Information System (INIS)

    Eskandarloo, Amir; Shokri, Abbas; Asl, Amin Mahdavi; Jalalzadeh, Mohsen; Tayari, Maryam; Hosseinipanah, Mohammad; Fardmal, Javad

    2016-01-01

    Accurate and early diagnosis of vertical root fractures (VRFs) is imperative to prevent extensive bone loss and unnecessary endodontic and prosthodontic treatments. The aim of this study was to assess the effect of time lapse on the diagnostic accuracy of cone beam computed tomography (CBCT) for VRFs in endodontically treated dog’s teeth. Forty eight incisors and premolars of three adult male dogs underwent root canal therapy. The teeth were assigned to two groups: VRFs were artificially induced in the first group (n=24) while the teeth in the second group remained intact (n=24). The CBCT scans were obtained by NewTom 3G unit immediately after inducing VRFs and after one, two, three, four, eight, 12 and 16 weeks. Three oral and maxillofacial radiologists blinded to the date of radiographs assessed the presence/absence of VRFs on CBCT scans. The sensitivity, specificity and accuracy values were calculated and data were analyzed using SPSS v.16 software and ANOVA. The total accuracy of detection of VRFs immediately after surgery, one, two, three, four, eight, 12 and 16 weeks was 67.3%, 68.7%, 66.6%, 64.6%, 64.5%, 69.4%, 68.7%, 68% respectively. The effect of time lapse on detection of VRFs was not significant (p>0.05). Overall sensitivity, specificity and accuracy of CBCT for detection of VRFs were 74.3%, 62.2%, 67.2% respectively. Cone beam computed tomography is a valuable tool for detection of VRFs. Time lapse (four months) had no effect on detection of VRFs on CBCT scans. (author)

  16. TIMED: a computer program for calculating cumulated activity of a radionuclide in the organs of the human body at a given time, t, after deposition

    International Nuclear Information System (INIS)

    Watson, S.B.; Snyder, W.S.; Ford, M.R.

    1976-12-01

    TIMED is a computer program designed to calculate cumulated radioactivity in the various source organs at various times after radionuclide deposition. TIMED embodies a system of differential equations which describes activity transfer in the lungs, gastrointestinal tract, and other organs of the body. This system accounts for delay of transfer of activity between compartments of the body and radioactive daughters

  17. The application of digital computers to near-real-time processing of flutter test data

    Science.gov (United States)

    Hurley, S. R.

    1976-01-01

    Procedures used in monitoring, analyzing, and displaying flight and ground flutter test data are presented. These procedures include three digital computer programs developed to process structural response data in near real time. Qualitative and quantitative modal stability data are derived from time history response data resulting from rapid sinusoidal frequency sweep forcing functions, tuned-mode quick stops, and pilot induced control pulses. The techniques have been applied to both fixed and rotary wing aircraft, during flight, whirl tower rotor systems tests, and wind tunnel flutter model tests. An hydraulically driven oscillatory aerodynamic vane excitation system utilized during the flight flutter test programs accomplished during Lockheed L-1011 and S-3A development is described.

  18. An Energy Efficient Neuromorphic Computing System Using Real Time Sensing Method

    DEFF Research Database (Denmark)

    Farkhani, Hooman; Tohidi, Mohammad; Farkhani, Sadaf

    2017-01-01

    In spintronic-based neuromorphic computing systems (NCS), the switching of magnetic moment in a magnetic tunnel junction (MTJ) is used to mimic neuron firing. However, the stochastic switching behavior of the MTJ and process variations effect leads to extra stimulation time. This leads to extra...... energy consumption and delay of such NCSs. In this paper, a new real-time sensing (RTS) circuit is proposed to track the MTJ state and terminate stimulation phase immediately after MTJ switching. This leads to significant degradation in energy consumption and delay of NCS. The simulation results using...... a 65-nm CMOS technology and a 40-nm MTJ technology confirm that the energy consumption of a RTS-based NCS is improved by 50% in comparison with a typical NCS. Moreover, utilizing RTS circuit improves the overall speed of an NCS by 2.75x....

  19. Some selection criteria for computers in real-time systems for high energy physics

    International Nuclear Information System (INIS)

    Kolpakov, I.F.

    1980-01-01

    The right choice of program source is for the organization of real-time systems of great importance as cost and reliability are decisive factors. Some selection criteria for program sources for high energy physics multiwire chamber spectrometers (MWCS) are considered in this report. MWCS's accept bits of information from event pattens. Large and small computers, microcomputers and intelligent controllers in CAMAC crates are compared with respect to the following characteristics: data exchange speed, number of addresses for peripheral devices, cost of interfacing a peripheral device, sizes of buffer and mass memory, configuration costs, and the mean time between failures (MTBF). The results of comparisons are shown by plots and histograms which allow the selection of program sources according to the above criteria. (Auth.)

  20. Hybrid automata models of cardiac ventricular electrophysiology for real-time computational applications.

    Science.gov (United States)

    Andalam, Sidharta; Ramanna, Harshavardhan; Malik, Avinash; Roop, Parthasarathi; Patel, Nitish; Trew, Mark L

    2016-08-01

    Virtual heart models have been proposed for closed loop validation of safety-critical embedded medical devices, such as pacemakers. These models must react in real-time to off-the-shelf medical devices. Real-time performance can be obtained by implementing models in computer hardware, and methods of compiling classes of Hybrid Automata (HA) onto FPGA have been developed. Models of ventricular cardiac cell electrophysiology have been described using HA which capture the complex nonlinear behavior of biological systems. However, many models that have been used for closed-loop validation of pacemakers are highly abstract and do not capture important characteristics of the dynamic rate response. We developed a new HA model of cardiac cells which captures dynamic behavior and we implemented the model in hardware. This potentially enables modeling the heart with over 1 million dynamic cells, making the approach ideal for closed loop testing of medical devices.