WorldWideScience

Sample records for models microcomputer techniques

  1. A MICROCOMPUTER MODEL FOR IRRIGATION SYSTEM EVALUATION

    OpenAIRE

    Williams, Jeffery R.; Buller, Orlan H.; Dvorak, Gary J.; Manges, Harry L.

    1988-01-01

    ICEASE (Irrigation Cost Estimator and System Evaluator) is a microcomputer model designed and developed to meet the need for conducting economic evaluation of adjustments to irrigation systems and management techniques to improve the use of irrigated water. ICEASE can calculate the annual operating costs for irrigation systems and has five options that can be used to economically evaluate improvements in the pumping plant or the way the irrigation system is used for crop production.

  2. Contour mapping using a microcomputer - techniques for improvement

    Energy Technology Data Exchange (ETDEWEB)

    Herkommer, M.A.

    1984-04-01

    There are 3 basic steps to contouring by microcomputer: control-point data analysis, generating a regular grid of estimated values, and contouring. Because each of these steps requires a significant amount of time and computer resource, techniques have been developed to improve grid parameter selection, speed gridding, and edit computer-generated contour maps. Grid cell size is a critical parameter used during the gridding stage of computer mapping and must be selected with consideration given to control-point distribution. A histogram showing control points per grid cell is a simple graphical presentation that illustrates before gridding is actually performed the effectiveness of the selected grid increment in producing the desirable grid characteristics of having one control point per grid cell. Gridding is accomplished in 2 steps at every node: gathering the control points for estimating, and actual calculation. Using collection and estimating techniques such as nearest-neighbor searches and inverse-distance moving weighted averaging, several thousands nodes can be calculated in a few minutes. Often, the generated grid (contour map) is satisfactory with exceptions in 1 or 2 areas. Using an interactive computer technique known as contour editing, geologic knowledge and experience can be infused into the map. When editing contour, the computer mapper indicates the way the contours should behave. The entered values are then used to calculate back to the grid nodes (without regridding the whole map area) so the when the grid is recontoured, the contours mimic the contour revisions.

  3. WHOLE FARM RISK-RATING MICROCOMPUTER MODEL

    OpenAIRE

    Anderson, Kim B.; Ikerd, John E.

    1985-01-01

    The Risk-Rating Model is designed to give extension specialists, teachers, and producers a method to analyze production, marketing, and financial risks. These risks may be analyzed either individually or simultaneously. The risk associated with each enterprise, for all combinations of enterprises, and for any combination of marketing strategies is estimated. Optimistic, expected, and pessimistic returns above variable cost and/or total cost are presented in the results. The probability that t...

  4. Micro-computed tomography and bond strength analysis of different root canal filling techniques

    Directory of Open Access Journals (Sweden)

    Juliane Nhata

    2014-01-01

    Full Text Available Introduction: The aim of this study was to evaluate the quality and bond strength of three root filling techniques (lateral compaction, continuous wave of condensation and Tagger′s Hybrid technique [THT] using micro-computed tomography (CT images and push-out tests, respectively. Materials and Methods: Thirty mandibular incisors were prepared using the same protocol and randomly divided into three groups (n = 10: Lateral condensation technique (LCT, continuous wave of condensation technique (CWCT, and THT. All specimens were filled with Gutta-percha (GP cones and AH Plus sealer. Five specimens of each group were randomly chosen for micro-CT analysis and all of them were sectioned into 1 mm slices and subjected to push-out tests. Results: Micro-CT analysis revealed less empty spaces when GP was heated within the root canals in CWCT and THT when compared to LCT. Push-out tests showed that LCT and THT had a significantly higher displacement resistance (P < 0.05 when compared to the CWCT. Bond strength was lower in apical and middle thirds than in the coronal thirds. Conclusions: It can be concluded that LCT and THT were associated with higher bond strengths to intraradicular dentine than CWCT. However, LCT was associated with more empty voids than the other techniques.

  5. The microcomputer scientific software series 2: general linear model--regression.

    Science.gov (United States)

    Harold M. Rauscher

    1983-01-01

    The general linear model regression (GLMR) program provides the microcomputer user with a sophisticated regression analysis capability. The output provides a regression ANOVA table, estimators of the regression model coefficients, their confidence intervals, confidence intervals around the predicted Y-values, residuals for plotting, a check for multicollinearity, a...

  6. The microcomputer scientific software series 3: general linear model--analysis of variance.

    Science.gov (United States)

    Harold M. Rauscher

    1985-01-01

    A BASIC language set of programs, designed for use on microcomputers, is presented. This set of programs will perform the analysis of variance for any statistical model describing either balanced or unbalanced designs. The program computes and displays the degrees of freedom, Type I sum of squares, and the mean square for the overall model, the error, and each factor...

  7. AUTOCASK (AUTOmatic Generation of 3-D CASK models). A microcomputer based system for shipping cask design review analysis

    International Nuclear Information System (INIS)

    Gerhard, M.A.; Sommer, S.C.

    1995-04-01

    AUTOCASK (AUTOmatic Generation of 3-D CASK models) is a microcomputer-based system of computer programs and databases developed at the Lawrence Livermore National Laboratory (LLNL) for the structural analysis of shipping casks for radioactive material. Model specification is performed on the microcomputer, and the analyses are performed on an engineering workstation or mainframe computer. AUTOCASK is based on 80386/80486 compatible microcomputers. The system is composed of a series of menus, input programs, display programs, a mesh generation program, and archive programs. All data is entered through fill-in-the-blank input screens that contain descriptive data requests

  8. AUTOCASK (AUTOmatic Generation of 3-D CASK models). A microcomputer based system for shipping cask design review analysis

    Energy Technology Data Exchange (ETDEWEB)

    Gerhard, M.A.; Sommer, S.C. [Lawrence Livermore National Lab., CA (United States)

    1995-04-01

    AUTOCASK (AUTOmatic Generation of 3-D CASK models) is a microcomputer-based system of computer programs and databases developed at the Lawrence Livermore National Laboratory (LLNL) for the structural analysis of shipping casks for radioactive material. Model specification is performed on the microcomputer, and the analyses are performed on an engineering workstation or mainframe computer. AUTOCASK is based on 80386/80486 compatible microcomputers. The system is composed of a series of menus, input programs, display programs, a mesh generation program, and archive programs. All data is entered through fill-in-the-blank input screens that contain descriptive data requests.

  9. Doing Physics with Microcomputers.

    Science.gov (United States)

    Bak, Per

    1983-01-01

    Describes how microcomputers can perform very demanding/large-scale physics calculations at speeds not much slower than those of modern, full-size computers. Among the examples provided are a Monte Carlo simulation of the three-dimensional Ising model and a program (for the Apple microcomputer) using the time-independent Schrodinger Equation. (JN)

  10. A Novel Model for Evaluating the Flow of Endodontic Materials Using Micro-computed Tomography.

    Science.gov (United States)

    Tanomaru-Filho, Mario; Torres, Fernanda Ferrari Esteves; Bosso-Martelo, Roberta; Chávez-Andrade, Gisselle Moraima; Bonetti-Filho, Idomeo; Guerreiro-Tanomaru, Juliane Maria

    2017-05-01

    Flow and filling ability are important properties of endodontic materials. The aim of this study was to propose a new technique for evaluating flow using micro-computed tomographic (μCT) imaging. A glass plate was manufactured with a central cavity and 4 grooves extending out horizontally and vertically. The flow of MTA-Angelus (Angelus, Londrina, PR, Brazil), zinc oxide eugenol (ZOE), and Biodentine (BIO) (Septodont, Saint Maur des Fossés, France) was evaluated using International Standards Organization (ISO) 6876/2002 and a new technique as follows: 0.05 ± 0.005 mL of each material was placed in the central cavity, and another glass plate and metal weight with a total mass of 120 g were placed over the material. The plate/material set was scanned using μCT imaging. The flow was calculated by linear measurement (mm) of the material in the grooves. Central cavity filling was calculated in mm 3 in the central cavity. Lateral cavity filling (LCF) was measured by LCF mean values up to 2 mm from the central cavity. Data were analyzed statistically using analysis of variance and Tukey tests with a 5% significance level. ZOE showed the highest flow rate determined by ISO methodology (P flow rates in the grooves. Central cavity filling was similar for the materials. However, LCF was higher for BIO versus ZOE. Although ZOE presented better flow determined by ISO methodology, BIO showed the best filling ability. The model of the technique proposed for evaluating flow using μCT imaging showed proper and reproducible results and could improve flow analysis. Copyright © 2017 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  11. Development and optimization of a high-throughput micro-computed tomography imaging method incorporating a novel analysis technique to evaluate bone mineral density of arthritic joints in a rodent model of collagen induced arthritis.

    Science.gov (United States)

    Sevilla, Raquel S; Cruz, Francisco; Chiu, Chi-Sung; Xue, Dahai; Bettano, Kimberly A; Zhu, Joe; Chakravarthy, Kalyan; Faltus, Robert; Wang, Shubing; Vanko, Amy; Robinson, Gain; Zielstorff, Mark; Miao, John; Leccese, Erica; Conway, Donald; Moy, Lily Y; Dogdas, Belma; Cicmil, Milenko; Zhang, Weisheng

    2015-04-01

    Rheumatoid arthritis (RA) is a chronic autoimmune disease resulting in joint inflammation, pain, and eventual bone loss. Bone loss and remodeling caused by symmetric polyarthritis, the hallmark of RA, is readily detectable by bone mineral density (BMD) measurement using micro-CT. Abnormalities in these measurements over time reflect the underlying pathophysiology of the bone. To evaluate the efficacy of anti-rheumatic agents in animal models of arthritis, we developed a high throughput knee and ankle joint imaging assay to measure BMD as a translational biomarker. A bone sample holder was custom designed for micro-CT scanning, which significantly increased assay throughput. Batch processing 3-dimensional image reconstruction, followed by automated image cropping, significantly reduced image processing time. In addition, we developed a novel, automated image analysis method to measure BMD and bone volume of knee and ankle joints. These improvements significantly increased the throughput of ex vivo bone sample analysis, reducing data turnaround from 5 days to 24 hours for a study with 200 rat hind limbs. Taken together, our data demonstrate that BMD, as quantified by micro-CT, is a robust efficacy biomarker with a high degree of sensitivity. Our innovative approach toward evaluation of BMD using optimized image acquisition and novel image processing techniques in preclinical models of RA enables high throughput assessment of anti-rheumatic agents offering a powerful tool for drug discovery. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.

  12. A microcomputer-controlled modulation technique for the detection of transient species in UV photoelectron spectroscopy

    International Nuclear Information System (INIS)

    Lonkhuyzen, H. van; Muller, H.G.; Lange, C.A. de

    1980-01-01

    A microcomputer-controlled modulation method is described to measure UV photoelectron spectra of transient species generated in a microwave discharge. Spectra at low and high microwave power levels are simultaneously recorded and afterwards linearly combined in order to remove parent compound signals. The method is applied to discharged oxygen where the transition O 2 + ( 2 PHIsub(u)) 2 ( 1 Δsub(g)) becomes visible without interference from the parent molecule O 2 ( 3 Σsub(g) - ), and to discharged sulphur dioxide where SO( 3 Σ - ) and S( 3 P) photoelectron spectra are obtained free from SO 2 bands. Finally the build-up of transient bands as a function of time is recorded. (orig.)

  13. Real-time interferometer phase detection using an LSI-11 microcomputer and high-speed digital techniques

    International Nuclear Information System (INIS)

    Mendell, D.S.

    1978-01-01

    This paper describes the basic design and philosophy of a real-time, interferometer phase-detection system used on the 2XIIB and TMX magnetic-fusion experiments at the Lawrence Livermore Laboratory. This diagnostics system is now a satellite to a host computer and uses high-speed, emitter-coupled logic techniques to derive data on real-time phase relationships. The system's input signals can be derived from interferometer outputs over a wide range of reference frequencies. An LSI-11 microcomputer is the interface between the high-speed phase-detection logic, buffer memory, human interaction, and host computer. Phase data on a storage CRT is immediately displayed after each experimental fusion shot. An operator can interrogate this phase data more closely from an interactive control panel, while the host computer is simultaneously examining the system's buffer memory or arming the system for the next shot

  14. A simple dynamic model and transient simulation of the nuclear power reactor on microcomputers

    Energy Technology Data Exchange (ETDEWEB)

    Han, Yang Gee; Park, Cheol [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1997-12-31

    A simple dynamic model is developed for the transient simulation of the nuclear power reactor. The dynamic model includes the normalized neutron kinetics model with reactivity feedback effects and the core thermal-hydraulics model. The main objective of this paper demonstrates the capability of the developed dynamic model to simulate various important variables of interest for a nuclear power reactor transient. Some representative results of transient simulations show the expected trends in all cases, even though no available data for comparison. In this work transient simulations are performed on a microcomputer using the DESIRE/N96T continuous system simulation language which is applicable to nuclear power reactor transient analysis. 3 refs., 9 figs. (Author)

  15. Microcomputer Testing Software Teachers Can Use.

    Science.gov (United States)

    Hsu, Tse-chi; Nitko, Anthony J.

    1983-01-01

    Microcomputer software for computer-assisted classroom testing is reviewed. The teacher and classroom are emphasized in applying computer technology. The major issues are identification of appropriate classroom testing microcomputer applications; identification of available microcomputer testing software; techniques for software evaluation; and…

  16. The Effect of Different Bone Marrow Stimulation Techniques on Human Talar Subchondral Bone: A Micro-Computed Tomography Evaluation.

    Science.gov (United States)

    Gianakos, Arianna L; Yasui, Youichi; Fraser, Ethan J; Ross, Keir A; Prado, Marcelo P; Fortier, Lisa A; Kennedy, John G

    2016-10-01

    To evaluate morphological alterations, microarchitectural disturbances, and the extent of bone marrow access to the subchondral bone marrow compartment using micro-computed tomography analysis in different bone marrow stimulation (BMS) techniques. Nine zones in a 3 × 3 grid pattern were assigned to 5 cadaveric talar dome articular surfaces. A 1.00-mm microfracture awl (s.MFX), a 2.00-mm standard microfracture awl (l.MFX), or a 1.25-mm Kirschner wire (K-wire) drill hole was used to penetrate the subchondral bone in each grid zone. Subchondral bone holes and adjacent tissue areas were assessed by micro-computed tomography to analyze adjacent bone area destruction and communicating channels to the bone marrow. Grades 1 to 3 were assigned, where 1 = minimal compression/sclerosis; 2 = moderate compression/sclerosis; 3 = severe compression/sclerosis. Bone volume/total tissue volume, bone surface area/bone volume, trabecular thickness, and trabecular number were calculated in the region of interest. Visual assessment revealed that the s.MFX had significantly more grade 1 holes (P < .001) and that the l.MFX had significantly more poor/grade 3 holes (P = .002). Bone marrow channel assessment showed a statistically significant increase in the number of channels in the s.MFX when compared with both K-wire and l.MFX holes (P < .001). Bone volume fraction for the s.MFX was significantly less than that of the l.MFX (P = .029). BMS techniques using instruments with larger diameters resulted in increased trabecular compaction and sclerosis in areas adjacent to the defect. K-wire and l.MFX techniques resulted in less open communicating bone marrow channels, denoting a reduction in bone marrow access. The results of this study indicate that BMS using larger diameter devices results in greater microarchitecture disturbances. The current study suggests that the choice of a BMS technique should be carefully considered as the results indicate that smaller diameter hole sizes may

  17. Modeling survival of juvenile salmon during downriver migration in the Columbia River on a microcomputer

    Energy Technology Data Exchange (ETDEWEB)

    Peloquin, R.A.; McKenzie, D.H.

    1994-10-01

    A compartmental model has been implemented on a microcomputer as an aid in the analysis of alternative solutions to a problem. The model, entitled Smolt Survival Simulator, simulates the survival of juvenile salmon during their downstream migration and passage of hydroelectric dams in the Columbia River. The model is designed to function in a workshop environment where resource managers and fisheries biologists can study alternative measures that may potentially increase juvenile anadromous fish survival during downriver migration. The potential application of the model has placed several requirements on the implementing software. It must be available for use in workshop settings. The software must be easily to use with minimal computer knowledge. Scenarios must be created and executed quickly and efficiently. Results must be immediately available. Software design emphasis vas placed on the user interface because of these requirements. The discussion focuses on methods used in the development of the SSS software user interface. These methods should reduce user stress and alloy thorough and easy parameter modification.

  18. A general DNA analysis program for the Hewlett-Packard Model 86/87 microcomputer.

    Science.gov (United States)

    Harley, E H

    1986-01-10

    A program is described to perform general DNA sequence analysis on the Hewlett-Packard Model 86/87 microcomputer operating on 128 K of RAM. The following analytical procedures can be performed: 1. display of the sequence, in whole or part, or its complement; 2. search for specified sequences e.g. restriction sites, and in the case of the latter give fragment sizes; 3. perform a comprehensive search for all known restriction enzyme sites; 4. map sites graphically; 5. perform editing functions; 6. base frequency analysis; 7. search for repeated sequences; 8. search for open reading frames or translate into the amino acid sequence and analyse for basic and acidic amino acids, hydrophobicity, and codon usage. Two sequences, or parts thereof, can be merged in various orientations to mimic recombination strategies, or can be compared for homologies. The program is written in HP BASIC and is designed principally as a tool for the laboratory investigator manipulating a defined set of vectors and recombinant DNA constructs.

  19. A Model Microcomputer Software Package for Educational Program-Oriented-Budgeting.

    Science.gov (United States)

    Leahy, Phillip E.

    The Microcomputer Program Oriented Budgeting System (MICROCOMP POB) is an accounts payable system intended for use in school districts. The system is based on the general principles of encumbrance accounting and the use of multiple account numbers. For each account number, the primary information provided consists of year to date balances for…

  20. Testing teletypes with a microcomputer. [TTYTST, for Intel Model 8008 processor

    Energy Technology Data Exchange (ETDEWEB)

    Byrd, J.S.

    1976-03-01

    An effective exerciser using a microcomputer was designed for ASR-33/35 Teletypes. The low-cost, portable instrument can be easily used by a maintenance technician to diagnose problems and to check out the Teletype after maintenance has been performed. The Teletype is operated under realistic, on-line conditions. System firmware includes programs for evaluation and diagnostic testing of all punching, reading, printing, and keying operations. 10 figures.

  1. Mechanical Preparation Showed Superior Shaping Ability than Manual Technique in Primary Molars - A Micro-Computed Tomography Study

    OpenAIRE

    Hidalgo, Lídia Regina da Costa; Silva, Léa Assed Bezerra da; Leoni, Graziela Bianchi; Mazzi-Chaves, Jardel Francisco; Carvalho, Emily Eduardo da Silva; Consolaro, Alberto; Sousa-Neto, Manoel Damião

    2017-01-01

    Abstract The aim of the study was to evaluate canal preparation in primary molars with hand files, ProTaper Next and Self-Adjusting File (SAF) by 2D and 3D micro-computed tomography (micro-CT) analysis. Canals of 24 primary molars were prepared with hand files (HF), ProTaper Next (PTN) and SAF (n=8/group). The teeth were scanned before and after root canal preparation and the pre- and postoperative micro-CT images were reconstructed. Changes in 2D (area, perimeter, roundness, minor and major ...

  2. Nuclear spectrum analysis by using microcomputer

    International Nuclear Information System (INIS)

    Sanyal, M.K.; Mukhopadhyay, P.K.; Rao, A.D.; Pethe, V.A.

    1984-01-01

    A method is presented for analysis of nuclear spectra by using microcomputer. A nonlinear least square fit of a mathematical model with observed spectrum is performed with variable metric method. The linear search procedure of the variable metric method has been modified so that the algorithm needs less program space and computational time both of which are important for microcomputer implementation. This widely used peak analysis method can now be made available in microcomputer based multichannel analysers. (author)

  3. Microcomputer simulation model for facility performance assessment: a case study of nuclear spent fuel handling facility operations

    Energy Technology Data Exchange (ETDEWEB)

    Chockie, A.D.; Hostick, C.J.; Otis, P.T.

    1985-10-01

    A microcomputer based simulation model was recently developed at the Pacific Northwest Laboratory (PNL) to assist in the evaluation of design alternatives for a proposed facility to receive, consolidate and store nuclear spent fuel from US commercial power plants. Previous performance assessments were limited to deterministic calculations and Gantt chart representations of the facility operations. To insure that the design of the facility will be adequate to meet the specified throughput requirements, the simulation model was used to analyze such factors as material flow, equipment capability and the interface between the MRS facility and the nuclear waste transportation system. The simulation analysis model was based on commercially available software and application programs designed to represent the MRS waste handling facility operations. The results of the evaluation were used by the design review team at PNL to identify areas where design modifications should be considered. 4 figs.

  4. Micro-computer simulation software: A review

    Directory of Open Access Journals (Sweden)

    P.S. Kruger

    2003-12-01

    Full Text Available Simulation modelling has proved to be one of the most powerful tools available to the Operations Research Analyst. The development of micro-computer technology has reached a state of maturity where the micro-computer can provide the necessary computing power and consequently various powerful and inexpensive simulation languages for micro-computers have became available. This paper will attempt to provide an introduction to the general philosophy and characteristics of some of the available micro-computer simulation languages. The emphasis will be on the characteristics of the specific micro-computer implementation rather than on a comparison of the modelling features of the various languages. Such comparisons may be found elsewhere.

  5. Detectability of Middle Mesial Root Canal Orifices by Troughing Technique in Mandibular Molars: A Micro-computed Tomographic Study.

    Science.gov (United States)

    Keleş, Ali; Keskin, Cangül

    2017-08-01

    The objective of the present study was to measure the orifice depth of middle mesial canals (MMCs) and evaluate the detectability of orifices using troughing preparation. For this study, 85 mandibular molar teeth with MMCs were selected from the scanned micro-computed tomographic images. The MMCs were categorized, and the distances between the MMC orifices and the cementoenamel junctions (CEJ) were recorded as the depth of the orifice. Data were evaluated with frequency analysis and a chi-square test using SPSS (SPSS Inc, Chicago, IL), and the results indicated a 5% significance level. It was found that 77.41% of the MMC orifices were at the CEJ level, whereas 5.38% and 9.69% of the MMC orifices were detectable within 1-mm and 2-mm depths from the CEJ, respectively. Of the specimens, 7.52% had MMC orifices deeper than 2 mm from the CEJ. Confluent anatomy was the most frequent configuration. No significant relation was detected between the orifice depth and MMC configuration (P > .05). It was concluded that 77.41% of the specimens did not require troughing preparation, the remaining 15.07% would require troughing, and 7.52% could not be accessed even with the troughing preparation. Copyright © 2017 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  6. Classifying variability modeling techniques

    NARCIS (Netherlands)

    Sinnema, Marco; Deelstra, Sybren

    Variability modeling is important for managing variability in software product families, especially during product derivation. In the past few years, several variability modeling techniques have been developed, each using its own concepts to model the variability provided by a product family. The

  7. Bubble Chamber Research Group Microcomputer Unit

    International Nuclear Information System (INIS)

    Bairstow, R.; Barlow, J.; Mace, P.R.; Seller, P.; Waters, M.; Watson, J.G.

    1982-05-01

    A distributed data acquisition system has been developed by the Bubble Chamber Research Group at the Rutherford Appleton laboratory for use with their film measuring machines. The system is based upon a set of microcomputers linked together with a VAX 11/780 computer, in a local area computer network. This network is of the star type and uses a packet switching technique. Each film measuring machine is equipped with a microcomputer which controls the function of the table, buffers data and enhances the interface between operators and machines. This paper provides a detailed description of each microcomputer and can be used as a reference manual for these computers. (author)

  8. Compression-recovery model of absorptive glass mat (AGM) separator guided by X-ray micro-computed tomography analysis

    Science.gov (United States)

    Kameswara Rao, P. V.; Rawal, Amit; Kumar, Vijay; Rajput, Krishn Gopal

    2017-10-01

    Absorptive glass mat (AGM) separators play a key role in enhancing the cycle life of the valve regulated lead acid (VRLA) batteries by maintaining the elastic characteristics under a defined level of compression force with the plates of the electrodes. Inevitably, there are inherent challenges to maintain the required level of compression characteristics of AGM separators during the charge and discharge of the battery. Herein, we report a three-dimensional (3D) analytical model for predicting the compression-recovery behavior of AGM separators by formulating a direct relationship with the constituent fiber and structural parameters. The analytical model of compression-recovery behavior of AGM separators has successfully included the fiber slippage criterion and internal friction losses. The presented work uses, for the first time, 3D data of fiber orientation from X-ray micro-computed tomography, for predicting the compression-recovery behavior of AGM separators. A comparison has been made between the theoretical and experimental results of compression-recovery behavior of AGM samples with defined fiber orientation characteristics. In general, the theory agreed reasonably well with the experimental results of AGM samples in both dry and wet states. Through theoretical modeling, fiber volume fraction was established as one of the key structural parameters that modulates the compression hysteresis of an AGM separator.

  9. Development of a microcomputed tomography scoring system to characterize disease progression in the Hartley guinea pig model of spontaneous osteoarthritis.

    Science.gov (United States)

    Radakovich, Lauren B; Marolf, Angela J; Shannon, John P; Pannone, Stephen C; Sherk, Vanessa D; Santangelo, Kelly S

    2017-12-11

    There is potential discrepancy between human and laboratory animal studies of osteoarthritis (OA), as radiographic assessment is the hallmark of the former and histopathology the standard for the latter. This suggests a need to evaluate OA in animal models in a manner similar to that utilized in people. Our study aimed to develop a whole joint grading scheme for microcomputed tomography (microCT) images in Hartley guinea pigs, a strain that recapitulates joint changes highlighted in human spontaneous OA. Knees from animals aged 2, 3, 5, 9, and 15 months were evaluated via whole joint microCT and standard histologic scoring. Quantitative microCT parameters, such as bone volume/total volume were also collected. Both whole joint microCT and histologic scores increased with advancing age and showed strong correlation (r = 0.89. p system in guinea pig studies of OA, as it provides important information regarding bony changes that occur at a different rate than articular cartilage changes. This grading scheme, in conjunction with histology and quantitative microCT measurements, may enhance the translational value of this animal model as it pertains to human work.

  10. Effect of Post Space Preparation on Apical Obturation Quality of Teeth Obturated with Different Techniques: A Micro-computed Tomographic Study.

    Science.gov (United States)

    Küçükkaya Eren, Selen; Askerbeyli Örs, Sevinc; Yılmaz, Zeliha

    2017-07-01

    The purpose of this study was to evaluate the obturation quality of root canals filled with different techniques and to determine whether post space preparation had an effect on the quality of apical obturation using micro-computed tomographic (micro-CT) imaging. The root canals of 30 human mandibular premolar teeth were instrumented, and the specimens were divided into 3 groups according to the obturation technique used: cold lateral compaction (CLC), warm vertical compaction (WVC), or single-cone (SC) techniques. The specimens were stored at 37°C and 100% humidity for 1 week. Then, the coronal root filling material was removed in order to create a post space. Micro-CT scans were performed before and after post space preparation for the volumetric analysis of voids and filling materials. Data were analyzed using repeated-measures analysis of variance and Bonferroni tests. The CLC and SC groups showed a significantly greater percentage volume of voids than the WVC group (P space preparation (P > .05). The post space preparation caused a significant increase in the percentage volume of voids in the CLC and SC groups (P space preparation (P > .05). No root fillings were void free. The WVC group presented the best obturation quality. The post space preparation negatively influenced the apical integrity of the filling materials in the CLC and SC groups, whereas it had no significant effect in the WVC group. Copyright © 2017 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  11. The effects of the Er:YAG laser on trabecular bone micro-architecture: Comparison with conventional dental drilling by micro-computed tomographic and histological techniques.

    Science.gov (United States)

    Zeitouni, Jihad; Clough, Bret; Zeitouni, Suzanne; Saleem, Mohammed; Al Aisami, Kenan; Gregory, Carl

    2017-01-01

    Background : The use of lasers has become increasingly common in the field of medicine and dentistry, and there is a growing need for a deeper understanding of the procedure and its effects on tissue. The aim of this study was to compare the erbium-doped yttrium aluminium garnet (Er:YAG) laser and conventional drilling techniques, by observing the effects on trabecular bone microarchitecture and the extent of thermal and mechanical damage. Methods : Ovine femoral heads were employed to mimic maxillofacial trabecular bone, and cylindrical osteotomies were generated to mimic implant bed preparation. Various laser parameters were tested, as well as a conventional dental drilling technique. The specimens were then subjected to micro-computed tomographic (μCT) histomorphometic analysis and histology. Results : Herein, we demonstrate that mCT measurements of trabecular porosity provide quantitative evidence that laser-mediated cutting preserves the trabecular architecture and reduces thermal and mechanical damage at the margins of the cut. We confirmed these observations with histological studies. In contrast with laser-mediated cutting, conventional drilling resulted in trabecular collapse, reduction of porosity at the margin of the cut and histological signs of thermal damage. Conclusions : This study has demonstrated, for the first time, that mCT and quantification of porosity at the margin of the cut provides a quantitative insight into damage caused by bone cutting techniques. We further show that with laser-mediated cutting, the marrow remains exposed to the margins of the cut, facilitating cellular infiltration and likely accelerating healing. However, with drilling, trabecular collapse and thermal damage is likely to delay healing by restricting the passage of cells to the site of injury and causing localized cell death.

  12. Airway remodeling in a mouse asthma model assessed by in-vivo respiratory-gated micro-computed tomography

    Energy Technology Data Exchange (ETDEWEB)

    Lederlin, Mathieu; Montaudon, Michel [Universite Bordeaux 2, Laboratoire de Physiologie Cellulaire Respiratoire, Bordeaux (France); Institut National de la Sante et de la Recherche Medicale (INSERM), U885, Bordeaux (France); CHU Bordeaux, Unite d' Imagerie Thoracique, Pessac (France); Ozier, Annaig; Begueret, Hugues; Ousova, Olga; Marthan, Roger; Berger, Patrick [Universite Bordeaux 2, Laboratoire de Physiologie Cellulaire Respiratoire, Bordeaux (France); Institut National de la Sante et de la Recherche Medicale (INSERM), U885, Bordeaux (France); Laurent, Francois [Universite Bordeaux 2, Laboratoire de Physiologie Cellulaire Respiratoire, Bordeaux (France); Institut National de la Sante et de la Recherche Medicale (INSERM), U885, Bordeaux (France); CHU Bordeaux, Unite d' Imagerie Thoracique, Pessac (France); CHU de Bordeaux, Hopital du Haut-Leveque, Hopital Cardiologique, Unite d' Imagerie Thoracique et Cardiovasculaire, Pessac (France)

    2010-01-15

    The aim of our study was to evaluate the feasibility of non-invasive respiratory-gated micro-computed tomography (micro-CT) for assessment of airway remodelling in a mouse asthma model. Six female BALB/c mice were challenged intranasally with ovalbumin. A control group of six mice received saline inhalation. All mice underwent plethysmographic study and micro-CT. For each mouse, peribronchial attenuation values of 12 bronchi were measured, from which a peribronchial density index (PBDI) was computed. Mice were then sacrificed and lungs examined histologically. Final analysis involved 10 out of 12 mice. Agreement of measurements across observers and over time was very good (intraclass correlation coefficients: 0.94-0.98). There was a significant difference in PBDI between asthmatic and control mice (-210 vs. -338.9 HU, P=0.008). PBDI values were correlated to bronchial muscle area (r=0.72, P=0.018). This study shows that respiratory-gated micro-CT may allow non-invasive monitoring of bronchial remodelling in asthmatic mice and evaluation of innovative treatment effects. (orig.)

  13. Combined microcomputed tomography, biomechanical and histomorphometric analysis of the peri-implant bone: a pilot study in minipig model.

    Science.gov (United States)

    Gramanzini, Matteo; Gargiulo, Sara; Zarone, Fernando; Megna, Rosario; Apicella, Antonio; Aversa, Raffaella; Salvatore, Marco; Mancini, Marcello; Sorrentino, Roberto; Brunetti, Arturo

    2016-06-01

    To present a practical approach that combines biomechanical tests, microcomputed tomography (μCT) and histomorphometry, providing quantitative results on bone structure and mechanical properties in a minipig model, in order to investigate the specific response to an innovative dental biomaterial. Titanium implants with innovative three-dimensional scaffolds were inserted in the tibias of 4 minipigs. Primary stability and osseointegration were investigated by means of insertion torque (IT) values, resonance frequency analysis (RFA), bone-to-implant contact (BIC), bone mineral density (BMD) and stereological measures of trabecular bone. A significant positive correlation was found between IT and RFA (r=0.980, p=0.0001). BMD at the implant sites was 18% less than the reference values (p=0.0156). Peri-implant Tb.Th was 50% higher, while Tb.N was 50% lower than the reference zone (pdestructive evaluation of bone architecture and mineral density, in combination with conventional analysis methods. The proposed multimodal approach allows to improve accuracy and reproducibility for peri-implant bone measurements and could support future investigations. Copyright © 2016 Academy of Dental Materials. Published by Elsevier Ltd. All rights reserved.

  14. Analysis of single point and continuous wave of condensation root filling techniques by micro-computed tomography

    Directory of Open Access Journals (Sweden)

    Daniele Angerame

    2012-01-01

    Full Text Available The aim of the present microtomographic study was to investigate the quality of root canal filling and the voids formation in canals of extracted teeth instrumented with a simultaneous technique and filled with two different methods. Twenty-four single-rooted teeth were assigned to two experimental groups (no. = 12; canals were shaped with NiTi rotary files, irrigated with NaOCl and filled either with the single point (group 1 or the continuous wave of condensation technique (group 2. Specimens underwent microtomographic scanning. Collected data were statistically analyzed by nonparametric methods. Void mean percentages were found to be limited and similar between the two groups; the single point technique led to greater sealer thickness in partially oval canals.

  15. Marginal adaptation of mineral trioxide aggregate (MTA) to root dentin surface with orthograde/retrograde application techniques: A microcomputed tomographic analysis.

    Science.gov (United States)

    Al Fouzan, Khalid; Awadh, Mohammed; Badwelan, Moahmmed; Gamal, Abeer; Geevarghese, Amrita; Babhair, Samar; Al-Rejaie, Mansour; Al Hezaimi, Khalid; Rotstein, Ilan

    2015-01-01

    Achieving a good apical seal for root canals is known to be associated with good mineral trioxide aggregate (MTA) adaptation to dentin. This study aims to compare the marginal adaptation of MTA with root dentin between orthograde and retrograde application techniques using microcomputed tomography (micro-CT) analysis. Fifty-two single-rooted human teeth were divided into four equal groups: (Group 1) Retrograde MTA (RMTA), (Group 2) Orthograde MTA (OMTA), (Group 3) Etched RMTA (ERMTA), and (Group 4) Etched OMTA (EOMTA). For Group 1, 3-mm retrograde cavities were prepared and filled with MTA. For Group 2, the apical 6 mm of the canals were filled with MTA and sealed with sealer cement and warm gutta-percha. In Groups 3 and 4, canals were treated the same as Groups 1 and 2, respectively, except that before placing the MTA, canals were irrigated with 17% ethylenediaminetetraacetic acid (EDTA). After 48 hours, all the teeth were analyzed using a micro-CT scanner. Mean dentin-MTA contact and the mean length and width of each gap was analysed using one-way analysis of variance (ANOVA). Statistical significance was set at an α level of 5%. No significant difference in gap volumes was observed in the dentin-MTA adaptation in both orthograde and retrograde application techniques. However, significant difference in the gap volumes was observed between RMTA and ERMTA (P = 0.045). Etching significantly improved the MTA-Dentin adaptation (P MTA adaptation, instead with the use of 17% EDTA, a significant improvement could be achieved. Within the limitations of the present study, it concludes that MTA adaptation to dentin tooth structure is not significantly different between an orthograde and retrograde approach. However, the use of EDTA significantly improved the MTA-Dentin adaptation.

  16. Tunable dynamic moduli of magnetic elastomers: from characterization by x-ray micro-computed tomography to mesoscopic modeling

    Science.gov (United States)

    Pessot, Giorgio; Schümann, Malte; Gundermann, Thomas; Odenbach, Stefan; Löwen, Hartmut; Menzel, Andreas M.

    2018-03-01

    Ferrogels and magnetorheological elastomers are composite materials obtained by embedding magnetic particles of mesoscopic size in a crosslinked polymeric matrix. They combine the reversible elastic deformability of polymeric materials with the high responsivity of ferrofluids to external magnetic fields. These materials stand out, for example, for significant magnetostriction as well as a pronounced increase of the elastic moduli in the presence of external magnetic fields. By means of x-ray micro-computed tomography, the position and size of each magnetic particle can be measured with a high degree of accuracy. We here use data extracted from real magnetoelastic samples as input for coarse-grained dipole-spring modeling and calculations to investigate internal restructuring, stiffening, and changes in the normal modes spectrum. More precisely, we assign to each particle a dipole moment proportional to its volume and set a randomized network of springs between them that mimics the behavior of the polymeric elastic matrix. Extending our previously developed methods, we compute the resulting structural changes in the systems as well as the frequency-dependent elastic moduli when magnetic interactions are turned on. Particularly, with increasing magnetization, we observe the formation of chain-like aggregates. Interestingly, the static elastic moduli can first show a slight decrease with growing amplitude of the magnetic interactions, before a pronounced increase appears upon the chain formation. The change of the dynamic moduli with increasing magnetization depends on the frequency and can even feature nonmonotonic behavior. Overall, we demonstrate how theory and experiments can complement each other to learn more about the dynamic behavior of this interesting class of materials.

  17. Tevatron extraction microcomputer

    International Nuclear Information System (INIS)

    Chapman, L.; Finley, D.A.; Harrison, M.; Merz, W.; Batavia, IL)

    1985-01-01

    Extraction in the Fermilab Tevatron is controlled by a multi-processor Multibus microcomputer system called QXR (Quad eXtraction Regulator). QXR monitors several analog beam signals and controls three sets of power supplies: the ''bucker'' and ''pulse'' magnets at a rate of 5760 Hz, and the ''QXR'' magnets at 720 Hz. QXR supports multiple slow spills (up to a total of 35 seconds) with multiple fast pulses intermixed. It linearizes the slow spill and bucks out the high frequency components. Fast extraction is done by outputting a variable pulse waveform. Closed loop learning techniques are used to improve performance from cycle to cycle for both slow and fast extraction. The system is connected to the Tevatron clock system so that it can track the machine cycle. QXR is also connected to the rest of the Fermilab control system, ACNET. Through ACNET, human operators and central computers can monitor and control extraction through communications with QXR. The controls hardware and software both employ some standard and some specialized components. This paper gives an overview of QXR as a control system; another paper (1) summarizes performance

  18. Tevatron extraction microcomputer

    International Nuclear Information System (INIS)

    Chapman, L.; Finley, D.A.; Harrison, M.; Merz, W.

    1985-06-01

    Extraction in the Fermilab Tevatron is controlled by a multi-processor Multibus microcomputer system called QXR (Quad eXtraction Regulator). QXR monitors several analog beam signals and controls three sets of power supplies: the ''bucker'' and ''pulse'' magnets at a rate of 5760 Hz, and the ''QXR'' magnets at 720 Hz. QXR supports multiple slow spills (up to a total of 35 seconds) with multiple fast pulses intermixed. It linearizes the slow spill and bucks out the high frequency components. Fast extraction is done by outputting a variable pulse waveform. Closed loop learning techniques are used to improve performance from cycle to cycle for both slow and fast extraction. The system is connected to the Tevatron clock system so that it can track the machine cycle. QXR is also connected to the rest of the Fermilab control system, ACNET. Through ACNET, human operators and central computers can monitor and control extraction through communications with QXR. The controls hardware and software both employ some standard and some specialized components. This paper gives an overview of QXR as a control system; another paper summarizes performance

  19. Microcomputer model for an analysis of the financial feasibility of a mining project

    International Nuclear Information System (INIS)

    Ciruelos, J.; Duchene, M.

    1983-01-01

    The model presented permits a simulation of the predicted profitability of a mining project at the stage of feasibility studies by making use of a simple individual computer, the Apple II. The model presented can be used to treat the following three areas: definition of the mode of financing the project and calculation of the financial flows which make it possible to evaluate the profitability of this project; analysis of sensitivity, which makes it possible to determine the most critical variables for the future of the project; analysis of the risk [fr

  20. Semiconductor Modeling Techniques

    CERN Document Server

    Xavier, Marie

    2012-01-01

    This book describes the key theoretical techniques for semiconductor research to quantitatively calculate and simulate the properties. It presents particular techniques to study novel semiconductor materials, such as 2D heterostructures, quantum wires, quantum dots and nitrogen containing III-V alloys. The book is aimed primarily at newcomers working in the field of semiconductor physics to give guidance in theory and experiment. The theoretical techniques for electronic and optoelectronic devices are explained in detail.

  1. Micro-computed tomography (CT) based assessment of dental regenerative therapy in the canine mandible model

    Science.gov (United States)

    Khobragade, P.; Jain, A.; Setlur Nagesh, S. V.; Andreana, S.; Dziak, R.; Sunkara, S. K.; Sunkara, S.; Bednarek, D. R.; Rudin, S.; Ionita, C. N.

    2015-03-01

    High-resolution 3D bone-tissue structure measurements may provide information critical to the understanding of the bone regeneration processes and to the bone strength assessment. Tissue engineering studies rely on such nondestructive measurements to monitor bone graft regeneration area. In this study, we measured bone yield, fractal dimension and trabecular thickness through micro-CT slices for different grafts and controls. Eight canines underwent surgery to remove a bone volume (defect) in the canine's jaw at a total of 44 different locations. We kept 11 defects empty for control and filled the remaining ones with three regenerative materials; NanoGen (NG), a FDA-approved material (n=11), a novel NanoCalcium Sulfate (NCS) material (n=11) and NCS alginate (NCS+alg) material (n=11). After a minimum of four and eight weeks, the canines were sacrificed and the jaw samples were extracted. We used a custombuilt micro-CT system to acquire the data volume and developed software to measure the bone yield, fractal dimension and trabecular thickness. The software used a segmentation algorithm based on histograms derived from volumes of interest indicated by the operator. Using bone yield and fractal dimension as indices we are able to differentiate between the control and regenerative material (p<0.005). Regenerative material NCS showed an average 63.15% bone yield improvement over the control sample, NCS+alg showed 55.55% and NanoGen showed 37.5%. The bone regeneration process and quality of bone were dependent upon the position of defect and time period of healing. This study presents one of the first quantitative comparisons using non-destructive Micro-CT analysis for bone regenerative material in a large animal with a critical defect model. Our results indicate that Micro-CT measurement could be used to monitor invivo bone regeneration studies for greater regenerative process understanding.

  2. Mathematical modelling techniques

    CERN Document Server

    Aris, Rutherford

    1995-01-01

    ""Engaging, elegantly written."" - Applied Mathematical ModellingMathematical modelling is a highly useful methodology designed to enable mathematicians, physicists and other scientists to formulate equations from a given nonmathematical situation. In this elegantly written volume, a distinguished theoretical chemist and engineer sets down helpful rules not only for setting up models but also for solving the mathematical problems they pose and for evaluating models.The author begins with a discussion of the term ""model,"" followed by clearly presented examples of the different types of mode

  3. Microcomputers and Preschoolers.

    Science.gov (United States)

    Evans, Dina

    Preschool children can benefit by working with microcomputers. Thinking skills are enhanced by software games that focus on logic, memory, problem solving, and pattern recognition. Counting, sequencing, and matching games develop mathematics skills, and word games focusing on basic letter symbol and word recognition develop language skills.…

  4. Boundary representation modelling techniques

    CERN Document Server

    2006-01-01

    Provides the most complete presentation of boundary representation solid modelling yet publishedOffers basic reference information for software developers, application developers and users Includes a historical perspective as well as giving a background for modern research.

  5. A microcomputer-based model for identifying urban and suburban roadways with critical large truck accident rates

    International Nuclear Information System (INIS)

    Brogan, J.D.; Cashwell, J.W.

    1992-01-01

    This paper presents an overview of techniques for merging highway accident record and roadway inventory files and employing the combined data set to identify spots or sections on highway facilities in urban and suburban areas with unusually high large truck accident rates. A statistical technique, the rate/quality control method, is used to calculate a critical rate for each location of interest. This critical rate may then be compared to the location's actual accident rate to identify locations for further study. Model enhancements and modifications are described to enable the technique to be employed in the evaluation of routing alternatives for the transport of radioactive material

  6. Graphics Tablet for the BBC Microcomputer.

    Science.gov (United States)

    Whale, R.

    1984-01-01

    Describes an inexpensive solution to the problem of transferring pictures onto a television screen using the analog part of the Model "B" BBC Microcomputer. Instruction for building the graphics tablet (which can easily be constructed by secondary students), program listing for required software, and documentation are included. (JN)

  7. Microcomputers and computer networks

    International Nuclear Information System (INIS)

    Owens, J.L.

    1976-01-01

    Computers, for all their speed and efficiency, have their foibles and failings. Until the advent of minicomputers, users often had to supervise their programs personally to make sure they executed correctly. Minicomputers could take over some of these chores, but they were too expensive to be dedicated to any but the most vital services. Inexpensive, easily programmed microcomputers are easing this limitation, and permitting a flood of new applications. 3 figures

  8. Microcomputer interfacing and applications

    CERN Document Server

    Mustafa, M A

    1990-01-01

    This is the applications guide to interfacing microcomputers. It offers practical non-mathematical solutions to interfacing problems in many applications including data acquisition and control. Emphasis is given to the definition of the objectives of the interface, then comparing possible solutions and producing the best interface for every situation. Dr Mustafa A Mustafa is a senior designer of control equipment and has written many technical articles and papers on the subject of computers and their application to control engineering.

  9. Survey of semantic modeling techniques

    Energy Technology Data Exchange (ETDEWEB)

    Smith, C.L.

    1975-07-01

    The analysis of the semantics of programing languages was attempted with numerous modeling techniques. By providing a brief survey of these techniques together with an analysis of their applicability for answering semantic issues, this report attempts to illuminate the state-of-the-art in this area. The intent is to be illustrative rather than thorough in the coverage of semantic models. A bibliography is included for the reader who is interested in pursuing this area of research in more detail.

  10. A General Introduction to Microcomputers.

    Science.gov (United States)

    Muiznieks, Viktors

    This basic introduction to microcomputers provides the neophyte with the terminology, definitions, and concepts that explain the microcomputer and computing technology in general. Mathematical operations with binary numbers, computer storage, controlling logic, and the concepts of stack and interrupt are explained. (RAO)

  11. Facility equipment performance evaluation using microcomputer simulation analysis

    Energy Technology Data Exchange (ETDEWEB)

    Chockie, A.D.; Hostick, C.J.

    1985-11-01

    The Pacific Northwest Laboratory (PNL) recently developed a facility performance assessment model as part of the US Department of Energy-sponsored monitored retrievable storage (MRS) program at PNL. The objective in the development of this model was to assist in the evaluation of the operational performance of the MRS facility design alternatives. The microcomputer-based simulation model provided a technique for the analysis of the design and performance of alternative MRS facility systems. The model was applied to the analysis of the material flow, equipment capability, and facility layout of various designs for a facility to receive and canister spent fuel from commercial nuclear power plants. Programs were also developed that evaluated alternative facility operating schedules and facility-equipment designs. The paper is a review of the facility performance assessment model and its advantages and benefits in the analysis of alternative facility designs employing varying degrees of remote handling capability.

  12. Microcomputer versus mainframe simulations: A case study

    Science.gov (United States)

    Bengtson, Neal M.

    1988-01-01

    The research was conducted to two parts. Part one consisted of a study of the feasibility of running the Space Transportation Model simulation on an office IBM-AT. The second part was to design simulation runs so as to study the effects of certain performance factors on the execution of the simulation model. The results of this research are given in the two reports which follow: Microcomputer vs. Mainframe Simulation: A Case Study and Fractional Factorial Designs of Simulation Runs for the Space Transportation System Operations Model. In the first part, a DOS batch job was written in order to simplify the execution of the simulation model on an office microcomputer. A comparison study was then performed of running the model on NASA-Langley's mainframe computer vs. running on the IBM-AT microcomputer. This was done in order to find the advantages and disadvantages of running the model on each machine with the objective of determining if running of the office PC was practical. The study concluded that it was. The large number of performance parameters in the Space Transportation model precluded running a full factorial design needed to determine the most significant design factors. The second report gives several suggested fractional factorial designs which require far fewer simulation runs in order to determine which factors have significant influence on results.

  13. Advanced Atmospheric Ensemble Modeling Techniques

    Energy Technology Data Exchange (ETDEWEB)

    Buckley, R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Chiswell, S. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Kurzeja, R. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Maze, G. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Viner, B. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Werth, D. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2017-09-29

    Ensemble modeling (EM), the creation of multiple atmospheric simulations for a given time period, has become an essential tool for characterizing uncertainties in model predictions. We explore two novel ensemble modeling techniques: (1) perturbation of model parameters (Adaptive Programming, AP), and (2) data assimilation (Ensemble Kalman Filter, EnKF). The current research is an extension to work from last year and examines transport on a small spatial scale (<100 km) in complex terrain, for more rigorous testing of the ensemble technique. Two different release cases were studied, a coastal release (SF6) and an inland release (Freon) which consisted of two release times. Observations of tracer concentration and meteorology are used to judge the ensemble results. In addition, adaptive grid techniques have been developed to reduce required computing resources for transport calculations. Using a 20- member ensemble, the standard approach generated downwind transport that was quantitatively good for both releases; however, the EnKF method produced additional improvement for the coastal release where the spatial and temporal differences due to interior valley heating lead to the inland movement of the plume. The AP technique showed improvements for both release cases, with more improvement shown in the inland release. This research demonstrated that transport accuracy can be improved when models are adapted to a particular location/time or when important local data is assimilated into the simulation and enhances SRNL’s capability in atmospheric transport modeling in support of its current customer base and local site missions, as well as our ability to attract new customers within the intelligence community.

  14. [Comparison of effectiveness and safety between Twisted File technique and ProTaper Universal rotary full sequence based on micro-computed tomography].

    Science.gov (United States)

    Chen, Xiao-bo; Chen, Chen; Liang, Yu-hong

    2016-02-18

    To evaluate the efficacy and security of two type of rotary nickel titanium system (Twisted File and ProTaper Universal) for root canal preparation based on micro-computed tomography(micro-CT). Twenty extracted molars (including 62 canals) were divided into two experimental groups and were respectively instrumented using Twisted File rotary nickel titanium system (TF) and ProTaper Universal rotary nickel titanium system (PU) to #25/0.08 following recommended protocol. Time for root canal instrumentation (accumulation of time for every single file) was recorded. The 0-3 mm root surface from apex was observed under an optical stereomicroscope at 25 × magnification. The presence of crack line was noted. The root canals were scanned with micro-CT before and after root canal preparation. Three-dimensional shape images of canals were reconstructed, calculated and evaluated. The amount of canal central transportation of the two groups was calculated and compared. The shorter preparation time [(0.53 ± 0.14) min] was observed in TF group, while the preparation time of PU group was (2.06 ± 0.39) min (Pvs. (0.097 ± 0.084) mm, P<0.05]. No instrument separation was observed in both the groups. Cracks were not found in both the groups either based in micro-CT images or observation under an optical stereomicroscope at 25 × magnification. Compared with ProTaper Universal, Twisted File took less time in root canal preparation and exhibited better shaping ability, and less canal transportation.

  15. The effects of the Er:YAG laser on trabecular bone micro-architecture: Comparison with conventional dental drilling by micro-computed tomographic and histological techniques [version 1; referees: 2 approved

    Directory of Open Access Journals (Sweden)

    Jihad Zeitouni

    2017-07-01

    Full Text Available Background: The use of lasers has become increasingly common in the field of medicine and dentistry, and there is a growing need for a deeper understanding of the procedure and its effects on tissue. The aim of this study was to compare the erbium-doped yttrium aluminium garnet (Er:YAG laser and conventional drilling techniques, by observing the effects on trabecular bone microarchitecture and the extent of thermal and mechanical damage. Methods: Ovine femoral heads were employed to mimic maxillofacial trabecular bone, and cylindrical osteotomies were generated to mimic implant bed preparation. Various laser parameters were tested, as well as a conventional dental drilling technique. The specimens were then subjected to micro-computed tomographic (μCT histomorphometic analysis and histology. Results: Herein, we demonstrate that mCT measurements of trabecular porosity provide quantitative evidence that laser-mediated cutting preserves the trabecular architecture and reduces thermal and mechanical damage at the margins of the cut. We confirmed these observations with histological studies. In contrast with laser-mediated cutting, conventional drilling resulted in trabecular collapse, reduction of porosity at the margin of the cut and histological signs of thermal damage. Conclusions: This study has demonstrated, for the first time, that mCT and quantification of porosity at the margin of the cut provides a quantitative insight into damage caused by bone cutting techniques. We further show that with laser-mediated cutting, the marrow remains exposed to the margins of the cut, facilitating cellular infiltration and likely accelerating healing. However, with drilling, trabecular collapse and thermal damage is likely to delay healing by restricting the passage of cells to the site of injury and causing localized cell death.

  16. Reactor physics using a microcomputer

    International Nuclear Information System (INIS)

    Murray, R.L.

    1983-01-01

    The object of the work reported is to develop educational computer modules for all aspects of reactor physics. The modules consist of a description of the theory, mathematical method, computer program listing, sample calculations, and problems for the student, along with a card deck. Modules were first written in FORTRAN for an IBM 360/75, then later in BASIC for microcomputers. Problems include: limitation of equipment, choice of format for the program, the variety of dialects of BASIC used in the different microcomputer and peripherals brands, and knowing when to quit in the process of developing a program

  17. Microcomputed tomography-based assessment of retrieved dental implants

    NARCIS (Netherlands)

    Narra, N.; Antalainen, A.K.; Zipprich, H.; Sándor, G.K.; Wolff, J.

    2015-01-01

    Purpose: The aim of this study was to demonstrate the potential of microcomputed tomography (micro-CT) technology in the assessment of retrieved dental implants. Cases are presented to illustrate the value of micro-CT imaging techniques in determining possible mechanical causes for dental implant

  18. Microcomputed Tomography-Based Assessment of Retrieved Dental Implants

    NARCIS (Netherlands)

    Narra, N.; Antalainen, A.K.; Zipprich, H.; Sandor, G.K.; Wolff, J.

    2015-01-01

    PURPOSE: The aim of this study was to demonstrate the potential of microcomputed tomography (micro-CT) technology in the assessment of retrieved dental implants. Cases are presented to illustrate the value of micro-CT imaging techniques in determining possible mechanical causes for dental implant

  19. A full automatic system controlled with IBM-PC/XT micro-computer for neutron activation analysis

    International Nuclear Information System (INIS)

    Song Quanxun

    1992-01-01

    A full automatic system controlled with micro-computers for NAA is described. All processes are automatically completed with an IBM-PC/XT micro-computer. The device is stable, reliable, flexible and convenient for use and has many functions and applications in automatical analysis of long, middle and short lived nuclides. Due to a high working efficiency of the instrument and micro-computers, both time and power can be saved. This method can be applied in other nuclear analysis techniques

  20. History Microcomputer Games: Update 2.

    Science.gov (United States)

    Sargent, James E.

    1985-01-01

    Provides full narrative reviews of B-1 Nuclear Bomber (Avalon, 1982); American History Adventure (Social Science Microcomputer Review Software, 1985); Government Simulations (Prentice-Hall, 1985); and The Great War, FDR and the New Deal, and Hitler's War, all from New Worlds Software, 1985. Lists additional information on five other history and…

  1. History Microcomputer Simulations: An Update.

    Science.gov (United States)

    Sargent, James E.

    1985-01-01

    This article summarizes seven currently available microcomputer simulations for history. The seven included are: Government Simulations (Prentice-Hall, 1984), Nomination (Brady Company, 1984), Tigers in the Snow (Strategic Simulations, 1981), Lincoln's Decisions (Educational Activities, 1982), Watergate Simulation (Social Science Research and…

  2. Prototype Microcomputer Courseware for Teaching High School Algebra. Final Report.

    Science.gov (United States)

    Dugdale, Sharon; Kibbey, David

    This project used an "intrinsic models" approach to develop a sample of microcomputer materials on graphing linear and quadratic equations and on recognition of general functional relationships. The "intrinsic models" approach uses motivating activities which are direct expressions of the underlying mathematics. The area of…

  3. The microcomputer scientific software series 6: ECOPHYS user's manual.

    Science.gov (United States)

    George E. Host; H. Michael Rauscher; J. G. Isebrands; Donald I. Dickmann; Richard E. Dickson; Thomas R. Crow; D.A. Michael

    1990-01-01

    ECOPHYS is an ecophysiological whole-tree growth process model designed to simulate the growth of poplar in the establishment year. This microcomputer-based model may be used to test the influence of genetically determined physiological or morphological attributes on plant growth. This manual describes the installation, file structures, and operational procedures for...

  4. A multi-channel microcomputer data acquisition system

    International Nuclear Information System (INIS)

    Loureiro, J.S.

    1987-01-01

    A data acquisition system was developed in order to transfer automatically to a 64 kb microcomputer the data generated by a nuclear spectroscopy system in a multichannel analyser. The data in the memory are stored in a floppy disk and will be further used as data entry for any spectrum analysis program, eliminating the tedious work of manually digitizing the spectrum and the possible mistakes associated with it. The developed system connected a POLYMAX 201 DP microcomputer, under CP/M operational system, to a NUCLEAR DATA MODEL ND-65 multichannel analyser and was planned for either local spectrum analysis in the microcomputer using a simplified program, or remote analysis in a mainframe using the sophisticated analysis program SAMPO. With the present system, the time spent between printing out of the 4096 channels with the multichannel analyser printer and its corresponding introduction in the analysis program has been reduced from about 6 hours to less than 2 minutes. (author)

  5. The microcomputer and image analysis in diagnostic pathology.

    Science.gov (United States)

    Jarvis, L R

    1992-06-01

    This paper presents a snapshot view of the influence and direction of microcomputer technology for image analysis techniques in diagnostic pathology. Microcomputers have had considerable impact in bringing image analysis to wider application. Semi-automated tracing techniques are a simple means of providing objective data and assist in a wide range of diagnostic problems. From the common theme of reducing subjectivity in diagnostic assessment, an extensive body of research has accrued. Some studies have addressed the need for quality control for reliable, routine application. Video digitizer cards bring digital image analysis within the reach of laboratory budgets, providing powerful tools for investigation of a wide range of cellular and tissue features. The use of staining procedures compatible with quantitative evaluation has become equally important. As well as assisting scene segmentation, cytochemical and immunochemical staining techniques relate the data to biological processes. With the present state of the art, practical use of microcomputer based image analysis is impaired by limitations of information extraction and specimen throughput. Recent advances in colour video imaging provide an extra dimension in the analysis of multi-spectral stains. Improvements will also be felt with predictable increase in speed of microprocessors, and with single chip devices which deliver video rate processing. If the full potential of this hardware is realized, high-speed, routine analysis becomes feasible. In addition, a microcomputer imaging system can play host to companion functions, such as image archiving and transmission. With this outlook, the use of microcomputers for image analysis in diagnostic pathology is certain to increase.(ABSTRACT TRUNCATED AT 250 WORDS)

  6. Microcomputer Techniques for Developing Hardwood Conservation Strategies

    Science.gov (United States)

    Steve McNiel

    1991-01-01

    Hardwoods are disappearing from the California landscape at alarming rates. This is due to a variety of influences, both natural and man made. It is clear that conservation and rehabilitation of hardwood resources will require a large effort on the part of research institutes, universities, government agencies, special interest groups, private developers, maintenance...

  7. Protected Microcomputers from VUVT-Engineering

    Directory of Open Access Journals (Sweden)

    Martin Sec

    2006-01-01

    Full Text Available The paper presents properties of new protected microcomputer prototypes that recently developed VÚVTEngineering a.s. Žilina. The five new prototypes of protected microcomputers and displays provide users with important protection against radiated information leaking from a computer and suppress unfavorable effects of electromagnetic radiation on a man. The first part of contribution describes technical parameters of protected microcomputers and displays. The results of optical characteristics measurement and measurement of electric field radiation from the protected microcomputers are discussed in the second part of the contribution.

  8. Sequence analysis on microcomputers.

    Science.gov (United States)

    Cannon, G C

    1987-10-02

    Overall, each of the program packages performed their tasks satisfactorily. For analyses where there was a well-defined answer, such as a search for a restriction site, there were few significant differences between the program sets. However, for tasks in which a degree of flexibility is desirable, such as homology or similarity determinations and database searches, DNASTAR consistently afforded the user more options in conducting the required analysis than did the other two packages. However, for laboratories where sequence analysis is not a major effort and the expense of a full sequence analysis workstation cannot be justified, MicroGenie and IBI-Pustell offer a satisfactory alternative. MicroGenie is a polished program system. Many may find that its user interface is more "user friendly" than the standard menu-driven interfaces. Its system of filing sequences under individual passwords facilitates use by more than one person. MicroGenie uses a hardware device for software protection that occupies a card slot in the computer on which it is used. Although I am sympathetic to the problem of software piracy, I feel that a less drastic solution is in order for a program likely to be sharing limited computer space with other software packages. The IBI-Pustell package performs the required analysis functions as accurately and quickly as MicroGenie but it lacks the clearness and ease of use. The menu system seems disjointed, and new or infrequent users often find themselves at apparent "dead-end menus" where the only clear alternative is to restart the entire program package. It is suggested from published accounts that the user interface is going to be upgraded and perhaps when that version is available, use of the system will be improved. The documentation accompanying each package was relatively clear as to how to run the programs, but all three packages assumed that the user was familiar with the computational techniques employed. MicroGenie and IBI-Pustell further

  9. Advanced structural equation modeling issues and techniques

    CERN Document Server

    Marcoulides, George A

    2013-01-01

    By focusing primarily on the application of structural equation modeling (SEM) techniques in example cases and situations, this book provides an understanding and working knowledge of advanced SEM techniques with a minimum of mathematical derivations. The book was written for a broad audience crossing many disciplines, assumes an understanding of graduate level multivariate statistics, including an introduction to SEM.

  10. Using Microcomputers in Schools: Some Initial Considerations.

    Science.gov (United States)

    Higgins, James E.; Linville, William J.

    Both instructional and administrative uses of microcomputers are considered in this monograph, which addresses the pressing questions of educators and administrators as indicated by requests for assistance directed to MICRONET, an Indiana school network that promotes microcomputer use in public schooling. The current condition in schools is…

  11. Integrating Mainframe Data Bases on a Microcomputer

    OpenAIRE

    Marciniak, Thomas A.

    1985-01-01

    Microcomputers support user-friendly software for interrogating their resident data bases. Many medical data bases currently consist of files on less accessible mainframe computers with more limited inquiry capabilities. We discuss the transferring and integrating of mainframe data into microcomputer data base systems in one medical environment.

  12. Implementing a Microcomputer Database Management System.

    Science.gov (United States)

    Manock, John J.; Crater, K. Lynne

    1985-01-01

    Current issues in selecting, structuring, and implementing microcomputer database management systems in research administration offices are discussed, and their capabilities are illustrated with the system used by the University of North Carolina at Wilmington. Trends in microcomputer technology and their likely impact on research administration…

  13. Microcomputer Database Management Systems for Bibliographic Data.

    Science.gov (United States)

    Pollard, Richard

    1986-01-01

    Discusses criteria for evaluating microcomputer database management systems (DBMS) used for storage and retrieval of bibliographic data. Two popular types of microcomputer DBMS--file management systems and relational database management systems--are evaluated with respect to these criteria. (Author/MBR)

  14. Verification of Orthogrid Finite Element Modeling Techniques

    Science.gov (United States)

    Steeve, B. E.

    1996-01-01

    The stress analysis of orthogrid structures, specifically with I-beam sections, is regularly performed using finite elements. Various modeling techniques are often used to simplify the modeling process but still adequately capture the actual hardware behavior. The accuracy of such 'Oshort cutso' is sometimes in question. This report compares three modeling techniques to actual test results from a loaded orthogrid panel. The finite element models include a beam, shell, and mixed beam and shell element model. Results show that the shell element model performs the best, but that the simpler beam and beam and shell element models provide reasonable to conservative results for a stress analysis. When deflection and stiffness is critical, it is important to capture the effect of the orthogrid nodes in the model.

  15. Microcomputer generated pipe support calculations

    International Nuclear Information System (INIS)

    Hankinson, R.F.; Czarnowski, P.; Roemer, R.E.

    1991-01-01

    The cost and complexity of pipe support design has been a continuing challenge to the construction and modification of commercial nuclear facilities. Typically, pipe support design or qualification projects have required large numbers of engineers centrally located with access to mainframe computer facilities. Much engineering time has been spent repetitively performing a sequence of tasks to address complex design criteria and consolidating the results of calculations into documentation packages in accordance with strict quality requirements. The continuing challenges of cost and quality, the need for support engineering services at operating plant sites, and the substantial recent advances in microcomputer systems suggested that a stand-alone microcomputer pipe support calculation generator was feasible and had become a necessity for providing cost-effective and high quality pipe support engineering services to the industry. This paper outlines the preparation for, and the development of, an integrated pipe support design/evaluation software system which maintains all computer programs in the same environment, minimizes manual performance of standard or repetitive tasks, and generates a high quality calculation which is consistent and easily followed

  16. A Strategy Modelling Technique for Financial Services

    OpenAIRE

    Heinrich, Bernd; Winter, Robert

    2004-01-01

    Strategy planning processes often suffer from a lack of conceptual models that can be used to represent business strategies in a structured and standardized form. If natural language is replaced by an at least semi-formal model, the completeness, consistency, and clarity of strategy descriptions can be drastically improved. A strategy modelling technique is proposed that is based on an analysis of modelling requirements, a discussion of related work and a critical analysis of generic approach...

  17. Possible Radiation-Induced Damage to the Molecular Structure of Wooden Artifacts Due to Micro-Computed Tomography, Handheld X-Ray Fluorescence, and X-Ray Photoelectron Spectroscopic Techniques

    Directory of Open Access Journals (Sweden)

    Madalena Kozachuk

    2016-05-01

    Full Text Available This study was undertaken to ascertain whether radiation produced by X-ray photoelectron spectroscopy (XPS, micro-computed tomography (μCT and/or portable handheld X-ray fluorescence (XRF equipment might damage wood artifacts during analysis. Changes at the molecular level were monitored by Fourier transform infrared (FTIR analysis. No significant changes in FTIR spectra were observed as a result of μCT or handheld XRF analysis. No substantial changes in the collected FTIR spectra were observed when XPS analytical times on the order of minutes were used. However, XPS analysis collected over tens of hours did produce significant changes in the FTIR spectra.

  18. "Hack" Is Not A Dirty Word--The Tenth Anniversary of Patron Access Microcomputer Centers in Libraries.

    Science.gov (United States)

    Dewey, Patrick R.

    1986-01-01

    The history of patron access microcomputers in libraries is described as carrying on a tradition that information and computer power should be shared. Questions that all types of libraries need to ask in planning microcomputer centers are considered and several model centers are described. (EM)

  19. Periodic precipitation a microcomputer analysis of transport and reaction processes in diffusion media, with software development

    CERN Document Server

    Henisch, H K

    1991-01-01

    Containing illustrations, worked examples, graphs and tables, this book deals with periodic precipitation (also known as Liesegang Ring formation) in terms of mathematical models and their logical consequences, and is entirely concerned with microcomputer analysis and software development. Three distinctive periodic precipitation mechanisms are included: binary diffusion-reaction; solubility modulation, and competitive particle growth. The book provides didactic illustrations of a valuable investigational procedure, in the form of hypothetical experimentation by microcomputer. The development

  20. A Method to Test Model Calibration Techniques

    Energy Technology Data Exchange (ETDEWEB)

    Judkoff, Ron; Polly, Ben; Neymark, Joel

    2016-08-26

    This paper describes a method for testing model calibration techniques. Calibration is commonly used in conjunction with energy retrofit audit models. An audit is conducted to gather information about the building needed to assemble an input file for a building energy modeling tool. A calibration technique is used to reconcile model predictions with utility data, and then the 'calibrated model' is used to predict energy savings from a variety of retrofit measures and combinations thereof. Current standards and guidelines such as BPI-2400 and ASHRAE-14 set criteria for 'goodness of fit' and assume that if the criteria are met, then the calibration technique is acceptable. While it is logical to use the actual performance data of the building to tune the model, it is not certain that a good fit will result in a model that better predicts post-retrofit energy savings. Therefore, the basic idea here is that the simulation program (intended for use with the calibration technique) is used to generate surrogate utility bill data and retrofit energy savings data against which the calibration technique can be tested. This provides three figures of merit for testing a calibration technique, 1) accuracy of the post-retrofit energy savings prediction, 2) closure on the 'true' input parameter values, and 3) goodness of fit to the utility bill data. The paper will also discuss the pros and cons of using this synthetic surrogate data approach versus trying to use real data sets of actual buildings.

  1. Peri-implant osseointegration after low-level laser therapy: micro-computed tomography and resonance frequency analysis in an animal model.

    Science.gov (United States)

    Mayer, Luciano; Gomes, Fernando Vacilotto; de Oliveira, Marília Gerhardt; de Moraes, João Feliz Duarte; Carlsson, Lennart

    2016-12-01

    The purpose of the present study is to evaluate the effects of low-level laser therapy on the osseointegration process by comparing resonance frequency analysis measurements performed at implant placement and after 30 days and micro-computed tomography images in irradiated vs nonirradiated rabbits. Fourteen male New Zealand rabbits were randomly divided into two groups of seven animals each, one control group (nonirradiated animals) and one experimental group that received low-level laser therapy (Thera Lase®, aluminum-gallium-arsenide laser diode, 10 J per spot, two spots per session, seven sessions, 830 nm, 50 mW, CW, Ø 0.0028 cm 2 ). The mandibular left incisor was surgically extracted in all animals, and one osseointegrated implant was placed immediately afterward (3.25ø × 11.5 mm; NanoTite, BIOMET 3i). Resonance frequency analysis was performed with the Osstell® device at implant placement and at 30 days (immediately before euthanasia). Micro-computed tomography analyses were then conducted using a high-resolution scanner (SkyScan 1172 X-ray Micro-CT) to evaluate the amount of newly formed bone around the implants. Irradiated animals showed significantly higher implant stability quotients at 30 days (64.286 ± 1.596; 95 % confidence interval (CI) 60.808-67.764) than controls (56.357 ± 1.596; 95 %CI 52.879-59.835) (P = .000). The percentage of newly formed bone around the implants was also significantly higher in irradiated animals (75.523 ± 8.510; 95 %CI 61.893-89.155) than in controls (55.012 ± 19.840; 95 %CI 41.380-68.643) (P = .027). Laser therapy, based on the irradiation protocol used in this study, was able to provide greater implant stability and increase the volume of peri-implant newly formed bone, indicating that laser irradiation effected an improvement in the osseointegration process.

  2. Microcomputer simulation of Stirling cryocoolers

    International Nuclear Information System (INIS)

    Walker, G.; Weiss, M.; Fauvel, O.R.

    1989-01-01

    A design aid for Stirling cryocoolers is described. The design aid is a digital simulation program for use on microcomputers. It is written in the Fortran 77 language and is especially designed to be easy to use with no specialist expertise required and minimal training. The program is versatile in that it can be used for all known types and variations of Striling cryocoolers. It is quick and, in the comparisons with practical machines accomplished to date, is as accurate in predicting cryocooler performance as the more sophisticated simulation programmes. Copies of the program including a graphics package and an accompanying manual are available at nominal cost. Technical support and consulting assistance to implement the program can be provided

  3. 275 C Downhole Microcomputer System

    Energy Technology Data Exchange (ETDEWEB)

    Chris Hutchens; Hooi Miin Soo

    2008-08-31

    An HC11 controller IC and along with serial SRAM and ROM support ICs chip set were developed to support a data acquisition and control for extreme temperature/harsh environment conditions greater than 275 C. The 68HC11 microprocessor is widely used in well logging tools for control, data acquisition, and signal processing applications and was the logical choice for a downhole controller. This extreme temperature version of the 68HC11 enables new high temperature designs and additionally allows 68HC11-based well logging tools and MWD tools to be upgraded for high temperature operation in deep gas reservoirs, The microcomputer chip consists of the microprocessor ALU, a small boot ROM, 4 kbyte data RAM, counter/timer unit, serial peripheral interface (SPI), asynchronous serial interface (SCI), and the A, B, C, and D parallel ports. The chip is code compatible with the single chip mode commercial 68HC11 except for the absence of the analog to digital converter system. To avoid mask programmed internal ROM, a boot program is used to load the microcomputer program from an external mask SPI ROM. A SPI RAM IC completes the chip set and allows data RAM to be added in 4 kbyte increments. The HC11 controller IC chip set is implemented in the Peregrine Semiconductor 0.5 micron Silicon-on-Sapphire (SOS) process using a custom high temperature cell library developed at Oklahoma State University. Yield data is presented for all, the HC11, SPI-RAM and ROM. The lessons learned in this project were extended to the successful development of two high temperature versions of the LEON3 and a companion 8 Kbyte SRAM, a 200 C version for the Navy and a 275 C version for the gas industry.

  4. 2750 C Downhole Microcomputer System

    International Nuclear Information System (INIS)

    Hutchens, Chris; Soo, Hooi Miin

    2008-01-01

    An HC11 controller IC and along with serial SRAM and ROM support ICs chip set were developed to support a data acquisition and control for extreme temperature/harsh environment conditions greater than 275 C. The 68HC11 microprocessor is widely used in well logging tools for control, data acquisition, and signal processing applications and was the logical choice for a downhole controller. This extreme temperature version of the 68HC11 enables new high temperature designs and additionally allows 68HC11-based well logging tools and MWD tools to be upgraded for high temperature operation in deep gas reservoirs, The microcomputer chip consists of the microprocessor ALU, a small boot ROM, 4 kbyte data RAM, counter/timer unit, serial peripheral interface (SPI), asynchronous serial interface (SCI), and the A, B, C, and D parallel ports. The chip is code compatible with the single chip mode commercial 68HC11 except for the absence of the analog to digital converter system. To avoid mask programmed internal ROM, a boot program is used to load the microcomputer program from an external mask SPI ROM. A SPI RAM IC completes the chip set and allows data RAM to be added in 4 kbyte increments. The HC11 controller IC chip set is implemented in the Peregrine Semiconductor 0.5 micron Silicon-on-Sapphire (SOS) process using a custom high temperature cell library developed at Oklahoma State University. Yield data is presented for all, the HC11, SPI-RAM and ROM. The lessons learned in this project were extended to the successful development of two high temperature versions of the LEON3 and a companion 8 Kbyte SRAM, a 200 C version for the Navy and a 275 C version for the gas industry

  5. The Microcomputer in the Clinical Nursing Research Unit

    Science.gov (United States)

    Schwirian, Patricia M.; Byers, Sandra R.

    1982-01-01

    This paper discusses the microcomputer in clinical nursing research. There are six general areas in which computers have been useful to nurses: nursing notes and charting; patient care plans; automated monitoring of high-tech nursing units; HIS and MIS systems; personnel distribution systems; and education. Three alternative models for the conduct of clinical nursing research in a hospital are described. The first is a centralized model relying on the bureaucratic structure of the hospital. Second is a decentralized network of professional nurses and research support personnel woven together by a Clinical Nurse Researcher, and third is a dedicated clinical nursing research unit. Microcomputers have five characteristics which make them vital tools for nurse researchers: user-friendliness; environment friendliness; low cost; ease of interface with other information systems; and range and quality of software.

  6. Model techniques for testing heated concrete structures

    International Nuclear Information System (INIS)

    Stefanou, G.D.

    1983-01-01

    Experimental techniques are described which may be used in the laboratory to measure strains of model concrete structures representing to scale actual structures of any shape or geometry, operating at elevated temperatures, for which time-dependent creep and shrinkage strains are dominant. These strains could be used to assess the distribution of stress in the scaled structure and hence to predict the actual behaviour of concrete structures used in nuclear power stations. Similar techniques have been employed in an investigation to measure elastic, thermal, creep and shrinkage strains in heated concrete models representing to scale parts of prestressed concrete pressure vessels for nuclear reactors. (author)

  7. Subantibiotic dose of azithromycin attenuates alveolar bone destruction and improves trabecular microarchitectures in a rat model of experimental periodontitis: A study using micro-computed tomography.

    Science.gov (United States)

    Park, Hye-Shin; Lee, Yong Sun; Choi, Eun-Young; Choi, Jeom-Il; Choi, In Soon; Kim, Sung-Jo

    2017-06-01

    Azithromycin, a macrolide antibiotic, has anti-inflammatory and immunomodulatory activities apart from its antibacterial properties. In this study, we examined the efficacy of subantibiotic dose of azithromycin on ligature-induced periodontitis in rats using micro-computed tomography (micro-CT) imaging and bone parameter analysis. Male Sprague-Dawley rats were allocated to the following four groups: non-ligation (NL) group; ligation-only (L) group; ligation-plus-subantibiotic dose azithromycin (SA) group; and 4) ligation-plus-antibiotic dose azithromycin (AA) group. The rats from Groups L, SA and AA were subjected to periodontitis by placing a ligature around lower right first molar. Immediately after ligation, the rats in SA and AA groups received daily intraperitoneal injections of azithromycin at a dosage of 3.5 or 10mg/kg body weight, respectively. The ligatures were maintained for 2weeks at which time the rats had their mandibles hemisected for micro-CT analysis. Subantibiotic dose of azithromycin strongly suppressed reductions in alveolar bone height and bone volume fraction caused by experimental periodontitis. When subantibiotic dosage of azithromycin was administered to rats, ligature-induced alterations in microarchitectural parameters of trabecular bone were significantly reversed. Rats treated with subantibiotic dose of azithromycin presented no significant difference compared to rats with antibiotic dosage in all parameters. While further studies are necessary, subantibiotic dose of azithromycin could be utilized as a host modulator for the treatment of periodontitis. Copyright © 2017 Elsevier B.V. All rights reserved.

  8. The effects of dental pulp stem cells on bone regeneration in rat calvarial defect model: micro-computed tomography and histomorphometric analysis.

    Science.gov (United States)

    Asutay, Fatih; Polat, Serkan; Gül, Mehmet; Subaşı, Cansu; Kahraman, Sevil Altundağ; Karaöz, Erdal

    2015-12-01

    Stem cell therapies may be applicable to all fields of medicine, including craniomaxillofacial surgery. Dental pulp stem cells also have significant osteogenic properties. This study aimed to evaluate the influence of dental pulp stem cells on bone regeneration and to ascertain whether or not there was any superiority over traditional methods. In this study, 15 non-immunodeficient Wistar albino rats were used. The rats were divided into three groups: (1) untreated control group; (2) hydroxyapatite tri-calcium-phosphate (HA/TCP) paste; (3) human dental pulp derived stem cells (DPSC) mixed with HA/TCP paste (HA/TCP+DSPC group, n=10). Two symmetrical full-thickness cranial defects were created on each parietal region (10 defects for each group). The animals were sacrificed 8 weeks post-surgery and samples were analyzed by microcomputer tomography (μ-CT) and histomorphometry. The calcification rate and bone mineral density (BMD) values in Group 3 were found to be significantly higher than in the other two groups. Radiographically, bone regeneration was greater in Group 2 compared with the control group. However, there was no significant difference between Groups 2 and 1 in respect of histological analysis. According to the results of the present study, DPSCs may be a suitable factor for bone tissue engineering because they can be easily obtained and differentiate into bone cells. Copyright © 2015 Elsevier Ltd. All rights reserved.

  9. A microcomputer software system for conformation therapy

    International Nuclear Information System (INIS)

    Akanuma, Atsuo; Aoki, Yukimasa; Nakagawa, Keiichi; Hosoi, Yoshio; Onogi, Yuzou; Muta, Nobuharu; Sakata, Koichi; Karasawa, Katsuyuki; Iio, Masahiro

    1987-01-01

    Effectivity of radiotherapy in the treatment of malignant tumors has gradually and constantly increased since the discovery of ionising radiation, which is greatly contributed by technological and industrial developments. Improved radiotherapy machines allowed higher and higher energy radiations. And the more penetrating radiation delivered the higher dose to a deep seated tumors with marked decreased integral dose, which rapidly increased the indications for malignant tumor therapy. The merits from the penetrating power of radiation appears recently saturated. Instead the developments in the automated processings provided easily computers for radiotherapy. Now applications of computers to radiotherapy potentiated the very frequent employment of conformation technique which is invented in this far east country. For conveniences on the computer application of radiotherapy, a set of microcomputer is chosen here and a software system on this set for conformation technique is being developed here. The system consists from a main program for maintenance and switching job programs. Digitizer input of body and inhomogenity contours is employed. Currently no dose distribution output is intended. Dose calculation at selected points is performed instead. (author)

  10. Workshop on Computational Modelling Techniques in Structural ...

    Indian Academy of Sciences (India)

    Home; Journals; Resonance – Journal of Science Education; Volume 22; Issue 6. Workshop on Computational Modelling Techniques in Structural Biology. Information and Announcements Volume 22 Issue 6 June 2017 pp 619-619. Fulltext. Click here to view fulltext PDF. Permanent link:

  11. Coping Strategies for Resistance to Microcomputers.

    Science.gov (United States)

    Townsend, Barbara; And Others

    1981-01-01

    Strategies which may be used to overcome opposition to microcomputers by faculty, students, and administrators are detailed. It is felt that, ultimately, all three groups will have to work together toward the establishment of the computer in the classroom. (MP)

  12. Evaluation of a polyetheretherketone (PEEK) titanium composite interbody spacer in an ovine lumbar interbody fusion model: biomechanical, microcomputed tomographic, and histologic analyses.

    Science.gov (United States)

    McGilvray, Kirk C; Waldorff, Erik I; Easley, Jeremiah; Seim, Howard B; Zhang, Nianli; Linovitz, Raymond J; Ryaby, James T; Puttlitz, Christian M

    2017-12-01

    The most commonly used materials used for interbody cages are titanium metal and polymer polyetheretherketone (PEEK). Both of these materials have demonstrated good biocompatibility. A major disadvantage associated with solid titanium cages is their radiopacity, limiting the postoperative monitoring of spinal fusion via standard imaging modalities. However, PEEK is radiolucent, allowing for a temporal assessment of the fusion mass by clinicians. On the other hand, PEEK is hydrophobic, which can limit bony ingrowth. Although both PEEK and titanium have demonstrated clinical success in obtaining a solid spinal fusion, innovations are being developed to improve fusion rates and to create stronger constructs using hybrid additive manufacturing approaches by incorporating both materials into a single interbody device. The purpose of this study was to examine the interbody fusion characteristic of a PEEK Titanium Composite (PTC) cage for use in lumbar fusion. Thirty-four mature female sheep underwent two-level (L 2 -L 3 and L 4 -L 5 ) interbody fusion using either a PEEK or a PTC cage (one of each per animal). Animals were sacrificed at 0, 8, 12, and 18 weeks post surgery. Post sacrifice, each surgically treated functional spinal unit underwent non-destructive kinematic testing, microcomputed tomography scanning, and histomorphometric analyses. Relative to the standard PEEK cages, the PTC constructs demonstrated significant reductions in ranges of motion and a significant increase in stiffness. These biomechanical findings were reinforced by the presence of significantly more bone at the fusion site as well as ingrowth into the porous end plates. Overall, the results indicate that PTC interbody devices could potentially lead to a more robust intervertebral fusion relative to a standard PEEK device in a clinical setting. Copyright © 2017 The Author(s). Published by Elsevier Inc. All rights reserved.

  13. Aerosol model selection and uncertainty modelling by adaptive MCMC technique

    Directory of Open Access Journals (Sweden)

    M. Laine

    2008-12-01

    Full Text Available We present a new technique for model selection problem in atmospheric remote sensing. The technique is based on Monte Carlo sampling and it allows model selection, calculation of model posterior probabilities and model averaging in Bayesian way.

    The algorithm developed here is called Adaptive Automatic Reversible Jump Markov chain Monte Carlo method (AARJ. It uses Markov chain Monte Carlo (MCMC technique and its extension called Reversible Jump MCMC. Both of these techniques have been used extensively in statistical parameter estimation problems in wide area of applications since late 1990's. The novel feature in our algorithm is the fact that it is fully automatic and easy to use.

    We show how the AARJ algorithm can be implemented and used for model selection and averaging, and to directly incorporate the model uncertainty. We demonstrate the technique by applying it to the statistical inversion problem of gas profile retrieval of GOMOS instrument on board the ENVISAT satellite. Four simple models are used simultaneously to describe the dependence of the aerosol cross-sections on wavelength. During the AARJ estimation all the models are used and we obtain a probability distribution characterizing how probable each model is. By using model averaging, the uncertainty related to selecting the aerosol model can be taken into account in assessing the uncertainty of the estimates.

  14. Structural design of SBWR reactor building complex using microcomputers

    International Nuclear Information System (INIS)

    Mandagi, K.; Rajagopal, R.S.; Sawhney, P.S.; Gou, P.F.

    1993-01-01

    The design concept of Simplified Boiling Water Reactor (SBWR) plant is based on simplicity and passive features to enhance safety and reliability, improve performance, and increase economic viability. The SBWR utilizes passive systems such as Gravity Driven Core-Cooling System (GDCS) and Passive Containment Cooling System (PCCS). To suit these design features the Reactor Building (RB) complex of the SBWR is configured as an integrated structure consisting of a cylindrical Reinforced Concrete Containment Vessel (RCCV) surrounded by square reinforced concrete safety envelope and outer box structures, all sharing a common reinforced concrete basemat. This paper describes the structural analysis and design aspects of the RB complex. A 3D STARDYNE finite element model has been developed for the structural analysis of the complex using a PC Compaq 486/33L microcomputer. The structural analysis is performed for service and factored load conditions for the applicable loading combinations. The dynamic responses of containment structures due to pool hydrodynamic loads have been calculated by an axisymmetric shell model using COSMOS/M program. The RCCV is designed in accordance with ASME Section 3, Division 2 Code. The rest of the RB which is classified as Seismic Category 1 structure is designed in accordance with the ACI 349 Code. This paper shows that microcomputers can be efficiently used for the analysis and design of large and complex structures such as RCCV and Reactor Building complex. The use of microcomputers can result in significant savings in the computational cost compared with that of mainframe computers

  15. Microcomputing Facilities: Objectives and Policies. The Microcomputing Program at Drexel University.

    Science.gov (United States)

    Drexel Univ., Philadelphia, PA. Microcomputing Program.

    Microcomputing facilities at Drexel University are described. In 1983 the university instituted a policy requiring all students to have personal access to a microcomputer. Students keep their computers in their dorm rooms or off-campus living quarters. To provide access to students while they are on campus, clusters of university-owned machines…

  16. Techniques to develop data for hydrogeochemical models

    Energy Technology Data Exchange (ETDEWEB)

    Thompson, C.M.; Holcombe, L.J.; Gancarz, D.H.; Behl, A.E. (Radian Corp., Austin, TX (USA)); Erickson, J.R.; Star, I.; Waddell, R.K. (Geotrans, Inc., Boulder, CO (USA)); Fruchter, J.S. (Battelle Pacific Northwest Lab., Richland, WA (USA))

    1989-12-01

    The utility industry, through its research and development organization, the Electric Power Research Institute (EPRI), is developing the capability to evaluate potential migration of waste constitutents from utility disposal sites to the environment. These investigations have developed computer programs to predict leaching, transport, attenuation, and fate of inorganic chemicals. To predict solute transport at a site, the computer programs require data concerning the physical and chemical conditions that affect solute transport at the site. This manual provides a comprehensive view of the data requirements for computer programs that predict the fate of dissolved materials in the subsurface environment and describes techniques to measure or estimate these data. In this manual, basic concepts are described first and individual properties and their associated measurement or estimation techniques are described later. The first three sections review hydrologic and geochemical concepts, discuss data requirements for geohydrochemical computer programs, and describe the types of information the programs produce. The remaining sections define and/or describe the properties of interest for geohydrochemical modeling and summarize available technique to measure or estimate values for these properties. A glossary of terms associated with geohydrochemical modeling and an index are provided at the end of this manual. 318 refs., 9 figs., 66 tabs.

  17. Small, microcomputer-based CAMAC controller

    International Nuclear Information System (INIS)

    Juras, R.C.

    1979-01-01

    The beam buncher necessary to condition the beam from the Oak Ridge National Laboratory 25 MV tandem accelerator for post-acceleration by the Oak Ridge Isochronous Cyclotron is CAMAC-based and will be controlled via one of the serial highways of the accelerator control system. However, prior to integration into the accelerator system, the buncher requires testing, including runs on the model EN tandem at Oak Ridge. In order to facilitate testing and initial operation of the buncher, a microcomputer-based controller was assembled. The controller consists of a CAMAC crate, several CAMAC modules, a touch panel display, a controller box, and software. The controller box contains one shaft encoder and two switches. One of the switches is a coarse/fine selector. The other switch is assignable via the touch panel display and is used, for example, to turn devices on and off. Operation of the controller is described. It can be quickly assembled to control any small CAMAC-based system. 2 figures

  18. Final report : evaluation of microcomputer applications in transportation engineering.

    Science.gov (United States)

    1984-01-01

    This study investigated areas where microcomputers can aid in the effectiveness of transportation engineering at state and local levels. A survey of the microcomputer needs of transportation professionals in state and local agencies in Virginia was c...

  19. [General-purpose microcomputer for medical laboratory instruments].

    Science.gov (United States)

    Vil'ner, G A; Dudareva, I E; Kurochkin, V E; Opalev, A A; Polek, A M

    1984-01-01

    Presented in the paper is the microcomputer based on the KP580 microprocessor set. Debugging of the hardware and the software by using the unique debugging stand developed on the basis of microcomputer "Electronica-60" is discussed.

  20. Improved modeling techniques for turbomachinery flow fields

    Energy Technology Data Exchange (ETDEWEB)

    Lakshminarayana, B.; Fagan, J.R. Jr.

    1995-12-31

    This program has the objective of developing an improved methodology for modeling turbomachinery flow fields, including the prediction of losses and efficiency. Specifically, the program addresses the treatment of the mixing stress tensor terms attributed to deterministic flow field mechanisms required in steady-state Computational Fluid Dynamic (CFD) models for turbomachinery flow fields. These mixing stress tensors arise due to spatial and temporal fluctuations (in an absolute frame of reference) caused by rotor-stator interaction due to various blade rows and by blade-to-blade variation of flow properties. This will be accomplished in a cooperative program by Penn State University and the Allison Engine Company. These tasks include the acquisition of previously unavailable experimental data in a high-speed turbomachinery environment, the use of advanced techniques to analyze the data, and the development of a methodology to treat the deterministic component of the mixing stress tenor.

  1. Advances in transgenic animal models and techniques.

    Science.gov (United States)

    Ménoret, Séverine; Tesson, Laurent; Remy, Séverine; Usal, Claire; Ouisse, Laure-Hélène; Brusselle, Lucas; Chenouard, Vanessa; Anegon, Ignacio

    2017-10-01

    On May 11th and 12th 2017 was held in Nantes, France, the international meeting "Advances in transgenic animal models and techniques" ( http://www.trm.univ-nantes.fr/ ). This biennial meeting is the fifth one of its kind to be organized by the Transgenic Rats ImmunoPhenomic (TRIP) Nantes facility ( http://www.tgr.nantes.inserm.fr/ ). The meeting was supported by private companies (SONIDEL, Scionics computer innovation, New England Biolabs, MERCK, genOway, Journal Disease Models and Mechanisms) and by public institutions (International Society for Transgenic Technology, University of Nantes, INSERM UMR 1064, SFR François Bonamy, CNRS, Région Pays de la Loire, Biogenouest, TEFOR infrastructure, ITUN, IHU-CESTI and DHU-Oncogeffe and Labex IGO). Around 100 participants, from France but also from different European countries, Japan and USA, attended the meeting.

  2. Facility/equipment performance evaluation using microcomputer simulation analysis

    Energy Technology Data Exchange (ETDEWEB)

    Chockie, A.D.; Hostick, C.J.

    1985-08-01

    A computer simulation analysis model was developed at the Pacific Northwest Laboratory to assist in assuring the adequacy of the Monitored Retrievable Storage facility design to meet the specified spent nuclear fuel throughput requirements. The microcomputer-based model was applied to the analysis of material flow, equipment capability and facility layout. The simulation analysis evaluated uncertainties concerning both facility throughput requirements and process duration times as part of the development of a comprehensive estimate of facility performance. The evaluations provided feedback into the design review task to identify areas where design modifications should be considered.

  3. Development and Evaluation of the SUMIT Microcomputer Module Entitled 'Predator Functional Response'.

    Science.gov (United States)

    Shaltz, Mark B.

    An experiment was conducted that compared the teaching effectiveness of a computer assisted instructional module and a lecture-discussion. The module, Predator Functional Response (PFR), was developed as part of the SUMIT (Single-concept User-adaptable Microcomputer-based Instructional Technique) project. A class of 30 students was randomly…

  4. Print Station Operation. Microcomputing Working Paper Series.

    Science.gov (United States)

    Wozny, Lucy Anne

    During the academic year 1983-84, Drexel University instituted a new policy requiring all incoming students to have access to a microcomputer. The computer chosen to fulfill this requirement was the Macintosh from Apple Computer, Inc. Although this requirement put an additional financial burden on the Drexel student, the university administration…

  5. Machine Distribution. Microcomputing Working Papers Series.

    Science.gov (United States)

    Drexel Univ., Philadelphia, PA. Microcomputing Program.

    During the academic year 1983-84, Drexel University instituted a new policy requiring all incoming students to have access to a microcomputer. The computer chosen to fulfill this requirement was the Macintosh from Apple Computer, Inc. This paper provides a brief description of the process undertaken to select the appropriate computer (i.e.,…

  6. Microcomputer Applications in Local Assessment Systems.

    Science.gov (United States)

    Harnisch, Delwyn L.; And Others

    The capabilities and hardware requirements of four microcomputer software packages produced by the Office of Educational Testing, Research and Service at the University of Illinois at Urbana-Champaign are described. These programs are: (1) the Scan-Tron Forms Analysis Package Version 2.0, an interface between an IBM-compatible and a Scan-Tron…

  7. Dictionary of microelectronics and microcomputer technology

    International Nuclear Information System (INIS)

    Attiyate, Y.H.; Shah, R.R.

    1984-01-01

    This bilingual dictionary (German-English and English-German) is to give the general public a clearer idea of the terminology of microelectronics, microcomputers, data processing, and computer science. Each part contains about 7500 terms frequently encountered in practice, about 2000 of which are supplemented by precise explanations. (orig./HP) [de

  8. The Microcomputer in the Acquisition Environment,

    Science.gov (United States)

    experience of our period of experimentation with Office Automation can aid other offices considering taking this course of action. We have had both positive...and negative result with our effort, but the overall conclusion is that (1) micro-computer office automation can not be avoided and (2) we have only scratched the surface of its applications in the acquisition environment. (Author)

  9. Enlist micros: Training science teachers to use microcomputers

    Science.gov (United States)

    Baird, William E.; Ellis, James D.; Kuerbis, Paul J.

    A National Science Foundation grant to the Biological Sciences Curriculum Study (BSCS) at The Colorado College supported the design and production of training materials to encourage literacy of science teachers in the use of microcomputers. ENLIST Micros is based on results of a national needs assessment that identified 22 compentencies needed by K-12 science teachers to use microcomputers for instruction. A writing team developed the 16-hour training program in the summer of 1985, and field-test coordinators tested it with 18 preservice or in-service groups during the 1985-86 academic year at 15 sites within the United States. The training materials consist of video programs, interactive computer disks for the Apple II series microcomputer, a training manual for participants, and a guide for the group leader. The experimental materials address major areas of educational computing: awareness, applications, implementation, evaluation, and resources. Each chapter contains activities developed for this program, such as viewing video segments of science teachers who are using computers effectively and running commercial science and training courseware. Role playing and small-group interaction help the teachers overcome their reluctance to use computers and plan for effective implementation of microcomputers in the school. This study examines the implementation of educational computing among 47 science teachers who completed the ENLIST Micros training at a southern university. We present results of formative evaluation for that site. Results indicate that both elementary and secondary teachers benefit from the training program and demonstrate gains in attitudes toward computer use. Participating teachers said that the program met its stated objectives and helped them obtain needed skills. Only 33 percent of these teachers, however, reported using computers one year after the training. In June 1986, the BSCS initiated a follow up to the ENLIST Micros curriculum to

  10. Three-dimensional analysis of the pulp cavity on surface models of molar teeth, using X-ray micro-computed tomography

    DEFF Research Database (Denmark)

    Markvart, Merete; Bjørndal, Lars; Darvann, Tron Andre

    2012-01-01

    im. The purpose of this study was to investigate the scanning and segmentation precision of surface models of molars for the detection of small volumes, such as the reduced pulp cavity; formation of mineral deposits; detection of narrow root canals and to improve the clinical and morphological un...

  11. Microcomputer for controlled substance record keeping.

    Science.gov (United States)

    Stein, R L; Motta, L J; Yee, A D

    1984-01-01

    The use of a microcomputer to maintain controlled substance inventory and record keeping is described. The system maintains perpetual inventories of the central narcotic vault and proof-of-use controlled drug records outstanding at nursing stations. The computerized system has eliminated (1) the delay previously encountered in the posting of transactions from the numeric log to perpetual inventory logs and (2) the potential addition and subtraction errors inherent in a manual system. Computerizing the controlled drug record-keeping system has saved approximately 166 minutes of labor per day, a cost savings of approximately $26. The new system also helps prevent diversion of controlled substances. The computer may also be used for other tasks while not running the controlled substance program. A microcomputer is well suited to the task of controlled-substance record-keeping functions, and the cost of the system (less than $4000) can be quickly recouped in labor savings.

  12. Versatile microcomputer-based temperature controller

    International Nuclear Information System (INIS)

    Yarberry, V.R.

    1980-09-01

    The wide range of thermal responses required in laboratory and scientific equipment requires a temperature controller with a great deal of flexibility. While a number of analog temperature controllers are commercially available, they have certain limitations, such as inflexible parameter control or insufficient precision. Most lack digital interface capabilities--a necessity when the temperature controller is part of a computer-controlled automatic data acquisition system. We have developed an extremely versatile microcomputer-based temperature controller to fulfill this need in a variety of equipment. The control algorithm used allows optimal tailoring of parameters to control overshoot, response time, and accuracy. This microcomputer-based temperature controller can be used as a standalone instrument (with a teletype used to enter para-meters), or it can be integrated into a data acquisition system

  13. Teaching Molecular Biology with Microcomputers.

    Science.gov (United States)

    Reiss, Rebecca; Jameson, David

    1984-01-01

    Describes a series of computer programs that use simulation and gaming techniques to present the basic principles of the central dogma of molecular genetics, mutation, and the genetic code. A history of discoveries in molecular biology is presented and the evolution of these computer assisted instructional programs is described. (MBR)

  14. An experimental comparison of modelling techniques for speaker ...

    Indian Academy of Sciences (India)

    Most of the existing modelling techniques for the speaker recognition task make an implicit assumption of sufficient data for speaker modelling and hence may lead to poor modelling under limited data condition. The present work gives an experimental evaluation of the modelling techniques like Crisp Vector Quantization ...

  15. An experimental comparison of modelling techniques for speaker ...

    Indian Academy of Sciences (India)

    Abstract. Most of the existing modelling techniques for the speaker recog- nition task make an implicit assumption of sufficient data for speaker modelling and hence may lead to poor modelling under limited data condition. The present work gives an experimental evaluation of the modelling techniques like Crisp.

  16. Computer Aided School Timetabling: Part 2--The Micro-Computer for School Timetabling.

    Science.gov (United States)

    Dowsland, W. B.; Lim, S.

    1983-01-01

    Use of a 56k Z80-based microcomputer for timetabling is described. Topics addressed include general program structure, data input, class assignment, timetable display, improvement program, possible enhancements to computer models, quality of timetable produced, and implications of data/program structure. Such a timetable is suitable for a…

  17. A Practical Application of Microcomputers to Control an Active Solar System.

    Science.gov (United States)

    Goldman, David S.; Warren, William

    1984-01-01

    Describes the design and implementation of a microcomputer-based model active solar heating system. Includes discussions of: (1) the active solar components (solar collector, heat exchanger, pump, and fan necessary to provide forced air heating); (2) software components; and (3) hardware components (in the form of sensors and actuators). (JN)

  18. Challenge in Numerical Software for Microcomputers

    Energy Technology Data Exchange (ETDEWEB)

    Cody, W J

    1977-09-02

    Microcomputers are now capable of serious numerical computation using programmed floating-point arithmetic and Basic compilers. Unless numerical software designers for these machines exploit experience gained in providing software for larger machines, history will repeat with the initial spread of treacherous software. This paper discusses good software, especially for the elementary functions, in terms of reliability and robustness. The emphasis. is on insight rather than detailed algorithms, to show why certain things are important and how they may be achieved.

  19. Multiprogrammation fast branch driver for microcomputer MICRAL

    International Nuclear Information System (INIS)

    Kaiser, Josef; Lacroix, Jean.

    1975-01-01

    This branch driver allows in association with the FIFO memories of the microcomputer Micral, very fast exchanges with the 7 crates of a CAMAC branch. A CAMAC programm (command, test, read, write) is loaded in the 1K FIFO buffer of the Micral before execution time and executed in sequence at a rate of 1,5μs per CAMAC command. After programm execution, data may be transferred directly on a magnetic tape [fr

  20. Accessing remote data bases using microcomputers

    OpenAIRE

    Saul, Peter D.

    1985-01-01

    General practitioners' access to remote data bases using microcomputers is increasing, making even the most obscure information readily available. Some of the systems available to general practitioners in the UK are described and the methods of access are outlined. General practitioners should be aware of the advances in technology; data bases are increasing in size, the cost of access is falling and their use is becoming easier.

  1. Microcomputer-based monitoring and control system

    International Nuclear Information System (INIS)

    Talaska, D.

    1979-03-01

    This report describes a microcomputer-based monitoring and control system devised within, and used by, the Cryogenic Operations group at SLAC. Presently, a version of it is operating at the one meter liquid hydrogen bubble chamber augmenting the conventional pneumatic and human feedback system. Its use has greatly improved the controlled tolerances of temperature and pulse shape, and it has nearly eliminated the need for operating personnel to adjust the conventional pneumatic control system. The latter is most important since the rapid cycling machine can demand attentions beyond the operator's skill. Similar microcomputer systems are being prepared to monitor and control cryogenic devices situated in regions of radiation which preclude human entry and at diverse locations which defy the dexterity of the few operators assigned to maintain them. An IMSAI 8080 microcomputer is basic to the system. The key to the use of the IMSAI 8080 in this system was in the development of unique interface circuitry, and the report is mostly concerned with this

  2. Respirometry techniques and activated sludge models

    NARCIS (Netherlands)

    Benes, O.; Spanjers, H.; Holba, M.

    2002-01-01

    This paper aims to explain results of respirometry experiments using Activated Sludge Model No. 1. In cases of insufficient fit of ASM No. 1, further modifications to the model were carried out and the so-called "Enzymatic model" was developed. The best-fit method was used to determine the effect of

  3. Formal modelling techniques in human-computer interaction

    NARCIS (Netherlands)

    de Haan, G.; de Haan, G.; van der Veer, Gerrit C.; van Vliet, J.C.

    1991-01-01

    This paper is a theoretical contribution, elaborating the concept of models as used in Cognitive Ergonomics. A number of formal modelling techniques in human-computer interaction will be reviewed and discussed. The analysis focusses on different related concepts of formal modelling techniques in

  4. UAV State Estimation Modeling Techniques in AHRS

    Science.gov (United States)

    Razali, Shikin; Zhahir, Amzari

    2017-11-01

    Autonomous unmanned aerial vehicle (UAV) system is depending on state estimation feedback to control flight operation. Estimation on the correct state improves navigation accuracy and achieves flight mission safely. One of the sensors configuration used in UAV state is Attitude Heading and Reference System (AHRS) with application of Extended Kalman Filter (EKF) or feedback controller. The results of these two different techniques in estimating UAV states in AHRS configuration are displayed through position and attitude graphs.

  5. Ambient temperature modelling with soft computing techniques

    Energy Technology Data Exchange (ETDEWEB)

    Bertini, Ilaria; Ceravolo, Francesco; Citterio, Marco; Di Pietra, Biagio; Margiotta, Francesca; Pizzuti, Stefano; Puglisi, Giovanni [Energy, New Technology and Environment Agency (ENEA), Via Anguillarese 301, 00123 Rome (Italy); De Felice, Matteo [Energy, New Technology and Environment Agency (ENEA), Via Anguillarese 301, 00123 Rome (Italy); University of Rome ' ' Roma 3' ' , Dipartimento di Informatica e Automazione (DIA), Via della Vasca Navale 79, 00146 Rome (Italy)

    2010-07-15

    This paper proposes a hybrid approach based on soft computing techniques in order to estimate monthly and daily ambient temperature. Indeed, we combine the back-propagation (BP) algorithm and the simple Genetic Algorithm (GA) in order to effectively train artificial neural networks (ANN) in such a way that the BP algorithm initialises a few individuals of the GA's population. Experiments concerned monthly temperature estimation of unknown places and daily temperature estimation for thermal load computation. Results have shown remarkable improvements in accuracy compared to traditional methods. (author)

  6. Moving objects management models, techniques and applications

    CERN Document Server

    Meng, Xiaofeng; Xu, Jiajie

    2014-01-01

    This book describes the topics of moving objects modeling and location tracking, indexing and querying, clustering, location uncertainty, traffic aware navigation and privacy issues as well as the application to intelligent transportation systems.

  7. Materials and techniques for model construction

    Science.gov (United States)

    Wigley, D. A.

    1985-01-01

    The problems confronting the designer of models for cryogenic wind tunnel models are discussed with particular reference to the difficulties in obtaining appropriate data on the mechanical and physical properties of candidate materials and their fabrication technologies. The relationship between strength and toughness of alloys is discussed in the context of maximizing both and avoiding the problem of dimensional and microstructural instability. All major classes of materials used in model construction are considered in some detail and in the Appendix selected numerical data is given for the most relevant materials. The stepped-specimen program to investigate stress-induced dimensional changes in alloys is discussed in detail together with interpretation of the initial results. The methods used to bond model components are considered with particular reference to the selection of filler alloys and temperature cycles to avoid microstructural degradation and loss of mechanical properties.

  8. [Evaluation of the efficiency of different instrumentation for root canal retreatment using micro-computed tomography].

    Science.gov (United States)

    Guo, Jing; Yin, Kaifeng; Yang, Haibing; Han, Xuan; Wang, Yan

    2012-06-01

    To establish the three-dimensional model of human permanent premolars based on Micro-computed tomography (Micro-CT) data, and evaluate the efficiency quantitatively of two different instrumentations for root canal retreatment. Forty extracted permanent premolars for the reason of orthodontic treatment were collected, prepared by using ProTaper Ni-Ti files and filled by cold gutta pertscha lateral condensation technique. The subjects were scanned by Micro-CT. Forty teeth were randomly divided into two groups and retreated by K-Flexo files and ProTaper Universal retreatment system respectively. Then all subjects were scanned again, and the mean percentage of remaining filling materials and the scores of remaining filling materials presented in upper, middle and apical 1/3 of the canal were calculated. Mean percentage of the remaining filling materials by K-Flexo files was lower than that by ProTaper Universal retreatment system (P=0.005). The scores demonstrated that K-Flexo files had greater efficiency than ProTaper Universal retreatment system when retreating the apical 1/3 of canal (P<0.05). This study shows that both techniques could not remove filling materials completely.

  9. Microcomputer-based digital image processing - A tutorial package for exploration geologists

    Science.gov (United States)

    Harrington, J. A., Jr.; Cartin, K. F.

    1985-01-01

    An Apple II microcomputer-based software package for analysis of digital data developed at the University of Oklahoma, the Digital Image Analysis System (DIAS), provides a relatively low-cost, portable alternative to large, dedicated minicomputers for digital image processing education. Digital processing techniques for analysis of Landsat MSS data and a series of tutorial exercises for exploration geologists are described and evaluated. DIAS allows in-house training that does not interfere with computer-based prospect analysis objectives.

  10. Advanced techniques for modeling avian nest survival

    Science.gov (United States)

    Dinsmore, S.J.; White, Gary C.; Knopf, F.L.

    2002-01-01

    Estimation of avian nest survival has traditionally involved simple measures of apparent nest survival or Mayfield constant-nest-survival models. However, these methods do not allow researchers to build models that rigorously assess the importance of a wide range of biological factors that affect nest survival. Models that incorporate greater detail, such as temporal variation in nest survival and covariates representative of individual nests represent a substantial improvement over traditional estimation methods. In an attempt to improve nest survival estimation procedures, we introduce the nest survival model now available in the program MARK and demonstrate its use on a nesting study of Mountain Plovers (Charadrius montanus Townsend) in Montana, USA. We modeled the daily survival of Mountain Plover nests as a function of the sex of the incubating adult, nest age, year, linear and quadratic time trends, and two weather covariates (maximum daily temperature and daily precipitation) during a six-year study (1995–2000). We found no evidence for yearly differences or an effect of maximum daily temperature on the daily nest survival of Mountain Plovers. Survival rates of nests tended by female and male plovers differed (female rate = 0.33; male rate = 0.49). The estimate of the additive effect for males on nest survival rate was 0.37 (95% confidence limits were 0.03, 0.71) on a logit scale. Daily survival rates of nests increased with nest age; the estimate of daily nest-age change in survival in the best model was 0.06 (95% confidence limits were 0.04, 0.09) on a logit scale. Daily precipitation decreased the probability that the nest would survive to the next day; the estimate of the additive effect of daily precipitation on the nest survival rate was −1.08 (95% confidence limits were −2.12, −0.13) on a logit scale. Our approach to modeling daily nest-survival rates allowed several biological factors of interest to be easily included in nest survival models

  11. Single-chip microcomputer application in nuclear radiation monitoring instruments

    International Nuclear Information System (INIS)

    Zhang Songshou

    1994-01-01

    The single-chip microcomputer has advantage in many respects i.e. multiple function, small size, low-power consumption,reliability etc. It is widely used now in industry, instrumentation, communication and machinery. The author introduced usage of single-chip microcomputer in nuclear radiation monitoring instruments for control, linear compensation, calculation, changeable parameter presetting and military training

  12. Enhancing a Mainframe Library System through Microcomputer Technology.

    Science.gov (United States)

    Breeding, Marshall

    1988-01-01

    Discusses ways a microcomputer system might enhance the mainframe library system at Vanderbilt University. Topics include: (1) types of microcomputers; (2) types of terminals; (3) characteristics of a dumb terminal; (4) characteristics of an intelligent terminal; (5) which terminals should be given additional features; and (6) designing an…

  13. Managing Microcomputer Applications: A Primer and Guide to Good Practice,

    Science.gov (United States)

    1988-03-01

    of individuals who with microcomputers, use (or do not wish to use) microcomputers: with different styles and * computer- phobes : individuals who fear...use of com- behaviors: puters. They may fear change, technology, be in- * computer- phobes timidated by supervisors, have unrealistic expectations

  14. Microcomputer based test system for charge coupled devices

    International Nuclear Information System (INIS)

    Sidman, S.

    1981-02-01

    A microcomputer based system for testing analog charge coupled integrated circuits has been developed. It measures device performance for three parameters: dynamic range, baseline shift due to leakage current, and transfer efficiency. A companion board tester has also been developed. The software consists of a collection of BASIC and assembly language routines developed on the test system microcomputer

  15. Model measurements for new accelerating techniques

    International Nuclear Information System (INIS)

    Aronson, S.; Haseroth, H.; Knott, J.; Willis, W.

    1988-06-01

    We summarize the work carried out for the past two years, concerning some different ways for achieving high-field gradients, particularly in view of future linear lepton colliders. These studies and measurements on low power models concern the switched power principle and multifrequency excitation of resonant cavities. 15 refs., 12 figs

  16. A microcomputer network for the control of digitising machines

    International Nuclear Information System (INIS)

    Seller, P.

    1981-01-01

    A distributed microcomputing network operates in the Bubble Chamber Research Group Scanning Laboratory at the Rutherford and Appleton Laboratories. A microcomputer at each digitising table buffers information, controls the functioning of the table and enhances the machine/operator interface. The system consists of fourteen microcomputers together with a VAX 11/780 computer used for data analysis. These are inter-connected via a packet switched network. This paper will describe the features of the combined system, including the distributed computing architecture and the packet switched method of communication. This paper will also describe in detail a high speed packet switching controller used as a central node of the network. This controller is a multiprocessor microcomputer system with eighteen central processor units, thirty-four direct memory access channels and thirty-four prioritorised and vectored interrupt channels. This microcomputer is of general interest as a communications controller due to its totally programmable nature. (orig.)

  17. Microcomputer-based video motion detection system

    International Nuclear Information System (INIS)

    Howington, L.C.

    1979-01-01

    This system was developed to enhance the volumetric intrusion detection capability of the Oak Ridge Y-12 Plant's security program. Not only does the system exhibit an extended range of detection over present infrared, microwave, and ultrasonic devices, it also provides an instantaneous assessment capability by providing the operator with a closed-circuit television (CCTV) image of the alarm scene as soon as motion is detected. The system consists of a custom-built, microcomputer-based, video processor which analyzes the signals received from a network of video cameras. The operator can view the camera images as they are displayed on a CCTV monitor while alarm scenes are displayed on a second monitor. Motion is detected by digitizing and comparing successive video frames and making an alarm decision based on the degree of mismatch. The software-based nature of the microcomputer lends a great deal of flexibility and adaptability in making the alarm decision. Alarm decision variables which are easily adjusted through software are the percent change in gray level required to label a pixel (picture element) as suspect, the number of suspect pixels required to generate an alarm, the pixel pattern to be sampled from the image, and the rate at which a new reference frame is taken. The system is currently being evaluated in a warehouse for potential application in several areas of the Plant. This paper discusses the hardware and software design of the system as well as problems encountered in its implementation and results obtained

  18. Implementation of linguistic models by holographic technique

    Science.gov (United States)

    Pavlov, Alexander V.; Shevchenko, Yanina Y.

    2004-01-01

    In this paper we consider linguistic model as an algebraic model and restrict our consideration to the semantics only. The concept allows "natural-like" language to be used by human-teacher to describe for machine the way of the problem solving, which is based on human"s knowledge and experience. Such imprecision words as "big", "very big", "not very big", etc can be used for human"s knowledge representation. Technically, the problem is to match metric scale, used by the technical device, with the linguistic scale, intuitively formed by the person. We develop an algebraic description of 4-f Fourier-holography setup by using triangular norms based approach. In the model we use the Fourier-duality of the t-norms and t-conorms, which is implemented by 4-f Fourier-holography setup. We demonstrate the setup is described adequately by De-Morgan"s law for involution. Fourier-duality of the t-norms and t-conorms leads to fuzzy-valued logic. We consider General Modus Ponens rule implementation to define the semantical operators, which are adequate to the setup. We consider scales, formed in both +1 and -1 orders of diffraction. We use representation of linguistic labels by fuzzy numbers to form the scale and discuss the dependence of the scale grading on the holographic recording medium operator. To implement reasoning with multi-parametric input variable we use Lorentz function to approximate linguistic labels. We use an example of medical diagnostics for experimental illustration of reasoning on the linguistic scale.

  19. Metamaterials modelling, fabrication, and characterisation techniques

    DEFF Research Database (Denmark)

    Malureanu, Radu; Zalkovskij, Maksim; Andryieuski, Andrei

    2012-01-01

    Metamaterials are artificially designed media that show averaged properties not yet encountered in nature. Among such properties, the possibility of obtaining optical magnetism and negative refraction are the ones mainly exploited but epsilon-near-zero and sub-unitary refraction index are also...... parameters that can be obtained. Such behaviour enables unprecedented applications. Within this work, we will present various aspects of metamaterials research field that we deal with at our department. From the modelling part, we will present tour approach for determining the field enhancement in slits...

  20. Metamaterials modelling, fabrication and characterisation techniques

    DEFF Research Database (Denmark)

    Malureanu, Radu; Zalkovskij, Maksim; Andryieuski, Andrei

    Metamaterials are artificially designed media that show averaged properties not yet encountered in nature. Among such properties, the possibility of obtaining optical magnetism and negative refraction are the ones mainly exploited but epsilon-near-zero and sub-unitary refraction index are also...... parameters that can be obtained. Such behaviour enables unprecedented applications. Within this work, we will present various aspects of metamaterials research field that we deal with at our department. From the modelling part, various approaches for determining the value of the refractive index...

  1. Model order reduction techniques with applications in finite element analysis

    CERN Document Server

    Qu, Zu-Qing

    2004-01-01

    Despite the continued rapid advance in computing speed and memory the increase in the complexity of models used by engineers persists in outpacing them. Even where there is access to the latest hardware, simulations are often extremely computationally intensive and time-consuming when full-blown models are under consideration. The need to reduce the computational cost involved when dealing with high-order/many-degree-of-freedom models can be offset by adroit computation. In this light, model-reduction methods have become a major goal of simulation and modeling research. Model reduction can also ameliorate problems in the correlation of widely used finite-element analyses and test analysis models produced by excessive system complexity. Model Order Reduction Techniques explains and compares such methods focusing mainly on recent work in dynamic condensation techniques: - Compares the effectiveness of static, exact, dynamic, SEREP and iterative-dynamic condensation techniques in producing valid reduced-order mo...

  2. Optimization using surrogate models - by the space mapping technique

    DEFF Research Database (Denmark)

    Søndergaard, Jacob

    2003-01-01

    Surrogate modelling and optimization techniques are intended for engineering design in the case where an expensive physical model is involved. This thesis provides a literature overview of the field of surrogate modelling and optimization. The space mapping technique is one such method for constr......Surrogate modelling and optimization techniques are intended for engineering design in the case where an expensive physical model is involved. This thesis provides a literature overview of the field of surrogate modelling and optimization. The space mapping technique is one such method...... conditions are satisfied. So hybrid methods, combining the space mapping technique with classical optimization methods, should be used if convergence to high accuracy is wanted. Approximation abilities of the space mapping surrogate are compared with those of a Taylor model of the expensive model. The space...... mapping surrogate has a lower approximation error for long steps. For short steps, however, the Taylor model of the expensive model is best, due to exact interpolation at the model origin. Five algorithms for space mapping optimization are presented and the numerical performance is evaluated. Three...

  3. A microcomputer-based gamma attenuation monitor for flowing liquids

    International Nuclear Information System (INIS)

    Everett, S.; Malcolme-Lawes, D.

    1984-01-01

    Many industrial processes involve large-scale flows of liquid mixtures or solutions in situations where it is desirable for the composition of the flow to be monitored. The attenuation (Isub(x)/I 0 ) of a beam of gamma photons can be used to provide a moderately sensitive means of monitoring mixture composition. Several applications of this technique to solid systems have been reported, and we have recently shown how the technique may be applied to monitoring the composition of a flowing liquid binary mixture. This paper describes a microcomputer-based apparatus which may be used to record the attenuation of a collimated beam of gamma photons by a liquid mixture flowing through a pipe. The apparatus and some example results obtained with trial mixtures are discussed. However, the sensitivity of the technique is determined both by the difference in u (the linear absorption coefficient of the matter concerned) for the materials which make up the mixture, and by the precision with which Isub(x)/I 0 may be monitored. For this reason we first consider the precision obtainable in the measurement of the intensity of a gamma photon beam. (author)

  4. An experimental comparison of modelling techniques for speaker ...

    Indian Academy of Sciences (India)

    Feature extraction involves extracting speaker-specific features from the speech signal at reduced data rate. The extracted features are further combined using modelling techniques to generate speaker models. The speaker models are then tested using the features extracted from the test speech signal. The improvement in ...

  5. Microcomputer-controlled ultrasonic data acquisition system

    International Nuclear Information System (INIS)

    Simpson, W.A. Jr.

    1978-11-01

    The large volume of ultrasonic data generated by computer-aided test procedures has necessitated the development of a mobile, high-speed data acquisition and storage system. This approach offers the decided advantage of on-site data collection and remote data processing. It also utilizes standard, commercially available ultrasonic instrumentation. This system is controlled by an Intel 8080A microprocessor. The MCS80-SDK microcomputer board was chosen, and magnetic tape is used as the storage medium. A detailed description is provided of both the hardware and software developed to interface the magnetic tape storage subsystem to Biomation 8100 and Biomation 805 waveform recorders. A boxcar integrator acquisition system is also described for use when signal averaging becomes necessary. Both assembly language and machine language listings are provided for the software

  6. VME bus based microcomputer system boards

    International Nuclear Information System (INIS)

    Chandra, A.K.; Ganesh, G.; Mayya, Anuradha; Chachondia, A.S.; Premraj, M.K.

    1991-01-01

    Several operator information systems for nuclear plants has been developed in the Division and these have involved extensive use of microcomputer boards for achieving various functions. Standard VME bus based boards have been developed to provide the most used functions. These boards have been fabricated and tested and used in several systems including Channel Temperature Monitoring systems, Disturbance Recording Systems etc. and are also proposed to be used in additional systems under developement. The use of standard bus and boards provides considerable savings in engineering time, prototyping, testing and evaluation costs, and maintenance support. This report desribes the various boards developed and the functions available on each. (author). 4 refs., 11 figs., 3 appendixes

  7. Structures of industrial robot controls with microcomputers

    Energy Technology Data Exchange (ETDEWEB)

    Haferkorn, P.; Schwarz, W.

    1982-01-01

    The authors explain the required tasks and basic structure of an industrial robot control system (an integrated component of the robot itself) and discusses three types of programming of process-flexible robots: direct teach-in programming; indirect teach-in programming; and textual programming where the program is first formulated, independent of the robot, in a suitable program language and then read-in to the working memory for use in the man-machine dialogue. The realization of controls for process-flexible industrial robots utilizing microcomputers is described in some detail, covering configurations for multi-computer coupling on the master-slave principle using the interconnecting electronics for the K1521 main frame unit. 8 references.

  8. A collection and information analysis of the experiment with microcomputer

    International Nuclear Information System (INIS)

    Mohd Ariffin bin Aton; Ler Leong Tat

    1985-01-01

    A microcomputer-based system for the continuous collection and analysis of data from a fermentor is described. The system was designed around commercially available hardware and interface and software packages written for microcomputers. Additional programmes were written in BASIC to allow the results to be printed in a specific format. The data read from the fermentor were automatically stored on a floppy disc and analysis on the data can be performed at our convenience. Such method for data collection is not limited to a bioreactor, however, since instruments that require continuous accurate reading, such as GLC, HPLC, etc., could be coupled to a microcomputer system. (author)

  9. Virtual 3d City Modeling: Techniques and Applications

    Science.gov (United States)

    Singh, S. P.; Jain, K.; Mandla, V. R.

    2013-08-01

    3D city model is a digital representation of the Earth's surface and it's related objects such as Building, Tree, Vegetation, and some manmade feature belonging to urban area. There are various terms used for 3D city models such as "Cybertown", "Cybercity", "Virtual City", or "Digital City". 3D city models are basically a computerized or digital model of a city contains the graphic representation of buildings and other objects in 2.5 or 3D. Generally three main Geomatics approach are using for Virtual 3-D City models generation, in first approach, researcher are using Conventional techniques such as Vector Map data, DEM, Aerial images, second approach are based on High resolution satellite images with LASER scanning, In third method, many researcher are using Terrestrial images by using Close Range Photogrammetry with DSM & Texture mapping. We start this paper from the introduction of various Geomatics techniques for 3D City modeling. These techniques divided in to two main categories: one is based on Automation (Automatic, Semi-automatic and Manual methods), and another is Based on Data input techniques (one is Photogrammetry, another is Laser Techniques). After details study of this, finally in short, we are trying to give the conclusions of this study. In the last, we are trying to give the conclusions of this research paper and also giving a short view for justification and analysis, and present trend for 3D City modeling. This paper gives an overview about the Techniques related with "Generation of Virtual 3-D City models using Geomatics Techniques" and the Applications of Virtual 3D City models. Photogrammetry, (Close range, Aerial, Satellite), Lasergrammetry, GPS, or combination of these modern Geomatics techniques play a major role to create a virtual 3-D City model. Each and every techniques and method has some advantages and some drawbacks. Point cloud model is a modern trend for virtual 3-D city model. Photo-realistic, Scalable, Geo-referenced virtual 3

  10. Single-chip microcomputer based protection, diagnostic and recording system for longwall shearers

    Energy Technology Data Exchange (ETDEWEB)

    Heyduk, A.; Krasucki, F. (Politechnika Slaska, Gliwice (Poland). Katedra Elektryfikacji i Automatyzacji Gornictwa)

    1993-05-01

    Presents a concept of microcomputer-aided operation, protection, diagnostics and recording for shearer loaders. A two-stage mathematical model is suggested and explained. The model represents the thermal processes that determine the overcurrent protection of drive motors. Circuits for monitoring fuses, supply voltages, contacts, relays, contactors and electro-hydraulic distributors with the use of transoptors are shown. Recording characteristic operation parameters of a shearer loader during the 5 minutes before a failure is proposed. Protection, diagnosis and control functions are suggested as additional functions to the microcomputer-aided system of shearer loader control being developed at the Silesian Technical University. The system is based on the NECmicroPD 78310 microprocessor. 10 refs.

  11. Circuit oriented electromagnetic modeling using the PEEC techniques

    CERN Document Server

    Ruehli, Albert; Jiang, Lijun

    2017-01-01

    This book provides intuitive solutions to electromagnetic problems by using the Partial Eelement Eequivalent Ccircuit (PEEC) method. This book begins with an introduction to circuit analysis techniques, laws, and frequency and time domain analyses. The authors also treat Maxwell's equations, capacitance computations, and inductance computations through the lens of the PEEC method. Next, readers learn to build PEEC models in various forms: equivalent circuit models, non orthogonal PEEC models, skin-effect models, PEEC models for dielectrics, incident and radiate field models, and scattering PEEC models. The book concludes by considering issues like such as stability and passivity, and includes five appendices some with formulas for partial elements.

  12. [Intestinal lengthening techniques: an experimental model in dogs].

    Science.gov (United States)

    Garibay González, Francisco; Díaz Martínez, Daniel Alberto; Valencia Flores, Alejandro; González Hernández, Miguel Angel

    2005-01-01

    To compare two intestinal lengthening procedures in an experimental dog model. Intestinal lengthening is one of the methods for gastrointestinal reconstruction used for treatment of short bowel syndrome. The modification to the Bianchi's technique is an alternative. The modified technique decreases the number of anastomoses to a single one, thus reducing the risk of leaks and strictures. To our knowledge there is not any clinical or experimental report that studied both techniques, so we realized the present report. Twelve creole dogs were operated with the Bianchi technique for intestinal lengthening (group A) and other 12 creole dogs from the same race and weight were operated by the modified technique (Group B). Both groups were compared in relation to operating time, difficulties in technique, cost, intestinal lengthening and anastomoses diameter. There were no statistical difference in the anastomoses diameter (A = 9.0 mm vs. B = 8.5 mm, p = 0.3846). Operating time (142 min vs. 63 min) cost and technique difficulties were lower in group B (p anastomoses (of Group B) and intestinal segments had good blood supply and were patent along their full length. Bianchi technique and the modified technique offer two good reliable alternatives for the treatment of short bowel syndrome. The modified technique improved operating time, cost and technical issues.

  13. On a Graphical Technique for Evaluating Some Rational Expectations Models

    DEFF Research Database (Denmark)

    Johansen, Søren; Swensen, Anders R.

    2011-01-01

    . In addition to getting a visual impression of the fit of the model, the purpose is to see if the two spreads are nevertheless similar as measured by correlation, variance ratio, and noise ratio. We extend these techniques to a number of rational expectation models and give a general definition of spread...

  14. Application of the numerical modelling techniques to the simulation ...

    African Journals Online (AJOL)

    The aquifer was modelled by the application of Finite Element Method (F.E.M), with appropriate initial and boundary conditions. The matrix solver technique adopted for the F.E.M. was that of the Conjugate Gradient Method. After the steady state calibration and transient verification, the model was used to predict the effect of ...

  15. Fuzzy Control Technique Applied to Modified Mathematical Model ...

    African Journals Online (AJOL)

    In this paper, fuzzy control technique is applied to the modified mathematical model for malaria control presented by the authors in an earlier study. Five Mamdani fuzzy controllers are constructed to control the input (some epidemiological parameters) to the malaria model simulated by 9 fully nonlinear ordinary differential ...

  16. A Microcomputer-based Contribution to Scientific and Technological Literacy.

    Science.gov (United States)

    Trumper, Ricardo; Gelbman, Moshe

    2001-01-01

    Attempts to validate the use of microcomputer-based laboratories for the teaching of physics on both theoretical and empirical grounds. Includes a laboratory that addresses the voltage-current characteristics of several components. (Author/MM)

  17. Microcomputer control of 3 MV Van de Graaff accelerator

    Science.gov (United States)

    Brendle, M.

    1985-05-01

    The high-voltage generator, the ion source, and the analyzing and beam-switching magnets of a 3 MV Van de Graaff accelerator are controlled by inexpensive microcomputers constructed from universal modules.

  18. Tractor performance monitor based on a single-chip microcomputer

    Energy Technology Data Exchange (ETDEWEB)

    Bedri, A.R.; Marley, S.J.; Buchelle, W.F.; Smay, T.A.

    1981-01-01

    A tractor performance monitor based on a single-chip microcomputer was developed to measure ground speed, slip, fuel consumption (rate and total), total area, theoretical time, and total time. Transducers used are presented in detail. 5 refs.

  19. Financial Planning in Transit : Use of Commercially Available Microcomputer Software

    Science.gov (United States)

    1983-11-01

    This report addresses the potential of using commercially available microcomputer software for transit financial planning activities. Discussions with transit operators identified the need for inexpensive, easy to use software for ridership and fare ...

  20. Investigating Electromagnetic Induction through a Microcomputer-Based Laboratory.

    Science.gov (United States)

    Trumper, Ricardo; Gelbman, Moshe

    2000-01-01

    Describes a microcomputer-based laboratory experiment designed for high school students that very accurately analyzes Faraday's law of electromagnetic induction, addressing each variable separately while the others are kept constant. (Author/CCM)

  1. Summary on several key techniques in 3D geological modeling.

    Science.gov (United States)

    Mei, Gang

    2014-01-01

    Several key techniques in 3D geological modeling including planar mesh generation, spatial interpolation, and surface intersection are summarized in this paper. Note that these techniques are generic and widely used in various applications but play a key role in 3D geological modeling. There are two essential procedures in 3D geological modeling: the first is the simulation of geological interfaces using geometric surfaces and the second is the building of geological objects by means of various geometric computations such as the intersection of surfaces. Discrete geometric surfaces that represent geological interfaces can be generated by creating planar meshes first and then spatially interpolating; those surfaces intersect and then form volumes that represent three-dimensional geological objects such as rock bodies. In this paper, the most commonly used algorithms of the key techniques in 3D geological modeling are summarized.

  2. X-ray diffraction identification of clay minerals by microcomputer

    International Nuclear Information System (INIS)

    Rodrigues, S.; Imasava, F.J.

    1988-01-01

    The identification of clay minerals by X-ray powder diffraction are done by searching an unknown pattern with a file of standard X-ray diffraction patterns. For this searching done by hand is necessary a long time. This paper shows a program in ''Basic'' language to be utilized in microcomputers for the math of the unknown pattern, using the high velocity of comparison of the microcomputer. A few minutes are used for the match. (author) [pt

  3. An Extended Microcomputer-Based Network Optimization Package.

    Science.gov (United States)

    1982-10-01

    for New Cycle ..... ............ 53 * 19. Four Node Cycle ...... ... .. ... .. ... .. 54 20. All Artificial Start .................. 61 21. Micronet ...package. For the sake of brevity, the microcomputer-based network optimization package herein described will be referred to as Micronet . The host...computer for Micronet is an Apple II Plus with 64K of memory. This is certainly not the most powerful microcomputer (with some competitors featuring

  4. Microcomputer spacecraft thermal analysis routines (MSTAR) Phase I: The user interface

    Energy Technology Data Exchange (ETDEWEB)

    Teti, N.M. [Swales and Associates, Inc., Beltsville, MD (United States)

    1993-12-31

    The Microcomputer Spacecraft Thermal Analysis Routines (MSTAR) software package is being developed for NASA/Goddard Space Flight Center by Swales and Associates, Inc. (S&AI). In December 1992, S&AI was awarded a phase I Small Business Inovative Research contract fronm NASA to develop a microcomputer based thermal analysis program to replace the current SSPTA and TRASYS programs. Phase I consists of a six month effort which will focus on developing geometric model generation and visualization capabilities using a graphical user interface (GUI). The information contained in this paper encompasses the work performed during the Phase I development cycle; with emphasis on the development of the graphical user interface (GUI). This includes both the theory behind and specific examples of how the MSTAR GUI was implemented. Furthermore, this report discusses new applications and enhancements which will improve the capabilities and commercialization of the MSTAR program.

  5. A Method to Test Model Calibration Techniques: Preprint

    Energy Technology Data Exchange (ETDEWEB)

    Judkoff, Ron; Polly, Ben; Neymark, Joel

    2016-09-01

    This paper describes a method for testing model calibration techniques. Calibration is commonly used in conjunction with energy retrofit audit models. An audit is conducted to gather information about the building needed to assemble an input file for a building energy modeling tool. A calibration technique is used to reconcile model predictions with utility data, and then the 'calibrated model' is used to predict energy savings from a variety of retrofit measures and combinations thereof. Current standards and guidelines such as BPI-2400 and ASHRAE-14 set criteria for 'goodness of fit' and assume that if the criteria are met, then the calibration technique is acceptable. While it is logical to use the actual performance data of the building to tune the model, it is not certain that a good fit will result in a model that better predicts post-retrofit energy savings. Therefore, the basic idea here is that the simulation program (intended for use with the calibration technique) is used to generate surrogate utility bill data and retrofit energy savings data against which the calibration technique can be tested. This provides three figures of merit for testing a calibration technique, 1) accuracy of the post-retrofit energy savings prediction, 2) closure on the 'true' input parameter values, and 3) goodness of fit to the utility bill data. The paper will also discuss the pros and cons of using this synthetic surrogate data approach versus trying to use real data sets of actual buildings.

  6. Numerical model updating technique for structures using firefly algorithm

    Science.gov (United States)

    Sai Kubair, K.; Mohan, S. C.

    2018-03-01

    Numerical model updating is a technique used for updating the existing experimental models for any structures related to civil, mechanical, automobiles, marine, aerospace engineering, etc. The basic concept behind this technique is updating the numerical models to closely match with experimental data obtained from real or prototype test structures. The present work involves the development of numerical model using MATLAB as a computational tool and with mathematical equations that define the experimental model. Firefly algorithm is used as an optimization tool in this study. In this updating process a response parameter of the structure has to be chosen, which helps to correlate the numerical model developed with the experimental results obtained. The variables for the updating can be either material or geometrical properties of the model or both. In this study, to verify the proposed technique, a cantilever beam is analyzed for its tip deflection and a space frame has been analyzed for its natural frequencies. Both the models are updated with their respective response values obtained from experimental results. The numerical results after updating show that there is a close relationship that can be brought between the experimental and the numerical models.

  7. Plasticity models of material variability based on uncertainty quantification techniques

    Energy Technology Data Exchange (ETDEWEB)

    Jones, Reese E. [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Rizzi, Francesco [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Boyce, Brad [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Templeton, Jeremy Alan [Sandia National Lab. (SNL-CA), Livermore, CA (United States); Ostien, Jakob [Sandia National Lab. (SNL-CA), Livermore, CA (United States)

    2017-11-01

    The advent of fabrication techniques like additive manufacturing has focused attention on the considerable variability of material response due to defects and other micro-structural aspects. This variability motivates the development of an enhanced design methodology that incorporates inherent material variability to provide robust predictions of performance. In this work, we develop plasticity models capable of representing the distribution of mechanical responses observed in experiments using traditional plasticity models of the mean response and recently developed uncertainty quantification (UQ) techniques. Lastly, we demonstrate that the new method provides predictive realizations that are superior to more traditional ones, and how these UQ techniques can be used in model selection and assessing the quality of calibrated physical parameters.

  8. Models and Techniques for Proving Data Structure Lower Bounds

    DEFF Research Database (Denmark)

    Larsen, Kasper Green

    In this dissertation, we present a number of new techniques and tools for proving lower bounds on the operational time of data structures. These techniques provide new lines of attack for proving lower bounds in both the cell probe model, the group model, the pointer machine model and the I....../O-model. In all cases, we push the frontiers further by proving lower bounds higher than what could possibly be proved using previously known techniques. For the cell probe model, our results have the following consequences: The rst (lg n) query time lower bound for linear space static data structures...... bound of tutq = (lgd􀀀1 n). For ball range searching, we get a lower bound of tutq = (n1􀀀1=d). The highest previous lower bound proved in the group model does not exceed ((lg n= lg lg n)2) on the maximum of tu and tq. Finally, we present a new technique for proving lower bounds...

  9. Micro-computer tomography and a renaissance of insect morphology

    Science.gov (United States)

    Friedrich, Frank; Beutel, Rolf G.

    2008-08-01

    The use of new technologies, especially computer-based three dimensional reconstructions and micro-computer tomography (μ-CT) have greatly improved and facilitated the detailed investigation of insect anatomy. Optimal results in morphological work aiming at phylogenetic reconstruction can be obtained with a combined application of different techniques such as histology, scanning electron microscopy, and μ-CT. The use of μ-CT greatly enhances the efficiency of the acquisition of detailed anatomical data and allows a broad taxon sampling in phylogenetic studies partly or entirely based on morphological characters. A disadvantage of phase contrasted μ-CT images is the poor differentiation of different tissue types. This problem can be overcome by the use of stable low energy photon beams as available at the beamline BW2 of the Deutsches Elektronen-Synchrotron in Hamburg (DESY). Synchrotron-radiation-based μ-CT data (SR μ-CT) obtained with this approach are an ideal basis for highly efficient three dimensional reconstructions of high quality.

  10. Modeling with data tools and techniques for scientific computing

    CERN Document Server

    Klemens, Ben

    2009-01-01

    Modeling with Data fully explains how to execute computationally intensive analyses on very large data sets, showing readers how to determine the best methods for solving a variety of different problems, how to create and debug statistical models, and how to run an analysis and evaluate the results. Ben Klemens introduces a set of open and unlimited tools, and uses them to demonstrate data management, analysis, and simulation techniques essential for dealing with large data sets and computationally intensive procedures. He then demonstrates how to easily apply these tools to the many threads of statistical technique, including classical, Bayesian, maximum likelihood, and Monte Carlo methods

  11. Microcomputer applications for NSSS data package review

    International Nuclear Information System (INIS)

    Eng, D.J.; Smith, F.M.

    1984-01-01

    Associated with each component of a Nuclear Steam Supply System are the necessary supporting documents which demonstrate compliance with applicable codes, specifications and design criteria. The documents are grouped into individual data packages, and may exceed 800 in number for a major installation. A complete review, initiated by a utility in response to federal code regulations (10 CFR 50), can involve a tremendous number of review transactions. A data management system to assist the reviewer in the tracking and resolution of discrepancy items is currently under development and use. The system uses microcomputer-based relational database management software and provides complete and flexible capabilities to process database components and attributes based on user specified criteria. The completed database is a ''portable'' system, and can be utilized on an as-needed basis after the review is completed. A discrepancy analysis is performed to identify relations between manufacturer and discrepancy occurrence, part type vs. discrepancy type, etc. These results may prove useful in subsequent reviews or application to existing QA procedures

  12. Plants status monitor: Modelling techniques and inherent benefits

    International Nuclear Information System (INIS)

    Breeding, R.J.; Lainoff, S.M.; Rees, D.C.; Prather, W.A.; Fickiessen, K.O.E.

    1987-01-01

    The Plant Status Monitor (PSM) is designed to provide plant personnel with information on the operational status of the plant and compliance with the plant technical specifications. The PSM software evaluates system models using a 'distributed processing' technique in which detailed models of individual systems are processed rather than by evaluating a single, plant-level model. In addition, development of the system models for PSM provides inherent benefits to the plant by forcing detailed reviews of the technical specifications, system design and operating procedures, and plant documentation. (orig.)

  13. Sensitivity analysis technique for application to deterministic models

    International Nuclear Information System (INIS)

    Ishigami, T.; Cazzoli, E.; Khatib-Rahbar, M.; Unwin, S.D.

    1987-01-01

    The characterization of sever accident source terms for light water reactors should include consideration of uncertainties. An important element of any uncertainty analysis is an evaluation of the sensitivity of the output probability distributions reflecting source term uncertainties to assumptions regarding the input probability distributions. Historically, response surface methods (RSMs) were developed to replace physical models using, for example, regression techniques, with simplified models for example, regression techniques, with simplified models for extensive calculations. The purpose of this paper is to present a new method for sensitivity analysis that does not utilize RSM, but instead relies directly on the results obtained from the original computer code calculations. The merits of this approach are demonstrated by application of the proposed method to the suppression pool aerosol removal code (SPARC), and the results are compared with those obtained by sensitivity analysis with (a) the code itself, (b) a regression model, and (c) Iman's method

  14. Selection of productivity improvement techniques via mathematical modeling

    Directory of Open Access Journals (Sweden)

    Mahassan M. Khater

    2011-07-01

    Full Text Available This paper presents a new mathematical model to select an optimal combination of productivity improvement techniques. The proposed model of this paper considers four-stage cycle productivity and the productivity is assumed to be a linear function of fifty four improvement techniques. The proposed model of this paper is implemented for a real-world case study of manufacturing plant. The resulted problem is formulated as a mixed integer programming which can be solved for optimality using traditional methods. The preliminary results of the implementation of the proposed model of this paper indicate that the productivity can be improved through a change on equipments and it can be easily applied for both manufacturing and service industries.

  15. Modeling and design techniques for RF power amplifiers

    CERN Document Server

    Raghavan, Arvind; Laskar, Joy

    2008-01-01

    The book covers RF power amplifier design, from device and modeling considerations to advanced circuit design architectures and techniques. It focuses on recent developments and advanced topics in this area, including numerous practical designs to back the theoretical considerations. It presents the challenges in designing power amplifiers in silicon and helps the reader improve the efficiency of linear power amplifiers, and design more accurate compact device models, with faster extraction routines, to create cost effective and reliable circuits.

  16. Modeling and Simulation Techniques for Large-Scale Communications Modeling

    National Research Council Canada - National Science Library

    Webb, Steve

    1997-01-01

    .... Tests of random number generators were also developed and applied to CECOM models. It was found that synchronization of random number strings in simulations is easy to implement and can provide significant savings for making comparative studies. If synchronization is in place, then statistical experiment design can be used to provide information on the sensitivity of the output to input parameters. The report concludes with recommendations and an implementation plan.

  17. Effects of Peer Modelling Technique in Reducing Substance Abuse ...

    African Journals Online (AJOL)

    The study investigated the effects of peer modelling techniques in reducing substance abuse among undergraduates in Nigeria. The participants were one hundred and twenty (120) undergraduate students in 100 and 400 levels respectively. There are two groups: one treatment group and one control group.

  18. Using of Structural Equation Modeling Techniques in Cognitive Levels Validation

    Directory of Open Access Journals (Sweden)

    Natalija Curkovic

    2012-10-01

    Full Text Available When constructing knowledge tests, cognitive level is usually one of the dimensions comprising the test specifications with each item assigned to measure a particular level. Recently used taxonomies of the cognitive levels most often represent some modification of the original Bloom’s taxonomy. There are many concerns in current literature about existence of predefined cognitive levels. The aim of this article is to investigate can structural equation modeling techniques confirm existence of different cognitive levels. For the purpose of the research, a Croatian final high-school Mathematics exam was used (N = 9626. Confirmatory factor analysis and structural regression modeling were used to test three different models. Structural equation modeling techniques did not support existence of different cognitive levels in this case. There is more than one possible explanation for that finding. Some other techniques that take into account nonlinear behaviour of the items as well as qualitative techniques might be more useful for the purpose of the cognitive levels validation. Furthermore, it seems that cognitive levels were not efficient descriptors of the items and so improvements are needed in describing the cognitive skills measured by items.

  19. AN OVERVIEW OF REDUCED ORDER MODELING TECHNIQUES FOR SAFETY APPLICATIONS

    Energy Technology Data Exchange (ETDEWEB)

    Mandelli, D.; Alfonsi, A.; Talbot, P.; Wang, C.; Maljovec, D.; Smith, C.; Rabiti, C.; Cogliati, J.

    2016-10-01

    The RISMC project is developing new advanced simulation-based tools to perform Computational Risk Analysis (CRA) for the existing fleet of U.S. nuclear power plants (NPPs). These tools numerically model not only the thermal-hydraulic behavior of the reactors primary and secondary systems, but also external event temporal evolution and component/system ageing. Thus, this is not only a multi-physics problem being addressed, but also a multi-scale problem (both spatial, µm-mm-m, and temporal, seconds-hours-years). As part of the RISMC CRA approach, a large amount of computationally-expensive simulation runs may be required. An important aspect is that even though computational power is growing, the overall computational cost of a RISMC analysis using brute-force methods may be not viable for certain cases. A solution that is being evaluated to assist the computational issue is the use of reduced order modeling techniques. During the FY2015, we investigated and applied reduced order modeling techniques to decrease the RISMC analysis computational cost by decreasing the number of simulation runs; for this analysis improvement we used surrogate models instead of the actual simulation codes. This article focuses on the use of reduced order modeling techniques that can be applied to RISMC analyses in order to generate, analyze, and visualize data. In particular, we focus on surrogate models that approximate the simulation results but in a much faster time (microseconds instead of hours/days).

  20. Space geodetic techniques for global modeling of ionospheric peak parameters

    Science.gov (United States)

    Alizadeh, M. Mahdi; Schuh, Harald; Schmidt, Michael

    The rapid development of new technological systems for navigation, telecommunication, and space missions which transmit signals through the Earth’s upper atmosphere - the ionosphere - makes the necessity of precise, reliable and near real-time models of the ionospheric parameters more crucial. In the last decades space geodetic techniques have turned into a capable tool for measuring ionospheric parameters in terms of Total Electron Content (TEC) or the electron density. Among these systems, the current space geodetic techniques, such as Global Navigation Satellite Systems (GNSS), Low Earth Orbiting (LEO) satellites, satellite altimetry missions, and others have found several applications in a broad range of commercial and scientific fields. This paper aims at the development of a three-dimensional integrated model of the ionosphere, by using various space geodetic techniques and applying a combination procedure for computation of the global model of electron density. In order to model ionosphere in 3D, electron density is represented as a function of maximum electron density (NmF2), and its corresponding height (hmF2). NmF2 and hmF2 are then modeled in longitude, latitude, and height using two sets of spherical harmonic expansions with degree and order 15. To perform the estimation, GNSS input data are simulated in such a way that the true position of the satellites are detected and used, but the STEC values are obtained through a simulation procedure, using the IGS VTEC maps. After simulating the input data, the a priori values required for the estimation procedure are calculated using the IRI-2012 model and also by applying the ray-tracing technique. The estimated results are compared with F2-peak parameters derived from the IRI model to assess the least-square estimation procedure and moreover, to validate the developed maps, the results are compared with the raw F2-peak parameters derived from the Formosat-3/Cosmic data.

  1. Inside marginal adaptation of crowns by X-ray micro-computed tomography

    Energy Technology Data Exchange (ETDEWEB)

    Dos Santos, T. M.; Lima, I.; Lopes, R. T. [Nuclear Instrumentation Laboratory, Nuclear Engineering Program, Federal University of Rio de Janeiro, RJ, (Brazil); Author, S. B. Jr. [Department of Physics, Colorado State University, Ft. Collins, CO 80523, (United States)

    2015-07-01

    The objective of this work was to access dental arcade by using X-ray micro-computed tomography. For this purpose high resolution system was used and three groups were studied: Zirkonzahn CAD-CAM system, IPS e.max Press, and metal ceramic. The three systems assessed in this study showed results of marginal and discrepancy gaps clinically accepted. The great result of 2D and 3D evaluations showed that the used technique is a powerful method to investigate quantitative characteristics of dental arcade. (authors)

  2. An automated microcomputer-controlled system for neutron activation and gamma-ray spectroscopy

    International Nuclear Information System (INIS)

    Edward, J.B.; Bennett, L.G.I.

    1990-01-01

    An automated instrumental neutron activation analysis (INAA) system has been constructed at the SLOWPOKE-2 reactor at the Royal Military College of Canada (RMC). Its pneumatic transfer system is controlled by an Apple IIe computer, linked in turn to an MS-DOS-compatible microcomputer which controls data acquisition. Custom software has been created for these computers and for off-line spectral analysis using programs that incorporate either peak boundary or Gaussian peak fitting methods of analysis. This system provides the gamut of INAA techniques for the analyst. The design and performance of the hardware and software are discussed. (orig.)

  3. Model-checking techniques based on cumulative residuals.

    Science.gov (United States)

    Lin, D Y; Wei, L J; Ying, Z

    2002-03-01

    Residuals have long been used for graphical and numerical examinations of the adequacy of regression models. Conventional residual analysis based on the plots of raw residuals or their smoothed curves is highly subjective, whereas most numerical goodness-of-fit tests provide little information about the nature of model misspecification. In this paper, we develop objective and informative model-checking techniques by taking the cumulative sums of residuals over certain coordinates (e.g., covariates or fitted values) or by considering some related aggregates of residuals, such as moving sums and moving averages. For a variety of statistical models and data structures, including generalized linear models with independent or dependent observations, the distributions of these stochastic processes tinder the assumed model can be approximated by the distributions of certain zero-mean Gaussian processes whose realizations can be easily generated by computer simulation. Each observed process can then be compared, both graphically and numerically, with a number of realizations from the Gaussian process. Such comparisons enable one to assess objectively whether a trend seen in a residual plot reflects model misspecification or natural variation. The proposed techniques are particularly useful in checking the functional form of a covariate and the link function. Illustrations with several medical studies are provided.

  4. Advanced techniques in reliability model representation and solution

    Science.gov (United States)

    Palumbo, Daniel L.; Nicol, David M.

    1992-01-01

    The current tendency of flight control system designs is towards increased integration of applications and increased distribution of computational elements. The reliability analysis of such systems is difficult because subsystem interactions are increasingly interdependent. Researchers at NASA Langley Research Center have been working for several years to extend the capability of Markov modeling techniques to address these problems. This effort has been focused in the areas of increased model abstraction and increased computational capability. The reliability model generator (RMG) is a software tool that uses as input a graphical object-oriented block diagram of the system. RMG uses a failure-effects algorithm to produce the reliability model from the graphical description. The ASSURE software tool is a parallel processing program that uses the semi-Markov unreliability range evaluator (SURE) solution technique and the abstract semi-Markov specification interface to the SURE tool (ASSIST) modeling language. A failure modes-effects simulation is used by ASSURE. These tools were used to analyze a significant portion of a complex flight control system. The successful combination of the power of graphical representation, automated model generation, and parallel computation leads to the conclusion that distributed fault-tolerant system architectures can now be analyzed.

  5. Use of hydrological modelling and isotope techniques in Guvenc basin

    International Nuclear Information System (INIS)

    Altinbilek, D.

    1991-07-01

    The study covers the work performed under Project No. 335-RC-TUR-5145 entitled ''Use of Hydrologic Modelling and Isotope Techniques in Guvenc Basin'' and is an initial part of a program for estimating runoff from Central Anatolia Watersheds. The study presented herein consists of mainly three parts: 1) the acquisition of a library of rainfall excess, direct runoff and isotope data for Guvenc basin; 2) the modification of SCS model to be applied to Guvenc basin first and then to other basins of Central Anatolia for predicting the surface runoff from gaged and ungaged watersheds; and 3) the use of environmental isotope technique in order to define the basin components of streamflow of Guvenc basin. 31 refs, figs and tabs

  6. Equivalence and Differences between Structural Equation Modeling and State-Space Modeling Techniques

    Science.gov (United States)

    Chow, Sy-Miin; Ho, Moon-ho R.; Hamaker, Ellen L.; Dolan, Conor V.

    2010-01-01

    State-space modeling techniques have been compared to structural equation modeling (SEM) techniques in various contexts but their unique strengths have often been overshadowed by their similarities to SEM. In this article, we provide a comprehensive discussion of these 2 approaches' similarities and differences through analytic comparisons and…

  7. Total laparoscopic gastrocystoplasty: experimental technique in a porcine model

    OpenAIRE

    Frederico R. Romero; Claudemir Trapp; Michael Muntener; Fabio A. Brito; Louis R. Kavoussi; Thomas W. Jarrett

    2007-01-01

    OBJECTIVE: Describe a unique simplified experimental technique for total laparoscopic gastrocystoplasty in a porcine model. MATERIAL AND METHODS: We performed laparoscopic gastrocystoplasty on 10 animals. The gastroepiploic arch was identified and carefully mobilized from its origin at the pylorus to the beginning of the previously demarcated gastric wedge. The gastric segment was resected with sharp dissection. Both gastric suturing and gastrovesical anastomosis were performed with absorbabl...

  8. A Bayesian Technique for Selecting a Linear Forecasting Model

    OpenAIRE

    Ramona L. Trader

    1983-01-01

    The specification of a forecasting model is considered in the context of linear multiple regression. Several potential predictor variables are available, but some of them convey little information about the dependent variable which is to be predicted. A technique for selecting the "best" set of predictors which takes into account the inherent uncertainty in prediction is detailed. In addition to current data, there is often substantial expert opinion available which is relevant to the forecas...

  9. Reactor protection system design using micro-computers

    International Nuclear Information System (INIS)

    Fairbrother, D.B.

    1976-01-01

    Reactor protection systems for nuclear power plants have traditionally been built using analog hardware. This hardware works quite well for single parameter trip functions; however, optimum protection against DNBR and KW/ft limits requires more complex trip functions than can easily be handled with analog hardware. For this reason, Babcock and Wilcox has introduced a Reactor Protection System, called the RPS-II, that utilizes a micro-computer to handle the more complex trip functions. The paper describes the design of the RPS-II and the operation of the micro-computer within the Reactor Protection System

  10. Reactor protection system design using micro-computers

    International Nuclear Information System (INIS)

    Fairbrother, D.B.

    1977-01-01

    Reactor Protection Systems for Nuclear Power Plants have traditionally been built using analog hardware. This hardware works quite well for single parameter trip functions; however, optimum protection against DNBR and KW/ft limits requires more complex trip functions than can easily be handled with analog hardware. For this reason, Babcock and Wilcox has introduced a Reactor Protection System, called the RPS-II, that utilizes a micro-computer to handle the more complex trip functions. This paper describes the design of the RPS-II and the operation of the micro-computer within the Reactor Protection System

  11. Teleradiology: an evaluation of a microcomputer-based system.

    Science.gov (United States)

    Gayler, B W; Gitlin, J N; Rappaport, W; Skinner, F L; Cerva, J

    1981-08-01

    A laboratory evaluation of a microcomputer-based teleradiology system is described. The system used state-of-the-art hardware, and provided for contrast and brightness control by the consulting radiologists. The transmitted, digitized images were compared with conventional radiographs. The radiologists' scores for findings, impressions, and confidence levels were significantly lower for radiographic images viewed on the teleradiology system. However, the quality of the images provided by the teleradiology system was high enough to warrant further study, especially since microcomputer-based teleradiology offers the advantage of providing access to radiology services for rural areas and small hospitals.

  12. Fuzzy techniques for subjective workload-score modeling under uncertainties.

    Science.gov (United States)

    Kumar, Mohit; Arndt, Dagmar; Kreuzfeld, Steffi; Thurow, Kerstin; Stoll, Norbert; Stoll, Regina

    2008-12-01

    This paper deals with the development of a computer model to estimate the subjective workload score of individuals by evaluating their heart-rate (HR) signals. The identification of a model to estimate the subjective workload score of individuals under different workload situations is too ambitious a task because different individuals (due to different body conditions, emotional states, age, gender, etc.) show different physiological responses (assessed by evaluating the HR signal) under different workload situations. This is equivalent to saying that the mathematical mappings between physiological parameters and the workload score are uncertain. Our approach to deal with the uncertainties in a workload-modeling problem consists of the following steps: 1) The uncertainties arising due the individual variations in identifying a common model valid for all the individuals are filtered out using a fuzzy filter; 2) stochastic modeling of the uncertainties (provided by the fuzzy filter) use finite-mixture models and utilize this information regarding uncertainties for identifying the structure and initial parameters of a workload model; and 3) finally, the workload model parameters for an individual are identified in an online scenario using machine learning algorithms. The contribution of this paper is to propose, with a mathematical analysis, a fuzzy-based modeling technique that first filters out the uncertainties from the modeling problem, analyzes the uncertainties statistically using finite-mixture modeling, and, finally, utilizes the information about uncertainties for adapting the workload model to an individual's physiological conditions. The approach of this paper, demonstrated with the real-world medical data of 11 subjects, provides a fuzzy-based tool useful for modeling in the presence of uncertainties.

  13. Sensitivity analysis techniques for models of human behavior.

    Energy Technology Data Exchange (ETDEWEB)

    Bier, Asmeret Brooke

    2010-09-01

    Human and social modeling has emerged as an important research area at Sandia National Laboratories due to its potential to improve national defense-related decision-making in the presence of uncertainty. To learn about which sensitivity analysis techniques are most suitable for models of human behavior, different promising methods were applied to an example model, tested, and compared. The example model simulates cognitive, behavioral, and social processes and interactions, and involves substantial nonlinearity, uncertainty, and variability. Results showed that some sensitivity analysis methods create similar results, and can thus be considered redundant. However, other methods, such as global methods that consider interactions between inputs, can generate insight not gained from traditional methods.

  14. Practical Techniques for Modeling Gas Turbine Engine Performance

    Science.gov (United States)

    Chapman, Jeffryes W.; Lavelle, Thomas M.; Litt, Jonathan S.

    2016-01-01

    The cost and risk associated with the design and operation of gas turbine engine systems has led to an increasing dependence on mathematical models. In this paper, the fundamentals of engine simulation will be reviewed, an example performance analysis will be performed, and relationships useful for engine control system development will be highlighted. The focus will be on thermodynamic modeling utilizing techniques common in industry, such as: the Brayton cycle, component performance maps, map scaling, and design point criteria generation. In general, these topics will be viewed from the standpoint of an example turbojet engine model; however, demonstrated concepts may be adapted to other gas turbine systems, such as gas generators, marine engines, or high bypass aircraft engines. The purpose of this paper is to provide an example of gas turbine model generation and system performance analysis for educational uses, such as curriculum creation or student reference.

  15. New techniques and models for assessing ischemic heart disease risks

    Directory of Open Access Journals (Sweden)

    I.N. Yakovina

    2017-09-01

    Full Text Available The paper focuses on tasks of creating and implementing a new technique aimed at assessing ischemic heart diseases risk. The techniques is based on a laboratory-diagnostic complex which includes oxidative, lipid-lipoprotein, inflammatory and metabolic biochemical parameters; s system of logic-mathematic models used for obtaining numeric risk assessments; and a program module which allows to calculate and analyze the results. we justified our models in the course of our re-search which included 172 patients suffering from ischemic heart diseases (IHD combined with coronary atherosclerosis verified by coronary arteriography and 167 patients who didn't have ischemic heart diseases. Our research program in-cluded demographic and social data, questioning on tobacco and alcohol addiction, questioning about dietary habits, chronic diseases case history and medications intake, cardiologic questioning as per Rose, anthropometry, 3-times meas-ured blood pressure, spirometry, and electrocardiogram taking and recording with decoding as per Minnesota code. We detected biochemical parameters of each patient and adjusted our task of creating techniques and models for assessing ischemic heart disease risks on the basis of inflammatory, oxidative, and lipid biological markers. We created a system of logic and mathematic models which is a universal scheme for laboratory parameters processing allowing for dissimilar data specificity. The system of models is universal, but a diagnostic approach to applied biochemical parameters is spe-cific. The created program module (calculator helps a physician to obtain a result on the basis of laboratory research data; the result characterizes numeric risks of coronary atherosclerosis and ischemic heart disease for a patient. It also allows to obtain a visual image of a system of parameters and their deviation from a conditional «standard – pathology» boundary. The complex is implemented into practice by the Scientific

  16. Mathematical and Numerical Techniques in Energy and Environmental Modeling

    Science.gov (United States)

    Chen, Z.; Ewing, R. E.

    Mathematical models have been widely used to predict, understand, and optimize many complex physical processes, from semiconductor or pharmaceutical design to large-scale applications such as global weather models to astrophysics. In particular, simulation of environmental effects of air pollution is extensive. Here we address the need for using similar models to understand the fate and transport of groundwater contaminants and to design in situ remediation strategies. Three basic problem areas need to be addressed in the modeling and simulation of the flow of groundwater contamination. First, one obtains an effective model to describe the complex fluid/fluid and fluid/rock interactions that control the transport of contaminants in groundwater. This includes the problem of obtaining accurate reservoir descriptions at various length scales and modeling the effects of this heterogeneity in the reservoir simulators. Next, one develops accurate discretization techniques that retain the important physical properties of the continuous models. Finally, one develops efficient numerical solution algorithms that utilize the potential of the emerging computing architectures. We will discuss recent advances and describe the contribution of each of the papers in this book in these three areas. Keywords: reservoir simulation, mathematical models, partial differential equations, numerical algorithms

  17. Improved ceramic slip casting technique. [application to aircraft model fabrication

    Science.gov (United States)

    Buck, Gregory M. (Inventor); Vasquez, Peter (Inventor)

    1993-01-01

    A primary concern in modern fluid dynamics research is the experimental verification of computational aerothermodynamic codes. This research requires high precision and detail in the test model employed. Ceramic materials are used for these models because of their low heat conductivity and their survivability at high temperatures. To fabricate such models, slip casting techniques were developed to provide net-form, precision casting capability for high-purity ceramic materials in aqueous solutions. In previous slip casting techniques, block, or flask molds made of plaster-of-paris were used to draw liquid from the slip material. Upon setting, parts were removed from the flask mold and cured in a kiln at high temperatures. Casting detail was usually limited with this technique -- detailed parts were frequently damaged upon separation from the flask mold, as the molded parts are extremely delicate in the uncured state, and the flask mold is inflexible. Ceramic surfaces were also marred by 'parting lines' caused by mold separation. This adversely affected the aerodynamic surface quality of the model as well. (Parting lines are invariably necessary on or near the leading edges of wings, nosetips, and fins for mold separation. These areas are also critical for flow boundary layer control.) Parting agents used in the casting process also affected surface quality. These agents eventually soaked into the mold, the model, or flaked off when releasing the case model. Different materials were tried, such as oils, paraffin, and even an algae. The algae released best, but some of it remained on the model and imparted an uneven texture and discoloration on the model surface when cured. According to the present invention, a wax pattern for a shell mold is provided, and an aqueous mixture of a calcium sulfate-bonded investment material is applied as a coating to the wax pattern. The coated wax pattern is then dried, followed by curing to vaporize the wax pattern and leave a shell

  18. Virtual casting of stab wounds in cartilage using micro-computed tomography.

    Science.gov (United States)

    Pounder, Derrick J; Sim, Louise J

    2011-06-01

    In homicidal stabbings using a serrated knife, stab wounds involving costal cartilage leave striations on the cut surface of the cartilage from the serration points on the blade edge. Class characteristics of the serrated blade can be determined from the striation marks, and individualizing characteristics may be seen also. The traditional method for recording the striation marks involves the pernickety technique of casting with dental impression material. We assessed the potential utility of micro-computed tomography scanning of the stab track as a technique for nondestructive recording of striation patterns and virtual casting of them. Stab tracks in porcine cartilage, produced with a coarsely serrated blade, were scanned with a bench-top micro-computed tomography scanner. The typical irregularly regular striation pattern could be demonstrated, and the images manipulated, using proprietary software to produce a virtual cast. Whether the technology will have sufficient resolution to image not only class characteristic tool marks but also the much finer individualizing tool marks remains to be evaluated, but the technology shows considerable promise.

  19. Cooperative cognitive radio networking system model, enabling techniques, and performance

    CERN Document Server

    Cao, Bin; Mark, Jon W

    2016-01-01

    This SpringerBrief examines the active cooperation between users of Cooperative Cognitive Radio Networking (CCRN), exploring the system model, enabling techniques, and performance. The brief provides a systematic study on active cooperation between primary users and secondary users, i.e., (CCRN), followed by the discussions on research issues and challenges in designing spectrum-energy efficient CCRN. As an effort to shed light on the design of spectrum-energy efficient CCRN, they model the CCRN based on orthogonal modulation and orthogonally dual-polarized antenna (ODPA). The resource allocation issues are detailed with respect to both models, in terms of problem formulation, solution approach, and numerical results. Finally, the optimal communication strategies for both primary and secondary users to achieve spectrum-energy efficient CCRN are analyzed.

  20. Numerical and modeling techniques used in the EPIC code

    International Nuclear Information System (INIS)

    Pizzica, P.A.; Abramson, P.B.

    1977-01-01

    EPIC models fuel and coolant motion which result from internal fuel pin pressure (from fission gas or fuel vapor) and/or from the generation of sodium vapor pressures in the coolant channel subsequent to pin failure in an LMFBR. The modeling includes the ejection of molten fuel from the pin into a coolant channel with any amount of voiding through a clad rip which may be of any length or which may expand with time. One-dimensional Eulerian hydrodynamics is used to model both the motion of fuel and fission gas inside a molten fuel cavity and the mixture of two-phase sodium and fission gas in the channel. Motion of molten fuel particles in the coolant channel is tracked with a particle-in-cell technique

  1. Teaching scientific concepts through simple models and social communication techniques

    International Nuclear Information System (INIS)

    Tilakaratne, K.

    2011-01-01

    For science education, it is important to demonstrate to students the relevance of scientific concepts in every-day life experiences. Although there are methods available for achieving this goal, it is more effective if cultural flavor is also added to the teaching techniques and thereby the teacher and students can easily relate the subject matter to their surroundings. Furthermore, this would bridge the gap between science and day-to-day experiences in an effective manner. It could also help students to use science as a tool to solve problems faced by them and consequently they would feel science is a part of their lives. In this paper, it has been described how simple models and cultural communication techniques can be used effectively in demonstrating important scientific concepts to the students of secondary and higher secondary levels by using two consecutive activities carried out at the Institute of Fundamental Studies (IFS), Sri Lanka. (author)

  2. An Iterative Uncertainty Assessment Technique for Environmental Modeling

    International Nuclear Information System (INIS)

    Engel, David W.; Liebetrau, Albert M.; Jarman, Kenneth D.; Ferryman, Thomas A.; Scheibe, Timothy D.; Didier, Brett T.

    2004-01-01

    The reliability of and confidence in predictions from model simulations are crucial--these predictions can significantly affect risk assessment decisions. For example, the fate of contaminants at the U.S. Department of Energy's Hanford Site has critical impacts on long-term waste management strategies. In the uncertainty estimation efforts for the Hanford Site-Wide Groundwater Modeling program, computational issues severely constrain both the number of uncertain parameters that can be considered and the degree of realism that can be included in the models. Substantial improvements in the overall efficiency of uncertainty analysis are needed to fully explore and quantify significant sources of uncertainty. We have combined state-of-the-art statistical and mathematical techniques in a unique iterative, limited sampling approach to efficiently quantify both local and global prediction uncertainties resulting from model input uncertainties. The approach is designed for application to widely diverse problems across multiple scientific domains. Results are presented for both an analytical model where the response surface is ''known'' and a simplified contaminant fate transport and groundwater flow model. The results show that our iterative method for approximating a response surface (for subsequent calculation of uncertainty estimates) of specified precision requires less computing time than traditional approaches based upon noniterative sampling methods

  3. Microcomputers--A Way for Teachers to Cope.

    Science.gov (United States)

    Squires, David; Coward, Geoffry

    1981-01-01

    An approach to teacher inservice training in the use of computer-based learning is described in this paper, with specific discussion devoted to the development of a videotape training module by the Hertfordshire Advisory Council for Educational Technology (HACET) to demonstrate the role of educational technology, and in particular microcomputers,…

  4. Planning for the Administrative Microcomputer Local Area Network.

    Science.gov (United States)

    Rohr, Theodore; Devoti, Bart

    An overview is provided of the methods used by the Forest Park campus of St. Louis Community College (SLCC) to plan and develop a local area network (LAN) for administrative microcomputers. The first three sections provide brief descriptions of the SLCC District, SLCC, and the Forest Park campus. Section IV looks at the organization of…

  5. A multi-microcomputer system for Monte Carlo calculations

    International Nuclear Information System (INIS)

    Hertzberger, L.O.; Berg, B.; Krasemann, H.

    1981-01-01

    We propose a microcomputer system which allows parallel processing for Monte Carlo calculations in lattice gauge theories, simulations of high energy physics experiments and presumably many other fields of current interest. The master-n-slave multiprocessor system is based on the Motorola MC 68000 microprocessor. One attraction if this processor is that it allows up to 16 M Byte random access memory. (orig.)

  6. The Use of Microcomputers in the Teaching of Calculus.

    Science.gov (United States)

    Hickernell, Fred; Proskurowski, Wlodek

    1985-01-01

    Discusses conception, design, implementation, and evaluation of a calculus course that utilizes a multi-user microcomputer system with graphics terminals. The numerical and utility routines of the software are briefly described, and examples of several types of calculus problems that can be solved utilizing the system are reviewed. (Author/MBR)

  7. Microcomputer-aided monitor for liquid hydrogen target system

    International Nuclear Information System (INIS)

    Kitami, T.; Watanabe, K.

    1983-03-01

    A microcomputer-aided monitor for a liquid hydrogen target system has been designed and tested. Various kinds of input data such as temperature, pressure, vacuum, etc. are scanned in a given time interval. Variation with time in any four items can be displayed on CRT and, if neccessary, printed out on a sheet of recording paper. (author)

  8. Machine Check-Out. Microcomputing Working Paper Series.

    Science.gov (United States)

    Drexel Univ., Philadelphia, PA. Microcomputing Program.

    During the academic year 1983-84, Drexel University instituted a new policy requiring all incoming students to have access to a microcomputer. The computer chosen to fulfill this requirement was the Macintosh from Apple Computer, Inc. Because Drexel University received one of the first large shipments of this new product, the degree to which these…

  9. Microcomputers in the Schools: A Case of Decentralized Diffusion.

    Science.gov (United States)

    Rogers, Everett M.; And Others

    Utilizing nine schools located in the San Francisco Bay area, a study was conducted to examine the patterns by which microcomputers are accepted and implemented in high schools. Rather than the educational or social effects of this new academic resource, the study focused on: (1) the process of the behaviors and decisions leading to the…

  10. Microcomputers as on-line catalogs in special libraries.

    Science.gov (United States)

    Faust, J B

    1986-01-01

    This article discusses the rationale for the conversion of a card catalog to an on-line system in a special library owning approximately 4000 titles. Equipment, software, and procedures are described. Pros and cons of the use of a microcomputer for such a project, as well as costs and personnel needs, are outlined.

  11. CC80-A microcomputer crate controller for CAMAC

    International Nuclear Information System (INIS)

    Walz, H.V.

    1976-11-01

    A microcomputer crate controller has been developed for use in CAMAC based instrumentation, control, and data processing systems at the Stanford Linear Accelerator Center. This unit may be used in multicrate branch highway systems to provide local processing capability within each crate, or in a single crate to form a stand-alone CAMAC system

  12. Learner/User Preferences for Fonts in Microcomputer Screen Displays.

    Science.gov (United States)

    Misanchuk, Earl R.

    1989-01-01

    Describes study that was conducted with 62 subjects who varied in age, educational background, and computer experience, to determine user preferences for fonts on a Macintosh microcomputer screen display. Topics discussed include the forced choice comparison mode used, a HyperCard stack designed for the study, and a Thurstone scale derived from…

  13. The Microcomputer in the Clinical Nursing Research Unit

    OpenAIRE

    Schwirian, Patricia M.; Byers, Sandra R.

    1982-01-01

    This paper discusses the microcomputer in clinical nursing research. There are six general areas in which computers have been useful to nurses: nursing notes and charting; patient care plans; automated monitoring of high-tech nursing units; HIS and MIS systems; personnel distribution systems; and education.

  14. A Microcomputer-Based Interactive Presentation Development System.

    Science.gov (United States)

    Moreau, Dennis R.; Dominick, Wayne D.

    1988-01-01

    Reviews research and development projects sponsored by the National Aeronautics and Space Administration (NASA) that address microcomputer-based support for instructional activities at the University of Southwestern Louisiana. Highlights include a graphics project, local area networks, and the Interactive Presentation Development System, which is…

  15. Dependent and Independent Use of Microcomputers in Aphasia Rehabilitation.

    Science.gov (United States)

    Mills, Russell H.

    1987-01-01

    Three modes of using the microcomputer in aphasia rehabilitation include the dependent mode (with the clinician in the clinic setting), the clinic-based independent mode, and the home-based mode. Also considered are hardware and software selection, patient selection, clinician responsibilities, and family support. A case study illustrates…

  16. Computers, Sex, and Society. The Microcomputing Program at Drexel University.

    Science.gov (United States)

    McCord, Joan

    Results of 1983 and 1984 surveys of faculty and students at Drexel University are reviewed to evaluate the effects of the requirement that incoming students must buy a microcomputer. Slightly more than half the faculty were computer-competent (i.e., had considerable experience with at least one class of computer, with using computers in teaching,…

  17. Multi-microcomputer system for Monte-Carlo calculations

    CERN Document Server

    Berg, B; Krasemann, H

    1981-01-01

    The authors propose a microcomputer system that allows parallel processing for Monte Carlo calculations in lattice gauge theories, simulations of high energy physics experiments and many other fields of current interest. The master-n-slave multiprocessor system is based on the Motorola MC 6800 microprocessor. One attraction of this processor is that it allows up to 16 M Byte random access memory.

  18. Evaluation of Three Microcomputer Teaching Modules. SUMIT Courseware Development Project.

    Science.gov (United States)

    Soldan, Ted

    The purpose of this series of experiments was to examine two questions related to the effectiveness of computer assisted instruction (CAI). Can microcomputer modules teach effectively, and do they enhance learning when used as a supplement to traditional teaching methods? Part 1 of this report addresses the former question and part 2 addresses the…

  19. Computer-Assisted Statistical Analysis: Mainframe or Microcomputer.

    Science.gov (United States)

    Shannon, David M.

    1993-01-01

    Describes a study that was designed to examine whether the computer attitudes of graduate students in a beginning statistics course differed based on their prior computer experience and the type of statistical analysis package used. Versions of statistical analysis packages using a mainframe and a microcomputer are compared. (14 references) (LRW)

  20. Application of microcomputer to X-ray radiometric ore separation

    International Nuclear Information System (INIS)

    Neverov, A.D.; Aleksandrov, P.S.; Kotler, N.I.

    1988-01-01

    The practical use of microcomputers as universal means for converting information for solving applied problems of X-ray radiometric ore separation method is considered. Laboratory tests of two metals - tungsten and tin manifested high efficiency of the developed system. X-ray radiometric separator software is developed

  1. Use of advanced modeling techniques to optimize thermal packaging designs.

    Science.gov (United States)

    Formato, Richard M; Potami, Raffaele; Ahmed, Iftekhar

    2010-01-01

    Through a detailed case study the authors demonstrate, for the first time, the capability of using advanced modeling techniques to correctly simulate the transient temperature response of a convective flow-based thermal shipper design. The objective of this case study was to demonstrate that simulation could be utilized to design a 2-inch-wall polyurethane (PUR) shipper to hold its product box temperature between 2 and 8 °C over the prescribed 96-h summer profile (product box is the portion of the shipper that is occupied by the payload). Results obtained from numerical simulation are in excellent agreement with empirical chamber data (within ±1 °C at all times), and geometrical locations of simulation maximum and minimum temperature match well with the corresponding chamber temperature measurements. Furthermore, a control simulation test case was run (results taken from identical product box locations) to compare the coupled conduction-convection model with a conduction-only model, which to date has been the state-of-the-art method. For the conduction-only simulation, all fluid elements were replaced with "solid" elements of identical size and assigned thermal properties of air. While results from the coupled thermal/fluid model closely correlated with the empirical data (±1 °C), the conduction-only model was unable to correctly capture the payload temperature trends, showing a sizeable error compared to empirical values (ΔT > 6 °C). A modeling technique capable of correctly capturing the thermal behavior of passively refrigerated shippers can be used to quickly evaluate and optimize new packaging designs. Such a capability provides a means to reduce the cost and required design time of shippers while simultaneously improving their performance. Another advantage comes from using thermal modeling (assuming a validated model is available) to predict the temperature distribution in a shipper that is exposed to ambient temperatures which were not bracketed

  2. Modelling galaxy formation with multi-scale techniques

    International Nuclear Information System (INIS)

    Hobbs, A.

    2011-01-01

    Full text: Galaxy formation and evolution depends on a wide variety of physical processes - star formation, gas cooling, supernovae explosions and stellar winds etc. - that span an enormous range of physical scales. We present a novel technique for modelling such massively multiscale systems. This has two key new elements: Lagrangian re simulation, and convergent 'sub-grid' physics. The former allows us to hone in on interesting simulation regions with very high resolution. The latter allows us to increase resolution for the physics that we can resolve, without unresolved physics spoiling convergence. We illustrate the power of our new approach by showing some new results for star formation in the Milky Way. (author)

  3. Prescribed wind shear modelling with the actuator line technique

    DEFF Research Database (Denmark)

    Mikkelsen, Robert Flemming; Sørensen, Jens Nørkær; Troldborg, Niels

    2007-01-01

    A method for prescribing arbitrary steady atmospheric wind shear profiles combined with CFD is presented. The method is furthermore combined with the actuator line technique governing the aerodynamic loads on a wind turbine. Computation are carried out on a wind turbine exposed to a representative...... steady atmospheric wind shear profile with and without wind direction changes up through the atmospheric boundary layer. Results show that the main impact on the turbine is captured by the model. Analysis of the wake behind the wind turbine, reveal the formation of a skewed wake geometry interacting...

  4. Validation techniques of agent based modelling for geospatial simulations

    Science.gov (United States)

    Darvishi, M.; Ahmadi, G.

    2014-10-01

    One of the most interesting aspects of modelling and simulation study is to describe the real world phenomena that have specific properties; especially those that are in large scales and have dynamic and complex behaviours. Studying these phenomena in the laboratory is costly and in most cases it is impossible. Therefore, Miniaturization of world phenomena in the framework of a model in order to simulate the real phenomena is a reasonable and scientific approach to understand the world. Agent-based modelling and simulation (ABMS) is a new modelling method comprising of multiple interacting agent. They have been used in the different areas; for instance, geographic information system (GIS), biology, economics, social science and computer science. The emergence of ABM toolkits in GIS software libraries (e.g. ESRI's ArcGIS, OpenMap, GeoTools, etc) for geospatial modelling is an indication of the growing interest of users to use of special capabilities of ABMS. Since ABMS is inherently similar to human cognition, therefore it could be built easily and applicable to wide range applications than a traditional simulation. But a key challenge about ABMS is difficulty in their validation and verification. Because of frequent emergence patterns, strong dynamics in the system and the complex nature of ABMS, it is hard to validate and verify ABMS by conventional validation methods. Therefore, attempt to find appropriate validation techniques for ABM seems to be necessary. In this paper, after reviewing on Principles and Concepts of ABM for and its applications, the validation techniques and challenges of ABM validation are discussed.

  5. Laparoscopic anterior resection: new anastomosis technique in a pig model.

    Science.gov (United States)

    Bedirli, Abdulkadir; Yucel, Deniz; Ekim, Burcu

    2014-01-01

    Bowel anastomosis after anterior resection is one of the most difficult tasks to perform during laparoscopic colorectal surgery. This study aims to evaluate a new feasible and safe intracorporeal anastomosis technique after laparoscopic left-sided colon or rectum resection in a pig model. The technique was evaluated in 5 pigs. The OrVil device (Covidien, Mansfield, Massachusetts) was inserted into the anus and advanced proximally to the rectum. A 0.5-cm incision was made in the sigmoid colon, and the 2 sutures attached to its delivery tube were cut. After the delivery tube was evacuated through the anus, the tip of the anvil was removed through the perforation. The sigmoid colon was transected just distal to the perforation with an endoscopic linear stapler. The rectosigmoid segment to be resected was removed through the anus with a grasper, and distal transection was performed. A 25-mm circular stapler was inserted and combined with the anvil, and end-to-side intracorporeal anastomosis was then performed. We performed the technique in 5 pigs. Anastomosis required an average of 12 minutes. We observed that the proximal and distal donuts were completely removed in all pigs. No anastomotic air leakage was observed in any of the animals. This study shows the efficacy and safety of intracorporeal anastomosis with the OrVil device after laparoscopic anterior resection.

  6. Advanced applications of numerical modelling techniques for clay extruder design

    Science.gov (United States)

    Kandasamy, Saravanakumar

    Ceramic materials play a vital role in our day to day life. Recent advances in research, manufacture and processing techniques and production methodologies have broadened the scope of ceramic products such as bricks, pipes and tiles, especially in the construction industry. These are mainly manufactured using an extrusion process in auger extruders. During their long history of application in the ceramic industry, most of the design developments of extruder systems have resulted from expensive laboratory-based experimental work and field-based trial and error runs. In spite of these design developments, the auger extruders continue to be energy intensive devices with high operating costs. Limited understanding of the physical process involved in the process and the cost and time requirements of lab-based experiments were found to be the major obstacles in the further development of auger extruders.An attempt has been made herein to use Computational Fluid Dynamics (CFD) and Finite Element Analysis (FEA) based numerical modelling techniques to reduce the costs and time associated with research into design improvement by experimental trials. These two techniques, although used widely in other engineering applications, have rarely been applied for auger extruder development. This had been due to a number of reasons including technical limitations of CFD tools previously available. Modern CFD and FEA software packages have much enhanced capabilities and allow the modelling of the flow of complex fluids such as clay.This research work presents a methodology in using Herschel-Bulkley's fluid flow based CFD model to simulate and assess the flow of clay-water mixture through the extruder and the die of a vacuum de-airing type clay extrusion unit used in ceramic extrusion. The extruder design and the operating parameters were varied to study their influence on the power consumption and the extrusion pressure. The model results were then validated using results from

  7. Biological modelling of pelvic radiotherapy. Potential gains from conformal techniques

    Energy Technology Data Exchange (ETDEWEB)

    Fenwick, J.D

    1999-07-01

    Models have been developed which describe the dose and volume dependences of various long-term rectal complications of radiotherapy; assumptions underlying the models are consistent with clinical and experimental descriptions of complication pathogenesis. In particular, rectal bleeding - perhaps the most common complication of modern external beam prostate radiotherapy, and which might be viewed as its principle dose-limiting toxicity - has been modelled as a parallel-type complication. Rectal dose-surface-histograms have been calculated for 79 patients treated, in the course of the Royal Marsden trial of pelvic conformal radiotherapy, for prostate cancer using conformal or conventional techniques; rectal bleeding data is also available for these patients. The maximum- likelihood fit of the parallel bleeding model to the dose-surface-histograms and complication data shows that the complication status of the patients analysed (most of whom received reference point doses of 64 Gy) was significantly dependent on, and almost linearly proportional to, the volume of highly dosed rectal wall: a 1% decrease in the fraction of rectal wall (outlined over an 11 cm rectal length) receiving a dose of 58 Gy or more lead to a reduction in the (RTOG) grade 1,2,3 bleeding rate of about 1.1% - 95% confidence interval [0.04%, 2.2%]. The parallel model fit to the bleeding data is only marginally biased by uncertainties in the calculated dose-surface-histograms (due to setup errors, rectal wall movement and absolute rectal surface area variability), causing the gradient of the observed volume-response curve to be slightly lower than that which would be seen in the absence of these uncertainties. An analysis of published complication data supports these single-centre findings and indicates that the reductions in highly dosed rectal wall volumes obtainable using conformal radiotherapy techniques can be exploited to allow escalation of the dose delivered to the prostate target volume, the

  8. A numerical integration approach suitable for simulating PWR dynamics using a microcomputer system

    International Nuclear Information System (INIS)

    Zhiwei, L.; Kerlin, T.W.

    1983-01-01

    It is attractive to use microcomputer systems to simulate nuclear power plant dynamics for the purpose of teaching and/or control system design. An analysis and a comparison of feasibility of existing numerical integration methods have been made. The criteria for choosing the integration step using various numerical integration methods including the matrix exponential method are derived. In order to speed up the simulation, an approach is presented using the Newton recursion calculus which can avoid convergence limitations in choosing the integration step size. The accuracy consideration will dominate the integration step limited. The advantages of this method have been demonstrated through a case study using CBM model 8032 microcomputer to simulate a reduced order linear PWR model under various perturbations. It has been proven theoretically and practically that the Runge-Kutta method and Adams-Moulton method are not feasible. The matrix exponential method is good at accuracy and fairly good at speed. The Newton recursion method can save 3/4 to 4/5 time compared to the matrix exponential method with reasonable accuracy. Vertical Barhis method can be expanded to deal with nonlinear nuclear power plant models and higher order models as well

  9. Techniques to Access Databases and Integrate Data for Hydrologic Modeling

    International Nuclear Information System (INIS)

    Whelan, Gene; Tenney, Nathan D.; Pelton, Mitchell A.; Coleman, Andre M.; Ward, Duane L.; Droppo, James G.; Meyer, Philip D.; Dorow, Kevin E.; Taira, Randal Y.

    2009-01-01

    This document addresses techniques to access and integrate data for defining site-specific conditions and behaviors associated with ground-water and surface-water radionuclide transport applicable to U.S. Nuclear Regulatory Commission reviews. Environmental models typically require input data from multiple internal and external sources that may include, but are not limited to, stream and rainfall gage data, meteorological data, hydrogeological data, habitat data, and biological data. These data may be retrieved from a variety of organizations (e.g., federal, state, and regional) and source types (e.g., HTTP, FTP, and databases). Available data sources relevant to hydrologic analyses for reactor licensing are identified and reviewed. The data sources described can be useful to define model inputs and parameters, including site features (e.g., watershed boundaries, stream locations, reservoirs, site topography), site properties (e.g., surface conditions, subsurface hydraulic properties, water quality), and site boundary conditions, input forcings, and extreme events (e.g., stream discharge, lake levels, precipitation, recharge, flood and drought characteristics). Available software tools for accessing established databases, retrieving the data, and integrating it with models were identified and reviewed. The emphasis in this review was on existing software products with minimal required modifications to enable their use with the FRAMES modeling framework. The ability of four of these tools to access and retrieve the identified data sources was reviewed. These four software tools were the Hydrologic Data Acquisition and Processing System (HDAPS), Integrated Water Resources Modeling System (IWRMS) External Data Harvester, Data for Environmental Modeling Environmental Data Download Tool (D4EM EDDT), and the FRAMES Internet Database Tools. The IWRMS External Data Harvester and the D4EM EDDT were identified as the most promising tools based on their ability to access and

  10. Techniques to Access Databases and Integrate Data for Hydrologic Modeling

    Energy Technology Data Exchange (ETDEWEB)

    Whelan, Gene; Tenney, Nathan D.; Pelton, Mitchell A.; Coleman, Andre M.; Ward, Duane L.; Droppo, James G.; Meyer, Philip D.; Dorow, Kevin E.; Taira, Randal Y.

    2009-06-17

    This document addresses techniques to access and integrate data for defining site-specific conditions and behaviors associated with ground-water and surface-water radionuclide transport applicable to U.S. Nuclear Regulatory Commission reviews. Environmental models typically require input data from multiple internal and external sources that may include, but are not limited to, stream and rainfall gage data, meteorological data, hydrogeological data, habitat data, and biological data. These data may be retrieved from a variety of organizations (e.g., federal, state, and regional) and source types (e.g., HTTP, FTP, and databases). Available data sources relevant to hydrologic analyses for reactor licensing are identified and reviewed. The data sources described can be useful to define model inputs and parameters, including site features (e.g., watershed boundaries, stream locations, reservoirs, site topography), site properties (e.g., surface conditions, subsurface hydraulic properties, water quality), and site boundary conditions, input forcings, and extreme events (e.g., stream discharge, lake levels, precipitation, recharge, flood and drought characteristics). Available software tools for accessing established databases, retrieving the data, and integrating it with models were identified and reviewed. The emphasis in this review was on existing software products with minimal required modifications to enable their use with the FRAMES modeling framework. The ability of four of these tools to access and retrieve the identified data sources was reviewed. These four software tools were the Hydrologic Data Acquisition and Processing System (HDAPS), Integrated Water Resources Modeling System (IWRMS) External Data Harvester, Data for Environmental Modeling Environmental Data Download Tool (D4EM EDDT), and the FRAMES Internet Database Tools. The IWRMS External Data Harvester and the D4EM EDDT were identified as the most promising tools based on their ability to access and

  11. Modeling and Forecasting Electricity Demand in Azerbaijan Using Cointegration Techniques

    Directory of Open Access Journals (Sweden)

    Fakhri J. Hasanov

    2016-12-01

    Full Text Available Policymakers in developing and transitional economies require sound models to: (i understand the drivers of rapidly growing energy consumption and (ii produce forecasts of future energy demand. This paper attempts to model electricity demand in Azerbaijan and provide future forecast scenarios—as far as we are aware this is the first such attempt for Azerbaijan using a comprehensive modelling framework. Electricity consumption increased and decreased considerably in Azerbaijan from 1995 to 2013 (the period used for the empirical analysis—it increased on average by about 4% per annum from 1995 to 2006 but decreased by about 4½% per annum from 2006 to 2010 and increased thereafter. It is therefore vital that Azerbaijani planners and policymakers understand what drives electricity demand and be able to forecast how it will grow in order to plan for future power production. However, modeling electricity demand for such a country has many challenges. Azerbaijan is rich in energy resources, consequently GDP is heavily influenced by oil prices; hence, real non-oil GDP is employed as the activity driver in this research (unlike almost all previous aggregate energy demand studies. Moreover, electricity prices are administered rather than market driven. Therefore, different cointegration and error correction techniques are employed to estimate a number of per capita electricity demand models for Azerbaijan, which are used to produce forecast scenarios for up to 2025. The resulting estimated models (in terms of coefficients, etc. and forecasts of electricity demand for Azerbaijan in 2025 prove to be very similar; with the Business as Usual forecast ranging from about of 19½ to 21 TWh.

  12. Automated Quantitative Bone Analysis in In Vivo X-ray Micro-Computed Tomography.

    Science.gov (United States)

    Behrooz, Ali; Kask, Peet; Meganck, Jeff; Kempner, Joshua

    2017-09-01

    Measurement and analysis of bone morphometry in 3D micro-computed tomography volumes using automated image processing and analysis improve the accuracy, consistency, reproducibility, and speed of preclinical osteological research studies. Automating segmentation and separation of individual bones in 3D micro-computed tomography volumes of murine models presents significant challenges considering partial volume effects and joints with thin spacing, i.e., 50 to [Formula: see text]. In this paper, novel hybrid splitting filters are presented to overcome the challenge of automated bone separation. This is achieved by enhancing joint contrast using rotationally invariant second-derivative operators. These filters generate split components that seed marker-controlled watershed segmentation. In addition, these filters can be used to separate metaphysis and epiphysis in long bones, e.g., femur, and remove the metaphyseal growth plate from the detected bone mask in morphometric measurements. Moreover, for slice-by-slice stereological measurements of long bones, particularly curved bones, such as tibia, the accuracy of the analysis can be improved if the planar measurements are guided to follow the longitudinal direction of the bone. In this paper, an approach is presented for characterizing the bone medial axis using morphological thinning and centerline operations. Building upon the medial axis, a novel framework is presented to automatically guide stereological measurements of long bones and enhance measurement accuracy and consistency. These image processing and analysis approaches are combined in an automated streamlined software workflow and applied to a range of in vivo micro-computed tomography studies for validation.

  13. Model assessment using a multi-metric ranking technique

    Science.gov (United States)

    Fitzpatrick, P. J.; Lau, Y.; Alaka, G.; Marks, F.

    2017-12-01

    Validation comparisons of multiple models presents challenges when skill levels are similar, especially in regimes dominated by the climatological mean. Assessing skill separation will require advanced validation metrics and identifying adeptness in extreme events, but maintain simplicity for management decisions. Flexibility for operations is also an asset. This work postulates a weighted tally and consolidation technique which ranks results by multiple types of metrics. Variables include absolute error, bias, acceptable absolute error percentages, outlier metrics, model efficiency, Pearson correlation, Kendall's Tau, reliability Index, multiplicative gross error, and root mean squared differences. Other metrics, such as root mean square difference and rank correlation were also explored, but removed when the information was discovered to be generally duplicative to other metrics. While equal weights are applied, weights could be altered depending for preferred metrics. Two examples are shown comparing ocean models' currents and tropical cyclone products, including experimental products. The importance of using magnitude and direction for tropical cyclone track forecasts instead of distance, along-track, and cross-track are discussed. Tropical cyclone intensity and structure prediction are also assessed. Vector correlations are not included in the ranking process, but found useful in an independent context, and will be briefly reported.

  14. Total laparoscopic gastrocystoplasty: experimental technique in a porcine model

    Directory of Open Access Journals (Sweden)

    Frederico R. Romero

    2007-02-01

    Full Text Available OBJECTIVE: Describe a unique simplified experimental technique for total laparoscopic gastrocystoplasty in a porcine model. MATERIAL AND METHODS: We performed laparoscopic gastrocystoplasty on 10 animals. The gastroepiploic arch was identified and carefully mobilized from its origin at the pylorus to the beginning of the previously demarcated gastric wedge. The gastric segment was resected with sharp dissection. Both gastric suturing and gastrovesical anastomosis were performed with absorbable running sutures. The complete procedure and stages of gastric dissection, gastric closure, and gastrovesical anastomosis were separately timed for each laparoscopic gastrocystoplasty. The end-result of the gastric suturing and the bladder augmentation were evaluated by fluoroscopy or endoscopy. RESULTS: Mean total operative time was 5.2 (range 3.5 - 8 hours: 84.5 (range 62 - 110 minutes for the gastric dissection, 56 (range 28 - 80 minutes for the gastric suturing, and 170.6 (range 70 to 200 minutes for the gastrovesical anastomosis. A cystogram showed a small leakage from the vesical anastomosis in the first two cases. No extravasation from gastric closure was observed in the postoperative gastrogram. CONCLUSIONS: Total laparoscopic gastrocystoplasty is a feasible but complex procedure that currently has limited clinical application. With the increasing use of laparoscopy in reconstructive surgery of the lower urinary tract, gastrocystoplasty may become an attractive option because of its potential advantages over techniques using small and large bowel segments.

  15. Mapping the Complexities of Online Dialogue: An Analytical Modeling Technique

    Directory of Open Access Journals (Sweden)

    Robert Newell

    2014-03-01

    Full Text Available The e-Dialogue platform was developed in 2001 to explore the potential of using the Internet for engaging diverse groups of people and multiple perspectives in substantive dialogue on sustainability. The system is online, text-based, and serves as a transdisciplinary space for bringing together researchers, practitioners, policy-makers and community leaders. The Newell-Dale Conversation Modeling Technique (NDCMT was designed for in-depth analysis of e-Dialogue conversations and uses empirical methodology to minimize observer bias during analysis of a conversation transcript. NDCMT elucidates emergent ideas, identifies connections between ideas and themes, and provides a coherent synthesis and deeper understanding of the underlying patterns of online conversations. Continual application and improvement of NDCMT can lead to powerful methodologies for empirically analyzing digital discourse and better capture of innovations produced through such discourse. URN: http://nbn-resolving.de/urn:nbn:de:0114-fqs140221

  16. Vector machine techniques for modeling of seismic liquefaction data

    Directory of Open Access Journals (Sweden)

    Pijush Samui

    2014-06-01

    Full Text Available This article employs three soft computing techniques, Support Vector Machine (SVM; Least Square Support Vector Machine (LSSVM and Relevance Vector Machine (RVM, for prediction of liquefaction susceptibility of soil. SVM and LSSVM are based on the structural risk minimization (SRM principle which seeks to minimize an upper bound of the generalization error consisting of the sum of the training error and a confidence interval. RVM is a sparse Bayesian kernel machine. SVM, LSSVM and RVM have been used as classification tools. The developed SVM, LSSVM and RVM give equations for prediction of liquefaction susceptibility of soil. A comparative study has been carried out between the developed SVM, LSSVM and RVM models. The results from this article indicate that the developed SVM gives the best performance for prediction of liquefaction susceptibility of soil.

  17. Demand Management Based on Model Predictive Control Techniques

    Directory of Open Access Journals (Sweden)

    Yasser A. Davizón

    2014-01-01

    Full Text Available Demand management (DM is the process that helps companies to sell the right product to the right customer, at the right time, and for the right price. Therefore the challenge for any company is to determine how much to sell, at what price, and to which market segment while maximizing its profits. DM also helps managers efficiently allocate undifferentiated units of capacity to the available demand with the goal of maximizing revenue. This paper introduces control system approach to demand management with dynamic pricing (DP using the model predictive control (MPC technique. In addition, we present a proper dynamical system analogy based on active suspension and a stability analysis is provided via the Lyapunov direct method.

  18. Enhancing photogrammetric 3d city models with procedural modeling techniques for urban planning support

    International Nuclear Information System (INIS)

    Schubiger-Banz, S; Arisona, S M; Zhong, C

    2014-01-01

    This paper presents a workflow to increase the level of detail of reality-based 3D urban models. It combines the established workflows from photogrammetry and procedural modeling in order to exploit distinct advantages of both approaches. The combination has advantages over purely automatic acquisition in terms of visual quality, accuracy and model semantics. Compared to manual modeling, procedural techniques can be much more time effective while maintaining the qualitative properties of the modeled environment. In addition, our method includes processes for procedurally adding additional features such as road and rail networks. The resulting models meet the increasing needs in urban environments for planning, inventory, and analysis

  19. Desk-top microcomputer for lab-scale process control

    International Nuclear Information System (INIS)

    Overman, R.F.; Byrd, J.S.; Goosey, M.H.; Sand, R.J.

    1981-01-01

    A desk-top microcomputer was programmed to acquire the data from various process control sensors installed in a laboratory scale liquid-liquid extraction, pulse column facility. The parameters monitored included valve positions, gamma spectra, alpha radioactivity, temperature, pH, density, and flow rates. The program for the microcomputer is written in BASIC and requires about 31000 8-bit bytes of memory. All data is stored on floppy discs, and can be displayed or printed. Unexpected data values are brought to the process operator's attention via CRT display or print-out. The general organization of the program and a few subroutines unique to polling instruments are explained. Some of the data acquisition devices were designed and built at the Savannah River Laboratory. These include a pulse height analyzer, a data multiplexer, and a data acquisition instrument. A general description of the electronics design of these instruments is also given with emphasis placed on data formatting and bus addressing

  20. A Method for Evaluation of Microcomputers for Tactical Applications.

    Science.gov (United States)

    1980-06-01

    to microcomputer systems. f. Ada Programming Language Ada , which is named after the world’s first com- puter programmer, Lady Ada Lovelace , is a very...HOL) Ada used as a program development tool for real-time processing including microprocessor applications. (2) Hosted on small development systems. 4...mainstream. The Defense Department’s Ada HOL is being developed to serve programming needs for real-time processing including 14 microprocessor

  1. A MICROCOMPUTER LINEAR PROGRAMMING PACKAGE: AN ALTERNATIVE TO MAINFRAMES

    OpenAIRE

    Laughlin, David H.

    1984-01-01

    This paper presents the capabilities and limitations of a microcomputer linear programming package. The solution algorithm is a version of the revised simplex. Rapid problem entry, user ease of operation, sensitivity analyses on objective function and right hand sides are advantages. A problem size of 150 activities and 64 constraints can be solved in present form. Due to problem size, limitations and lack of parametric and integer programming routines, this package is thought to have the mos...

  2. MultiSimPC: a multilevel logic simulator for microcomputers

    OpenAIRE

    Kelly, John S.

    1986-01-01

    Approved for public release; distribution is unlimited This thesis describes extensions to a multilevel VLSI logic simulator named MultiSim. Originally developed by Dr. Ausif Mahmood of the Washington State University for large minicomputers such as the VAX-11/780; MultiSim is now operational on desktop microcomputers costing only a few thousand dollars. In addition, MultiSim has been expanded to include provisions for adding user-defined primitive cells to the circuit library, true mu...

  3. Simulation of Heat Transfer and Electromagnetic Fields of Protected Microcomputers

    Directory of Open Access Journals (Sweden)

    Josef Lakatos

    2006-01-01

    Full Text Available The paper presents results of collaboration between Department of mechatronics and electronics at University of Žilina and VÚVT Engineering a.s. Žilina in area of heat transfer simulations and disturbing electromagnetic radiation simulations in computer construction. The simulations results were used in development of protected microcomputer prototypes in frame of applied research at both of workplaces.

  4. The Microcomputer Knowledge Base: Introduction and User Instructions.

    Science.gov (United States)

    1984-01-01

    for storing existing facts. In understanding how the knowledge access system actually works, it is often helpful to think of it as a "tree-type...is saved. Example System Response abort Cancels the comment entered, exits the comment editor, and returns the user to the Knowledge Access System...Ai37 694 THE MICROCOMPUTER KNOWLEDGE BASE. INTRODUCTION AND USER 1/1 INSTRUCTIONS(U) CONSTRUCTION ENGINEERING RESEARCH LAB (ARMY) CHAMPAIGN IL F

  5. Biocompatibility property of 100% strontium-substituted SiO2 -Al2 O3 -P2 O5 -CaO-CaF2 glass ceramics over 26 weeks implantation in rabbit model: Histology and micro-Computed Tomography analysis.

    Science.gov (United States)

    Basu, Bikramjit; Sabareeswaran, A; Shenoy, S J

    2015-08-01

    One of the desired properties for any new biomaterial composition is its long-term stability in a suitable animal model and such property cannot be appropriately assessed by performing short-term implantation studies. While hydroxyapatite (HA) or bioglass coated metallic biomaterials are being investigated for in vivo biocompatibility properties, such study is not extensively being pursued for bulk glass ceramics. In view of their inherent brittle nature, the implant stability as well as impact of long-term release of metallic ions on bone regeneration have been a major concern. In this perspective, the present article reports the results of the in vivo implantation experiments carried out using 100% strontium (Sr)-substituted glass ceramics with the nominal composition of 4.5 SiO2 -3Al2 O3 -1.5P2 O5 -3SrO-2SrF2 for 26 weeks in cylindrical bone defects in rabbit model. The combination of histological and micro-computed tomography analysis provided a qualitative and quantitative understanding of the bone regeneration around the glass ceramic implants in comparison to the highly bioactive HA bioglass implants (control). The sequential polychrome labeling of bone during in vivo osseointegration using three fluorochromes followed by fluorescence microscopy observation confirmed homogeneous bone formation around the test implants. The results of the present study unequivocally confirm the long-term implant stability as well as osteoconductive property of 100% Sr-substituted glass ceramics, which is comparable to that of a known bioactive implant, that is, HA-based bioglass. © 2014 Wiley Periodicals, Inc.

  6. The phase field technique for modeling multiphase materials

    Science.gov (United States)

    Singer-Loginova, I.; Singer, H. M.

    2008-10-01

    This paper reviews methods and applications of the phase field technique, one of the fastest growing areas in computational materials science. The phase field method is used as a theory and computational tool for predictions of the evolution of arbitrarily shaped morphologies and complex microstructures in materials. In this method, the interface between two phases (e.g. solid and liquid) is treated as a region of finite width having a gradual variation of different physical quantities, i.e. it is a diffuse interface model. An auxiliary variable, the phase field or order parameter \\phi(\\vec{x}) , is introduced, which distinguishes one phase from the other. Interfaces are identified by the variation of the phase field. We begin with presenting the physical background of the phase field method and give a detailed thermodynamical derivation of the phase field equations. We demonstrate how equilibrium and non-equilibrium physical phenomena at the phase interface are incorporated into the phase field methods. Then we address in detail dendritic and directional solidification of pure and multicomponent alloys, effects of natural convection and forced flow, grain growth, nucleation, solid-solid phase transformation and highlight other applications of the phase field methods. In particular, we review the novel phase field crystal model, which combines atomistic length scales with diffusive time scales. We also discuss aspects of quantitative phase field modeling such as thin interface asymptotic analysis and coupling to thermodynamic databases. The phase field methods result in a set of partial differential equations, whose solutions require time-consuming large-scale computations and often limit the applicability of the method. Subsequently, we review numerical approaches to solve the phase field equations and present a finite difference discretization of the anisotropic Laplacian operator.

  7. The microcomputer: A tool for personal language learning

    Directory of Open Access Journals (Sweden)

    David H. Wyatt

    2013-02-01

    Full Text Available Computer-assisted methods of teaching and learning languages have been surrounded by controversy and debate for over a decade. In 1979, however, microcomputers began to appear in a form suitable for educational applications, offering for the first time an alternative to both the cost and the approach of large computer systems. The impact of the microcomputer has been limited by a number of factors, and microcomputerassisted learning is still in a relative state of infancy. The main implications for language teaching and learning are only now beginning to be understood, just as the limiting factors are starting to disappear. This paper will assess the present situation and outline some likely future developments in the use of microcomputers in language learning. Rekenaargesteunde metodes by die onderrig en aanleer van tale is reeds meer as 'n dekade lank omgewe deur meningsverskil. In 1979 egter het mikrorekenaars hulle verskyning begin maak in 'n toepaslike vorm vir opvoedkundige doeleindes. Vir die eerste keer was daar 'n alternatief vir die koste verbonde aan en die benaderingswyse van groot rekenaarstelsels. Die trefkrag van die mikrorekenaar is deur 'n aantal faktore gekortwiek en rekenaargesteunde onderrig is steeds in sy kinderskoene. Die belangrikste implikasies vir die aanleer en onderrig van tale begin nou eers deurdring soos die beperkende faktore begin verdwyn. Hierdie artikel takseer die huidige situasie en omlyn moontlike toekomstige ontwikkelings vir die gebruik van mikrorekenaars by taalaanleer.

  8. Graphics of (X,Y) spectrum for microcomputer

    International Nuclear Information System (INIS)

    Macias B, L.R.

    1991-08-01

    When carrying out diffraction works is frequently required to visualize the spectra of the data obtained in order to analyzing them. The design for the obtaining of data in the neutron diffractometer by means of the microcomputer allows to store them in a file by means of the one which transferring to the CYBER system so that by means of its utilities the mentioned spectrum is observed in a graph. In diffraction works, it is sought to study crystalline materials by means of the execution of the Bragg law by that the mounted sample on the diffractometer is subjected to a scanning of the sample with a radiation of a well-known wavelength and this way varying the angles, the corresponding interplanar distances are determined. The main objective of this work, is starting of a data set generated by the diffractometer, to generate the graph of the corresponding (X,Y) spectra in visual form in the screen of a microcomputer and if it is required, to obtain the graph in printed form by means of the same computer program for microcomputer. (Author)

  9. Establishment of reproducible osteosarcoma rat model using orthotopic implantation technique.

    Science.gov (United States)

    Yu, Zhe; Sun, Honghui; Fan, Qingyu; Long, Hua; Yang, Tongtao; Ma, Bao'an

    2009-05-01

    In experimental musculoskeletal oncology, there remains a need for animal models that can be used to assess the efficacy of new and innovative treatment methodologies for bone tumors. Rat plays a very important role in the bone field especially in the evaluation of metabolic bone diseases. The objective of this study was to develop a rat osteosarcoma model for evaluation of new surgical and molecular methods of treatment for extremity sarcoma. One hundred male SD rats weighing 125.45+/-8.19 g were divided into 5 groups and anesthetized intraperitoneally with 10% chloral hydrate. Orthotopic implantation models of rat osteosarcoma were performed by injecting directly into the SD rat femur with a needle for inoculation with SD tumor cells. In the first step of the experiment, 2x10(5) to 1x10(6) UMR106 cells in 50 microl were injected intraosseously into median or distal part of the femoral shaft and the tumor take rate was determined. The second stage consisted of determining tumor volume, correlating findings from ultrasound with findings from necropsia and determining time of survival. In the third stage, the orthotopically implanted tumors and lung nodules were resected entirely, sectioned, and then counter stained with hematoxylin and eosin for histopathologic evaluation. The tumor take rate was 100% for implants with 8x10(5) tumor cells or more, which was much less than the amount required for subcutaneous implantation, with a high lung metastasis rate of 93.0%. Ultrasound and necropsia findings matched closely (r=0.942; p<0.01), which demonstrated that Doppler ultrasonography is a convenient and reliable technique for measuring cancer at any stage. Tumor growth curve showed that orthotopically implanted tumors expanded vigorously with time-lapse, especially in the first 3 weeks. The median time of survival was 38 days and surgical mortality was 0%. The UMR106 cell line has strong carcinogenic capability and high lung metastasis frequency. The present rat

  10. Software development assurance for special control computer sets on the base of the Electronica-60M microcomputer in the stations of fuel element nondestructive testing

    International Nuclear Information System (INIS)

    Vasil'kova, A.D.; Grachev, S.N.; Salakatova, L.S.

    1987-01-01

    A special computational control complex (SCCC) including the ''Elektronika-60M'' microcomputer, a device for communication with an object (DCO) based on the typical microprocessing set (TMS) based on LIUS-2 KTS means and an adapter for communication of the ''Elektronika-60M'' microcomputer bus with IK1 TMS bus developed for their conjugation is used for development of nondestructive control post software. An instrumental SCCC including SM-4 microcomputer, TMS units as DCO, an adapter for communication of the common bus of the SM-4 minicomputer with the IK1 TMS bus developed for their conjugation and devices for simulation of the facility operation is suggested to increase labour productivity of programmers. The SM-4 microcomputer permits to develop programs in the FORTRAN language and to carry out their effective adjustment. A subprogram library for communication with TMS units is compiled, and the technique for testing FORTRAN programs in programmable read-only memory TMS as well as start of these programs are developed

  11. Derandomizing buffer and microcomputer memory for the use of a fast MWPC image memory

    International Nuclear Information System (INIS)

    Skvaril, J.

    1986-01-01

    We have developed a special derandomizing buffer memory for MWPC imaging which makes it possible to use a microcomputer memory in DMA mode. The buffer regulates an input data stream (X and Y coordinates of a event) into a microcomputer memory with practically no data losses. The advantages of this approach are (a) no special histogramming memory is needed and (b) a resultant image is in the memory space of the microcomputer used and can be immediately processed. (orig.)

  12. Verification Techniques for Parameter Selection and Bayesian Model Calibration Presented for an HIV Model

    Science.gov (United States)

    Wentworth, Mami Tonoe

    Uncertainty quantification plays an important role when making predictive estimates of model responses. In this context, uncertainty quantification is defined as quantifying and reducing uncertainties, and the objective is to quantify uncertainties in parameter, model and measurements, and propagate the uncertainties through the model, so that one can make a predictive estimate with quantified uncertainties. Two of the aspects of uncertainty quantification that must be performed prior to propagating uncertainties are model calibration and parameter selection. There are several efficient techniques for these processes; however, the accuracy of these methods are often not verified. This is the motivation for our work, and in this dissertation, we present and illustrate verification frameworks for model calibration and parameter selection in the context of biological and physical models. First, HIV models, developed and improved by [2, 3, 8], describe the viral infection dynamics of an HIV disease. These are also used to make predictive estimates of viral loads and T-cell counts and to construct an optimal control for drug therapy. Estimating input parameters is an essential step prior to uncertainty quantification. However, not all the parameters are identifiable, implying that they cannot be uniquely determined by the observations. These unidentifiable parameters can be partially removed by performing parameter selection, a process in which parameters that have minimal impacts on the model response are determined. We provide verification techniques for Bayesian model calibration and parameter selection for an HIV model. As an example of a physical model, we employ a heat model with experimental measurements presented in [10]. A steady-state heat model represents a prototypical behavior for heat conduction and diffusion process involved in a thermal-hydraulic model, which is a part of nuclear reactor models. We employ this simple heat model to illustrate verification

  13. Ex Vivo Methods for Informing Computational Models of the Mitral Valve

    OpenAIRE

    Bloodworth, Charles H.; Pierce, Eric L.; Easley, Thomas F.; Drach, Andrew; Khalighi, Amir H.; Toma, Milan; Jensen, Morten O.; Sacks, Michael S.; Yoganathan, Ajit P.

    2016-01-01

    Computational modeling of the mitral valve (MV) has potential applications for determining optimal MV repair techniques and risk of recurrent mitral regurgitation. Two key concerns for informing these models are (1) sensitivity of model performance to the accuracy of the input geometry, and, (2) acquisition of comprehensive data sets against which the simulation can be validated across clinically relevant geometries. Addressing the first concern, ex vivo micro-computed tomography (microCT) wa...

  14. Radiation treatment planning using a microcomputer

    International Nuclear Information System (INIS)

    Lunsqui, A.R.; Calil, S.J.; Rocha, J.R.O.; Alexandre, A.C.

    1990-01-01

    The radiation treatment planning requires a lenght manipulation of data from isodose charts to obtain the best irradiation technique. Over the past 25 years this tedious operation has been replaced by computerized methods. These can reduce the working time by at least 20 times. It is being developed at the Biomedical Engineering Center a software to generate a polychromatic image of dose distribution. By means of a digitizing board, the patient contour and the beam data are transfered to the computer and stored as polinomial and Fourier series respectively. To calculate the dose distribution, the irradiated region is represented by a variable size bidimensional dot matrix. The dose at each point is calculated by correcting and adding the stored data for each beam. An algorithm for color definition according to the dose intensity was developed to display on a computer monitor the resultant matrix. A hard copy can be obtained be means of a six color plotter. (author)

  15. A Critical Review of Model-Based Economic Studies of Depression: Modelling Techniques, Model Structure and Data Sources

    OpenAIRE

    Hossein Haji Ali Afzali; Jonathan Karnon; Jodi Gray

    2012-01-01

    Depression is the most common mental health disorder and is recognized as a chronic disease characterized by multiple acute episodes/relapses. Although modelling techniques play an increasingly important role in the economic evaluation of depression interventions, comparatively little attention has been paid to issues around modelling studies with a focus on potential biases. This, however, is important as different modelling approaches, variations in model structure and input parameters may ...

  16. A novel approach for studying the temporal modulation of embryonic skeletal development using organotypic bone cultures and microcomputed tomography.

    Science.gov (United States)

    Kanczler, Janos M; Smith, Emma L; Roberts, Carol A; Oreffo, Richard O C

    2012-10-01

    Understanding the structural development of embryonic bone in a three dimensional framework is fundamental to developing new strategies for the recapitulation of bone tissue in latter life. We present an innovative combined approach of an organotypic embryonic femur culture model, microcomputed tomography (μCT) and immunohistochemistry to examine the development and modulation of the three dimensional structures of the developing embryonic femur. Isolated embryonic chick femurs were organotypic (air/liquid interface) cultured for 10 days in either basal, chondrogenic, or osteogenic supplemented culture conditions. The growth development and modulating effects of basal, chondrogenic, or osteogenic culture media of the embryonic chick femurs was investigated using μCT, immunohistochemistry, and histology. The growth and development of noncultured embryonic chick femur stages E10, E11, E12, E13, E15, and E17 were very closely correlated with increased morphometric indices of bone formation as determined by μCT. After 10 days in the organotpyic culture set up, the early aged femurs (E10 and E11) demonstrated a dramatic response to the chondrogenic or osteogenic culture conditions compared to the basal cultured femurs as determined by a change in μCT morphometric indices and modified expression of chondrogenic and osteogenic markers. Although the later aged femurs (E12 and E13) increased in size and structure after 10 days organotpypic culture, the effects of the osteogenic and chondrogenic organotypic cultures on these femurs were not significantly altered compared to basal conditions. We have demonstrated that the embryonic chick femur organotpyic culture model combined with the μCT and immunohistochemical analysis can provide an integral methodology for investigating the modulation of bone development in an ex vivo culture setting. Hence, these interdisciplinary techniques of μCT and whole organ bone cultures will enable us to delineate some of the temporal

  17. Adaptive Atmospheric Modeling Key Techniques in Grid Generation, Data Structures, and Numerical Operations with Applications

    CERN Document Server

    Behrens, Jörn

    2006-01-01

    Gives an overview and guidance in the development of adaptive techniques for atmospheric modeling. This book covers paradigms of adaptive techniques, such as error estimation and adaptation criteria. Considering applications, it demonstrates several techniques for discretizing relevant conservation laws from atmospheric modeling.

  18. Frequency Weighted Model Order Reduction Technique and Error Bounds for Discrete Time Systems

    Directory of Open Access Journals (Sweden)

    Muhammad Imran

    2014-01-01

    for whole frequency range. However, certain applications (like controller reduction require frequency weighted approximation, which introduce the concept of using frequency weights in model reduction techniques. Limitations of some existing frequency weighted model reduction techniques include lack of stability of reduced order models (for two sided weighting case and frequency response error bounds. A new frequency weighted technique for balanced model reduction for discrete time systems is proposed. The proposed technique guarantees stable reduced order models even for the case when two sided weightings are present. Efficient technique for frequency weighted Gramians is also proposed. Results are compared with other existing frequency weighted model reduction techniques for discrete time systems. Moreover, the proposed technique yields frequency response error bounds.

  19. Constitutional Model and Rationality in Judicial Decisions from Proportionality Technique

    OpenAIRE

    Feio, Thiago Alves

    2016-01-01

    In the current legal systems, the content of the Constitutions consists of values that serve to limit state action. The department in charge of the control of this system is, usually, the Judiciary. This choice leads to two major problems, the tension between democracy and constitutionalism and the subjectivity that control. One of the solutions to subjectivity is weighting of principles through the proportionality technique, which aims to give rational decisions. This technique doesn’t elimi...

  20. Experiments with microcomputer-based artificial intelligence environments

    Science.gov (United States)

    Summers, E.G.; MacDonald, R.A.

    1988-01-01

    The U.S. Geological Survey (USGS) has been experimenting with the use of relatively inexpensive microcomputers as artificial intelligence (AI) development environments. Several AI languages are available that perform fairly well on desk-top personal computers, as are low-to-medium cost expert system packages. Although performance of these systems is respectable, their speed and capacity limitations are questionable for serious earth science applications foreseen by the USGS. The most capable artificial intelligence applications currently are concentrated on what is known as the "artificial intelligence computer," and include Xerox D-series, Tektronix 4400 series, Symbolics 3600, VAX, LMI, and Texas Instruments Explorer. The artificial intelligence computer runs expert system shells and Lisp, Prolog, and Smalltalk programming languages. However, these AI environments are expensive. Recently, inexpensive 32-bit hardware has become available for the IBM/AT microcomputer. USGS has acquired and recently completed Beta-testing of the Gold Hill Systems 80386 Hummingboard, which runs Common Lisp on an IBM/AT microcomputer. Hummingboard appears to have the potential to overcome many of the speed/capacity limitations observed with AI-applications on standard personal computers. USGS is a Beta-test site for the Gold Hill Systems GoldWorks expert system. GoldWorks combines some high-end expert system shell capabilities in a medium-cost package. This shell is developed in Common Lisp, runs on the 80386 Hummingboard, and provides some expert system features formerly available only on AI-computers including frame and rule-based reasoning, on-line tutorial, multiple inheritance, and object-programming. ?? 1988 International Association for Mathematical Geology.

  1. Pedicle screw "hubbing" in the immature thoracic spine: a biomechanical and micro-computed tomography evaluation.

    Science.gov (United States)

    Kang, Daniel G; Lehman, Ronald A; Bevevino, Adam J; Gaume, Rachel E; Purcell, Richard L; Dmitriev, Anton E; Lenke, Lawrence G

    2014-01-01

    A previous biomechanical study using adult thoracic vertebrae (both normal and osteoporotic bone density) demonstrated the deleterious effect of the pedicle screw hubbing technique. Pedicle screw "hubbing" involves seating and engaging the ventral aspect of the screw head onto the dorsal lamina cortex. This technique is postulated to provide a load-sharing effect by improving pullout resistance, as well as decreasing cephalocaudad toggling and implant loosening. We hypothesized the elastic properties of immature bone may mitigate, and perhaps enhance the purported benefits of the hubbing technique. We set out to evaluate pullout strength of fixed-head pedicle screws after hubbing versus standard insertion in the immature thoracic calf spine. Twenty-two (n=22) single-level disarticulated fresh-frozen immature calf thoracic vertebra specimens (ranging from T2 to T13) were prepared. Twelve specimens were instrumented with pedicle screws in group I (nonhubbed) and group II (hubbed) in the opposite pedicle. Cyclic loading in a cephalocaudad direction was applied for 2000 cycles at a rate of 1 Hz. Pullout testing was performed in-line with the midline of the vertebra and peak pullout strength was measured in Newtons. Ten different specimens underwent micro-computed tomography evaluation to assess for trabecular architecture and incidence of iatrogenic microfractures. Hubbed screws resulted in significantly lower pullout strength (747±197 vs. 922±112 N, P=0.01). With the hubbing technique, the dorsal cortex demonstrated plastic deformation and conformed to the screw head in 83% of cases compared with no visible plastic deformation in the control group. Micro-computed tomography demonstrated microfractures of the dorsal cortex in 10/10 for the hubbed group compared with 1/10 for the control group. This is the largest study ever performed on immature thoracic vertebra to evaluate this topic. Hubbed pedicle screws have significantly decreased pullout strength and

  2. Sand and gravel mine operations and reclamation planning using microcomputers

    Energy Technology Data Exchange (ETDEWEB)

    Ariffin, J.B.

    1990-02-01

    The purpose of this study is to focus on the application of microcomputers, also known as personal computers, in planning for sand and gravel mine operations and reclamation at a site in Story County, Iowa. This site, called the Arrasmith Pit, is operated by Martin Marietta Aggregates, Inc. The Arrasmith site, which encompasses an area of about 25 acres, is a relatively small site for aggregate mining. However, planning for the concurrent mine operation and reclamation program at this site is just as critical as with larger sites and the planning process is the same.

  3. Microcomputer simulation of nuclear magnetic resonance imaging contrasts

    International Nuclear Information System (INIS)

    Le Bihan, D.

    1985-01-01

    The high information content of magnetic resonance images is due to the multiplicity of its parameters. However, this advantage introduces a difficulty in the interpretation of the contrast: an image is strongly modified according to the visualised parameters. The author proposes a micro-computer simulation program. After recalling the main intrinsic and extrinsic parameters, he shows how the program works and its interest as a pedagogic tool and as an aid for contrast optimisation of images as a function of the suspected pathology [fr

  4. Reactor core protection system using a 4-channel microcomputer

    International Nuclear Information System (INIS)

    Mertens, U.

    1982-12-01

    A four channel microcomputer system was fitted in Grafenrheinfeld NPP for local core protection. This system performs continuous on-line monitoring of peak power density, departure from nucleate boiling ratio and fuel duty. The system implements limitation functions with more sophisticated criteria and improved accuracy. The Grafenrheinfeld system points the way to the employment of computer based limitation system, particularly in the field of programming language, demarkation of tasks, commissioning and documentation aids, streamlining of qualification and structuring of the system. (orig.) [de

  5. Program pseudo-random number generator for microcomputers

    International Nuclear Information System (INIS)

    Ososkov, G.A.

    1980-01-01

    Program pseudo-random number generators (PNG) intended for the test of control equipment and communication channels are considered. In the case of 8-bit microcomputers it is necessary to assign 4 words of storage to allocate one random number. The proposed economical algorithms of the random number generation are based on the idea of the ''mixing'' of such quarters of the preceeding random number to obtain the next one. Test results of the PNG are displayed for two such generators. A FORTRAN variant of the PNG is presented along with a program realizing the PNG made on the base of the INTEL-8080 autocode

  6. Quantification of lung fibrosis and emphysema in mice using automated micro-computed tomography.

    Directory of Open Access Journals (Sweden)

    Ellen De Langhe

    Full Text Available BACKGROUND: In vivo high-resolution micro-computed tomography allows for longitudinal image-based measurements in animal models of lung disease. The combination of repetitive high resolution imaging with fully automated quantitative image analysis in mouse models of lung fibrosis lung benefits preclinical research. This study aimed to develop and validate such an automated micro-computed tomography analysis algorithm for quantification of aerated lung volume in mice; an indicator of pulmonary fibrosis and emphysema severity. METHODOLOGY: Mice received an intratracheal instillation of bleomycin (n = 8, elastase (0.25 U elastase n = 9, 0.5 U elastase n = 8 or saline control (n = 6 for fibrosis, n = 5 for emphysema. A subset of mice was scanned without intervention, to evaluate potential radiation-induced toxicity (n = 4. Some bleomycin-instilled mice were treated with imatinib for proof of concept (n = 8. Mice were scanned weekly, until four weeks after induction, when they underwent pulmonary function testing, lung histology and collagen quantification. Aerated lung volumes were calculated with our automated algorithm. PRINCIPAL FINDINGS: Our automated image-based aerated lung volume quantification method is reproducible with low intra-subject variability. Bleomycin-treated mice had significantly lower scan-derived aerated lung volumes, compared to controls. Aerated lung volume correlated with the histopathological fibrosis score and total lung collagen content. Inversely, a dose-dependent increase in lung volume was observed in elastase-treated mice. Serial scanning of individual mice is feasible and visualized dynamic disease progression. No radiation-induced toxicity was observed. Three-dimensional images provided critical topographical information. CONCLUSIONS: We report on a high resolution in vivo micro-computed tomography image analysis algorithm that runs fully automated and allows quantification of aerated lung volume in mice. This

  7. Electromagnetic interference modeling and suppression techniques in variable-frequency drive systems

    Science.gov (United States)

    Yang, Le; Wang, Shuo; Feng, Jianghua

    2017-11-01

    Electromagnetic interference (EMI) causes electromechanical damage to the motors and degrades the reliability of variable-frequency drive (VFD) systems. Unlike fundamental frequency components in motor drive systems, high-frequency EMI noise, coupled with the parasitic parameters of the trough system, are difficult to analyze and reduce. In this article, EMI modeling techniques for different function units in a VFD system, including induction motors, motor bearings, and rectifierinverters, are reviewed and evaluated in terms of applied frequency range, model parameterization, and model accuracy. The EMI models for the motors are categorized based on modeling techniques and model topologies. Motor bearing and shaft models are also reviewed, and techniques that are used to eliminate bearing current are evaluated. Modeling techniques for conventional rectifierinverter systems are also summarized. EMI noise suppression techniques, including passive filter, Wheatstone bridge balance, active filter, and optimized modulation, are reviewed and compared based on the VFD system models.

  8. Review of air quality modeling techniques. Volume 8

    International Nuclear Information System (INIS)

    Rosen, L.C.

    1977-01-01

    Air transport and diffusion models which are applicable to the assessment of the environmental effects of nuclear, geothermal, and fossil-fuel electric generation are reviewed. The general classification of models and model inputs are discussed. A detailed examination of the statistical, Gaussian plume, Gaussian puff, one-box and species-conservation-of-mass models is given. Representative models are discussed with attention given to the assumptions, input data requirement, advantages, disadvantages and applicability of each

  9. Modeling technique for the process of liquid film disintegration

    Science.gov (United States)

    Modorskii, V. Ya.; Sipatov, A. M.; Babushkina, A. V.; Kolodyazhny, D. Yu.; Nagorny, V. S.

    2016-10-01

    In the course of numerical experiments the method of calculation of two-phase flows was developed by solving a model problem. The results of the study were compared between the two models that describe the processes of two-phase flow and the collapse of the liquid jet into droplets. VoF model and model QMOM - two mathematical models were considered the implementation of the spray.

  10. Using Microcomputer Software to Score Placement Tests--An Example from the University of California, Irvine.

    Science.gov (United States)

    Shoemaker, Judith S.; And Others

    This article describes the placement testing program, the use of microcomputer software for scoring and analyzing test results, and the integration of the computerized test results into a comprehensive microcomputer-based student information system at the University of California, Irvine (UCI). UCI offers placement tests in five academic fields:…

  11. MICROsearch--A Project to Extend the ERIC Database to Microcomputers.

    Science.gov (United States)

    Clark, W. Bruce

    A project at the ERIC Clearinghouse on Information Resources has resulted in the extension of the ERIC database to microcomputer. Small subject-related files for current-awareness purposes are downloaded from the ERIC tapes and are made available for searching with Apple II Plus microcomputers. Since the target audience for this system is…

  12. User-Adaptable Microcomputer Graphics Software for Life Science Instruction. Final Project Report.

    Science.gov (United States)

    Spain, James D.

    The objectives of the SUMIT project was to develop, evaluate, and disseminate 20 course modules (microcomputer programs) for instruction in general biology and ecology. To encourage broad utilization, the programs were designed for the Apple II microcomputer and written in Applesoft Basic with a user-adaptable format. Each package focused on a key…

  13. Microcomputers versus Mainframe Computers: Selecting the Appropriate Computing Tool in Institutional Research.

    Science.gov (United States)

    Mannering, Linda W.; McKinney, Richard L.

    The paper sets the stage for the effective and efficient use of computers in institutional research. The microcomputer and the mainframe computer are contrasted for those researchers who have a choice of where to process data. General computer use, specific microcomputer and mainframe computing guidelines, and an application utilizing both…

  14. Micro-computed tomography: Introducing new dimensions to taxonomy.

    Science.gov (United States)

    Faulwetter, Sarah; Vasileiadou, Aikaterini; Kouratoras, Michail; Thanos Dailianis; Arvanitidis, Christos

    2013-01-01

    Continuous improvements in the resolution of three-dimensional imaging have led to an increased application of these techniques in conventional taxonomic research in recent years. Coupled with an ever increasing research effort in cybertaxonomy, three-dimensional imaging could give a boost to the development of virtual specimen collections, allowing rapid and simultaneous access to accurate virtual representations of type material. This paper explores the potential of micro-computed tomography (X-ray micro-tomography), a non-destructive three-dimensional imaging technique based on mapping X-ray attenuation in the scanned object, for supporting research in systematics and taxonomy. The subsequent use of these data as virtual type material, so-called "cybertypes", and the creation of virtual collections lie at the core of this potential. Sample preparation, image acquisition, data processing and presentation of results are demonstrated using polychaetes (bristle worms), a representative taxon of macro-invertebrates, as a study object. Effects of the technique on the morphological, anatomical and molecular identity of the specimens are investigated. The paper evaluates the results and discusses the potential and the limitations of the technique for creating cybertypes. It also discusses the challenges that the community might face to establish virtual collections. Potential future applications of three-dimensional information in taxonomic research are outlined, including an outlook to new ways of producing, disseminating and publishing taxonomic information.

  15. Modelling tick abundance using machine learning techniques and satellite imagery

    DEFF Research Database (Denmark)

    Kjær, Lene Jung; Korslund, L.; Kjelland, V.

    satellite images to run Boosted Regression Tree machine learning algorithms to predict overall distribution (presence/absence of ticks) and relative tick abundance of nymphs and larvae in southern Scandinavia. For nymphs, the predicted abundance had a positive correlation with observed abundance...... the predicted distribution of larvae was mostly even throughout Denmark, it was primarily around the coastlines in Norway and Sweden. Abundance was fairly low overall except in some fragmented patches corresponding to forested habitats in the region. Machine learning techniques allow us to predict for larger...... the collected ticks for pathogens and using the same machine learning techniques to develop prevalence maps of the ScandTick region....

  16. Adaptive subdomain modeling: A multi-analysis technique for ocean circulation models

    Science.gov (United States)

    Altuntas, Alper; Baugh, John

    2017-07-01

    Many coastal and ocean processes of interest operate over large temporal and geographical scales and require a substantial amount of computational resources, particularly when engineering design and failure scenarios are also considered. This study presents an adaptive multi-analysis technique that improves the efficiency of these computations when multiple alternatives are being simulated. The technique, called adaptive subdomain modeling, concurrently analyzes any number of child domains, with each instance corresponding to a unique design or failure scenario, in addition to a full-scale parent domain providing the boundary conditions for its children. To contain the altered hydrodynamics originating from the modifications, the spatial extent of each child domain is adaptively adjusted during runtime depending on the response of the model. The technique is incorporated in ADCIRC++, a re-implementation of the popular ADCIRC ocean circulation model with an updated software architecture designed to facilitate this adaptive behavior and to utilize concurrent executions of multiple domains. The results of our case studies confirm that the method substantially reduces computational effort while maintaining accuracy.

  17. The visualization of hepatic vasculature by x-ray micro-computed tomography

    International Nuclear Information System (INIS)

    Ananda, Shalini; Marsden, Valerie; Korkmaz, Emine; Tsafnat, Naomi; Soon, Lilian; Jones, Allan; Braet, Filip; Vekemans, Katrien

    2006-01-01

    In this study, X-ray micro-computed tomography (CT) was used to reconstruct the fine structure macro- and microvasculature in three dimensions in contrast-enhanced rat liver samples. The subsequent application in the experimental CC531s colorectal caner model was concurrent with results obtained from confocal microscopy in earlier studies. The en bloc stains osmium tetroxide in combination with uranyl acetate provided an excellent contrasting result for hepatic tissue after a trial of several contrasting agents. X-ray micro-CT allowed us to image the large blood vessels together with the branching sinusoids of hepatic tissue in three dimensions. Furthermore, interruption of the microvasculature was noted when rats were injected with CC531s colorectal cancer cells indicating the presence of hepatic metastases. (author)

  18. Three-dimensional arrangement of β-tricalcium phosphate granules evaluated by microcomputed tomography and fractal analysis.

    Science.gov (United States)

    Ndiaye, Mambaye; Terranova, Lisa; Mallet, Romain; Mabilleau, Guillaume; Chappard, Daniel

    2015-01-01

    The macrophysical properties of granular biomaterials used to fill bone defects have rarely been considered. Granules of a given biomaterial occupy three-dimensional (3-D) space when packed together and create a macroporosity suitable for the invasion of vascular and bone cells. Granules of β-tricalcium phosphate were prepared using polyurethane foam technology and increasing the amount of material powder in the slurry (10, 11, 15, 18, 21 and 25 g). After sintering, granules of 1000-2000 μm were prepared by sieving. They were analyzed morphologically by scanning electron microscopy and placed in polyethylene test tubes to produce 3-D scaffolds. Microcomputed tomography (microCT) was used to image the scaffolds and to determine porosity and fractal dimension in three dimensions. Two-dimensional sections of the microCT models were binarized and used to compute classical morphometric parameters describing porosity (interconnectivity index, strut analysis and star volumes) and fractal dimensions. In addition, two newly important fractal parameters (lacunarity and succolarity) were measured. Compression analysis of the stacks of granules was done. Porosity decreased as the amount of material in the slurry increased but non-linear relationships were observed between microarchitectural parameters describing the pores and porosity. Lacunarity increased in the series of granules but succolarity (reflecting the penetration of a fluid) was maximal in the 15-18 g groups and decreased noticeably in the 25 g group. The 3-D arrangement of biomaterial granules studied by these new fractal techniques allows the optimal formulation to be derived based on the lowest amount of material, suitable mechanical resistance during crushing and the creation of large interconnected pores. Copyright © 2014 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.

  19. Application of integrated modeling technique for data services ...

    African Journals Online (AJOL)

    This paper, therefore, describes the application of the integrated simulation technique for deriving the optimum resources required for data services in an asynchronous transfer mode (ATM) based private wide area network (WAN) to guarantee specific QoS requirement. The simulation tool drastically cuts the simulation ...

  20. Simulation technique for hard-disk models in two dimensions

    DEFF Research Database (Denmark)

    Fraser, Diane P.; Zuckermann, Martin J.; Mouritsen, Ole G.

    1990-01-01

    A method is presented for studying hard-disk systems by Monte Carlo computer-simulation techniques within the NpT ensemble. The method is based on the Voronoi tesselation, which is dynamically maintained during the simulation. By an analysis of the Voronoi statistics, a quantity is identified...

  1. (NHIS) using data mining technique as a statistical model

    African Journals Online (AJOL)

    kofi.mereku

    2014-05-23

    May 23, 2014 ... Scheme (NHIS) claims in the Awutu-Effutu-Senya District using data mining techniques, with a specific focus on .... transform them into a format that is friendly to data mining algorithms, such as .... many groups to access the data, facilitate updating the data, and improve the efficiency of checking the data for ...

  2. Use of System Dynamics Techniques in the Garrison Health Modelling Tool

    Science.gov (United States)

    2010-11-01

    Joint Health Command (JHC) tasked DSTO to develop techniques for modelling Defence health service delivery both in a Garrison environment in Australia ...UNCLASSIFIED UNCLASSIFIED Use of System Dynamics Techniques in the Garrison Health Modelling Tool Mark Burnett, Kerry Clifford and...Garrison Health Modelling Tool, a prototype software package designed to provide decision-support to JHC health officers and managers in a garrison

  3. On a Numerical and Graphical Technique for Evaluating some Models Involving Rational Expectations

    DEFF Research Database (Denmark)

    Johansen, Søren; Swensen, Anders Rygh

    Campbell and Shiller (1987) proposed a graphical technique for the present value model which consists of plotting the spread and theoretical spread as calculated from the cointegrated vector autoregressive model. We extend these techniques to a number of rational expectation models and give...

  4. On a numerical and graphical technique for evaluating some models involving rational expectations

    DEFF Research Database (Denmark)

    Johansen, Søren; Swensen, Anders Rygh

    Campbell and Shiller (1987) proposed a graphical technique for the present value model which consists of plotting the spread and theoretical spread as calculated from the cointegrated vector autoregressive model. We extend these techniques to a number of rational expectation models and give...

  5. OFF-LINE HANDWRITING RECOGNITION USING VARIOUS HYBRID MODELING TECHNIQUES AND CHARACTER N-GRAMS

    NARCIS (Netherlands)

    Brakensiek, A.; Rottland, J.; Kosmala, A.; Rigoll, G.

    2004-01-01

    In this paper a system for on-line cursive handwriting recognition is described. The system is based on Hidden Markov Models (HMMs) using discrete and hybrid modeling techniques. Here, we focus on two aspects of the recognition system. First, we present different hybrid modeling techniques, whereas

  6. SUPER CAVIAR: Memory mapping the general-purpose microcomputer

    International Nuclear Information System (INIS)

    Cittolin, S.; Taylor, B.G.

    1981-01-01

    Over the past 3 years, CAVIAR (CAMAC Video Autonomous Read-out) microcomputers have been applied in growing numbers at CERN and related institutes. As typical user programs expanded in size, and the incorporated firmware libraries were enlarged also, the microprocessor addressing limit of 64 Kbytes became a serious constraint. An enhanced microcomputer, SUPER CAVIAR, has now been created by the incorporation of memory mapping to expand the physical address space to 344 Kbytes. The new facility provides independent firmware and RAM maps, dynamic allocation of common RAM, automatic inter-page transfer modes, and a RAM/EPROM overlay. A memory-based file system has been implemented, and control and data can be interchanged between separate programs in different RAM maps. 84 Kbytes of EPROM are incorporated on the mapper card itself, as well as an ADLC serial data link. In addition to providing more space for consolidated user programs and data, SUPER CAVIAR has allowed the introduction of several improvements to the BAMBI interpreter and extensions to the CAVIAR libraries. A context editor and enhanced debug monitor have been added, as well as new data types and extended array-handling and graphics routines, including isoline plotting, line-fitting and FFT operations. A SUPER CAVIAR converter has been developed which allows a standard CAVIAR to be upgraded to incorporate the new facilities without loss of the existing investment. (orig.)

  7. The application of neural networks with artificial intelligence technique in the modeling of industrial processes

    International Nuclear Information System (INIS)

    Saini, K. K.; Saini, Sanju

    2008-01-01

    Neural networks are a relatively new artificial intelligence technique that emulates the behavior of biological neural systems in digital software or hardware. These networks can 'learn', automatically, complex relationships among data. This feature makes the technique very useful in modeling processes for which mathematical modeling is difficult or impossible. The work described here outlines some examples of the application of neural networks with artificial intelligence technique in the modeling of industrial processes.

  8. A microcomputer program for energy assessment and aggregation using the triangular probability distribution

    Science.gov (United States)

    Crovelli, R.A.; Balay, R.H.

    1991-01-01

    A general risk-analysis method was developed for petroleum-resource assessment and other applications. The triangular probability distribution is used as a model with an analytic aggregation methodology based on probability theory rather than Monte-Carlo simulation. Among the advantages of the analytic method are its computational speed and flexibility, and the saving of time and cost on a microcomputer. The input into the model consists of a set of components (e.g. geologic provinces) and, for each component, three potential resource estimates: minimum, most likely (mode), and maximum. Assuming a triangular probability distribution, the mean, standard deviation, and seven fractiles (F100, F95, F75, F50, F25, F5, and F0) are computed for each component, where for example, the probability of more than F95 is equal to 0.95. The components are aggregated by combining the means, standard deviations, and respective fractiles under three possible siutations (1) perfect positive correlation, (2) complete independence, and (3) any degree of dependence between these two polar situations. A package of computer programs named the TRIAGG system was written in the Turbo Pascal 4.0 language for performing the analytic probabilistic methodology. The system consists of a program for processing triangular probability distribution assessments and aggregations, and a separate aggregation routine for aggregating aggregations. The user's documentation and program diskette of the TRIAGG system are available from USGS Open File Services. TRIAGG requires an IBM-PC/XT/AT compatible microcomputer with 256kbyte of main memory, MS-DOS 3.1 or later, either two diskette drives or a fixed disk, and a 132 column printer. A graphics adapter and color display are optional. ?? 1991.

  9. Models, Web-Based Simulations, and Integrated Analysis Techniques for Improved Logistical Performance

    National Research Council Canada - National Science Library

    Hill, Raymond

    2001-01-01

    ... Laboratory, Logistics Research Division, Logistics Readiness Branch to propose a research agenda entitled, "Models, Web-based Simulations, and Integrated Analysis Techniques for Improved Logistical Performance...

  10. Modelling Data Mining Dynamic Code Attributes with Scheme Definition Technique

    OpenAIRE

    Sipayung, Evasaria M; Fiarni, Cut; Tanudjaja, Randy

    2014-01-01

    Data mining is a technique used in differentdisciplines to search for significant relationships among variablesin large data sets. One of the important steps on data mining isdata preparation. On these step, we need to transform complexdata with more than one attributes into representative format fordata mining algorithm. In this study, we concentrated on thedesigning a proposed system to fetch attributes from a complexdata such as product ID. Then the proposed system willdetermine the basic ...

  11. Wave propagation in fluids models and numerical techniques

    CERN Document Server

    Guinot, Vincent

    2012-01-01

    This second edition with four additional chapters presents the physical principles and solution techniques for transient propagation in fluid mechanics and hydraulics. The application domains vary including contaminant transport with or without sorption, the motion of immiscible hydrocarbons in aquifers, pipe transients, open channel and shallow water flow, and compressible gas dynamics. The mathematical formulation is covered from the angle of conservation laws, with an emphasis on multidimensional problems and discontinuous flows, such as steep fronts and shock waves. Finite

  12. An eigenexpansion technique for modelling plasma start-up

    International Nuclear Information System (INIS)

    Pillsbury, R.D.

    1989-01-01

    An algorithm has been developed and implemented in a computer program that allows the estimation of PF coil voltages required to start-up an axisymmetric plasma in a tokamak in the presence of eddy currents in toroidally continuous conducting structures. The algorithm makes use of an eigen-expansion technique to solve the lumped parameter circuit loop voltage equations associated with the PF coils and passive (conducting) structures. An example of start-up for CIT (Compact Ignition Tokamak) is included

  13. TESTING DIFFERENT SURVEY TECHNIQUES TO MODEL ARCHITECTONIC NARROW SPACES

    Directory of Open Access Journals (Sweden)

    A. Mandelli

    2017-08-01

    Full Text Available In the architectural survey field, there has been the spread of a vast number of automated techniques. However, it is important to underline the gap that exists between the technical specification sheet of a particular instrument and its usability, accuracy and level of automation reachable in real cases scenario, especially speaking about Cultural Heritage (CH field. In fact, even if the technical specifications (range, accuracy and field of view are known for each instrument, their functioning and features are influenced by the environment, shape and materials of the object. The results depend more on how techniques are employed than the nominal specifications of the instruments. The aim of this article is to evaluate the real usability, for the 1:50 architectonic restitution scale, of common and not so common survey techniques applied to the complex scenario of dark, intricate and narrow spaces such as service areas, corridors and stairs of Milan’s cathedral indoors. Tests have shown that the quality of the results is strongly affected by side-issues like the impossibility of following the theoretical ideal methodology when survey such spaces. The tested instruments are: the laser scanner Leica C10, the GeoSLAM ZEB1, the DOT DPI 8 and two photogrammetric setups, a full frame camera with a fisheye lens and the NCTech iSTAR, a panoramic camera. Each instrument presents advantages and limits concerning both the sensors themselves and the acquisition phase.

  14. Testing Different Survey Techniques to Model Architectonic Narrow Spaces

    Science.gov (United States)

    Mandelli, A.; Fassi, F.; Perfetti, L.; Polari, C.

    2017-08-01

    In the architectural survey field, there has been the spread of a vast number of automated techniques. However, it is important to underline the gap that exists between the technical specification sheet of a particular instrument and its usability, accuracy and level of automation reachable in real cases scenario, especially speaking about Cultural Heritage (CH) field. In fact, even if the technical specifications (range, accuracy and field of view) are known for each instrument, their functioning and features are influenced by the environment, shape and materials of the object. The results depend more on how techniques are employed than the nominal specifications of the instruments. The aim of this article is to evaluate the real usability, for the 1:50 architectonic restitution scale, of common and not so common survey techniques applied to the complex scenario of dark, intricate and narrow spaces such as service areas, corridors and stairs of Milan's cathedral indoors. Tests have shown that the quality of the results is strongly affected by side-issues like the impossibility of following the theoretical ideal methodology when survey such spaces. The tested instruments are: the laser scanner Leica C10, the GeoSLAM ZEB1, the DOT DPI 8 and two photogrammetric setups, a full frame camera with a fisheye lens and the NCTech iSTAR, a panoramic camera. Each instrument presents advantages and limits concerning both the sensors themselves and the acquisition phase.

  15. Artificial intelligence techniques for modeling database user behavior

    Science.gov (United States)

    Tanner, Steve; Graves, Sara J.

    1990-01-01

    The design and development of the adaptive modeling system is described. This system models how a user accesses a relational database management system in order to improve its performance by discovering use access patterns. In the current system, these patterns are used to improve the user interface and may be used to speed data retrieval, support query optimization and support a more flexible data representation. The system models both syntactic and semantic information about the user's access and employs both procedural and rule-based logic to manipulate the model.

  16. Reduction of thermal models of buildings: improvement of techniques using meteorological influence models; Reduction de modeles thermiques de batiments: amelioration des techniques par modelisation des sollicitations meteorologiques

    Energy Technology Data Exchange (ETDEWEB)

    Dautin, S.

    1997-04-01

    This work concerns the modeling of thermal phenomena inside buildings for the evaluation of energy exploitation costs of thermal installations and for the modeling of thermal and aeraulic transient phenomena. This thesis comprises 7 chapters dealing with: (1) the thermal phenomena inside buildings and the CLIM2000 calculation code, (2) the ETNA and GENEC experimental cells and their modeling, (3) the techniques of model reduction tested (Marshall`s truncature, Michailesco aggregation method and Moore truncature) with their algorithms and their encoding in the MATRED software, (4) the application of model reduction methods to the GENEC and ETNA cells and to a medium size dual-zone building, (5) the modeling of meteorological influences classically applied to buildings (external temperature and solar flux), (6) the analytical expression of these modeled meteorological influences. The last chapter presents the results of these improved methods on the GENEC and ETNA cells and on a lower inertia building. These new methods are compared to classical methods. (J.S.) 69 refs.

  17. Multiparous Ewe as a Model for Teaching Vaginal Hysterectomy Techniques.

    Science.gov (United States)

    Kerbage, Yohan; Cosson, Michel; Hubert, Thomas; Giraudet, Géraldine

    2017-12-01

    Despite being linked to improving patient outcomes and limiting costs, the use of vaginal hysterectomy is on the wane. Although a combination of reasons might explain this trend, one cause is a lack of practical training. An appropriate teaching model must therefore be devised. Currently, only low-fidelity simulators exist. Ewes provide an appropriate model for pelvic anatomy and are well-suited for testing vaginal mesh properties. This article sets out a vaginal hysterectomy procedure for use as an education and training model. A multiparous ewe was the model. Surgery was performed under general anesthesia. The ewe was in a lithotomy position resembling that assumed by women on the operating table. Two vaginal hysterectomies were performed on two ewes, following every step precisely as if the model were human. Each surgical step of vaginal hysterectomy performed on the ewe and on a woman were compared side by side. We identified that all surgical steps were particularly similar. The main limitations of this model are costs ($500/procedure), logistic problems (housing large animals), and public opposition to animal training models. The ewe appears to be an appropriate model for teaching and training of vaginal hysterectomy.

  18. Detecting and visualizing internal 3D oleoresin in agarwood by means of micro-computed tomography

    International Nuclear Information System (INIS)

    Khairiah Yazid; Roslan Yahya; Mat Rosol Awang

    2012-01-01

    Detection and analysis of oleoresin is particularly significant since the commercial value of agarwood is related to the quantity of oleoresins that are present. A modern technique of non-destructive may reach the interior region of the wood. Currently, tomographic image data in particular is most commonly visualized in three dimensions using volume rendering. The aim of this paper is to explore the potential of high resolution non-destructive 3D visualization technique, X-ray micro-computed tomography, as imaging tools to visualize micro-structure oleoresin in agarwood. Investigations involving desktop X-ray micro-tomography system on high grade agarwood sample, performed at the Centre of Tomography in Nuclear Malaysia, demonstrate the applicability of the method. Prior to experiments, a reference test was conducted to stimulate the attenuation of oleoresin in agarwood. Based on the experiment results, micro-CT imaging with voxel size 7.0 μm is capable to of detecting oleoresin and pores in agarwood. This imaging technique, although sophisticated can be used for standard development especially in grading of agarwood for commercial activities. (author)

  19. Household water use and conservation models using Monte Carlo techniques

    Directory of Open Access Journals (Sweden)

    R. Cahill

    2013-10-01

    Full Text Available The increased availability of end use measurement studies allows for mechanistic and detailed approaches to estimating household water demand and conservation potential. This study simulates water use in a single-family residential neighborhood using end-water-use parameter probability distributions generated from Monte Carlo sampling. This model represents existing water use conditions in 2010 and is calibrated to 2006–2011 metered data. A two-stage mixed integer optimization model is then developed to estimate the least-cost combination of long- and short-term conservation actions for each household. This least-cost conservation model provides an estimate of the upper bound of reasonable conservation potential for varying pricing and rebate conditions. The models were adapted from previous work in Jordan and are applied to a neighborhood in San Ramon, California in the eastern San Francisco Bay Area. The existing conditions model produces seasonal use results very close to the metered data. The least-cost conservation model suggests clothes washer rebates are among most cost-effective rebate programs for indoor uses. Retrofit of faucets and toilets is also cost-effective and holds the highest potential for water savings from indoor uses. This mechanistic modeling approach can improve understanding of water demand and estimate cost-effectiveness of water conservation programs.

  20. Line impedance estimation using model based identification technique

    DEFF Research Database (Denmark)

    Ciobotaru, Mihai; Agelidis, Vassilios; Teodorescu, Remus

    2011-01-01

    The estimation of the line impedance can be used by the control of numerous grid-connected systems, such as active filters, islanding detection techniques, non-linear current controllers, detection of the on/off grid operation mode. Therefore, estimating the line impedance can add extra functions......-passive behaviour of the proposed method comes from the combination of the non intrusive behaviour of the passive methods with a better accuracy of the active methods. The simulation results reveal the good accuracy of the proposed method....

  1. Size reduction techniques for vital compliant VHDL simulation models

    Science.gov (United States)

    Rich, Marvin J.; Misra, Ashutosh

    2006-08-01

    A method and system select delay values from a VHDL standard delay file that correspond to an instance of a logic gate in a logic model. Then the system collects all the delay values of the selected instance and builds super generics for the rise-time and the fall-time of the selected instance. Then, the system repeats this process for every delay value in the standard delay file (310) that correspond to every instance of every logic gate in the logic model. The system then outputs a reduced size standard delay file (314) containing the super generics for every instance of every logic gate in the logic model.

  2. Ground-water Education System (for microcomputers). Model-Simulation

    Energy Technology Data Exchange (ETDEWEB)

    1990-01-01

    The Groundwater Education System is a program designed on the concept of hypermedia. Users operating the program using hypertext can determine how much information they receive from the program by selecting what topics they want to see. The user selects hypertext links by clicking on a highlighted word or phrase. The program then sends the user to another section regarding their selection. In this way, more material can be interwoven. This allows a broader scope for education and applications.

  3. Use of machine learning techniques for modeling of snow depth

    Directory of Open Access Journals (Sweden)

    G. V. Ayzel

    2017-01-01

    Full Text Available Snow exerts significant regulating effect on the land hydrological cycle since it controls intensity of heat and water exchange between the soil-vegetative cover and the atmosphere. Estimating of a spring flood runoff or a rain-flood on mountainous rivers requires understanding of the snow cover dynamics on a watershed. In our work, solving a problem of the snow cover depth modeling is based on both available databases of hydro-meteorological observations and easily accessible scientific software that allows complete reproduction of investigation results and further development of this theme by scientific community. In this research we used the daily observational data on the snow cover and surface meteorological parameters, obtained at three stations situated in different geographical regions: Col de Porte (France, Sodankyla (Finland, and Snoquamie Pass (USA.Statistical modeling of the snow cover depth is based on a complex of freely distributed the present-day machine learning models: Decision Trees, Adaptive Boosting, Gradient Boosting. It is demonstrated that use of combination of modern machine learning methods with available meteorological data provides the good accuracy of the snow cover modeling. The best results of snow cover depth modeling for every investigated site were obtained by the ensemble method of gradient boosting above decision trees – this model reproduces well both, the periods of snow cover accumulation and its melting. The purposeful character of learning process for models of the gradient boosting type, their ensemble character, and use of combined redundancy of a test sample in learning procedure makes this type of models a good and sustainable research tool. The results obtained can be used for estimating the snow cover characteristics for river basins where hydro-meteorological information is absent or insufficient.

  4. Modeling and Control of Multivariable Process Using Intelligent Techniques

    Directory of Open Access Journals (Sweden)

    Subathra Balasubramanian

    2010-10-01

    Full Text Available For nonlinear dynamic systems, the first principles based modeling and control is difficult to implement. In this study, a fuzzy controller and recurrent fuzzy controller are developed for MIMO process. Fuzzy logic controller is a model free controller designed based on the knowledge about the process. In fuzzy controller there are two types of rule-based fuzzy models are available: one the linguistic (Mamdani model and the other is Takagi–Sugeno model. Of these two, Takagi-Sugeno model (TS has attracted most attention. The fuzzy controller application is limited to static processes due to their feedforward structure. But, most of the real-time processes are dynamic and they require the history of input/output data. In order to store the past values a memory unit is needed, which is introduced by the recurrent structure. The proposed recurrent fuzzy structure is used to develop a controller for the two tank heating process. Both controllers are designed and implemented in a real time environment and their performance is compared.

  5. A study on the modeling techniques using LS-INGRID

    Energy Technology Data Exchange (ETDEWEB)

    Ku, J. H.; Park, S. W

    2001-03-01

    For the development of radioactive material transport packages, the verification of structural safety of a package against the free drop impact accident should be carried out. The use of LS-DYNA, which is specially developed code for impact analysis, is essential for impact analysis of the package. LS-INGRID is a pre-processor for LS-DYNA with considerable capability to deal with complex geometries and allows for parametric modeling. LS-INGRID is most effective in combination with LS-DYNA code. Although the usage of LS-INGRID seems very difficult relative to many commercial mesh generators, the productivity of users performing parametric modeling tasks with LS-INGRID can be much higher in some cases. Therefore, LS-INGRID has to be used with LS-DYNA. This report presents basic explanations for the structure and commands, basic modelling examples and advanced modelling of LS-INGRID to use it for the impact analysis of various packages. The new users can build the complex model easily, through a study for the basic examples presented in this report from the modelling to the loading and constraint conditions.

  6. A study on the modeling techniques using LS-INGRID

    International Nuclear Information System (INIS)

    Ku, J. H.; Park, S. W.

    2001-03-01

    For the development of radioactive material transport packages, the verification of structural safety of a package against the free drop impact accident should be carried out. The use of LS-DYNA, which is specially developed code for impact analysis, is essential for impact analysis of the package. LS-INGRID is a pre-processor for LS-DYNA with considerable capability to deal with complex geometries and allows for parametric modeling. LS-INGRID is most effective in combination with LS-DYNA code. Although the usage of LS-INGRID seems very difficult relative to many commercial mesh generators, the productivity of users performing parametric modeling tasks with LS-INGRID can be much higher in some cases. Therefore, LS-INGRID has to be used with LS-DYNA. This report presents basic explanations for the structure and commands, basic modelling examples and advanced modelling of LS-INGRID to use it for the impact analysis of various packages. The new users can build the complex model easily, through a study for the basic examples presented in this report from the modelling to the loading and constraint conditions

  7. GENERALIZATION TECHNIQUE FOR 2D+SCALE DHE DATA MODEL

    Directory of Open Access Journals (Sweden)

    H. Karim

    2016-10-01

    Full Text Available Different users or applications need different scale model especially in computer application such as game visualization and GIS modelling. Some issues has been raised on fulfilling GIS requirement of retaining the details while minimizing the redundancy of the scale datasets. Previous researchers suggested and attempted to add another dimension such as scale or/and time into a 3D model, but the implementation of scale dimension faces some problems due to the limitations and availability of data structures and data models. Nowadays, various data structures and data models have been proposed to support variety of applications and dimensionality but lack research works has been conducted in terms of supporting scale dimension. Generally, the Dual Half Edge (DHE data structure was designed to work with any perfect 3D spatial object such as buildings. In this paper, we attempt to expand the capability of the DHE data structure toward integration with scale dimension. The description of the concept and implementation of generating 3D-scale (2D spatial + scale dimension for the DHE data structure forms the major discussion of this paper. We strongly believed some advantages such as local modification and topological element (navigation, query and semantic information in scale dimension could be used for the future 3D-scale applications.

  8. Fusing Observations and Model Results for Creation of Enhanced Ozone Spatial Fields: Comparison of Three Techniques

    Science.gov (United States)

    This paper presents three simple techniques for fusing observations and numerical model predictions. The techniques rely on model/observation bias being considered either as error free, or containing some uncertainty, the latter mitigated with a Kalman filter approach or a spati...

  9. Increasing the reliability of ecological models using modern software engineering techniques

    Science.gov (United States)

    Robert M. Scheller; Brian R. Sturtevant; Eric J. Gustafson; Brendan C. Ward; David J. Mladenoff

    2009-01-01

    Modern software development techniques are largely unknown to ecologists. Typically, ecological models and other software tools are developed for limited research purposes, and additional capabilities are added later, usually in an ad hoc manner. Modern software engineering techniques can substantially increase scientific rigor and confidence in ecological models and...

  10. The impact of applying product-modelling techniques in configurator projects

    DEFF Research Database (Denmark)

    Hvam, Lars; Kristjansdottir, Katrin; Shafiee, Sara

    2018-01-01

    This paper aims to increase understanding of the impact of using product-modelling techniques to structure and formalise knowledge in configurator projects. Companies that provide customised products increasingly apply configurators in support of sales and design activities, reaping benefits...... though extant literature has shown the importance of formal modelling techniques, the impact of utilising these techniques remains relatively unknown. Therefore, this article studies three main areas: (1) the impact of using modelling techniques based on Unified Modelling Language (UML), in which...... ability to reduce the number of product variants. This paper contributes to an increased understanding of what companies can gain from using more formalised modelling techniques in configurator projects, and under what circumstances they should be used....

  11. Biliary System Architecture: Experimental Models and Visualization Techniques

    Czech Academy of Sciences Publication Activity Database

    Sarnová, Lenka; Gregor, Martin

    2017-01-01

    Roč. 66, č. 3 (2017), s. 383-390 ISSN 0862-8408 R&D Projects: GA MŠk(CZ) LQ1604; GA ČR GA15-23858S Institutional support: RVO:68378050 Keywords : Biliary system * Mouse model * Cholestasis * Visualisation * Morphology Subject RIV: EB - Genetics ; Molecular Biology OBOR OECD: Cell biology Impact factor: 1.461, year: 2016

  12. Discovering Process Reference Models from Process Variants Using Clustering Techniques

    NARCIS (Netherlands)

    Li, C.; Reichert, M.U.; Wombacher, Andreas

    2008-01-01

    In today's dynamic business world, success of an enterprise increasingly depends on its ability to react to changes in a quick and flexible way. In response to this need, process-aware information systems (PAIS) emerged, which support the modeling, orchestration and monitoring of business processes

  13. Suitability of sheet bending modelling techniques in CAPP applications

    NARCIS (Netherlands)

    Streppel, A.H.; de Vin, L.J.; de Vin, L.J.; Brinkman, J.; Brinkman, J.; Kals, H.J.J.

    1993-01-01

    The use of CNC machine tools, together with decreasing lot sizes and stricter tolerance prescriptions, has led to changes in sheet-metal part manufacturing. In this paper, problems introduced by the difference between the actual material behaviour and the results obtained from analytical models and

  14. Air quality modelling using chemometric techniques | Azid | Journal ...

    African Journals Online (AJOL)

    DA shows all seven parameters (CO, O3, PM10, SO2, NOx, NO and NO2) gave the most significant variables after stepwise backward mode. PCA identifies the major source of air pollution is due to combustion of fossil fuels in motor vehicles and industrial activities. The ANN model shows a better prediction compared to the ...

  15. Teaching Behavioral Modeling and Simulation Techniques for Power Electronics Courses

    Science.gov (United States)

    Abramovitz, A.

    2011-01-01

    This paper suggests a pedagogical approach to teaching the subject of behavioral modeling of switch-mode power electronics systems through simulation by general-purpose electronic circuit simulators. The methodology is oriented toward electrical engineering (EE) students at the undergraduate level, enrolled in courses such as "Power…

  16. A multifunctional data acquisition and processing system based on microcomputer

    International Nuclear Information System (INIS)

    Chen Huaide; Pan Dajin; Zhu Cuiqin; Wang Guangyu; Wang Shibo; Niu Wanfu; Gu Jinglan; Tang Yunli

    1988-01-01

    A data acquisition and processing system based on microcomputer TRS-80 has been developed since 1983 and has been operated for years. The system is a multifunctional MCA-Computer system which can be operated as a foreground and a background with the capabilities of high resolution in γ-ray spectrum analysis, multiplex multi-scale, multi-scale and the display system having the resolution up to 8192 points. MCA is 8192 channel with 20 bits 8 k buffer memory. First, the acquisition data is stored into the buffer memory. Then, it is transfered into computer for processing. The system software consists of the system management, spectrum processing and multiplex multi-scale measurement softwares. This article gives the block diagram of the system, the specifications, application and results

  17. A differential approach to microcomputer test battery development and implementation

    Science.gov (United States)

    Kennedy, R. S.; Baltzley, D. R.; Osteen, M. K.; Turnage, J. J.

    1988-01-01

    The present microcomputer-based performance test battery emphasizes psychometric theory and utility for repeated-measures applications during extended exposure to various environmental stressors. In the menu that has been defined at the current state of this system's development, there are more than 30 'qualified' mental tests which stabilize in less than 10 min and possess test-retest reliabilities greater than 0.7 for a three-minute test/work period. The battery encompasses tests of cognition, information processing, psychomotor skill, memory, mood, etc. Several of the tests have demonstrated sensitivity to chemoradiotherapy, sleep loss, hypoxia, amphetamines, thermal stress, sensory deprivation, altitude, fatigue, and alcohol use. Recommendations are presented for 6-, 12-, and 22-min batteries.

  18. A mini/microcomputer-based land use information system

    Science.gov (United States)

    Seitz, R. N.; Keefer, R. L.; Britton, L. J.; Wilson, J. M.

    1977-01-01

    The paper describes the Multipurpose Interactive NASA Information System (MINIS), a data management system for land-use applications. MINIS is written nearly entirely in FORTRAN IV, and has a full range of conditional, Boolean and arithmetic commands, as well as extensive format control and the capability of interactive file creation and updating. It requires a mini or microcomputer with at least 64 K of core or semiconductor memory. MINIS has its own equation-oriented query language for retrieval from different kinds of data bases. It features a graphics output which permits output of overlay maps. Some experience of the U.S. Department of Agriculture and the Tennessee State Planning Office with MINIS is discussed.

  19. Page: a program for gamma spectra analysis in PC microcomputers

    International Nuclear Information System (INIS)

    Goncalves, M.A.; Yamaura, M.; Costa, G.J.C.; Carvalho, E.I. de; Matsuda, H.T.; Araujo, B.F. de.

    1991-04-01

    PAGE is a software package, written in BASIC language, to perform gamma spectra analysis. It was developed to be used in a high-purity intrinsic germanium detector-multichannel analyser-PC microcomputer system. The analysis program of PAGE package accomplishes functions as follows: peak location; gamma nuclides identification; activity determination. Standard nuclides sources were used to calibrate the system. To perform the efficiency x energy calibration a logarithmic fit was applied. Analysis of nuclides with overlapping peaks is allowed by PAGE program. PAGE has additional auxiliary programs for: building and list of isotopic nuclear data libraries; data acquisition from multichannel analyser; spectrum display with automatic area and FWHM determinations. This software is to be applied in analytical process control where time response is a very important parameter. PAGE takes ca. 1.5 minutes to analyse a complex spectrum from a 4096 channels MCA. (author)

  20. Application of DEC PRO-380 microcomputer for treatment planning. 129

    International Nuclear Information System (INIS)

    Chung-Bin, D.; Bernard, D.; Chung-Bin, A.

    1987-01-01

    Since 1974 we have used a PDP 11/45 computer manufactured by Digital Equipment Corp. for radiation therapy treatment planning. The Cunningham TAR method has been used for developing algorithms for external beam treatment planning. The Stovell and Shalek method has been used for intracavitary and interstitial implants. Recently we have implemented all this softwre on a DEC Pro 380 microcomputer with graphic input and output. This inexpensive machine has computer power comparable to that of a PDP 11/45. All the hardware is standard DEC equipment except the digitizer. This system is adequate to meet all the treatment planning needs of any moderate size radiation therapy center. It lacks the ability to input CT data directly from magnetic tape, but it can enter the contour, tumor volumes and inhomogeneity from a CT film using the back-lighted digitizer. 2 refs

  1. Management of quality assurance in diagnostic radiology by microcomputer

    International Nuclear Information System (INIS)

    Evans, S.H.

    1985-01-01

    Software has been written for the calculation, interpretation and presentation of quality control measurements of X-ray machines. The programs run on a Z80 based microcomputer system comprising a Comart Communicator CP500 interfaced to a Volker Craig 4404 visual display unit (VDU) and an EpsonMX80F/T printer. The software has been written in the dbaseII database language (Ashton Tate 1975) and it runs under the CP/M operating system. The programs guide the user through each routine required and can be operated without knowledge of computers. The programs calculate the results from the raw data obtained from the test equipment and these results are then stored and analysed. (U.K.)

  2. Microcomputer-based workforce scheduling for hospital porters.

    Science.gov (United States)

    Lin, C K

    1999-01-01

    This paper focuses on labour scheduling for hospital porters who are the major workforce providing routine cleansing of wards, transportation and messenger services. Generating an equitable monthly roster for porters while meeting the daily minimum demand is a tedious task scheduled manually by a supervisor. In considering a variety of constraints and goals, a manual schedule was usually produced in seven to ten days. To be in line with the strategic goal of scientific management of an acute care regional hospital in Hong Kong, a microcomputer-based algorithm was developed to schedule the monthly roster. The algorithm, coded in Digital Visual Fortran 5.0 Professional, could generate a monthly roster in seconds. Implementation has been carried out since September 1998 and the results proved to be useful to hospital administrators and porters. This paper discusses both the technical and human issues involved during the computerization process.

  3. Enhancing the usability of safety software for microcomputers.

    Science.gov (United States)

    Greenberg, L

    1984-07-01

    With the proliferation of microcomputers, increasing effort is being made to write programs dealing with occupational safety and health. So as to assure the widespread use of this material, it is important that it be prepared in a manner easy to use. Ignoring this consideration may cause the potential users to be repelled by it from the start, or that its use will be difficult and error-prone. This article dwells on some of the points that need to be borne in mind when writing such software to make it easy and pleasant to use. Additionally, a project, carried out by the author, is described, which makes use of a free-form program access approach, as opposed to a purely menu-driven one. Two potential extensions of this approach are then described.

  4. Fault tolerant microcomputer based alarm annunciator for Dhruva reactor

    International Nuclear Information System (INIS)

    Chandra, A.K.

    1988-01-01

    The Dhruva alarm annunciator displays the status of 624 alarm points on an array of display windows using the standard ringback sequence. Recognizing the need for a very high availability, the system is implemented as a fault tolerant configuration. The annunciator is partitioned into three identical units; each unit is implemented using two microcomputers wired in a hot standby mode. In the event of one computer malfunctioning, the standby computer takes over control in a bouncefree transfer. The use of microprocessors has helped built-in flexibility in the system. The system also provides built-in capability to resolve the sequence of occurrence of events and conveys this information to another system for display on a CRT. This report describes the system features, fault tolerant organisation used and the hardware and software developed for the annunciation function. (author). 8 figs

  5. Data flow methods for dynamic system simulation - A CSSL-IV microcomputer network interface

    Science.gov (United States)

    Makoui, A.; Karplus, W. J.

    1983-01-01

    A major problem in employing networks of microcomputers for the real-time simulation of complex systems is to allocate computational tasks to the various microcomputers in such a way that idle time and time lost in interprocess communication is minimized. The research reported in this paper is directed to the development of a software interface between a higher-level simulation language and a network of microcomputers. A CSSL-IV source program is translated to a data flow graph. This graph is then analyzed automatically so as to allocate computing tasks to the various processors.

  6. Application of a 24-digit microcomputer for construction of the VEPP-4 complex distributed control systems

    International Nuclear Information System (INIS)

    Aleshaev, A.N.; Belov, S.D.; Levicheva, B.V.

    1985-01-01

    In connection with moral and physical ageing of ODRA 1325-ODRA 1305 computers, used in the VEPP-4 control system, a possibility and ways of fully distributed control system creating at this facility are considered. ODRENOK microcomputer which is the CAMAC module, placed into the crate controller position is developed within the framework of these activities. The microcomputer is made on 130 integral circuits and realizes the complete ODRA-1300 computer command set. The main parameters of ODRENOK microcomputer are: mass memory - 64 K of 24 digit words; the average speed is 200-300 thousand operationsa second; microprogram operation with a floating point

  7. Reliability analysis of microcomputer boards and computer based systems important to safety of nuclear plants

    International Nuclear Information System (INIS)

    Shrikhande, S.V.; Patil, V.K.; Ganesh, G.; Biswas, B.; Patil, R.K.

    2010-01-01

    Computer Based Systems (CBS) are employed in Indian nuclear plants for protection, control and monitoring purpose. For forthcoming CBS, Reactor Control Division has designed and developed a new standardized family of microcomputer boards qualified to stringent requirements of nuclear industry. These boards form the basic building blocks of CBS. Reliability analysis of these boards is being carried out using analysis package based on MIL-STD-217Plus methodology. The estimated failure rate values of these standardized microcomputer boards will be useful for reliability assessment of these systems. The paper presents reliability analysis of microcomputer boards and case study of a CBS system built using these boards. (author)

  8. FECAP - FINITE ELEMENT COMPOSITE ANALYSIS PROGRAM FOR A MICROCOMPUTER

    Science.gov (United States)

    Bowles, D. E.

    1994-01-01

    Advanced composite materials have gained use in the aerospace industry over the last 20 years because of their high specific strength and stiffness, and low coefficient of thermal expansion. Design of composite structures requires the analysis of composite material behavior. The Finite Element Composite Analysis Program, FECAP, is a special purpose finite element analysis program for analyzing composite material behavior with a microcomputer. Composite materials, in regard to this program, are defined as the combination of at least two distinct materials to form one nonhomogeneous anisotropic material. FECAP assumes a state of generalized plane strain exists in a material consisting of two or more orthotropic phases, subjected to mechanical and/or thermal loading. The finite element formulation used in FECAP is displacement based and requires the minimization of the total potential energy for each element with respect to the unknown variables. This procedure leads to a set of linear simultaneous equations relating the unknown nodal displacements to the applied loads. The equations for each element are assembled into a global system, the boundary conditions are applied, and the system is solved for the nodal displacements. The analysis may be performed using either 4-mode linear or 8-mode quadratic isoparametric elements. Output includes the nodal displacements, and the element stresses and strains. FECAP was written for a Hewlett Packard HP9000 Series 200 Microcomputer with the HP Basic operating system. It was written in HP BASIC 3.0 and requires approximately 0.5 Mbytes of RAM in addition to what is required for the operating system. A math coprocessor card is highly recommended. FECAP was developed in 1988.

  9. Ionospheric scintillation forecasting model based on NN-PSO technique

    Science.gov (United States)

    Sridhar, M.; Venkata Ratnam, D.; Padma Raju, K.; Sai Praharsha, D.; Saathvika, K.

    2017-09-01

    The forecasting and modeling of ionospheric scintillation effects are crucial for precise satellite positioning and navigation applications. In this paper, a Neural Network model, trained using Particle Swarm Optimization (PSO) algorithm, has been implemented for the prediction of amplitude scintillation index (S4) observations. The Global Positioning System (GPS) and Ionosonde data available at Darwin, Australia (12.4634° S, 130.8456° E) during 2013 has been considered. The correlation analysis between GPS S4 and Ionosonde drift velocities (hmf2 and fof2) data has been conducted for forecasting the S4 values. The results indicate that forecasted S4 values closely follow the measured S4 values for both the quiet and disturbed conditions. The outcome of this work will be useful for understanding the ionospheric scintillation phenomena over low latitude regions.

  10. Techniques for studies of unbinned model independent CP violation

    Energy Technology Data Exchange (ETDEWEB)

    Bedford, Nicholas; Weisser, Constantin; Parkes, Chris; Gersabeck, Marco; Brodzicka, Jolanta; Chen, Shanzhen [University of Manchester (United Kingdom)

    2016-07-01

    Charge-Parity (CP) violation is a known part of the Standard Model and has been observed and measured in both the B and K meson systems. The observed levels, however, are insufficient to explain the observed matter-antimatter asymmetry in the Universe, and so other sources need to be found. One area of current investigation is the D meson system, where predicted levels of CP violation are much lower than in the B and K meson systems. This means that more sensitive methods are required when searching for CP violation in this system. Several unbinned model independent methods have been proposed for this purpose, all of which need to be optimised and their sensitivities compared.

  11. Suitability of sheet bending modelling techniques in CAPP applications

    OpenAIRE

    Streppel, A.H.; de Vin, L.J.; de Vin, L.J.; Brinkman, J.; Brinkman, J.; Kals, H.J.J.

    1993-01-01

    The use of CNC machine tools, together with decreasing lot sizes and stricter tolerance prescriptions, has led to changes in sheet-metal part manufacturing. In this paper, problems introduced by the difference between the actual material behaviour and the results obtained from analytical models and FEM simulations are discussed against the background of the required predictable accuracy in small-batch part manufacturing and FMS environments. The topics are limited to those relevant to bending...

  12. Solving microwave heating model using Hermite-Pade approximation technique

    International Nuclear Information System (INIS)

    Makinde, O.D.

    2005-11-01

    We employ the Hermite-Pade approximation method to explicitly construct the approximate solution of steady state reaction- diffusion equations with source term that arises in modeling microwave heating in an infinite slab with isothermal walls. In particular, we consider the case where the source term decreases spatially and increases with temperature. The important properties of the temperature fields including bifurcations and thermal criticality are discussed. (author)

  13. PLATO: PSF modelling using a micro-scanning technique

    Directory of Open Access Journals (Sweden)

    Ouazzani R-M.

    2015-01-01

    Full Text Available The PLATO space mission is designed to detect telluric planets in the habitable zone of solar type stars, and simultaneously characterise the host star using ultra high precision photometry. The photometry will be performed on board using weighted masks. However, to reach the required precision, corrections will have to be performed by the ground segment and will rely on precise knowledge of the instrument PSF (Point Spread Function. We here propose to model the PSF using a microscanning method.

  14. Three-dimensional imaging using microcomputed tomography for studying tooth macromorphology.

    Science.gov (United States)

    Plotino, Gianluca; Grande, Nicola M; Pecci, Raffaella; Bedini, Rossella; Pameijer, Cornelis H; Somma, Francesco

    2006-11-01

    The authors conducted a study to demonstrate potential applications of microcomputed tomography (microCT) in the analysis of tooth morphology. The authors selected for microCT analysis five maxillary first molars with a second canal in the mesiobuccal (MB) root, five mandibular first molars with a mesial root possessing a considerable curvature and five single-canal premolars with complicated apical anatomy. The hardware device used in this study was a desktop X-ray microfocus CT scanner (SkyScan 1072, SkyScan bvba, Aartselaar, Belgium). The authors obtained a three-dimensional image from each of the 15 teeth. In three cases, the MB canals coalesced into one canal, while in the other two molars the canals were separate. Four of the five mandibular molars exhibited a single canal in the mesial root, which had a broad, flat appearance in a mesiodistal dimension. In the premolar teeth, the canals were independent; however, the apical delta and ramifications of the root canals were obvious, yet intricate. MicroCT offers a reproducible technique for 3-D noninvasive assessment of root canal systems. While this technique is not suitable for clinical use, it can be applied to improve preclinical training and analysis of fundamental procedures in endodontic and restorative treatment.

  15. Modern modelling techniques are data hungry: a simulation study for predicting dichotomous endpoints.

    Science.gov (United States)

    van der Ploeg, Tjeerd; Austin, Peter C; Steyerberg, Ewout W

    2014-12-22

    Modern modelling techniques may potentially provide more accurate predictions of binary outcomes than classical techniques. We aimed to study the predictive performance of different modelling techniques in relation to the effective sample size ("data hungriness"). We performed simulation studies based on three clinical cohorts: 1282 patients with head and neck cancer (with 46.9% 5 year survival), 1731 patients with traumatic brain injury (22.3% 6 month mortality) and 3181 patients with minor head injury (7.6% with CT scan abnormalities). We compared three relatively modern modelling techniques: support vector machines (SVM), neural nets (NN), and random forests (RF) and two classical techniques: logistic regression (LR) and classification and regression trees (CART). We created three large artificial databases with 20 fold, 10 fold and 6 fold replication of subjects, where we generated dichotomous outcomes according to different underlying models. We applied each modelling technique to increasingly larger development parts (100 repetitions). The area under the ROC-curve (AUC) indicated the performance of each model in the development part and in an independent validation part. Data hungriness was defined by plateauing of AUC and small optimism (difference between the mean apparent AUC and the mean validated AUC techniques. The RF, SVM and NN models showed instability and a high optimism even with >200 events per variable. Modern modelling techniques such as SVM, NN and RF may need over 10 times as many events per variable to achieve a stable AUC and a small optimism than classical modelling techniques such as LR. This implies that such modern techniques should only be used in medical prediction problems if very large data sets are available.

  16. A titration model for evaluating calcium hydroxide removal techniques

    Directory of Open Access Journals (Sweden)

    Mark PHILLIPS

    2015-02-01

    Full Text Available Objective Calcium hydroxide (Ca(OH2 has been used in endodontics as an intracanal medicament due to its antimicrobial effects and its ability to inactivate bacterial endotoxin. The inability to totally remove this intracanal medicament from the root canal system, however, may interfere with the setting of eugenol-based sealers or inhibit bonding of resin to dentin, thus presenting clinical challenges with endodontic treatment. This study used a chemical titration method to measure residual Ca(OH2 left after different endodontic irrigation methods. Material and Methods Eighty-six human canine roots were prepared for obturation. Thirty teeth were filled with known but different amounts of Ca(OH2 for 7 days, which were dissolved out and titrated to quantitate the residual Ca(OH2 recovered from each root to produce a standard curve. Forty-eight of the remaining teeth were filled with equal amounts of Ca(OH2 followed by gross Ca(OH2 removal using hand files and randomized treatment of either: 1 Syringe irrigation; 2 Syringe irrigation with use of an apical file; 3 Syringe irrigation with added 30 s of passive ultrasonic irrigation (PUI, or 4 Syringe irrigation with apical file and PUI (n=12/group. Residual Ca(OH2 was dissolved with glycerin and titrated to measure residual Ca(OH2 left in the root. Results No method completely removed all residual Ca(OH2. The addition of 30 s PUI with or without apical file use removed Ca(OH2 significantly better than irrigation alone. Conclusions This technique allowed quantification of residual Ca(OH2. The use of PUI (with or without apical file resulted in significantly lower Ca(OH2 residue compared to irrigation alone.

  17. Modern EMC analysis techniques II models and applications

    CERN Document Server

    Kantartzis, Nikolaos V

    2008-01-01

    The objective of this two-volume book is the systematic and comprehensive description of the most competitive time-domain computational methods for the efficient modeling and accurate solution of modern real-world EMC problems. Intended to be self-contained, it performs a detailed presentation of all well-known algorithms, elucidating on their merits or weaknesses, and accompanies the theoretical content with a variety of applications. Outlining the present volume, numerical investigations delve into printed circuit boards, monolithic microwave integrated circuits, radio frequency microelectro

  18. Robust image modeling technique with a bioluminescence image segmentation application

    Science.gov (United States)

    Zhong, Jianghong; Wang, Ruiping; Tian, Jie

    2009-02-01

    A robust pattern classifier algorithm for the variable symmetric plane model, where the driving noise is a mixture of a Gaussian and an outlier process, is developed. The veracity and high-speed performance of the pattern recognition algorithm is proved. Bioluminescence tomography (BLT) has recently gained wide acceptance in the field of in vivo small animal molecular imaging. So that it is very important for BLT to how to acquire the highprecision region of interest in a bioluminescence image (BLI) in order to decrease loss of the customers because of inaccuracy in quantitative analysis. An algorithm in the mode is developed to improve operation speed, which estimates parameters and original image intensity simultaneously from the noise corrupted image derived from the BLT optical hardware system. The focus pixel value is obtained from the symmetric plane according to a more realistic assumption for the noise sequence in the restored image. The size of neighborhood is adaptive and small. What's more, the classifier function is base on the statistic features. If the qualifications for the classifier are satisfied, the focus pixel intensity is setup as the largest value in the neighborhood.Otherwise, it will be zeros.Finally,pseudo-color is added up to the result of the bioluminescence segmented image. The whole process has been implemented in our 2D BLT optical system platform and the model is proved.

  19. Advancing botnet modeling techniques for military and security simulations

    Science.gov (United States)

    Banks, Sheila B.; Stytz, Martin R.

    2011-06-01

    Simulation environments serve many purposes, but they are only as good as their content. One of the most challenging and pressing areas that call for improved content is the simulation of bot armies (botnets) and their effects upon networks and computer systems. Botnets are a new type of malware, a type that is more powerful and potentially dangerous than any other type of malware. A botnet's power derives from several capabilities including the following: 1) the botnet's capability to be controlled and directed throughout all phases of its activity, 2) a command and control structure that grows increasingly sophisticated, and 3) the ability of a bot's software to be updated at any time by the owner of the bot (a person commonly called a bot master or bot herder.) Not only is a bot army powerful and agile in its technical capabilities, a bot army can be extremely large, can be comprised of tens of thousands, if not millions, of compromised computers or it can be as small as a few thousand targeted systems. In all botnets, their members can surreptitiously communicate with each other and their command and control centers. In sum, these capabilities allow a bot army to execute attacks that are technically sophisticated, difficult to trace, tactically agile, massive, and coordinated. To improve our understanding of their operation and potential, we believe that it is necessary to develop computer security simulations that accurately portray bot army activities, with the goal of including bot army simulations within military simulation environments. In this paper, we investigate issues that arise when simulating bot armies and propose a combination of the biologically inspired MSEIR infection spread model coupled with the jump-diffusion infection spread model to portray botnet propagation.

  20. Precision and trueness of dental models manufactured with different 3-dimensional printing techniques.

    Science.gov (United States)

    Kim, Soo-Yeon; Shin, Yoo-Seok; Jung, Hwi-Dong; Hwang, Chung-Ju; Baik, Hyoung-Seon; Cha, Jung-Yul

    2018-01-01

    In this study, we assessed the precision and trueness of dental models printed with 3-dimensional (3D) printers via different printing techniques. Digital reference models were printed 5 times using stereolithography apparatus (SLA), digital light processing (DLP), fused filament fabrication (FFF), and the PolyJet technique. The 3D printed models were scanned and evaluated for tooth, arch, and occlusion measurements. Precision and trueness were analyzed with root mean squares (RMS) for the differences in each measurement. Differences in measurement variables among the 3D printing techniques were analyzed by 1-way analysis of variance (α = 0.05). Except in trueness of occlusion measurements, there were significant differences in all measurements among the 4 techniques (P techniques exhibited significantly different mean RMS values of precision than the SLA (88 ± 14 μm) and FFF (99 ± 14 μm) techniques (P techniques (P techniques (P techniques: SLA (107 ± 11 μm), DLP (143 ± 8 μm), FFF (188 ± 14 μm), and PolyJet (78 ± 9 μm) (P techniques exhibited significantly different mean RMS values of trueness than DLP (469 ± 49 μm) and FFF (409 ± 36 μm) (P techniques showed significant differences in precision of all measurements and in trueness of tooth and arch measurements. The PolyJet and DLP techniques were more precise than the FFF and SLA techniques, with the PolyJet technique having the highest accuracy. Copyright © 2017 American Association of Orthodontists. Published by Elsevier Inc. All rights reserved.

  1. The microcomputer workstation - An alternate hardware architecture for remotely sensed image analysis

    Science.gov (United States)

    Erickson, W. K.; Hofman, L. B.; Donovan, W. E.

    1984-01-01

    Difficulties regarding the digital image analysis of remotely sensed imagery can arise in connection with the extensive calculations required. In the past, an expensive large to medium mainframe computer system was needed for performing these calculations. For image-processing applications smaller minicomputer-based systems are now used by many organizations. The costs for such systems are still in the range from $100K to $300K. Recently, as a result of new developments, the use of low-cost microcomputers for image processing and display systems appeared to have become feasible. These developments are related to the advent of the 16-bit microprocessor and the concept of the microcomputer workstation. Earlier 8-bit microcomputer-based image processing systems are briefly examined, and a computer workstation architecture is discussed. Attention is given to a microcomputer workstation developed by Stanford University, and the design and implementation of a workstation network.

  2. Medical and administrative management of a nuclear medicine department with a microcomputer

    International Nuclear Information System (INIS)

    Legras, B.; Kohler, F.

    1984-01-01

    The use of a microcomputer for data management in a department of Nuclear Medicine has allowed to reduce considerably office work, and supplies the physicians with very useful statistics on the investigations carried out [fr

  3. Digital TV-echelle spectrograph for simultaneous multielemental analysis using microcomputer control

    International Nuclear Information System (INIS)

    Davidson, J.B.; Case, A.L.

    1980-12-01

    A digital TV-echelle spectrograph with microcomputer control was developed for simultaneous multielemental analysis. The optical system is a commercially available unit originally equipped for film and photomultiplier (single element) readout. The film port was adapted for the intensifier camera. The camera output is digitized and stored in a microcomputer-controlled, 512 x 512 x 12 bit memory and image processor. Multiple spectra over the range of 200 to 800 nm are recorded in a single exposure. Spectra lasting from nanoseconds to seconds are digitized and stored in 0.033 s and displayed on a TV monitor. An inexpensive microcomputer controls the exposure, reads and displays the intensity of predetermined spectral lines, and calculates wavelengths of unknown lines. The digital addresses of unknown lines are determined by superimposing a cursor on the TV display. The microcomputer also writes into memory wavelength fiducial marks for alignment of the TV camera

  4. Multivariate moment closure techniques for stochastic kinetic models

    Energy Technology Data Exchange (ETDEWEB)

    Lakatos, Eszter, E-mail: e.lakatos13@imperial.ac.uk; Ale, Angelique; Kirk, Paul D. W.; Stumpf, Michael P. H., E-mail: m.stumpf@imperial.ac.uk [Department of Life Sciences, Centre for Integrative Systems Biology and Bioinformatics, Imperial College London, London SW7 2AZ (United Kingdom)

    2015-09-07

    Stochastic effects dominate many chemical and biochemical processes. Their analysis, however, can be computationally prohibitively expensive and a range of approximation schemes have been proposed to lighten the computational burden. These, notably the increasingly popular linear noise approximation and the more general moment expansion methods, perform well for many dynamical regimes, especially linear systems. At higher levels of nonlinearity, it comes to an interplay between the nonlinearities and the stochastic dynamics, which is much harder to capture correctly by such approximations to the true stochastic processes. Moment-closure approaches promise to address this problem by capturing higher-order terms of the temporally evolving probability distribution. Here, we develop a set of multivariate moment-closures that allows us to describe the stochastic dynamics of nonlinear systems. Multivariate closure captures the way that correlations between different molecular species, induced by the reaction dynamics, interact with stochastic effects. We use multivariate Gaussian, gamma, and lognormal closure and illustrate their use in the context of two models that have proved challenging to the previous attempts at approximating stochastic dynamics: oscillations in p53 and Hes1. In addition, we consider a larger system, Erk-mediated mitogen-activated protein kinases signalling, where conventional stochastic simulation approaches incur unacceptably high computational costs.

  5. Using simulation models to evaluate ape nest survey techniques.

    Directory of Open Access Journals (Sweden)

    Ryan H Boyko

    Full Text Available BACKGROUND: Conservationists frequently use nest count surveys to estimate great ape population densities, yet the accuracy and precision of the resulting estimates are difficult to assess. METHODOLOGY/PRINCIPAL FINDINGS: We used mathematical simulations to model nest building behavior in an orangutan population to compare the quality of the population size estimates produced by two of the commonly used nest count methods, the 'marked recount method' and the 'matrix method.' We found that when observers missed even small proportions of nests in the first survey, the marked recount method produced large overestimates of the population size. Regardless of observer reliability, the matrix method produced substantial overestimates of the population size when surveying effort was low. With high observer reliability, both methods required surveying approximately 0.26% of the study area (0.26 km(2 out of 100 km(2 in this simulation to achieve an accurate estimate of population size; at or above this sampling effort both methods produced estimates within 33% of the true population size 50% of the time. Both methods showed diminishing returns at survey efforts above 0.26% of the study area. The use of published nest decay estimates derived from other sites resulted in widely varying population size estimates that spanned nearly an entire order of magnitude. The marked recount method proved much better at detecting population declines, detecting 5% declines nearly 80% of the time even in the first year of decline. CONCLUSIONS/SIGNIFICANCE: These results highlight the fact that neither nest surveying method produces highly reliable population size estimates with any reasonable surveying effort, though either method could be used to obtain a gross population size estimate in an area. Conservation managers should determine if the quality of these estimates are worth the money and effort required to produce them, and should generally limit surveying effort to

  6. Full Semantics Preservation in Model Transformation - A Comparison of Proof Techniques

    NARCIS (Netherlands)

    Hülsbusch, Mathias; König, Barbara; Rensink, Arend; Semenyak, Maria; Soltenborn, Christian; Wehrheim, Heike

    Model transformation is a prime technique in modern, model-driven software design. One of the most challenging issues is to show that the semantics of the models is not affected by the transformation. So far, there is hardly any research into this issue, in particular in those cases where the source

  7. Development of Reservoir Characterization Techniques and Production Models for Exploiting Naturally Fractured Reservoirs

    Energy Technology Data Exchange (ETDEWEB)

    Wiggins, Michael L.; Brown, Raymon L.; Civan, Frauk; Hughes, Richard G.

    2001-08-15

    Research continues on characterizing and modeling the behavior of naturally fractured reservoir systems. Work has progressed on developing techniques for estimating fracture properties from seismic and well log data, developing naturally fractured wellbore models, and developing a model to characterize the transfer of fluid from the matrix to the fracture system for use in the naturally fractured reservoir simulator.

  8. Techniques to maximize software reliability in radiation fields

    International Nuclear Information System (INIS)

    Eichhorn, G.; Piercey, R.B.

    1986-01-01

    Microprocessor system failures due to memory corruption by single event upsets (SEUs) and/or latch-up in RAM or ROM memory are common in environments where there is high radiation flux. Traditional methods to harden microcomputer systems against SEUs and memory latch-up have usually involved expensive large scale hardware redundancy. Such systems offer higher reliability, but they tend to be more complex and non-standard. At the Space Astronomy Laboratory the authors have developed general programming techniques for producing software which is resistant to such memory failures. These techniques, which may be applied to standard off-the-shelf hardware, as well as custom designs, include an implementation of Maximally Redundant Software (MRS) model, error detection algorithms and memory verification and management

  9. Personal recommender systems for learners in lifelong learning: requirements, techniques and model

    NARCIS (Netherlands)

    Drachsler, Hendrik; Hummel, Hans; Koper, Rob

    2007-01-01

    Drachsler, H., Hummel, H. G. K., & Koper, R. (2008). Personal recommender systems for learners in lifelong learning: requirements, techniques and model. International Journal of Learning Technology, 3(4), 404-423.

  10. Prediction of intracranial findings on CT-scans by alternative modelling techniques

    NARCIS (Netherlands)

    T. van der Ploeg (Tjeerd); M. Smits (Marion); D.W.J. Dippel (Diederik); M.G.M. Hunink (Myriam); E.W. Steyerberg (Ewout)

    2011-01-01

    textabstractBackground: Prediction rules for intracranial traumatic findings in patients with minor head injury are designed to reduce the use of computed tomography (CT) without missing patients at risk for complications. This study investigates whether alternative modelling techniques might

  11. A Shell/3D Modeling Technique for the Analysis of Delaminated Composite Laminates

    Science.gov (United States)

    Krueger, Ronald; OBrien, T. Kevin

    2000-01-01

    A shell/3D modeling technique was developed for which a local solid finite element model is used only in the immediate vicinity of the delamination front. The goal was to combine the accuracy of the full three-dimensional solution with the computational efficiency of a shell finite element model. Multi-point constraints provided a kinematically compatible interface between the local 3D model and the global structural model which has been meshed with shell finite elements. Double Cantilever Beam, End Notched Flexure, and Single Leg Bending specimens were analyzed first using full 3D finite element models to obtain reference solutions. Mixed mode strain energy release rate distributions were computed using the virtual crack closure technique. The analyses were repeated using the shell/3D technique to study the feasibility for pure mode I, mode II and mixed mode I/II cases. Specimens with a unidirectional layup and with a multidirectional layup were simulated. For a local 3D model, extending to a minimum of about three specimen thicknesses on either side of the delamination front, the results were in good agreement with mixed mode strain energy release rates obtained from computations where the entire specimen had been modeled with solid elements. For large built-up composite structures the shell/3D modeling technique offers a great potential for reducing the model size, since only a relatively small section in the vicinity of the delamination front needs to be modeled with solid elements.

  12. System Response Analysis and Model Order Reduction, Using Conventional Method, Bond Graph Technique and Genetic Programming

    Directory of Open Access Journals (Sweden)

    Lubna Moin

    2009-04-01

    Full Text Available This research paper basically explores and compares the different modeling and analysis techniques and than it also explores the model order reduction approach and significance. The traditional modeling and simulation techniques for dynamic systems are generally adequate for single-domain systems only, but the Bond Graph technique provides new strategies for reliable solutions of multi-domain system. They are also used for analyzing linear and non linear dynamic production system, artificial intelligence, image processing, robotics and industrial automation. This paper describes a unique technique of generating the Genetic design from the tree structured transfer function obtained from Bond Graph. This research work combines bond graphs for model representation with Genetic programming for exploring different ideas on design space tree structured transfer function result from replacing typical bond graph element with their impedance equivalent specifying impedance lows for Bond Graph multiport. This tree structured form thus obtained from Bond Graph is applied for generating the Genetic Tree. Application studies will identify key issues and importance for advancing this approach towards becoming on effective and efficient design tool for synthesizing design for Electrical system. In the first phase, the system is modeled using Bond Graph technique. Its system response and transfer function with conventional and Bond Graph method is analyzed and then a approach towards model order reduction is observed. The suggested algorithm and other known modern model order reduction techniques are applied to a 11th order high pass filter [1], with different approach. The model order reduction technique developed in this paper has least reduction errors and secondly the final model retains structural information. The system response and the stability analysis of the system transfer function taken by conventional and by Bond Graph method is compared and

  13. Micro-computed tomography of false starts produced on bone by different hand-saws.

    Science.gov (United States)

    Pelletti, Guido; Viel, Guido; Fais, Paolo; Viero, Alessia; Visentin, Sindi; Miotto, Diego; Montisci, Massimo; Cecchetto, Giovanni; Giraudo, Chiara

    2017-05-01

    The analysis of macro- and microscopic characteristics of saw marks on bones can provide useful information about the class of the tool utilized to produce the injury. The aim of the present study was to test micro-computed tomography (micro-CT) for the analysis of false starts experimentally produced on 32 human bone sections using 4 different hand-saws in order to verify the potential utility of micro-CT for distinguishing false starts produced by different saws and to correlate the morphology of the tool with that of the bone mark. Each sample was analysed through stereomicroscopy and micro-CT. Stereomicroscopic analysis allowed the identification of the false starts and the detection of the number of tool marks left by each saw. Micro-CT scans, through the integration of 3D renders and multiplanar reconstructions (MPR), allowed the identification of the shape of each false start correlating it to the injuring tool. Our results suggest that micro-CT could be a useful technique for assessing false starts produced by different classes of saws, providing accurate morphological profiles of the bone marks with all the advantages of high resolution 3D imaging (e.g., high accuracy, non-destructive analysis, preservation and documentation of evidence). However, further studies are necessary to integrate qualitative data with quantitative metrical analysis in order to further characterize the false start and the related injuring tool. Copyright © 2017 Elsevier B.V. All rights reserved.

  14. Seismic loadings and measured parameters combined by power spectral density methods on a microcomputer

    International Nuclear Information System (INIS)

    Christian, J.T.; Harizi, P.D.; Hall, J.R. Jr.

    1987-01-01

    The qualification of equipment to resist the design seismic loading may require determining what motions will be felt at the points of support for the equipment, and many other calculations of seismic and vibrational effects require that dynamic loadings be established at points other than those at which they have been prescribed or calculated in the course of structural analysis. These calculations have always presented a number of experimental and theoretical difficulties. In particular the engineer needs a procedure for expressing the seismic excitation in a form that is convenient for calculation, a technique for determining the amplification functions for the structures and equipment through which the seismic loadings are transmitted, and a computational implementation that is convenient and efficient. This paper describes the use of previously published algorithms to develop the power spectral density function for a seismic excitation previously described in terms of its response spectrum and to compute the amplification function for a system whose modal shapes, but not its mass distribution, are known. The resulting computer program, which runs on an IBM PC/XT microcomputer, includes graphical screen output that permits the engineer to determine the current state of the calculations and to modify the course of the calculations

  15. Prototype particulate stack sampler with single-cut nozzle and microcomputer calculating/display system

    International Nuclear Information System (INIS)

    Eler, J.C.; Littlefield, L.G.; Tillery, M.I.

    1979-01-01

    A prototype particulate stack sampler (PPSS) has been developed to improve on the existing EPA Method 5 sampling apparatus. Its primary features are (1) higher sampling rate (56 1/min); (2) display (on demand) of all required variables and calculated values by a microcomputer-based calculating and display system; (3) continuous stack gas moisture determination; (4) a virtual impactor nozzle with 3 μm mass median diameter cutpoint which collects fine and coarse particle fractions on separate glass fiber filters; (5) a variable-area inlet to maintain isokinetic sampling conditions; and (6) conversion to stainless steel components from the glass specified by EPA Method 5. The basic sampling techniques of EPA Method 5 have been retained; however, versatility in the form of optional in-stack filters and general modernization of the stack sampler have been provided in the prototype design. Laboratory testing with monodisperse dye aerosols has shown the present variable inlet, virtual impactor nozzle to have a collection efficiency which is less than 77% and significant wall losses. This is primarily due to lack of symmetry in this rectangular jet impactor and short transition lengths dictated by physical design constraints (required passage of the nozzle through a 7.6 cm (3 in) diameter stack port). Electronic components have shown acceptable service in laboratory testing although no field testing of the prototype under a broad range of temperature, humidity, and SO 2 concentration has been undertaken

  16. Propulsion modeling techniques and applications for the NASA Dryden X-30 real-time simulator

    Science.gov (United States)

    Hicks, John W.

    1991-01-01

    An overview is given of the flight planning activities to date in the current National Aero-Space Plane (NASP) program. The government flight-envelope expansion concept and other design flight operational assessments are discussed. The NASA Dryden NASP real-time simulator configuration is examined and hypersonic flight planning simulation propulsion modeling requirements are described. The major propulsion modeling techniques developed by the Edwards flight test team are outlined, and the application value of techniques for developmental hypersonic vehicles are discussed.

  17. Modelling the effects of the sterile insect technique applied to Eldana saccharina Walker in sugarcane

    Directory of Open Access Journals (Sweden)

    L Potgieter

    2012-12-01

    Full Text Available A mathematical model is formulated for the population dynamics of an Eldana saccharina Walker infestation of sugarcane under the influence of partially sterile released insects. The model describes the population growth of and interaction between normal and sterile E.saccharina moths in a temporally variable, but spatially homogeneous environment. The model consists of a deterministic system of difference equations subject to strictly positive initial data. The primary objective of this model is to determine suitable parameters in terms of which the above population growth and interaction may be quantified and according to which E.saccharina infestation levels and the associated sugarcane damage may be measured. Although many models have been formulated in the past describing the sterile insect technique, few of these models describe the technique for Lepidopteran species with more than one life stage and where F1-sterility is relevant. In addition, none of these models consider the technique when fully sterile females and partially sterile males are being released. The model formulated is also the first to describe the technique applied specifically to E.saccharina, and to consider the economic viability of applying the technique to this species. Pertinent decision support is provided to farm managers in terms of the best timing for releases, release ratios and release frequencies.

  18. Construction of Eh-pH and other stability diagrams of uranium in a multicomponent system with a microcomputer

    International Nuclear Information System (INIS)

    Haung, H.; Cuentas, L.

    1989-01-01

    Stability diagrams for a multicomponent system in aqueous chemistry provide important information for hydrometallurgy, corrosion science, geochemistry and environmental science. Two distinct types of diagrams constructed are predominance diagrams and distribution diagrams. The ability to construct stability diagrams easily, quickly and accurately is most helpful in research and development and in academic programs. The use of a microcomputer is handicapped by slow speed and limited memory. Developing program methods that promote easy calculation and plot the diagram directly on a CRT or a plotter is a primary concern. As presented in this paper, the calculation of equilibrium and boundary constraints, combined with isolation of stability areas, works well for constructing predominance diagrams. Equilibrium constraints can be obtained based on free energies of formation. Boundary constraints for the ligand component are the boundary of the diagram, and constraints for the main component are the surrounding lines of each dominant ligand. Other considerations regarding the chemical model, mathematics computation and the use of microcomputers pertaining to diagram construction are discussed. The uranium in a multicomponent system is used for demonstration

  19. Validation of a musculoskeletal model of lifting and its application for biomechanical evaluation of lifting techniques.

    Science.gov (United States)

    Mirakhorlo, Mojtaba; Azghani, Mahmood Reza; Kahrizi, Sedighe

    2014-01-01

    Lifting methods, including standing stance and techniques have wide effects on spine loading and stability. Previous studies explored lifting techniques in many biomechanical terms and documented changes in muscular and postural response of body as a function of techniques .However, the impact of standing stance and lifting technique on human musculoskeletal had not been investigated concurrently. A whole body musculoskeletal model of lifting had been built in order to evaluate standing stance impact on muscle activation patterns and spine loading during each distinctive lifting technique. Verified model had been used in different stances width during squat, stoop and semi-squat lifting for examining the effect of standing stance on each lifting technique. The model muscle's activity was validated by experimental muscle EMGs resulting in Pearson's coefficients of greater than 0.8. Results from analytical analyses show that the effect of stance width on biomechanical parameters consists in the lifting technique, depending on what kind of standing stance was used. Standing stance in each distinctive lifting technique exhibit positive and negative aspects and it can't be recommended either one as being better in terms of biomechanical parameters.

  20. Inexpensive remote video surveillance system with microcomputer and solar cells

    International Nuclear Information System (INIS)

    Guevara Betancourt, Edder

    2013-01-01

    A low-cost prototype is developed with a RPI plate for remote video surveillance. Additionally, the theoretical basis to provide energy independence have developed through solar cells and a battery bank. Some existing commercial monitoring systems are studied and analyzed, components such as: cameras, communication devices (WiFi and 3G), free software packages for video surveillance, control mechanisms and theory remote photovoltaic systems. A number of steps are developed to implement the module and install, configure and test each of the elements of hardware and software that make up the module, exploring the feasibility of providing intelligence to the system using the software chosen. Events that have been generated by motion detection have been simple, intuitive way to view, archive and extract. The implementation of the module by a microcomputer video surveillance and motion detection software (Zoneminder) has been an option for a lot of potential; as the platform for monitoring and recording data has provided all the tools to make a robust and secure surveillance. (author) [es

  1. Monte Carlo calculations of electron transport on microcomputers

    International Nuclear Information System (INIS)

    Chung, Manho; Jester, W.A.; Levine, S.H.; Foderaro, A.H.

    1990-01-01

    In the work described in this paper, the Monte Carlo program ZEBRA, developed by Berber and Buxton, was converted to run on the Macintosh computer using Microsoft BASIC to reduce the cost of Monte Carlo calculations using microcomputers. Then the Eltran2 program was transferred to an IBM-compatible computer. Turbo BASIC and Microsoft Quick BASIC have been used on the IBM-compatible Tandy 4000SX computer. The paper shows the running speed of the Monte Carlo programs on the different computers, normalized to one for Eltran2 on the Macintosh-SE or Macintosh-Plus computer. Higher values refer to faster running times proportionally. Since Eltran2 is a one-dimensional program, it calculates energy deposited in a semi-infinite multilayer slab. Eltran2 has been modified to a two-dimensional program called Eltran3 to computer more accurately the case with a point source, a small detector, and a short source-to-detector distance. The running time of Eltran3 is about twice as long as that of Eltran2 for a similar case

  2. Dose projection for nuclear emergency response on a microcomputer

    International Nuclear Information System (INIS)

    Vold, E.; Mauro, J.; Michlewicz, D.

    1984-01-01

    A set of microcomputer programs were developed to implement the offsite steady-state dose projection calculations given in the Standard Operating Procedures for Nuclear Power Plant Emergency Response of the Radiologic Health Branch, California Department of Health Services. The dose projection system is intended for training, education and drill exercises for offsite nuclear emergency response staff. The simplicity of use and accuracy would be adequate for use as a back-up calculation to assist in determining emergency response in the event of an actual nuclear emergency. Nuclide release data and current meteorology is input by the user in response to program prompting. The output is specified from the options: (1) doses, dose rates and duration until the Protective Action Guides (PAGs) are exceeded at 2, 5 and 10 miles downwind; (2) dose and dose rates at any specified location; (3) isopleuths plotted on a site map; (4) user-specified isodoses plotted on a site map; (5) isodoses plotted equal to the PAGs. Six site maps to characterize the three current nuclear reactor sites in California can be displayed in full color for the plotting programs. Additional site maps can be added with an available low-cost digitizing tablet. Five separate dose pathways which cover the expected range in critical exposure pathways during a nuclear power plant accident can be calculated or plotted on a selected site map. Some dose projection results are presented, discussed and compared to other simple dose projection methods

  3. Practical use of a microcomputer for ventilation calculations

    International Nuclear Information System (INIS)

    D'Albrand, N.; Froger, C.; Josien, J.P.

    1983-01-01

    Calculations of ventilation networks are necessary to elaborate the projects and to handle the gradual development of a ventilation network. The means which have been used up to the present to tackle these problems are the simulator and the computer, each with its own advantages and disadvantages. These means can be improved by considering the following needs of the user: short response time in the calculation of the state of a network; easy data input and presentation of the results (diagram, visual display); keeping records of the results. Cerchar has developed a program for the calculation of ventilation networks for use with a microcomputer psi 80 (Kontron) which has been constructed around a microprocessor Z 80. The data of the network to be calculated are entered by a keyboard and stored on a small disc; the data and the results can be displayed on a screen or produced as a print-out. This program is suitable for the calculation of networks composed of 400 branches and 300 nodes and comprising up to 9 different curves of characteristics as far as the mine fans are concerned. Cerchar investigates at present the visual display of a network on an interactive graphic terminal which, when perfected, will put at the disposal of the mine operators a tool for use by people other than data processing experts; this enables them to have permanently on the site their network for immediate access to consult it or make new calculations [fr

  4. Microcomputer-controlled flow meter used on a water loop

    International Nuclear Information System (INIS)

    Haniger, L.

    1982-01-01

    The report describes a microcomputer-controlled instrument intended for operational measurement on an experimental water loop. On the basis of pressure and temperature input signals the instrument calculates the specific weight, and for ten operator-selectable measuring channels it calculates the mass flow G(kp/s), or the voluminal flow Q(m 3 /h). On pressing the appropriate push-buttons the built-in display indicates the values of pressure (p) and temperature (t), as well as the values of specific weight γ calculated therefrom. For ten individually selectable channels the instrument displays either the values of the pressure differences of the measuring throttling elements (√Δpsub(i)), or the values of Gsub(i) or Qsub(i) as obtained by calculation. In addition, on pressing the Σ-push-button it summarizes the values of Gsub(i) and Qsub(i) for the selected channels. The device is controlled by an 8085 microprocessor, the analog unit MP 6812 being used as the A/D convertor. The instrument algorithm indicates some possible errors which may concern faults of input signals or mistakes in calculation. (author)

  5. The FORTRAN NALAP code adapted to a microcomputer compiler

    International Nuclear Information System (INIS)

    Lobo, Paulo David de Castro; Borges, Eduardo Madeira; Braz Filho, Francisco Antonio; Guimaraes, Lamartine Nogueira Frutuoso

    2010-01-01

    The Nuclear Energy Division of the Institute for Advanced Studies (IEAv) is conducting the TERRA project (TEcnologia de Reatores Rapidos Avancados), Technology for Advanced Fast Reactors project, aimed at a space reactor application. In this work, to attend the TERRA project, the NALAP code adapted to a microcomputer compiler called Compaq Visual Fortran (Version 6.6) is presented. This code, adapted from the light water reactor transient code RELAP 3B, simulates thermal-hydraulic responses for sodium cooled fast reactors. The strategy to run the code in a PC was divided in some steps mainly to remove unnecessary routines, to eliminate old statements, to introduce new ones and also to include extension precision mode. The source program was able to solve three sample cases under conditions of protected transients suggested in literature: the normal reactor shutdown, with a delay of 200 ms to start the control rod movement and a delay of 500 ms to stop the pumps; reactor scram after transient of loss of flow; and transients protected from overpower. Comparisons were made with results from the time when the NALAP code was acquired by the IEAv, back in the 80's. All the responses for these three simulations reproduced the calculations performed with the CDC compiler in 1985. Further modifications will include the usage of gas as coolant for the nuclear reactor to allow a Closed Brayton Cycle Loop - CBCL - to be used as a heat/electric converter. (author)

  6. The FORTRAN NALAP code adapted to a microcomputer compiler

    Energy Technology Data Exchange (ETDEWEB)

    Lobo, Paulo David de Castro; Borges, Eduardo Madeira; Braz Filho, Francisco Antonio; Guimaraes, Lamartine Nogueira Frutuoso, E-mail: plobo.a@uol.com.b, E-mail: eduardo@ieav.cta.b, E-mail: fbraz@ieav.cta.b, E-mail: guimarae@ieav.cta.b [Instituto de Estudos Avancados (IEAv/CTA), Sao Jose dos Campos, SP (Brazil)

    2010-07-01

    The Nuclear Energy Division of the Institute for Advanced Studies (IEAv) is conducting the TERRA project (TEcnologia de Reatores Rapidos Avancados), Technology for Advanced Fast Reactors project, aimed at a space reactor application. In this work, to attend the TERRA project, the NALAP code adapted to a microcomputer compiler called Compaq Visual Fortran (Version 6.6) is presented. This code, adapted from the light water reactor transient code RELAP 3B, simulates thermal-hydraulic responses for sodium cooled fast reactors. The strategy to run the code in a PC was divided in some steps mainly to remove unnecessary routines, to eliminate old statements, to introduce new ones and also to include extension precision mode. The source program was able to solve three sample cases under conditions of protected transients suggested in literature: the normal reactor shutdown, with a delay of 200 ms to start the control rod movement and a delay of 500 ms to stop the pumps; reactor scram after transient of loss of flow; and transients protected from overpower. Comparisons were made with results from the time when the NALAP code was acquired by the IEAv, back in the 80's. All the responses for these three simulations reproduced the calculations performed with the CDC compiler in 1985. Further modifications will include the usage of gas as coolant for the nuclear reactor to allow a Closed Brayton Cycle Loop - CBCL - to be used as a heat/electric converter. (author)

  7. Effects of mastication on mandibular growth evaluated by microcomputed tomography.

    Science.gov (United States)

    Enomoto, Akiko; Watahiki, Junichi; Yamaguchi, Tetsutaro; Irie, Tarou; Tachikawa, Tetsuhiko; Maki, Koutaro

    2010-02-01

    It is well known that mastication has a significant influence on mandibular growth and development, but the mechanism behind this effect has not yet been clarified. Furthermore, no studies have examined the effects of changes in mastication on the three-dimensional (3D) morphometry of the mandible. The aim of the present study was to investigate the influences of changes in mastication on mandibular growth and morphology. Twenty-five 3-week-old (at the time of weaning) imprinting control region mice were randomly divided into three groups: mice fed a hard diet (HD), mice fed a soft diet (SD), and mice alternately fed hard and soft diets (HSDs) every week for 4 weeks. The morphometry of the mandible was analysed using 3D microcomputed tomography (muCT). Statistical analysis was undertaken using a t-test. muCT analysis showed that the condylar width was significantly greater in the HD group than in the SD group after 1 week. After 4 weeks, mandibular length was significantly longer and ramus height was greater in the HSD group than in the other two groups. Bone volume was significantly less in the SD group than in the other two groups after 4 weeks. These findings suggest that changes in mastication markedly affect mandibular condylar cartilage growth and mandibular morphology. It is considered that dietary education at an early age is important in order to prevent disruption of the development of the mandible.

  8. New sunshine-based models for predicting global solar radiation using PSO (particle swarm optimization) technique

    International Nuclear Information System (INIS)

    Behrang, M.A.; Assareh, E.; Noghrehabadi, A.R.; Ghanbarzadeh, A.

    2011-01-01

    PSO (particle swarm optimization) technique is applied to estimate monthly average daily GSR (global solar radiation) on horizontal surface for different regions of Iran. To achieve this, five new models were developed as well as six models were chosen from the literature. First, for each city, the empirical coefficients for all models were separately determined using PSO technique. The results indicate that new models which are presented in this study have better performance than existing models in the literature for 10 cities from 17 considered cities in this study. It is also shown that the empirical coefficients found for a given latitude can be generalized to estimate solar radiation in cities at similar latitude. Some case studies are presented to demonstrate this generalization with the result showing good agreement with the measurements. More importantly, these case studies further validate the models developed, and demonstrate the general applicability of the models developed. Finally, the obtained results of PSO technique were compared with the obtained results of SRTs (statistical regression techniques) on Angstrom model for all 17 cities. The results showed that obtained empirical coefficients for Angstrom model based on PSO have more accuracy than SRTs for all 17 cities. -- Highlights: → The first study to apply an intelligent optimization technique to more accurately determine empirical coefficients in solar radiation models. → New models which are presented in this study have better performance than existing models. → The empirical coefficients found for a given latitude can be generalized to estimate solar radiation in cities at similar latitude. → A fair comparison between the performance of PSO and SRTs on GSR modeling.

  9. Measuring the efficacy of flunixin meglumine and meloxicam for lame sows using a GAITFour pressure mat and an embedded microcomputer-based force plate system.

    Science.gov (United States)

    Pairis-Garcia, M D; Johnson, A K; Abell, C A; Coetzee, J F; Karriker, L A; Millman, S T; Stalder, K J

    2015-05-01

    Pain associated with lameness on farm is a negative affective state and has a detrimental impact on individual farm animal welfare. Animal pain can be managed utilizing husbandry tools and through pharmacological approaches. Nonsteroidal anti-inflammatory drugs including meloxicam and flunixin meglumine are compounds used in many species for pain management because they are easy to administer, long lasting, and cost-effective. Assessing an animal's biomechanical parameters using such tools as the embedded microcomputer-based force plate system and GAITFour pressure mat gait analysis walkway system provides an objective, sensitive, and precise means to detect animals in lame states. The objectives of this study were to determine the efficacy of meloxicam and flunixin meglumine for pain mitigation in lame sows using the embedded microcomputer-based force plate system and GAITFour pressure mat gait analysis walkway system. Lameness was induced in 24 mature mixed-parity sows using a chemical synovitis model and compared 3 treatments: meloxicam (1.0 mg/kg per os), flunixin meglumine (2.2 mg/kg intramuscular) and sterile saline (intramuscular). Weight distribution (kg) for each foot was collected twice per second for a total of 5 min for each time point using the embedded microcomputer-based force plate system. Stride time, stride length, maximum pressure, activated sensors, and stance time were collected using 3 quality walks (readings) for each time point using the GAITFour pressure mat gait analysis walkway system. Sows administered flunixin meglumine or meloxicam tolerated more weight on their lame leg compared with saline sows (P embedded microcomputer-based force plate system and GAITFour pressure mat gait analysis walkway system. Analgesic drugs may be a key tool to manage negative pain affective states associated with lameness.

  10. Modelling of ground penetrating radar data in stratified media using the reflectivity technique

    International Nuclear Information System (INIS)

    Sena, Armando R; Sen, Mrinal K; Stoffa, Paul L

    2008-01-01

    Horizontally layered media are often encountered in shallow exploration geophysics. Ground penetrating radar (GPR) data in these environments can be modelled by techniques that are more efficient than finite difference (FD) or finite element (FE) schemes because the lateral homogeneity of the media allows us to reduce the dependence on the horizontal spatial variables through Fourier transforms on these coordinates. We adapt and implement the invariant embedding or reflectivity technique used to model elastic waves in layered media to model GPR data. The results obtained with the reflectivity and FDTD modelling techniques are in excellent agreement and the effects of the air–soil interface on the radiation pattern are correctly taken into account by the reflectivity technique. Comparison with real wide-angle GPR data shows that the reflectivity technique can satisfactorily reproduce the real GPR data. These results and the computationally efficient characteristics of the reflectivity technique (compared to FD or FE) demonstrate its usefulness in interpretation and possible model-based inversion schemes of GPR data in stratified media

  11. MIPROPS - INTERACTIVE FORTRAN PROGRAMS FOR MICROCOMPUTERS TO CALCULATE THE THERMAL PHYSICAL PROPERTIES OF TWELVE FLUIDS

    Science.gov (United States)

    Cleghorn, T. F.

    1994-01-01

    MIPROPS is a set of programs which gives the thermophysical and transport properties of selected fluids. Although these programs are written in FORTRAN 77 for implementation on microcomputers, they are direct translations of interactive FORTRAN IV programs which were originally developed for large mainframes. MIPROPS calculates the properties of fluids in both the liquid and vapor states over a wide range of temperatures and pressures. The fluids included are: helium, parahydrogen, nitrogen, oxygen, argon, nitrogen trifluoride, methane, ethylene, ethane, propane, and iso- and normal butane. All of the programs except for the helium program utilize the same mathematical model of the equation of state. A separate program was necessary for helium, as the model for the helium thermodynamic surface is of a different form. The input variables are any two of pressure, density, or temperature for the single phase regions, and either pressure or temperature for the saturated liquid or vapor states. The output is pressure, density, temperature, internal energy, enthalpy, entropy, specific heat capacities, and speed of sound. In addition, viscosity, thermal conductivity, and dielectric constants are calculated for most of the fluids. The user can select either a single point or a table of output values for a specified temperature range, and can display the data either in engineering or metric units. This machine independent FORTRAN 77 program was implemented on an IBM PC XT with an MS-DOS 3.21 operating system. It has a memory requirement of approximately 100K. The program was developed in 1986.

  12. BREEDER: a microcomputer program for financial analysis of a large-scale prototype breeder reactor

    International Nuclear Information System (INIS)

    Giese, R.F.

    1984-04-01

    This report describes a microcomputer-based, single-project financial analysis program: BREEDER. BREEDER is a user-friendly model designed to facilitate frequent and rapid analyses of the financial implications associated with alternative design and financing strategies for electric generating plants and large-scale prototype breeder (LSPB) reactors in particular. The model has proved to be a useful tool in establishing cost goals for LSPB reactors. The program is available on floppy disks for use on an IBM personal computer (or IBM look-a-like) running under PC-DOS or a Kaypro II transportable computer running under CP/M (and many other CP/M machines). The report documents version 1.5 of BREEDER and contains a user's guide. The report also includes a general overview of BREEDER, a summary of hardware requirements, a definition of all required program inputs, a description of all algorithms used in performing the construction-period and operation-period analyses, and a summary of all available reports. The appendixes contain a complete source-code listing, a cross-reference table, a sample interactive session, several sample runs, and additional documentation of the net-equity program option

  13. Modelling techniques for predicting the long term consequences of radiation on natural aquatic populations

    International Nuclear Information System (INIS)

    Wallis, I.G.

    1978-01-01

    The purpose of this working paper is to describe modelling techniques for predicting the long term consequences of radiation on natural aquatic populations. Ideally, it would be possible to use aquatic population models: (1) to predict changes in the health and well-being of all aquatic populations as a result of changing the composition, amount and location of radionuclide discharges; (2) to compare the effects of steady, fluctuating and accidental releases of radionuclides; and (3) to evaluate the combined impact of the discharge of radionuclides and other wastes, and natural environmental stresses on aquatic populations. At the onset it should be stated that there is no existing model which can achieve this ideal performance. However, modelling skills and techniques are available to develop useful aquatic population models. This paper discusses the considerations involved in developing these models and briefly describes the various types of population models which have been developed to date

  14. Large wind power plants modeling techniques for power system simulation studies

    Energy Technology Data Exchange (ETDEWEB)

    Larose, Christian; Gagnon, Richard; Turmel, Gilbert; Giroux, Pierre; Brochu, Jacques [IREQ Hydro-Quebec Research Institute, Varennes, QC (Canada); McNabb, Danielle; Lefebvre, Daniel [Hydro-Quebec TransEnergie, Montreal, QC (Canada)

    2009-07-01

    This paper presents efficient modeling techniques for the simulation of large wind power plants in the EMT domain using a parallel supercomputer. Using these techniques, large wind power plants can be simulated in detail, with each wind turbine individually represented, as well as the collector and receiving network. The simulation speed of the resulting models is fast enough to perform both EMT and transient stability studies. The techniques are applied to develop an EMT detailed model of a generic wind power plant consisting of 73 x 1.5-MW doubly-fed induction generator (DFIG) wind turbine. Validation of the modeling techniques is presented using a comparison with a Matlab/SimPowerSystems simulation. To demonstrate the simulation capabilities using these modeling techniques, simulations involving a 120-bus receiving network with two generic wind power plants (146 wind turbines) are performed. The complete system is modeled using the Hypersim simulator and Matlab/SimPowerSystems. The simulations are performed on a 32-processor supercomputer using an EMTP-like solution with a time step of 18.4 {mu}s. The simulation performance is 10 times slower than in real-time, which is a huge gain in performance compared to traditional tools. The simulation is designed to run in real-time so it never stops, resulting in a capability to perform thousand of tests via automatic testing tools. (orig.)

  15. Real-time reservoir geological model updating using the hybrid EnKF and geostatistical technique

    Energy Technology Data Exchange (ETDEWEB)

    Li, H.; Chen, S.; Yang, D. [Regina Univ., SK (Canada). Petroleum Technology Research Centre

    2008-07-01

    Reservoir simulation plays an important role in modern reservoir management. Multiple geological models are needed in order to analyze the uncertainty of a given reservoir development scenario. Ideally, dynamic data should be incorporated into a reservoir geological model. This can be done by using history matching and tuning the model to match the past performance of reservoir history. This study proposed an assisted history matching technique to accelerate and improve the matching process. The Ensemble Kalman Filter (EnKF) technique, which is an efficient assisted history matching method, was integrated with a conditional geostatistical simulation technique to dynamically update reservoir geological models. The updated models were constrained to dynamic data, such as reservoir pressure and fluid saturations, and approaches geologically realistic at each time step by using the EnKF technique. The new technique was successfully applied in a heterogeneous synthetic reservoir. The uncertainty of the reservoir characterization was significantly reduced. More accurate forecasts were obtained from the updated models. 3 refs., 2 figs.

  16. Uncertainty analysis in rainfall-runoff modelling : Application of machine learning techniques

    NARCIS (Netherlands)

    Shrestha, D.l.

    2009-01-01

    This thesis presents powerful machine learning (ML) techniques to build predictive models of uncertainty with application to hydrological models. Two different methods are developed and tested. First one focuses on parameter uncertainty analysis by emulating the results of Monte Carlo simulations of

  17. Uncertainty Analysis in Rainfall-Runoff Modelling: Application of Machine Learning Techniques

    NARCIS (Netherlands)

    Shrestha, D.L.

    2009-01-01

    This thesis presents powerful machine learning (ML) techniques to build predictive models of uncertainty with application to hydrological models. Two different methods are developed and tested. First one focuses on parameter uncertainty analysis by emulating the results of Monte Carlo simulations of

  18. Using Game Theory Techniques and Concepts to Develop Proprietary Models for Use in Intelligent Games

    Science.gov (United States)

    Christopher, Timothy Van

    2011-01-01

    This work is about analyzing games as models of systems. The goal is to understand the techniques that have been used by game designers in the past, and to compare them to the study of mathematical game theory. Through the study of a system or concept a model often emerges that can effectively educate students about making intelligent decisions…

  19. Application of Soft Computing Techniques and Multiple Regression Models for CBR prediction of Soils

    Directory of Open Access Journals (Sweden)

    Fatimah Khaleel Ibrahim

    2017-08-01

    Full Text Available The techniques of soft computing technique such as Artificial Neutral Network (ANN have improved the predicting capability and have actually discovered application in Geotechnical engineering. The aim of this research is to utilize the soft computing technique and Multiple Regression Models (MLR for forecasting the California bearing ratio CBR( of soil from its index properties. The indicator of CBR for soil could be predicted from various soils characterizing parameters with the assist of MLR and ANN methods. The data base that collected from the laboratory by conducting tests on 86 soil samples that gathered from different projects in Basrah districts. Data gained from the experimental result were used in the regression models and soft computing techniques by using artificial neural network. The liquid limit, plastic index , modified compaction test and the CBR test have been determined. In this work, different ANN and MLR models were formulated with the different collection of inputs to be able to recognize their significance in the prediction of CBR. The strengths of the models that were developed been examined in terms of regression coefficient (R2, relative error (RE% and mean square error (MSE values. From the results of this paper, it absolutely was noticed that all the proposed ANN models perform better than that of MLR model. In a specific ANN model with all input parameters reveals better outcomes than other ANN models.

  20. Application of separable parameter space techniques to multi-tracer PET compartment modeling.

    Science.gov (United States)

    Zhang, Jeff L; Michael Morey, A; Kadrmas, Dan J

    2016-02-07

    Multi-tracer positron emission tomography (PET) can image two or more tracers in a single scan, characterizing multiple aspects of biological functions to provide new insights into many diseases. The technique uses dynamic imaging, resulting in time-activity curves that contain contributions from each tracer present. The process of separating and recovering separate images and/or imaging measures for each tracer requires the application of kinetic constraints, which are most commonly applied by fitting parallel compartment models for all tracers. Such multi-tracer compartment modeling presents challenging nonlinear fits in multiple dimensions. This work extends separable parameter space kinetic modeling techniques, previously developed for fitting single-tracer compartment models, to fitting multi-tracer compartment models. The multi-tracer compartment model solution equations were reformulated to maximally separate the linear and nonlinear aspects of the fitting problem, and separable least-squares techniques were applied to effectively reduce the dimensionality of the nonlinear fit. The benefits of the approach are then explored through a number of illustrative examples, including characterization of separable parameter space multi-tracer objective functions and demonstration of exhaustive search fits which guarantee the true global minimum to within arbitrary search precision. Iterative gradient-descent algorithms using Levenberg-Marquardt were also tested, demonstrating improved fitting speed and robustness as compared to corresponding fits using conventional model formulations. The proposed technique overcomes many of the challenges in fitting simultaneous multi-tracer PET compartment models.

  1. Generation of 3-D finite element models of restored human teeth using micro-CT techniques.

    NARCIS (Netherlands)

    Verdonschot, N.J.J.; Fennis, W.M.M.; Kuys, R.H.; Stolk, J.; Kreulen, C.M.; Creugers, N.H.J.

    2001-01-01

    PURPOSE: This article describes the development of a three-dimensional finite element model of a premolar based on a microscale computed tomographic (CT) data-acquisition technique. The development of the model is part of a project studying the optimal design and geometry of adhesive tooth-colored

  2. Application of separable parameter space techniques to multi-tracer PET compartment modeling

    Science.gov (United States)

    Zhang, Jeff L.; Morey, A. Michael; Kadrmas, Dan J.

    2016-02-01

    Multi-tracer positron emission tomography (PET) can image two or more tracers in a single scan, characterizing multiple aspects of biological functions to provide new insights into many diseases. The technique uses dynamic imaging, resulting in time-activity curves that contain contributions from each tracer present. The process of separating and recovering separate images and/or imaging measures for each tracer requires the application of kinetic constraints, which are most commonly applied by fitting parallel compartment models for all tracers. Such multi-tracer compartment modeling presents challenging nonlinear fits in multiple dimensions. This work extends separable parameter space kinetic modeling techniques, previously developed for fitting single-tracer compartment models, to fitting multi-tracer compartment models. The multi-tracer compartment model solution equations were reformulated to maximally separate the linear and nonlinear aspects of the fitting problem, and separable least-squares techniques were applied to effectively reduce the dimensionality of the nonlinear fit. The benefits of the approach are then explored through a number of illustrative examples, including characterization of separable parameter space multi-tracer objective functions and demonstration of exhaustive search fits which guarantee the true global minimum to within arbitrary search precision. Iterative gradient-descent algorithms using Levenberg-Marquardt were also tested, demonstrating improved fitting speed and robustness as compared to corresponding fits using conventional model formulations. The proposed technique overcomes many of the challenges in fitting simultaneous multi-tracer PET compartment models.

  3. Application of separable parameter space techniques to multi-tracer PET compartment modeling

    International Nuclear Information System (INIS)

    Zhang, Jeff L; Michael Morey, A; Kadrmas, Dan J

    2016-01-01

    Multi-tracer positron emission tomography (PET) can image two or more tracers in a single scan, characterizing multiple aspects of biological functions to provide new insights into many diseases. The technique uses dynamic imaging, resulting in time-activity curves that contain contributions from each tracer present. The process of separating and recovering separate images and/or imaging measures for each tracer requires the application of kinetic constraints, which are most commonly applied by fitting parallel compartment models for all tracers. Such multi-tracer compartment modeling presents challenging nonlinear fits in multiple dimensions. This work extends separable parameter space kinetic modeling techniques, previously developed for fitting single-tracer compartment models, to fitting multi-tracer compartment models. The multi-tracer compartment model solution equations were reformulated to maximally separate the linear and nonlinear aspects of the fitting problem, and separable least-squares techniques were applied to effectively reduce the dimensionality of the nonlinear fit. The benefits of the approach are then explored through a number of illustrative examples, including characterization of separable parameter space multi-tracer objective functions and demonstration of exhaustive search fits which guarantee the true global minimum to within arbitrary search precision. Iterative gradient-descent algorithms using Levenberg–Marquardt were also tested, demonstrating improved fitting speed and robustness as compared to corresponding fits using conventional model formulations. The proposed technique overcomes many of the challenges in fitting simultaneous multi-tracer PET compartment models. (paper)

  4. Techniques to extract physical modes in model-independent analysis of rings

    International Nuclear Information System (INIS)

    Wang, C.-X.

    2004-01-01

    A basic goal of Model-Independent Analysis is to extract the physical modes underlying the beam histories collected at a large number of beam position monitors so that beam dynamics and machine properties can be deduced independent of specific machine models. Here we discuss techniques to achieve this goal, especially the Principal Component Analysis and the Independent Component Analysis.

  5. A fully blanketed early B star LTE model atmosphere using an opacity sampling technique

    International Nuclear Information System (INIS)

    Phillips, A.P.; Wright, S.L.

    1980-01-01

    A fully blanketed LTE model of a stellar atmosphere with Tsub(e) = 21914 K (thetasub(e) = 0.23), log g = 4 is presented. The model includes an explicit representation of the opacity due to the strongest lines, and uses a statistical opacity sampling technique to represent the weaker line opacity. The sampling technique is subjected to several tests and the model is compared with an atmosphere calculated using the line-distribution function method. The limitations of the distribution function method and the particular opacity sampling method used here are discussed in the light of the results obtained. (author)

  6. Modeling techniques used in the communications link analysis and simulation system (CLASS)

    Science.gov (United States)

    Braun, W. R.; Mckenzie, T. M.

    1985-01-01

    CLASS (Communications Link Analysis and Simulation System) is a software package developed for NASA to predict the communication and tracking performance of the Tracking and Data Relay Satellite System (TDRSS) services. The modeling techniques used in CLASS are described. The components of TDRSS and the performance parameters to be computed by CLASS are too diverse to permit the use of a single technique to evaluate all performance measures. Hence, each CLASS module applies the modeling approach best suited for a particular subsystem and/or performance parameter in terms of model accuracy and computational speed.

  7. Comparison of bag-valve-mask hand-sealing techniques in a simulated model.

    Science.gov (United States)

    Otten, David; Liao, Michael M; Wolken, Robert; Douglas, Ivor S; Mishra, Ramya; Kao, Amanda; Barrett, Whitney; Drasler, Erin; Byyny, Richard L; Haukoos, Jason S

    2014-01-01

    Bag-valve-mask ventilation remains an essential component of airway management. Rescuers continue to use both traditional 1- or 2-handed mask-face sealing techniques, as well as a newer modified 2-handed technique. We compare the efficacy of 1-handed, 2-handed, and modified 2-handed bag-valve-mask technique. In this prospective, crossover study, health care providers performed 1-handed, 2-handed, and modified 2-handed bag-valve-mask ventilation on a standardized ventilation model. Subjects performed each technique for 5 minutes, with 3 minutes' rest between techniques. The primary outcome was expired tidal volume, defined as percentage of total possible expired tidal volume during a 5-minute bout. A specialized inline monitor measured expired tidal volume. We compared 2-handed versus modified 2-handed and 2-handed versus 1-handed techniques. We enrolled 52 subjects: 28 (54%) men, 32 (62%) with greater than or equal to 5 actual emergency bag-valve-mask situations. Median expired tidal volume percentage for 1-handed technique was 31% (95% confidence interval [CI] 17% to 51%); for 2-handed technique, 85% (95% CI 78% to 91%); and for modified 2-handed technique, 85% (95% CI 82% to 90%). Both 2-handed (median difference 47%; 95% CI 34% to 62%) and modified 2-handed technique (median difference 56%; 95% CI 29% to 65%) resulted in significantly higher median expired tidal volume percentages compared with 1-handed technique. The median expired tidal volume percentages between 2-handed and modified 2-handed techniques did not significantly differ from each other (median difference 0; 95% CI -2% to 2%). In a simulated model, both 2-handed mask-face sealing techniques resulted in higher ventilatory tidal volumes than 1-handed technique. Tidal volumes from 2-handed and modified 2-handed techniques did not differ. Rescuers should perform bag-valve-mask ventilation with 2-handed techniques. Copyright © 2013 American College of Emergency Physicians. Published by Mosby

  8. A novel model surgery technique for LeFort III advancement.

    Science.gov (United States)

    Vachiramon, Amornpong; Yen, Stephen L-K; Lypka, Michael; Bindignavale, Vijay; Hammoudeh, Jeffrey; Reinisch, John; Urata, Mark M

    2007-09-01

    Current techniques for model surgery and occlusal splint fabrication lack the ability to mark, measure and plan the position of the orbital rim for LeFort III and Monobloc osteotomies. This report describes a model surgery technique for planning the three dimensional repositioning of the orbital rims. Dual orbital pointers were used to mark the infraorbital rim during the facebow transfer. These pointer positions were transferred onto the surgical models in order to follow splint-determined movements. Case reports are presented to illustrate how the model surgery technique was used to differentiate the repositioning of the orbital rim from the occlusal correction in single segment and combined LeFort III/LeFort I osteotomies.

  9. The Effect of Learning Based on Technology Model and Assessment Technique toward Thermodynamic Learning Achievement

    Science.gov (United States)

    Makahinda, T.

    2018-02-01

    The purpose of this research is to find out the effect of learning model based on technology and assessment technique toward thermodynamic achievement by controlling students intelligence. This research is an experimental research. The sample is taken through cluster random sampling with the total respondent of 80 students. The result of the research shows that the result of learning of thermodynamics of students who taught the learning model of environmental utilization is higher than the learning result of student thermodynamics taught by simulation animation, after controlling student intelligence. There is influence of student interaction, and the subject between models of technology-based learning with assessment technique to student learning result of Thermodynamics, after controlling student intelligence. Based on the finding in the lecture then should be used a thermodynamic model of the learning environment with the use of project assessment technique.

  10. Experimental investigation of the predictive capabilities of data driven modeling techniques in hydrology - Part 2: Application

    Directory of Open Access Journals (Sweden)

    A. Elshorbagy

    2010-10-01

    Full Text Available In this second part of the two-part paper, the data driven modeling (DDM experiment, presented and explained in the first part, is implemented. Inputs for the five case studies (half-hourly actual evapotranspiration, daily peat soil moisture, daily till soil moisture, and two daily rainfall-runoff datasets are identified, either based on previous studies or using the mutual information content. Twelve groups (realizations were randomly generated from each dataset by randomly sampling without replacement from the original dataset. Neural networks (ANNs, genetic programming (GP, evolutionary polynomial regression (EPR, Support vector machines (SVM, M5 model trees (M5, K-nearest neighbors (K-nn, and multiple linear regression (MLR techniques are implemented and applied to each of the 12 realizations of each case study. The predictive accuracy and uncertainties of the various techniques are assessed using multiple average overall error measures, scatter plots, frequency distribution of model residuals, and the deterioration rate of prediction performance during the testing phase. Gamma test is used as a guide to assist in selecting the appropriate modeling technique. Unlike two nonlinear soil moisture case studies, the results of the experiment conducted in this research study show that ANNs were a sub-optimal choice for the actual evapotranspiration and the two rainfall-runoff case studies. GP is the most successful technique due to its ability to adapt the model complexity to the modeled data. EPR performance could be close to GP with datasets that are more linear than nonlinear. SVM is sensitive to the kernel choice and if appropriately selected, the performance of SVM can improve. M5 performs very well with linear and semi linear data, which cover wide range of hydrological situations. In highly nonlinear case studies, ANNs, K-nn, and GP could be more successful than other modeling techniques. K-nn is also successful in linear situations, and it

  11. Experimental investigation of the predictive capabilities of data driven modeling techniques in hydrology - Part 2: Application

    Science.gov (United States)

    Elshorbagy, A.; Corzo, G.; Srinivasulu, S.; Solomatine, D. P.

    2010-10-01

    In this second part of the two-part paper, the data driven modeling (DDM) experiment, presented and explained in the first part, is implemented. Inputs for the five case studies (half-hourly actual evapotranspiration, daily peat soil moisture, daily till soil moisture, and two daily rainfall-runoff datasets) are identified, either based on previous studies or using the mutual information content. Twelve groups (realizations) were randomly generated from each dataset by randomly sampling without replacement from the original dataset. Neural networks (ANNs), genetic programming (GP), evolutionary polynomial regression (EPR), Support vector machines (SVM), M5 model trees (M5), K-nearest neighbors (K-nn), and multiple linear regression (MLR) techniques are implemented and applied to each of the 12 realizations of each case study. The predictive accuracy and uncertainties of the various techniques are assessed using multiple average overall error measures, scatter plots, frequency distribution of model residuals, and the deterioration rate of prediction performance during the testing phase. Gamma test is used as a guide to assist in selecting the appropriate modeling technique. Unlike two nonlinear soil moisture case studies, the results of the experiment conducted in this research study show that ANNs were a sub-optimal choice for the actual evapotranspiration and the two rainfall-runoff case studies. GP is the most successful technique due to its ability to adapt the model complexity to the modeled data. EPR performance could be close to GP with datasets that are more linear than nonlinear. SVM is sensitive to the kernel choice and if appropriately selected, the performance of SVM can improve. M5 performs very well with linear and semi linear data, which cover wide range of hydrological situations. In highly nonlinear case studies, ANNs, K-nn, and GP could be more successful than other modeling techniques. K-nn is also successful in linear situations, and it should

  12. Towards a Business Process Modeling Technique for Agile Development of Case Management Systems

    Directory of Open Access Journals (Sweden)

    Ilia Bider

    2017-12-01

    Full Text Available A modern organization needs to adapt its behavior to changes in the business environment by changing its Business Processes (BP and corresponding Business Process Support (BPS systems. One way of achieving such adaptability is via separation of the system code from the process description/model by applying the concept of executable process models. Furthermore, to ease introduction of changes, such process model should separate different perspectives, for example, control-flow, human resources, and data perspectives, from each other. In addition, for developing a completely new process, it should be possible to start with a reduced process model to get a BPS system quickly running, and then continue to develop it in an agile manner. This article consists of two parts, the first sets requirements on modeling techniques that could be used in the tools that supports agile development of BPs and BPS systems. The second part suggests a business process modeling technique that allows to start modeling with the data/information perspective which would be appropriate for processes supported by Case or Adaptive Case Management (CM/ACM systems. In a model produced by this technique, called data-centric business process model, a process instance/case is defined as sequence of states in a specially designed instance database, while the process model is defined as a set of rules that set restrictions on allowed states and transitions between them. The article details the background for the project of developing the data-centric process modeling technique, presents the outline of the structure of the model, and gives formal definitions for a substantial part of the model.

  13. Validation of micro-computed tomography for occlusal caries detection: an in vitro study.

    Science.gov (United States)

    Özkan, Gökhan; Kanli, Aydan; Başeren, Nurdan Meserret; Arslan, Umut; Tatar, İlkan

    2015-01-01

    Conventional methods of caries detection, including the gold standard of histological examination, have certain disadvantages that must be addressed prior to validating any other diagnostic technique-current or new. Here we evaluated the validity of micro-computed tomography (micro-CT) as an alternative gold-standard technique for caries detection. Sixty teeth with suspected occlusal caries were chosen from a pool of teeth extracted for orthodontic, periodontal, or surgical reasons. Identical reference points were marked on photographs taken for teeth and were used to evaluate each method. Dimensions of caries were assessed by two calibrated examiners using the ICDAS-II visual examination system, bitewing radiographs, and micro-CT. The teeth included in the study were selected randomly from solution before all measurements. For micro-CT, the device was set to 50 kV, 800 µA, pixel size 15 µm (at 1024 × 1024 resolution), and 1° rotation step. NRecon software (SkyScan) was used to obtain reconstructed images. For each diagnostic method, results were compared with histology results using the McNemar test. Receiver operator characteristic (ROC) analysis was also performed for each method (Z-test; p < 0.05). Besides showing a high correlation with histology results, micro-CT yielded the greatest values at the D3 threshold; moreover, accuracy and area under the ROC curve (AUC) values were greatest at the D1threshold. Our results indicate that micro-CT performs as well as or better than histological examination for the purpose of comparing methods for caries detection.

  14. Reliability modeling of digital component in plant protection system with various fault-tolerant techniques

    International Nuclear Information System (INIS)

    Kim, Bo Gyung; Kang, Hyun Gook; Kim, Hee Eun; Lee, Seung Jun; Seong, Poong Hyun

    2013-01-01

    Highlights: • Integrated fault coverage is introduced for reflecting characteristics of fault-tolerant techniques in the reliability model of digital protection system in NPPs. • The integrated fault coverage considers the process of fault-tolerant techniques from detection to fail-safe generation process. • With integrated fault coverage, the unavailability of repairable component of DPS can be estimated. • The new developed reliability model can reveal the effects of fault-tolerant techniques explicitly for risk analysis. • The reliability model makes it possible to confirm changes of unavailability according to variation of diverse factors. - Abstract: With the improvement of digital technologies, digital protection system (DPS) has more multiple sophisticated fault-tolerant techniques (FTTs), in order to increase fault detection and to help the system safely perform the required functions in spite of the possible presence of faults. Fault detection coverage is vital factor of FTT in reliability. However, the fault detection coverage is insufficient to reflect the effects of various FTTs in reliability model. To reflect characteristics of FTTs in the reliability model, integrated fault coverage is introduced. The integrated fault coverage considers the process of FTT from detection to fail-safe generation process. A model has been developed to estimate the unavailability of repairable component of DPS using the integrated fault coverage. The new developed model can quantify unavailability according to a diversity of conditions. Sensitivity studies are performed to ascertain important variables which affect the integrated fault coverage and unavailability

  15. Two-dimensional gel electrophoresis image registration using block-matching techniques and deformation models.

    Science.gov (United States)

    Rodriguez, Alvaro; Fernandez-Lozano, Carlos; Dorado, Julian; Rabuñal, Juan R

    2014-06-01

    Block-matching techniques have been widely used in the task of estimating displacement in medical images, and they represent the best approach in scenes with deformable structures such as tissues, fluids, and gels. In this article, a new iterative block-matching technique-based on successive deformation, search, fitting, filtering, and interpolation stages-is proposed to measure elastic displacements in two-dimensional polyacrylamide gel electrophoresis (2D-PAGE) images. The proposed technique uses different deformation models in the task of correlating proteins in real 2D electrophoresis gel images, obtaining an accuracy of 96.6% and improving the results obtained with other techniques. This technique represents a general solution, being easy to adapt to different 2D deformable cases and providing an experimental reference for block-matching algorithms. Copyright © 2014 Elsevier Inc. All rights reserved.

  16. A Shell/3D Modeling Technique for the Analyses of Delaminated Composite Laminates

    Science.gov (United States)

    Krueger, Ronald; OBrien, T. Kevin

    2001-01-01

    A shell/3D modeling technique was developed for which a local three-dimensional solid finite element model is used only in the immediate vicinity of the delamination front. The goal was to combine the accuracy of the full three-dimensional solution with the computational efficiency of a plate or shell finite element model. Multi-point constraints provided a kinematically compatible interface between the local three-dimensional model and the global structural model which has been meshed with plate or shell finite elements. Double Cantilever Beam (DCB), End Notched Flexure (ENF), and Single Leg Bending (SLB) specimens were modeled using the shell/3D technique to study the feasibility for pure mode I (DCB), mode II (ENF) and mixed mode I/II (SLB) cases. Mixed mode strain energy release rate distributions were computed across the width of the specimens using the virtual crack closure technique. Specimens with a unidirectional layup and with a multidirectional layup where the delamination is located between two non-zero degree plies were simulated. For a local three-dimensional model, extending to a minimum of about three specimen thicknesses on either side of the delamination front, the results were in good agreement with mixed mode strain energy release rates obtained from computations where the entire specimen had been modeled with solid elements. For large built-up composite structures modeled with plate elements, the shell/3D modeling technique offers a great potential for reducing the model size, since only a relatively small section in the vicinity of the delamination front needs to be modeled with solid elements.

  17. Modelling techniques for underwater noise generated by tidal turbines in shallow water

    OpenAIRE

    Lloyd, Thomas P.; Turnock, Stephen R.; Humphrey, Victor F.

    2011-01-01

    The modelling of underwater noise sources and their potential impact on the marine environment is considered, focusing on tidal turbines in shallow water. The requirement for device noise prediction as part of environmental impact assessment is outlined and the limited amount of measurement data and modelling research identified. Following the identification of potential noise sources, the dominant flowgenerated sources are modelled using empirical techniques. The predicted sound pressure lev...

  18. Study on ABCD Analysis Technique for Business Models, business strategies, Operating Concepts & Business Systems

    OpenAIRE

    Sreeramana Aithal

    2016-01-01

    Studying the implications of a business model, choosing success strategies, developing viable operational concepts or evolving a functional system, it is important to analyse it in all dimensions. For this purpose, various analysing techniques/frameworks are used. This paper is a discussion on how to use an innovative analysing framework called ABCD model on a given business model, or on a business strategy or an operational concept/idea or business system. Based on four constructs Advantages...

  19. Car sharing demand estimation and urban transport demand modelling using stated preference techniques

    OpenAIRE

    Catalano, Mario; Lo Casto, Barbara; Migliore, Marco

    2008-01-01

    The research deals with the use of the stated preference technique (SP) and transport demand modelling to analyse travel mode choice behaviour for commuting urban trips in Palermo, Italy. The principal aim of the study was the calibration of a demand model to forecast the modal split of the urban transport demand, allowing for the possibility of using innovative transport systems like car sharing and car pooling. In order to estimate the demand model parameters, a specific survey was carried ...

  20. Transfer of physics detector models into CAD systems using modern techniques

    International Nuclear Information System (INIS)

    Dach, M.; Vuoskoski, J.

    1996-01-01

    Designing high energy physics detectors for future experiments requires sophisticated computer aided design and simulation tools. In order to satisfy the future demands in this domain, modern techniques, methods, and standards have to be applied. We present an interface application, designed and implemented using object-oriented techniques, for the widely used GEANT physics simulation package. It converts GEANT detector models into the future industrial standard, STEP. (orig.)

  1. Finite element analysis of the mechanical properties of cellular aluminium based on micro-computed tomography

    International Nuclear Information System (INIS)

    Veyhl, C.; Belova, I.V.; Murch, G.E.; Fiedler, T.

    2011-01-01

    Research highlights: → Elastic and plastic anisotropy is observed for both materials → Both show qualitatively similar characteristics with quantitative differences → Distinctly higher mechanical properties for closed-cell foam → The 'big' and 'small' models show good agreement for the closed-cell foam. - Abstract: In the present paper, the macroscopic mechanical properties of open-cell M-Pore sponge (porosity of 91-93%) and closed-cell Alporas foam (porosity of 80-86%) are investigated. The complex geometry of these cellular materials is scanned by micro-computed tomography and used in finite element (FE) analysis. The mechanical properties are determined by uni-axial compression simulations in three perpendicular directions (x-, y- and z-direction). M-Pore and Alporas exhibit the same qualitative mechanical characteristics but with quantitative differences. In both cases, strong anisotropy is observed for Young's modulus and the 0.002 offset yield stress. Furthermore, for the investigated relative density range a linear dependence between relative density and mechanical properties is found. Finally, a distinctly higher Young's modulus and 0.002 offset yield stress is observed for Alporas.

  2. A microcomputer controlled snow ski binding system--II. Release decision theories.

    Science.gov (United States)

    MacGregor, D; Hull, M L

    1985-01-01

    A hierarchy of release decision theories for both tibia fracture and knee ligamentous injury are defined and simulated on a computer. Moment loading data, recorded during actual skiing by the microcomputer-based ski binding system described in Part I, are processed by the various release decision theories. At the bottom of the hierarchy is the simplest theory which treats boot loading as quasi-static and compares moment components to threshold levels. Another stage of the hierarchy defines an analytic expression for a combined loading failure locus. Note that this is the first formulation of a combined loading release decision theory. Yet another stage of the hierarchy computes bone moments via dynamic system leg models. The various release decision theories are evaluated by comparing processed results to both pain and bone failure limits. For the data generated by the field tests conducted to date, the simplest release decision theory satisfied the retention requirement for pain limits in the presence of muscle activity for both torsion and forward bending. For pain limits in the absence of muscle activity the retention requirement was not satisfied however. Another result is that leg dynamics are significant. A final result is that combined loading considerations lead to a more conservative theory.

  3. Visualization of Three-Dimensional Nephron Structure With Microcomputed Tomography

    International Nuclear Information System (INIS)

    Bentley, M.; Jorgensen, S.; Lerman, L.; Ritman, E.; Romero, J.

    2007-01-01

    The three-dimensional architecture of nephrons in situ and their interrelationship with other nephrons are difficult to visualize by microscopic methods. The present study uses microcomputed X-ray tomography (micro-CT) to visualize intact nephrons in situ. Rat kidneys were perfusion-fixed with buffered formalin and their vasculature was subsequently perfused with radiopaque silicone. Cortical tissue was stained en bloc with osmium tetroxide, embedded in plastic, scanned, and reconstructed at voxel resolutions of 6, 2, and 1 μm. At 6 μm resolution, large blood vessels and glomeruli could be visualized but nephrons and their lumens were small and difficult to visualize. Optimal images were obtained using a synchrotron radiation source at 2 μm resolution where nephron components could be identified, correlated with histological sections, and traced. Proximal tubules had large diameters and opaque walls, whereas distal tubules, connecting tubules, and collecting ducts had smaller diameters and less opaque walls. Blood vessels could be distinguished from nephrons by the luminal presence of radiopaque silicone. Proximal tubules were three times longer than distal tubules. Proximal and distal tubules were tightly coiled in the outer cortex but were loosely coiled in the middle and inner cortex. The connecting tubules had the narrowest diameters of the tubules and converged to form arcades that paralleled the radial vessels as they extended to the outer cortex. These results illustrate a potential use of micro-CT to obtain three-dimensional information about nephron architecture and nephron interrelationships, which could be useful in evaluating experimental tubular hypertrophy, atrophy, and necrosis

  4. Determination of Complex-Valued Parametric Model Coefficients Using Artificial Neural Network Technique

    Directory of Open Access Journals (Sweden)

    A. M. Aibinu

    2010-01-01

    Full Text Available A new approach for determining the coefficients of a complex-valued autoregressive (CAR and complex-valued autoregressive moving average (CARMA model coefficients using complex-valued neural network (CVNN technique is discussed in this paper. The CAR and complex-valued moving average (CMA coefficients which constitute a CARMA model are computed simultaneously from the adaptive weights and coefficients of the linear activation functions in a two-layered CVNN. The performance of the proposed technique has been evaluated using simulated complex-valued data (CVD with three different types of activation functions. The results show that the proposed method can accurately determine the model coefficients provided that the network is properly trained. Furthermore, application of the developed CVNN-based technique for MRI K-space reconstruction results in images with improve resolution.

  5. A novel CT acquisition and analysis technique for breathing motion modeling

    International Nuclear Information System (INIS)

    Low, Daniel A; White, Benjamin M; Lee, Percy P; Thomas, David H; Gaudio, Sergio; Jani, Shyam S; Wu, Xiao; Lamb, James M

    2013-01-01

    To report on a novel technique for providing artifact-free quantitative four-dimensional computed tomography (4DCT) image datasets for breathing motion modeling. Commercial clinical 4DCT methods have difficulty managing irregular breathing. The resulting images contain motion-induced artifacts that can distort structures and inaccurately characterize breathing motion. We have developed a novel scanning and analysis method for motion-correlated CT that utilizes standard repeated fast helical acquisitions, a simultaneous breathing surrogate measurement, deformable image registration, and a published breathing motion model. The motion model differs from the CT-measured motion by an average of 0.65 mm, indicating the precision of the motion model. The integral of the divergence of one of the motion model parameters is predicted to be a constant 1.11 and is found in this case to be 1.09, indicating the accuracy of the motion model. The proposed technique shows promise for providing motion-artifact free images at user-selected breathing phases, accurate Hounsfield units, and noise characteristics similar to non-4D CT techniques, at a patient dose similar to or less than current 4DCT techniques. (fast track communication)

  6. Optimizing Availability of a Framework in Series Configuration Utilizing Markov Model and Monte Carlo Simulation Techniques

    Directory of Open Access Journals (Sweden)

    Mansoor Ahmed Siddiqui

    2017-06-01

    Full Text Available This research work is aimed at optimizing the availability of a framework comprising of two units linked together in series configuration utilizing Markov Model and Monte Carlo (MC Simulation techniques. In this article, effort has been made to develop a maintenance model that incorporates three distinct states for each unit, while taking into account their different levels of deterioration. Calculations are carried out using the proposed model for two distinct cases of corrective repair, namely perfect and imperfect repairs, with as well as without opportunistic maintenance. Initially, results are accomplished using an analytical technique i.e., Markov Model. Validation of the results achieved is later carried out with the help of MC Simulation. In addition, MC Simulation based codes also work well for the frameworks that follow non-exponential failure and repair rates, and thus overcome the limitations offered by the Markov Model.

  7. Kerf modelling in abrasive waterjet milling using evolutionary computation and ANOVA techniques

    Science.gov (United States)

    Alberdi, A.; Rivero, A.; Carrascal, A.; Lamikiz, A.

    2012-04-01

    Many researchers demonstrated the capability of Abrasive Waterjet (AWJ) technology for precision milling operations. However, the concurrence of several input parameters along with the stochastic nature of this technology leads to a complex process control, which requires a work focused in process modelling. This research work introduces a model to predict the kerf shape in AWJ slot milling in Aluminium 7075-T651 in terms of four important process parameters: the pressure, the abrasive flow rate, the stand-off distance and the traverse feed rate. A hybrid evolutionary approach was employed for kerf shape modelling. This technique allowed characterizing the profile through two parameters: the maximum cutting depth and the full width at half maximum. On the other hand, based on ANOVA and regression techniques, these two parameters were also modelled as a function of process parameters. Combination of both models resulted in an adequate strategy to predict the kerf shape for different machining conditions.

  8. A Review of Domain Modelling and Domain Imaging Techniques in Ferroelectric Crystals.

    Science.gov (United States)

    Potnis, Prashant R; Tsou, Nien-Ti; Huber, John E

    2011-02-16

    The present paper reviews models of domain structure in ferroelectric crystals, thin films and bulk materials. Common crystal structures in ferroelectric materials are described and the theory of compatible domain patterns is introduced. Applications to multi-rank laminates are presented. Alternative models employing phase-field and related techniques are reviewed. The paper then presents methods of observing ferroelectric domain structure, including optical, polarized light, scanning electron microscopy, X-ray and neutron diffraction, atomic force microscopy and piezo-force microscopy. Use of more than one technique for unambiguous identification of the domain structure is also described.

  9. A Review of Domain Modelling and Domain Imaging Techniques in Ferroelectric Crystals

    Directory of Open Access Journals (Sweden)

    John E. Huber

    2011-02-01

    Full Text Available The present paper reviews models of domain structure in ferroelectric crystals, thin films and bulk materials. Common crystal structures in ferroelectric materials are described and the theory of compatible domain patterns is introduced. Applications to multi-rank laminates are presented. Alternative models employing phase-field and related techniques are reviewed. The paper then presents methods of observing ferroelectric domain structure, including optical, polarized light, scanning electron microscopy, X-ray and neutron diffraction, atomic force microscopy and piezo-force microscopy. Use of more than one technique for unambiguous identification of the domain structure is also described.

  10. Application of rapid prototyping techniques for modelling of anatomical structures in medical training and education.

    Science.gov (United States)

    Torres, K; Staśkiewicz, G; Śnieżyński, M; Drop, A; Maciejewski, R

    2011-02-01

    Rapid prototyping has become an innovative method of fast and cost-effective production of three-dimensional models for manufacturing. Wide access to advanced medical imaging methods allows application of this technique for medical training purposes. This paper presents the feasibility of rapid prototyping technologies: stereolithography, selective laser sintering, fused deposition modelling, and three-dimensional printing for medical education. Rapid prototyping techniques are a promising method for improvement of anatomical education in medical students but also a valuable source of training tools for medical specialists.

  11. A Shell/3D Modeling Technique for Delaminations in Composite Laminates

    Science.gov (United States)

    Krueger, Ronald

    1999-01-01

    A shell/3D modeling technique was developed for which a local solid finite element model is used only in the immediate vicinity of the delamination front. The goal was to combine the accuracy of the full three-dimensional solution with the computational efficiency of a plate or shell finite element model. Multi-point constraints provide a kinematically compatible interface between the local 3D model and the global structural model which has been meshed with plate or shell finite elements. For simple double cantilever beam (DCB), end notched flexure (ENF), and single leg bending (SLB) specimens, mixed mode energy release rate distributions were computed across the width from nonlinear finite element analyses using the virtual crack closure technique. The analyses served to test the accuracy of the shell/3D technique for the pure mode I case (DCB), mode II case (ENF) and a mixed mode I/II case (SLB). Specimens with a unidirectional layup where the delamination is located between two 0 plies, as well as a multidirectional layup where the delamination is located between two non-zero degree plies, were simulated. For a local 3D model extending to a minimum of about three specimen thicknesses in front of and behind the delamination front, the results were in good agreement with mixed mode strain energy release rates obtained from computations where the entire specimen had been modeled with solid elements. For large built-up composite structures modeled with plate elements, the shell/3D modeling technique offers a great potential, since only a relatively small section in the vicinity of the delamination front needs to be modeled with solid elements.

  12. Integrated approach to model decomposed flow hydrograph using artificial neural network and conceptual techniques

    Science.gov (United States)

    Jain, Ashu; Srinivasulu, Sanaga

    2006-02-01

    This paper presents the findings of a study aimed at decomposing a flow hydrograph into different segments based on physical concepts in a catchment, and modelling different segments using different technique viz. conceptual and artificial neural networks (ANNs). An integrated modelling framework is proposed capable of modelling infiltration, base flow, evapotranspiration, soil moisture accounting, and certain segments of the decomposed flow hydrograph using conceptual techniques and the complex, non-linear, and dynamic rainfall-runoff process using ANN technique. Specifically, five different multi-layer perceptron (MLP) and two self-organizing map (SOM) models have been developed. The rainfall and streamflow data derived from the Kentucky River catchment were employed to test the proposed methodology and develop all the models. The performance of all the models was evaluated using seven different standard statistical measures. The results obtained in this study indicate that (a) the rainfall-runoff relationship in a large catchment consists of at least three or four different mappings corresponding to different dynamics of the underlying physical processes, (b) an integrated approach that models the different segments of the decomposed flow hydrograph using different techniques is better than a single ANN in modelling the complex, dynamic, non-linear, and fragmented rainfall runoff process, (c) a simple model based on the concept of flow recession is better than an ANN to model the falling limb of a flow hydrograph, and (d) decomposing a flow hydrograph into the different segments corresponding to the different dynamics based on the physical concepts is better than using the soft decomposition employed using SOM.

  13. Microcomputer control of infrared detector arrays used in direct imaging and in Fabry-Perot spectroscopy

    International Nuclear Information System (INIS)

    Rossano, G.S.

    1989-01-01

    A microcomputer based data acquisition system has been developed for astronomical observing with two-dimensional infrared detector arrays operating at high pixel rates. The system is based on a 16-bit 8086/8087 microcomputer operating at 10 MHz. Data rates of up to 560,000 pixels/sec from arrays of up to 4096 elements are supported using the microcomputer system alone. A hardware co-adder the authors are developing permits data accumulation at rates of up to 1.67 million pixels/sec in both staring and chopped data acquisition modes. The system has been used for direct imaging and for data acquisition in a Fabry-Perot Spectrometer developed by NRL. The hardware is operated using interactive software which supports the several available modes of data acquisition, and permits data display and reduction during observing sessions

  14. Hybrid models for hydrological forecasting: Integration of data-driven and conceptual modelling techniques

    NARCIS (Netherlands)

    Corzo Perez, G.A.

    2009-01-01

    This book presents the investigation of different architectures of integrating hydrological knowledge and models with data-driven models for the purpose of hydrological flow forecasting. The models resulting from such integration are referred to as hybrid models. The book addresses the following

  15. Hybrid models for hydrological forecasting : Integration of data-driven and conceptual modelling techniques

    NARCIS (Netherlands)

    Corzo Perez, G.A.

    2009-01-01

    This book presents the investigation of different architectures of integrating hydrological knowledge and models with data-driven models for the purpose of hydrological flow forecasting. The models resulting from such integration are referred to as hybrid models. The book addresses the following

  16. 48 CFR 1552.239-103 - Acquisition of Energy Star Compliant Microcomputers, Including Personal Computers, Monitors and...

    Science.gov (United States)

    2010-10-01

    ... Compliant Microcomputers, Including Personal Computers, Monitors and Printers. 1552.239-103 Section 1552.239... Star Compliant Microcomputers, Including Personal Computers, Monitors and Printers. As prescribed in... Personal Computers, Monitors, and Printers (APR 1996) (a) The Contractor shall provide computer products...

  17. Development of a Microcomputer/Videodisc Aided Math Instructional Management System for Mildly Handicapped Children. Final Report.

    Science.gov (United States)

    Hofmeister, Alan M.

    This final report describes activities and accomplishments of a project which developed, implemented, and evaluated the effectiveness of a microcomputer/videodisc math instructional management system for grades K-4. The system was designed to operate on an APPLE II microcomputer, videodisc player, and input-output devices. It included three…

  18. NEW TECHNIQUE FOR OBESITY SURGERY: INTERNAL GASTRIC PLICATION TECHNIQUE USING INTRAGASTRIC SINGLE-PORT (IGS-IGP) IN EXPERIMENTAL MODEL.

    Science.gov (United States)

    Müller, Verena; Fikatas, Panagiotis; Gül, Safak; Noesser, Maximilian; Fuehrer, Kirs Ten; Sauer, Igor; Pratschke, Johann; Zorron, Ricardo

    2017-01-01

    Bariatric surgery is currently the most effective method to ameliorate co-morbidities as consequence of morbidly obese patients with BMI over 35 kg/m2. Endoscopic techniques have been developed to treat patients with mild obesity and ameliorate comorbidities, but endoscopic skills are needed, beside the costs of the devices. To report a new technique for internal gastric plication using an intragastric single port device in an experimental swine model. Twenty experiments using fresh pig cadaver stomachs in a laparoscopic trainer were performed. The procedure was performed as follow in ten pigs: 1) volume measure; 2) insufflation of the stomach with CO2; 3) extroversion of the stomach through the simulator and installation of the single port device (Gelpoint Applied Mini) through a gastrotomy close to the pylorus; 4) performance of four intragastric handsewn 4-point sutures with Prolene 2-0, from the gastric fundus to the antrum; 5) after the performance, the residual volume was measured. Sleeve gastrectomy was also performed in further ten pigs and pre- and post-procedure gastric volume were measured. The internal gastric plication technique was performed successfully in the ten swine experiments. The mean procedure time was 27±4 min. It produced a reduction of gastric volume of a mean of 51%, and sleeve gastrectomy, a mean of 90% in this swine model. The internal gastric plication technique using an intragastric single port device required few skills to perform, had low operative time and achieved good reduction (51%) of gastric volume in an in vitro experimental model. A cirurgia bariátrica é atualmente o método mais efetivo para melhorar as co-morbidades decorrentes da obesidade mórbida com IMC acima de 35 kg/m2. Técnicas endoscópicas foram desenvolvidas para tratar pacientes com obesidade leve e melhorar as comorbidades, mas habilidades endoscópicas são necessárias, além dos custos. Relatar uma nova técnica para a plicatura gástrica interna

  19. Electricity market price spike analysis by a hybrid data model and feature selection technique

    International Nuclear Information System (INIS)

    Amjady, Nima; Keynia, Farshid

    2010-01-01

    In a competitive electricity market, energy price forecasting is an important activity for both suppliers and consumers. For this reason, many techniques have been proposed to predict electricity market prices in the recent years. However, electricity price is a complex volatile signal owning many spikes. Most of electricity price forecast techniques focus on the normal price prediction, while price spike forecast is a different and more complex prediction process. Price spike forecasting has two main aspects: prediction of price spike occurrence and value. In this paper, a novel technique for price spike occurrence prediction is presented composed of a new hybrid data model, a novel feature selection technique and an efficient forecast engine. The hybrid data model includes both wavelet and time domain variables as well as calendar indicators, comprising a large candidate input set. The set is refined by the proposed feature selection technique evaluating both relevancy and redundancy of the candidate inputs. The forecast engine is a probabilistic neural network, which are fed by the selected candidate inputs of the feature selection technique and predict price spike occurrence. The efficiency of the whole proposed method for price spike occurrence forecasting is evaluated by means of real data from the Queensland and PJM electricity markets. (author)

  20. Machine Learning Techniques for Modelling Short Term Land-Use Change

    Directory of Open Access Journals (Sweden)

    Mileva Samardžić-Petrović

    2017-11-01

    Full Text Available The representation of land use change (LUC is often achieved by using data-driven methods that include machine learning (ML techniques. The main objectives of this research study are to implement three ML techniques, Decision Trees (DT, Neural Networks (NN, and Support Vector Machines (SVM for LUC modeling, in order to compare these three ML techniques and to find the appropriate data representation. The ML techniques are applied on the case study of LUC in three municipalities of the City of Belgrade, the Republic of Serbia, using historical geospatial data sets and considering nine land use classes. The ML models were built and assessed using two different time intervals. The information gain ranking technique and the recursive attribute elimination procedure were implemented to find the most informative attributes that were related to LUC in the study area. The results indicate that all three ML techniques can be used effectively for short-term forecasting of LUC, but the SVM achieved the highest agreement of predicted changes.

  1. Low level waste management: a compilation of models and monitoring techniques. Volume 1

    International Nuclear Information System (INIS)

    Mosier, J.E.; Fowler, J.R.; Barton, C.J.

    1980-04-01

    In support of the National Low-Level Waste (LLW) Management Research and Development Program being carried out at Oak Ridge National Laboratory, Science Applications, Inc., conducted a survey of models and monitoring techniques associated with the transport of radionuclides and other chemical species from LLW burial sites. As a result of this survey, approximately 350 models were identified. For each model the purpose and a brief description are presented. To the extent possible, a point of contact and reference material are identified. The models are organized into six technical categories: atmospheric transport, dosimetry, food chain, groundwater transport, soil transport, and surface water transport. About 4% of the models identified covered other aspects of LLW management and are placed in a miscellaneous category. A preliminary assessment of all these models was performed to determine their ability to analyze the transport of other chemical species. The models that appeared to be applicable are identified. A brief survey of the state-of-the-art techniques employed to monitor LLW burial sites is also presented, along with a very brief discussion of up-to-date burial techniques

  2. Low level waste management: a compilation of models and monitoring techniques. Volume 1

    Energy Technology Data Exchange (ETDEWEB)

    Mosier, J.E.; Fowler, J.R.; Barton, C.J. (comps.)

    1980-04-01

    In support of the National Low-Level Waste (LLW) Management Research and Development Program being carried out at Oak Ridge National Laboratory, Science Applications, Inc., conducted a survey of models and monitoring techniques associated with the transport of radionuclides and other chemical species from LLW burial sites. As a result of this survey, approximately 350 models were identified. For each model the purpose and a brief description are presented. To the extent possible, a point of contact and reference material are identified. The models are organized into six technical categories: atmospheric transport, dosimetry, food chain, groundwater transport, soil transport, and surface water transport. About 4% of the models identified covered other aspects of LLW management and are placed in a miscellaneous category. A preliminary assessment of all these models was performed to determine their ability to analyze the transport of other chemical species. The models that appeared to be applicable are identified. A brief survey of the state-of-the-art techniques employed to monitor LLW burial sites is also presented, along with a very brief discussion of up-to-date burial techniques.

  3. A review of techniques for spatial modeling in geographical, conservation and landscape genetics.

    Science.gov (United States)

    Diniz-Filho, José Alexandre Felizola; Nabout, João Carlos; de Campos Telles, Mariana Pires; Soares, Thannya Nascimento; Rangel, Thiago Fernando L V B

    2009-04-01

    Most evolutionary processes occur in a spatial context and several spatial analysis techniques have been employed in an exploratory context. However, the existence of autocorrelation can also perturb significance tests when data is analyzed using standard correlation and regression techniques on modeling genetic data as a function of explanatory variables. In this case, more complex models incorporating the effects of autocorrelation must be used. Here we review those models and compared their relative performances in a simple simulation, in which spatial patterns in allele frequencies were generated by a balance between random variation within populations and spatially-structured gene flow. Notwithstanding the somewhat idiosyncratic behavior of the techniques evaluated, it is clear that spatial autocorrelation affects Type I errors and that standard linear regression does not provide minimum variance estimators. Due to its flexibility, we stress that principal coordinate of neighbor matrices (PCNM) and related eigenvector mapping techniques seem to be the best approaches to spatial regression. In general, we hope that our review of commonly used spatial regression techniques in biology and ecology may aid population geneticists towards providing better explanations for population structures dealing with more complex regression problems throughout geographic space.

  4. Evaluation of mesh morphing and mapping techniques in patient specific modeling of the human pelvis.

    Science.gov (United States)

    Salo, Zoryana; Beek, Maarten; Whyne, Cari Marisa

    2013-01-01

    Robust generation of pelvic finite element models is necessary to understand the variation in mechanical behaviour resulting from differences in gender, aging, disease and injury. The objective of this study was to apply and evaluate mesh morphing and mapping techniques to facilitate the creation and structural analysis of specimen-specific finite element (FE) models of the pelvis. A specimen-specific pelvic FE model (source mesh) was generated following a traditional user-intensive meshing scheme. The source mesh was morphed onto a computed tomography scan generated target surface of a second pelvis using a landmarked-based approach, in which exterior source nodes were shifted to target surface vertices, while constrained along a normal. A second copy of the morphed model was further refined through mesh mapping, in which surface nodes of the initial morphed model were selected in patches and remapped onto the surfaces of the target model. Computed tomography intensity based material properties were assigned to each model. The source, target, morphed and mapped models were analyzed under axial compression using linear static FE analysis and their strain distributions evaluated. Morphing and mapping techniques were effectively applied to generate good quality geometrically complex specimen-specific pelvic FE models. Mapping significantly improved strain concurrence with the target pelvis FE model. Copyright © 2012 John Wiley & Sons, Ltd.

  5. Evaluation of mesh morphing and mapping techniques in patient specific modelling of the human pelvis.

    Science.gov (United States)

    Salo, Zoryana; Beek, Maarten; Whyne, Cari Marisa

    2012-08-01

    Robust generation of pelvic finite element models is necessary to understand variation in mechanical behaviour resulting from differences in gender, aging, disease and injury. The objective of this study was to apply and evaluate mesh morphing and mapping techniques to facilitate the creation and structural analysis of specimen-specific finite element (FE) models of the pelvis. A specimen-specific pelvic FE model (source mesh) was generated following a traditional user-intensive meshing scheme. The source mesh was morphed onto a computed tomography scan generated target surface of a second pelvis using a landmarked-based approach, in which exterior source nodes were shifted to target surface vertices, while constrained along a normal. A second copy of the morphed model was further refined through mesh mapping, in which surface nodes of the initial morphed model were selected in patches and remapped onto the surfaces of the target model. Computed tomography intensity-based material properties were assigned to each model. The source, target, morphed and mapped models were analyzed under axial compression using linear static FE analysis, and their strain distributions were evaluated. Morphing and mapping techniques were effectively applied to generate good quality and geometrically complex specimen-specific pelvic FE models. Mapping significantly improved strain concurrence with the target pelvis FE model. Copyright © 2012 John Wiley & Sons, Ltd.

  6. New techniques for the analysis of manual control systems. [mathematical models of human operator behavior

    Science.gov (United States)

    Bekey, G. A.

    1971-01-01

    Studies are summarized on the application of advanced analytical and computational methods to the development of mathematical models of human controllers in multiaxis manual control systems. Specific accomplishments include the following: (1) The development of analytical and computer methods for the measurement of random parameters in linear models of human operators. (2) Discrete models of human operator behavior in a multiple display situation were developed. (3) Sensitivity techniques were developed which make possible the identification of unknown sampling intervals in linear systems. (4) The adaptive behavior of human operators following particular classes of vehicle failures was studied and a model structure proposed.

  7. Modeling of PV Systems Based on Inflection Points Technique Considering Reverse Mode

    Directory of Open Access Journals (Sweden)

    Bonie J. Restrepo-Cuestas

    2013-11-01

    Full Text Available This paper proposes a methodology for photovoltaic (PV systems modeling, considering their behavior in both direct and reverse operating mode and considering mismatching conditions. The proposed methodology is based on the inflection points technique with a linear approximation to model the bypass diode and a simplified PV model. The proposed mathematical model allows to evaluate the energetic performance of a PV system, exhibiting short simulation times in large PV systems. In addition, this methodology allows to estimate the condition of the modules affected by the partial shading since it is possible to know the power dissipated due to its operation at the second quadrant.

  8. Control System Design for Cylindrical Tank Process Using Neural Model Predictive Control Technique

    Directory of Open Access Journals (Sweden)

    M. Sridevi

    2010-10-01

    Full Text Available Chemical manufacturing and process industry requires innovative technologies for process identification. This paper deals with model identification and control of cylindrical process. Model identification of the process was done using ARMAX technique. A neural model predictive controller was designed for the identified model. The performance of the controllers was evaluated using MATLAB software. The performance of NMPC controller was compared with Smith Predictor controller and IMC controller based on rise time, settling time, overshoot and ISE and it was found that the NMPC controller is better suited for this process.

  9. Coupled Numerical Methods to Analyze Interacting Acoustic-Dynamic Models by Multidomain Decomposition Techniques

    Directory of Open Access Journals (Sweden)

    Delfim Soares

    2011-01-01

    Full Text Available In this work, coupled numerical analysis of interacting acoustic and dynamic models is focused. In this context, several numerical methods, such as the finite difference method, the finite element method, the boundary element method, meshless methods, and so forth, are considered to model each subdomain of the coupled model, and multidomain decomposition techniques are applied to deal with the coupling relations. Two basic coupling algorithms are discussed here, namely the explicit direct coupling approach and the implicit iterative coupling approach, which are formulated based on explicit/implicit time-marching techniques. Completely independent spatial and temporal discretizations among the interacting subdomains are permitted, allowing optimal discretization for each sub-domain of the model to be considered. At the end of the paper, numerical results are presented, illustrating the performance and potentialities of the discussed methodologies.

  10. Animal models in bariatric surgery--a review of the surgical techniques and postsurgical physiology.

    Science.gov (United States)

    Rao, Raghavendra S; Rao, Venkatesh; Kini, Subhash

    2010-09-01

    Bariatric surgery is considered the most effective current treatment for morbid obesity. Since the first publication of an article by Kremen, Linner, and Nelson, many experiments have been performed using animal models. The initial experiments used only malabsorptive procedures like intestinal bypass which have largely been abandoned now. These experimental models have been used to assess feasibility and safety as well as to refine techniques particular to each procedure. We will discuss the surgical techniques and the postsurgical physiology of the four major current bariatric procedures (namely, Roux-en-Y gastric bypass, gastric banding, sleeve gastrectomy, and biliopancreatic diversion). We have also reviewed the anatomy and physiology of animal models. We have reviewed the literature and presented it such that it would be a reference to an investigator interested in animal experiments in bariatric surgery. Experimental animal models are further divided into two categories: large mammals that include dogs, cats, rabbits, and pig and small mammals that include rats and mice.

  11. The application and development on radiation monitoring microcomputer management system in design of a certain project

    International Nuclear Information System (INIS)

    Zhang Hongyou

    1993-01-01

    A scheme of a radiation monitoring system with a RMMMS (Radiation Monitoring Microcomputer Management System) has been designed for the first time in the radiation protection design of a certain nuclear projection undertaken by the BINE (Beijing Institute of Nuclear Engineering). Meanwhile, we accepted the research task of the RMMMS that can manage 40 monitoring channel. The key factors of radiation monitors, microcomputer, information management and systematic design method are considered in the development of the RMMMS. This paper presents briefly the scheme and functions of the RMMMS

  12. Independent multi-channel analyzer with the color display and microcomputer in the CAMAC standard

    International Nuclear Information System (INIS)

    Gyunter, Z.; Elizarov, O.I.; Zhukov, G.P.; Mikhaehlis, B.; Shul'tts, K.Kh.

    1983-01-01

    An independent multi-channel time-of-flight analyzer developed for a spectrometer of polarised neutrons and used in experiments with the IBR-2 pulse reactor is described. Different cyclic modes of measuring are realized in the analyzer, 4 kwords of analyzer momory being assigned for each mode. The spectra are displayed on a colour screen during measurements. Simultaneous displaying of up to 8 spectra is possible. A microcomputer transfers the spectra from the buffer analyzer memory to the microcomputer memory. The accumulated information is transferred to the PDP-11/70 central computer

  13. Analysis of Multipath Mitigation Techniques with Land Mobile Satellite Channel Model

    Directory of Open Access Journals (Sweden)

    M. Z. H. Bhuiyan J. Zhang

    2012-12-01

    Full Text Available Multipath is undesirable for Global Navigation Satellite System (GNSS receivers, since the reception of multipath can create a significant distortion to the shape of the correlation function leading to an error in the receivers’ position estimate. Many multipath mitigation techniques exist in the literature to deal with the multipath propagation problem in the context of GNSS. The multipath studies in the literature are often based on optimistic assumptions, for example, assuming a static two-path channel or a fading channel with a Rayleigh or a Nakagami distribution. But, in reality, there are a lot of channel modeling issues, for example, satellite-to-user geometry, variable number of paths, variable path delays and gains, Non Line-Of-Sight (NLOS path condition, receiver movements, etc. that are kept out of consideration when analyzing the performance of these techniques. Therefore, this is of utmost importance to analyze the performance of different multipath mitigation techniques in some realistic measurement-based channel models, for example, the Land Multipath is undesirable for Global Navigation Satellite System (GNSS receivers, since the reception of multipath can create a significant distortion to the shape of the correlation function leading to an error in the receivers’ position estimate. Many multipath mitigation techniques exist in the literature to deal with the multipath propagation problem in the context of GNSS. The multipath studies in the literature are often based on optimistic assumptions, for example, assuming a static two-path channel or a fading channel with a Rayleigh or a Nakagami distribution. But, in reality, there are a lot of channel modeling issues, for example, satellite-to-user geometry, variable number of paths, variable path delays and gains, Non Line-Of-Sight (NLOS path condition, receiver movements, etc. that are kept out of consideration when analyzing the performance of these techniques. Therefore, this

  14. Modeling and comparative study of various detection techniques for FMCW LIDAR using optisystem

    Science.gov (United States)

    Elghandour, Ahmed H.; Ren, Chen D.

    2013-09-01

    In this paper we investigated the different detection techniques especially direct detection, coherent heterodyne detection and coherent homodyne detection on FMCW LIDAR system using Optisystem package. A model for target, propagation channel and various detection techniques were developed using Optisystem package and then a comparative study among various detection techniques for FMCW LIDAR systems is done analytically and simulated using the developed model. Performance of direct detection, heterodyne detection and homodyne detection for FMCW LIDAR system was calculated and simulated using Optisystem package. The output simulated performance was checked using simulated results of MATLAB simulator. The results shows that direct detection is sensitive to the intensity of the received electromagnetic signal and has low complexity system advantage over the others detection architectures at the expense of the thermal noise is the dominant noise source and the sensitivity is relatively poor. In addition to much higher detection sensitivity can be achieved using coherent optical mixing which is performed by heterodyne and homodyne detection.

  15. Assessing the validity of two indirect questioning techniques: A Stochastic Lie Detector versus the Crosswise Model.

    Science.gov (United States)

    Hoffmann, Adrian; Musch, Jochen

    2016-09-01

    Estimates of the prevalence of sensitive attributes obtained through direct questions are prone to being distorted by untruthful responding. Indirect questioning procedures such as the Randomized Response Technique (RRT) aim to control for the influence of social desirability bias. However, even on RRT surveys, some participants may disobey the instructions in an attempt to conceal their true status. In the present study, we experimentally compared the validity of two competing indirect questioning techniques that presumably offer a solution to the problem of nonadherent respondents: the Stochastic Lie Detector and the Crosswise Model. For two sensitive attributes, both techniques met the "more is better" criterion. Their application resulted in higher, and thus presumably more valid, prevalence estimates than a direct question. Only the Crosswise Model, however, adequately estimated the known prevalence of a nonsensitive control attribute.

  16. Stakeholder approach, Stakeholders mental model: A visualization test with cognitive mapping technique

    Directory of Open Access Journals (Sweden)

    Garoui Nassreddine

    2012-04-01

    Full Text Available The idea of this paper is to determine the mental models of actors in the firm with respect to the stakeholder approach of corporate governance. The use of the cognitive map to view these diagrams to show the ways of thinking and conceptualization of the stakeholder approach. The paper takes a corporate governance perspective, discusses stakeholder model. It takes also a cognitive mapping technique.

  17. Similitude Conditions Modeling Geosynthetic-Reinforced Piled Embankments Using FEM and FDM Techniques

    OpenAIRE

    Jennings, Keith; Naughton, Patrick J.

    2012-01-01

    The numerical modelling of geosynthetic-reinforced piled embankments using both the finite element method (FEM) and finite difference method (FDM) are compared. Plaxis 2D (FEM) was utilized to replicate FLAC (FDM) analysis originally presented by Han and Gabr on a unit cell axisymmetric model within a geosynthetic reinforced piled embankment (GRPE). The FEM and FED techniques were found to be in reasonable agreement, in both characteristic trend and absolute value. FEM consistently replicated...

  18. An open microcomputer-based laboratory system for perceptional experimentality

    Science.gov (United States)

    Hamalainen, Ari

    A computer, equipped with hardware for acquiring data about the properties of a physical system and programs for processing that data, is a powerful tool for physics research and instruction. There is strong evidence that utilizing microcomputer-based laboratories (MBLs) in instruction can lead to significantly improved learning. The perceptional approach is a method for physics instruction, developed at the Department of Physics, University of Helsinki. Its main arguments are that the meanings of the concepts must be learnt before their formal definitions and adoption, and that learning and research are fundamentally similar concept formation processes. Applying the perceptional approach requires the ability to perform quantitative experiments, either as students' laboratory exercises or as lecture demonstrations, and to process their results. MBL tools are essential for this. In student's laboratory exercises, they reduce the routine work and leave more time for the actual learning. In lecture demonstrations, they make it possible to perform the experiments in the tight time limits. At a previous stage of the research, a set of requirements was found that the perceptional approach places on MBL systems. The primary goal of this thesis is to build a prototype of a MBL system that would fulfil these requirements. A secondary goal is to describe technical aspects of a computerized measurement system from the standpoint of educational use. The prototype was built using mostly commercial sensors and data acquisition units. The software was written with a visual programming language, designed for instrumentation applications. The prototype system was developed and tested with a set of demonstrations of various topics in the Finnish high school physics curriculum, which were implemented according to the perceptional approach. Limited usability tests were also performed. The prototype was improved, until it could perform the test demonstrations. It was found to meet the

  19. Micro-computed tomography of fatigue microdamage in cortical bone using a barium sulfate contrast agent.

    Science.gov (United States)

    Leng, Huijie; Wang, Xiang; Ross, Ryan D; Niebur, Glen L; Roeder, Ryan K

    2008-01-01

    Accumulation of microdamage during fatigue can lead to increased fracture susceptibility in bone. Current techniques for imaging microdamage in bone are inherently destructive and two-dimensional. Therefore, the objective of this study was to image the accumulation of fatigue microdamage in cortical bone using micro-computed tomography (micro-CT) with a barium sulfate (BaSO(4)) contrast agent. Two symmetric notches were machined on the tensile surface of bovine cortical bone beams in order to generate damage ahead of the stress concentrations during four-point bending fatigue. Specimens were loaded to a specified number of cycles or until one notch fractured, such that the other notch exhibited the accumulation of microdamage prior to fracture. Microdamage ahead of the notch was stained in vitro by precipitation of BaSO(4) and imaged using micro-CT. Reconstructed images showed a distinct region of bright voxels around the notch tip or along propagating cracks due to the presence of BaSO(4), which was verified by backscattered electron imaging and energy dispersive spectroscopy. The shape of the stained region ahead of the notch tip was consistent with principal strain contours calculated by finite element analysis. The relative volume of the stained region was correlated with the number of loading cycles by non-linear regression using a power-law. This study demonstrates new methods for the non-destructive and three-dimensional detection of fatigue microdamage accumulation in cortical bone in vitro, which may be useful to gain further understanding into the role of microdamage in bone fragility.

  20. Validation of micro-computed tomography for occlusal caries detection: an in vitro study

    Directory of Open Access Journals (Sweden)

    Gökhan ÖZKAN

    2015-01-01

    Full Text Available Conventional methods of caries detection, including the gold standard of histological examination, have certain disadvantages that must be addressed prior to validating any other diagnostic technique—current or new. Here we evaluated the validity of micro-computed tomography (micro-CT as an alternative gold-standard technique for caries detection. Sixty teeth with suspected occlusal caries were chosen from a pool of teeth extracted for orthodontic, periodontal, or surgical reasons. Identical reference points were marked on photographs taken for teeth and were used to evaluate each method. Dimensions of caries were assessed by two calibrated examiners using the ICDAS-II visual examination system, bitewing radiographs, and micro-CT. The teeth included in the study were selected randomly from solution before all measurements. For micro-CT, the device was set to 50 kV, 800 µA, pixel size 15 µm (at 1024 × 1024 resolution, and 1° rotation step. NRecon software (SkyScan was used to obtain reconstructed images. For each diagnostic method, results were compared with histology results using the McNemar test. Receiver operator characteristic (ROC analysis was also performed for each method (Z-test; p < 0.05. Besides showing a high correlation with histology results, micro-CT yielded the greatest values at the D3 threshold; moreover, accuracy and area under the ROC curve (AUC values were greatest at the D1threshold. Our results indicate that micro-CT performs as well as or better than histological examination for the purpose of comparing methods for caries detection.

  1. Finding of Correction Factor and Dimensional Error in Bio-AM Model by FDM Technique

    Science.gov (United States)

    Manmadhachary, Aiamunoori; Ravi Kumar, Yennam; Krishnanand, Lanka

    2016-06-01

    Additive Manufacturing (AM) is the swift manufacturing process, in which input data can be provided from various sources like 3-Dimensional (3D) Computer Aided Design (CAD), Computed Tomography (CT), Magnetic Resonance Imaging (MRI) and 3D scanner data. From the CT/MRI data can be manufacture Biomedical Additive Manufacturing (Bio-AM) models. The Bio-AM model gives a better lead on preplanning of oral and maxillofacial surgery. However manufacturing of the accurate Bio-AM model is one of the unsolved problems. The current paper demonstrates error between the Standard Triangle Language (STL) model to Bio-AM model of dry mandible and found correction factor in Bio-AM model with Fused Deposition Modelling (FDM) technique. In the present work dry mandible CT images are acquired by CT scanner and supplied into a 3D CAD model in the form of STL model. Further the data is sent to FDM machine for fabrication of Bio-AM model. The difference between Bio-AM to STL model dimensions is considered as dimensional error and the ratio of STL to Bio-AM model dimensions considered as a correction factor. This correction factor helps to fabricate the AM model with accurate dimensions of the patient anatomy. These true dimensional Bio-AM models increasing the safety and accuracy in pre-planning of oral and maxillofacial surgery. The correction factor for Dimension SST 768 FDM AM machine is 1.003 and dimensional error is limited to 0.3 %.

  2. Extending the reach of strong-coupling: an iterative technique for Hamiltonian lattice models

    International Nuclear Information System (INIS)

    Alberty, J.; Greensite, J.; Patkos, A.

    1983-12-01

    The authors propose an iterative method for doing lattice strong-coupling-like calculations in a range of medium to weak couplings. The method is a modified Lanczos scheme, with greatly improved convergence properties. The technique is tested on the Mathieu equation and on a Hamiltonian finite-chain XY model, with excellent results. (Auth.)

  3. Particle Markov Chain Monte Carlo Techniques of Unobserved Component Time Series Models Using Ox

    DEFF Research Database (Denmark)

    Nonejad, Nima

    This paper details Particle Markov chain Monte Carlo techniques for analysis of unobserved component time series models using several economic data sets. PMCMC combines the particle filter with the Metropolis-Hastings algorithm. Overall PMCMC provides a very compelling, computationally fast...

  4. Analysis of fluidized bed granulation process using conventional and novel modeling techniques.

    Science.gov (United States)

    Petrović, Jelena; Chansanroj, Krisanin; Meier, Brigitte; Ibrić, Svetlana; Betz, Gabriele

    2011-10-09

    Various modeling techniques have been applied to analyze fluidized-bed granulation process. Influence of various input parameters (product, inlet and outlet air temperature, consumption of liquid-binder, granulation liquid-binder spray rate, spray pressure, drying time) on granulation output properties (granule flow rate, granule size determined using light scattering method and sieve analysis, granules Hausner ratio, porosity and residual moisture) has been assessed. Both conventional and novel modeling techniques were used, such as screening test, multiple regression analysis, self-organizing maps, artificial neural networks, decision trees and rule induction. Diverse testing of developed models (internal and external validation) has been discussed. Good correlation has been obtained between the predicted and the experimental data. It has been shown that nonlinear methods based on artificial intelligence, such as neural networks, are far better in generalization and prediction in comparison to conventional methods. Possibility of usage of SOMs, decision trees and rule induction technique to monitor and optimize fluidized-bed granulation process has also been demonstrated. Obtained findings can serve as guidance to implementation of modeling techniques in fluidized-bed granulation process understanding and control. Copyright © 2011 Elsevier B.V. All rights reserved.

  5. Application of modelling techniques in the food industry: determination of shelf-life for chilled foods

    NARCIS (Netherlands)

    Membré, J.M.; Johnston, M.D.; Bassett, J.; Naaktgeboren, G.; Blackburn, W.; Gorris, L.G.M.

    2005-01-01

    Microbiological modelling techniques (predictive microbiology, the Bayesian Markov Chain Monte Carlo method and a probability risk assessment approach) were combined to assess the shelf-life of an in-pack heat-treated, low-acid sauce intended to be marketed under chilled conditions. From a safety

  6. Reduced order modelling techniques for mesh movement strategies as applied to fluid structure interactions

    CSIR Research Space (South Africa)

    Bogaers, Alfred EJ

    2010-01-01

    Full Text Available In this paper, we implement the method of Proper Orthogonal Decomposition (POD) to generate a reduced order model (ROM) of an optimization based mesh movement technique. In the study it is shown that POD can be used effectively to generate a ROM...

  7. New model reduction technique for a class of parabolic partial differential equations

    NARCIS (Netherlands)

    Vajta, Miklos

    1991-01-01

    A model reduction (or lumping) technique for a class of parabolic-type partial differential equations is given, and its application is discussed. The frequency response of the temperature distribution in any multilayer solid is developed and given by a matrix expression. The distributed transfer

  8. Prediction of Monthly Summer Monsoon Rainfall Using Global Climate Models Through Artificial Neural Network Technique

    Science.gov (United States)

    Nair, Archana; Singh, Gurjeet; Mohanty, U. C.

    2018-01-01

    The monthly prediction of summer monsoon rainfall is very challenging because of its complex and chaotic nature. In this study, a non-linear technique known as Artificial Neural Network (ANN) has been employed on the outputs of Global Climate Models (GCMs) to bring out the vagaries inherent in monthly rainfall prediction. The GCMs that are considered in the study are from the International Research Institute (IRI) (2-tier CCM3v6) and the National Centre for Environmental Prediction (Coupled-CFSv2). The ANN technique is applied on different ensemble members of the individual GCMs to obtain monthly scale prediction over India as a whole and over its spatial grid points. In the present study, a double-cross-validation and simple randomization technique was used to avoid the over-fitting during training process of the ANN model. The performance of the ANN-predicted rainfall from GCMs is judged by analysing the absolute error, box plots, percentile and difference in linear error in probability space. Results suggest that there is significant improvement in prediction skill of these GCMs after applying the ANN technique. The performance analysis reveals that the ANN model is able to capture the year to year variations in monsoon months with fairly good accuracy in extreme years as well. ANN model is also able to simulate the correct signs of rainfall anomalies over different spatial points of the Indian domain.

  9. Flexible multibody simulation of automotive systems with non-modal model reduction techniques

    Science.gov (United States)

    Shiiba, Taichi; Fehr, Jörg; Eberhard, Peter

    2012-12-01

    The stiffness of the body structure of an automobile has a strong relationship with its noise, vibration, and harshness (NVH) characteristics. In this paper, the effect of the stiffness of the body structure upon ride quality is discussed with flexible multibody dynamics. In flexible multibody simulation, the local elastic deformation of the vehicle has been described traditionally with modal shape functions. Recently, linear model reduction techniques from system dynamics and mathematics came into the focus to find more sophisticated elastic shape functions. In this work, the NVH-relevant states of a racing kart are simulated, whereas the elastic shape functions are calculated with modern model reduction techniques like moment matching by projection on Krylov-subspaces, singular value decomposition-based reduction techniques, and combinations of those. The whole elastic multibody vehicle model consisting of tyres, steering, axle, etc. is considered, and an excitation with a vibration characteristics in a wide frequency range is evaluated in this paper. The accuracy and the calculation performance of those modern model reduction techniques is investigated including a comparison of the modal reduction approach.

  10. The commercial use of segmentation and predictive modeling techniques for database marketing in the Netherlands

    NARCIS (Netherlands)

    Verhoef, PC; Spring, PN; Hoekstra, JC; Leeflang, PSH

    Although the application of segmentation and predictive modeling is an important topic in the database marketing (DBM) literature, no study has yet investigated the extent of adoption of these techniques. We present the results of a Dutch survey involving 228 database marketing companies. We find

  11. Dimensional Analysis: an Elegant Technique for Facilitating the Teaching of Mathematical Modelling.

    Science.gov (United States)

    Fay, Temple H.; Joubert, Stephan V.

    2002-01-01

    Dimension analysis is promoted as a technique that promotes better understanding of the role of units and dimensions in mathematical modeling problems. Dimensional analysis is shown to lead to interesting systems of linear equations to solve, and can point the way to more quantitative analysis. Two student problems are discussed. (Author/MM)

  12. Nonlinear modelling of polymer electrolyte membrane fuel cell stack using nonlinear cancellation technique

    International Nuclear Information System (INIS)

    Barus, R. P. P.; Tjokronegoro, H. A.; Leksono, E.; Ismunandar

    2014-01-01

    Fuel cells are promising new energy conversion devices that are friendly to the environment. A set of control systems are required in order to operate a fuel cell based power plant system optimally. For the purpose of control system design, an accurate fuel cell stack model in describing the dynamics of the real system is needed. Currently, linear model are widely used for fuel cell stack control purposes, but it has limitations in narrow operation range. While nonlinear models lead to nonlinear control implemnetation whos more complex and hard computing. In this research, nonlinear cancellation technique will be used to transform a nonlinear model into a linear form while maintaining the nonlinear characteristics. The transformation is done by replacing the input of the original model by a certain virtual input that has nonlinear relationship with the original input. Then the equality of the two models is tested by running a series of simulation. Input variation of H2, O2 and H2O as well as disturbance input I (current load) are studied by simulation. The error of comparison between the proposed model and the original nonlinear model are less than 1 %. Thus we can conclude that nonlinear cancellation technique can be used to represent fuel cell nonlinear model in a simple linear form while maintaining the nonlinear characteristics and therefore retain the wide operation range

  13. Bone diagnosis by X-ray techniques

    Energy Technology Data Exchange (ETDEWEB)

    Lima, I. [Nuclear Engineering Program/COPPE/UFRJ, P.O. Box 68509, Av. Horacio Macedo, 2030, Sala I-133, Cidade Universitaria, Zip Code: 21941-972 Rio de Janeiro, RJ (Brazil)], E-mail: inaya@lin.ufrj.br; Anjos, M.J. [Nuclear Engineering Program/COPPE/UFRJ, P.O. Box 68509, Av. Horacio Macedo, 2030, Sala I-133, Cidade Universitaria, Zip Code: 21941-972 Rio de Janeiro, RJ (Brazil); Physics Institute, UERJ (Brazil); Farias, M.L.F. [University Hospital, UFRJ (Brazil); Parcegoni, N.; Rosenthal, D. [Biophysics Institute, UFRJ (Brazil); Duarte, M.E.L. [Histologic and Embriology Department, UFRJ (Brazil); Lopes, R.T. [Nuclear Engineering Program/COPPE/UFRJ, P.O. Box 68509, Av. Horacio Macedo, 2030, Sala I-133, Cidade Universitaria, Zip Code: 21941-972 Rio de Janeiro, RJ (Brazil)

    2008-12-15

    In this work, two X-ray techniques used were 3D microcomputed tomography (micro-CT) and X-ray microfluorescence (micro-XRF) in order to investigate the internal structure of the bone samples. Those two techniques work together, e.g. as a complement to each other, to characterize bones structure and composition. Initially, the specimens were used to do the scan procedure in the microcomputer tomography system and the second step consists of doing the X-ray microfluorescence analysis. The results show that both techniques are powerful methods for analyzing, inspecting and characterizing bone samples: they are alternative procedures for examining bone structures and compositions and they are complementary.

  14. A characteristic study of CCF modeling techniques and optimization of CCF defense strategies

    International Nuclear Information System (INIS)

    Kim, Min Chull

    2000-02-01

    Common Cause Failures (CCFs ) are among the major contributors to risk and core damage frequency (CDF ) from operating nuclear power plants (NPPs ). Our study on CCF focused on the following aspects : 1) a characteristic study on the CCF modeling techniques and 2) development of the optimal CCF defense strategy. Firstly, the characteristics of CCF modeling techniques were studied through sensitivity study of CCF occurrence probability upon system redundancy. The modeling techniques considered in this study include those most widely used worldwide, i.e., beta factor, MGL, alpha factor, and binomial failure rate models. We found that MGL and alpha factor models are essentially identical in terms of the CCF probability. Secondly, in the study for CCF defense, the various methods identified in the previous studies for defending against CCF were classified into five different categories. Based on these categories, we developed a generic method by which the optimal CCF defense strategy can be selected. The method is not only qualitative but also quantitative in nature: the selection of the optimal strategy among candidates is based on the use of analytic hierarchical process (AHP). We applied this method to two motor-driven valves for containment sump isolation in Ulchin 3 and 4 nuclear power plants. The result indicates that the method for developing an optimal CCF defense strategy is effective

  15. A MECHANISTIC MODEL FOR PARTICLE DEPOSITION IN DIESEL PARTICLUATE FILTERS USING THE LATTICE-BOLTZMANN TECHNIQUE

    Energy Technology Data Exchange (ETDEWEB)

    Stewart, Mark L.; Rector, David R.; Muntean, George G.; Maupin, Gary D.

    2004-08-01

    Cordierite diesel particulate filters (DPFs) offer one of the most promising aftertreatment technologies to meet the quickly approaching EPA 2007 heavy-duty emissions regulations. A critical, yet poorly understood, component of particulate filter modeling is the representation of soot deposition. The structure and distribution of soot deposits upon and within the ceramic substrate directly influence many of the macroscopic phenomenon of interest, including filtration efficiency, back pressure, and filter regeneration. Intrinsic soot cake properties such as packing density and permeability coefficients remain inadequately characterized. The work reported in this paper involves subgrid modeling techniques which may prove useful in resolving these inadequacies. The technique involves the use of a lattice Boltzmann modeling approach. This approach resolves length scales which are orders of magnitude below those typical of a standard computational fluid dynamics (CFD) representation of an aftertreatment device. Individual soot particles are introduced and tracked as they move through the flow field and are deposited on the filter substrate or previously deposited particles. Electron micrographs of actual soot deposits were taken and compared to the model predictions. Descriptions of the modeling technique and the development of the computational domain are provided. Preliminary results are presented, along with some comparisons with experimental observations.

  16. A New Profile Learning Model for Recommendation System based on Machine Learning Technique

    Directory of Open Access Journals (Sweden)

    Shereen H. Ali

    2016-03-01

    Full Text Available Recommender systems (RSs have been used to successfully address the information overload problem by providing personalized and targeted recommendations to the end users. RSs are software tools and techniques providing suggestions for items to be of use to a user, hence, they typically apply techniques and methodologies from Data Mining. The main contribution of this paper is to introduce a new user profile learning model to promote the recommendation accuracy of vertical recommendation systems. The proposed profile learning model employs the vertical classifier that has been used in multi classification module of the Intelligent Adaptive Vertical Recommendation (IAVR system to discover the user’s area of interest, and then build the user’s profile accordingly. Experimental results have proven the effectiveness of the proposed profile learning model, which accordingly will promote the recommendation accuracy.

  17. Reduced technique for modeling electromagnetic immunity on braid shielding cable bundles

    International Nuclear Information System (INIS)

    Xiao Pei; Du Ping-An; Nie Bao-Lin; Ren Dan

    2017-01-01

    In this paper, an efficient multi-conductor simplification technique is proposed to model the electromagnetic immunity on cable bundles within a braid shielding structure over a large frequency range. By grouping together the conductors based on the knowledge of Z -Smith chart, the required computation time is markedly reduced and the complexity of modeling the completely shielding cable bundles is significantly simplified with a good accuracy. After a brief description of the immunity problems in shielding structure, a six-phase procedure is detailed to generate the geometrical characteristics of the reduced cable bundles. Numerical simulation is carried out by using a commercial software CST to validate the efficiency and advantages of the proposed approach. The research addressed in this paper is considered as a simplified modeling technique for the electromagnetic immunity within a shielding structure. (paper)

  18. A Multi-Model Reduction Technique for Optimization of Coupled Structural-Acoustic Problems

    DEFF Research Database (Denmark)

    Creixell Mediante, Ester; Jensen, Jakob Søndergaard; Brunskog, Jonas

    2016-01-01

    Finite Element models of structural-acoustic coupled systems can become very large for complex structures with multiple connected parts. Optimization of the performance of the structure based on harmonic analysis of the system requires solving the coupled problem iteratively and for several...... frequencies, which can become highly time consuming. Several modal-based model reduction techniques for structure-acoustic interaction problems have been developed in the literature. The unsymmetric nature of the pressure-displacement formulation of the problem poses the question of how the reduction modal...... base should be formed, given that the modal vectors are not orthogonal due to the asymmetry of the system matrices. In this paper, a multi-model reduction (MMR) technique for structure-acoustic interaction problems is developed. In MMR, the reduction base is formed with the modal vectors of a family...

  19. Finite-element-model updating using computational intelligence techniques applications to structural dynamics

    CERN Document Server

    Marwala, Tshilidzi

    2010-01-01

    Finite element models (FEMs) are widely used to understand the dynamic behaviour of various systems. FEM updating allows FEMs to be tuned better to reflect measured data and may be conducted using two different statistical frameworks: the maximum likelihood approach and Bayesian approaches. Finite Element Model Updating Using Computational Intelligence Techniques applies both strategies to the field of structural mechanics, an area vital for aerospace, civil and mechanical engineering. Vibration data is used for the updating process. Following an introduction a number of computational intelligence techniques to facilitate the updating process are proposed; they include: • multi-layer perceptron neural networks for real-time FEM updating; • particle swarm and genetic-algorithm-based optimization methods to accommodate the demands of global versus local optimization models; • simulated annealing to put the methodologies into a sound statistical basis; and • response surface methods and expectation m...

  20. [Eco-value level classification model of forest ecosystem based on modified projection pursuit technique].

    Science.gov (United States)

    Wu, Chengzhen; Hong, Wei; Hong, Tao

    2006-03-01

    To optimize the projection function and direction of projection pursuit technique, predigest its realization process, and overcome the shortcomings in long time calculation and in the difficulty of optimizing projection direction and computer programming, this paper presented a modified simplex method (MSM), and based on it, brought forward the eco-value level classification model (EVLCM) of forest ecosystem, which could integrate the multidimensional classification index into one-dimensional projection value, with high projection value denoting high ecosystem services value. Examples of forest ecosystem could be reasonably classified by the new model according to their projection value, suggesting that EVLCM driven directly by samples data of forest ecosystem was simple and feasible, applicable, and maneuverable. The calculating time and value of projection function were 34% and 143% of those with the traditional projection pursuit technique, respectively. This model could be applied extensively to classify and estimate all kinds of non-linear and multidimensional data in ecology, biology, and regional sustainable development.

  1. Comparison of different uncertainty techniques in urban stormwater quantity and quality modelling

    DEFF Research Database (Denmark)

    Dotto, C. B.; Mannina, G.; Kleidorfer, M.

    2012-01-01

    it is rarely practiced. The International Working Group on Data and Models, which works under the IWA/IAHR Joint Committee on Urban Drainage, has been working on the development of a framework for defining and assessing uncertainties in the field of urban drainage modelling. A part of that work...... techniques, common criteria have been set for the likelihood formulation, defining the number of simulations, and the measure of uncertainty bounds. Moreover, all the uncertainty techniques were implemented for the same case study, in which the same stormwater quantity and quality model was used alongside...... the specific advantages and disadvantages of each method. In relation to computational efficiency (i.e. number of iterations required to generate the probability distribution of parameters), it was found that SCEM-UA and AMALGAM produce results quicker than GLUE in terms of required number of simulations...

  2. Optimization models and techniques for implementation and pricing of electricity markets

    International Nuclear Information System (INIS)

    Madrigal Martinez, M.

    2001-01-01

    The operation and planning of vertically integrated electric power systems can be optimized using models that simulate solutions to problems. As the electric power industry is going through a period of restructuring, there is a need for new optimization tools. This thesis describes the importance of optimization tools and presents techniques for implementing them. It also presents methods for pricing primary electricity markets. Three modeling groups are studied. The first considers a simplified continuous and discrete model for power pool auctions. The second considers the unit commitment problem, and the third makes use of a new type of linear network-constrained clearing system model for daily markets for power and spinning reserve. The newly proposed model considers bids for supply and demand and bilateral contracts. It is a direct current model for the transmission network

  3. A Modified Particle Swarm Optimization Technique for Finding Optimal Designs for Mixture Models

    Science.gov (United States)

    Wong, Weng Kee; Chen, Ray-Bing; Huang, Chien-Chih; Wang, Weichung

    2015-01-01

    Particle Swarm Optimization (PSO) is a meta-heuristic algorithm that has been shown to be successful in solving a wide variety of real and complicated optimization problems in engineering and computer science. This paper introduces a projection based PSO technique, named ProjPSO, to efficiently find different types of optimal designs, or nearly optimal designs, for mixture models with and without constraints on the components, and also for related models, like the log contrast models. We also compare the modified PSO performance with Fedorov's algorithm, a popular algorithm used to generate optimal designs, Cocktail algorithm, and the recent algorithm proposed by [1]. PMID:26091237

  4. Floating Node Method and Virtual Crack Closure Technique for Modeling Matrix Cracking-Delamination Migration

    Science.gov (United States)

    DeCarvalho, Nelson V.; Chen, B. Y.; Pinho, Silvestre T.; Baiz, P. M.; Ratcliffe, James G.; Tay, T. E.

    2013-01-01

    A novel approach is proposed for high-fidelity modeling of progressive damage and failure in composite materials that combines the Floating Node Method (FNM) and the Virtual Crack Closure Technique (VCCT) to represent multiple interacting failure mechanisms in a mesh-independent fashion. In this study, the approach is applied to the modeling of delamination migration in cross-ply tape laminates. Delamination, matrix cracking, and migration are all modeled using fracture mechanics based failure and migration criteria. The methodology proposed shows very good qualitative and quantitative agreement with experiments.

  5. Accuracy Enhanced Stability and Structure Preserving Model Reduction Technique for Dynamical Systems with Second Order Structure

    DEFF Research Database (Denmark)

    Tahavori, Maryamsadat; Shaker, Hamid Reza

    gramians within the time interval to build the appropriate Petrov-Galerkin projection for dynamical systems within the time interval of interest. The bound on approximation error is also derived. The numerical results are compared with the counterparts from other techniques. The results confirm......A method for model reduction of dynamical systems with the second order structure is proposed in this paper. The proposed technique preserves the second order structure of the system, and also preserves the stability of the original systems. The method uses the controllability and observability...

  6. Comparing modelling techniques when designing VPH gratings for BigBOSS

    Science.gov (United States)

    Poppett, Claire; Edelstein, Jerry; Lampton, Michael; Jelinsky, Patrick; Arns, James

    2012-09-01

    BigBOSS is a Stage IV Dark Energy instrument based on the Baryon Acoustic Oscillations (BAO) and Red Shift Distortions (RSD) techniques using spectroscopic data of 20 million ELG and LRG galaxies at 0.5VPH) gratings have been identified as a key technology which will enable the efficiency requirement to be met, however it is important to be able to accurately predict their performance. In this paper we quantitatively compare different modelling techniques in order to assess the parameter space over which they are more capable of accurately predicting measured performance. Finally we present baseline parameters for grating designs that are most suitable for the BigBOSS instrument.

  7. River suspended sediment modelling using the CART model: A comparative study of machine learning techniques.

    Science.gov (United States)

    Choubin, Bahram; Darabi, Hamid; Rahmati, Omid; Sajedi-Hosseini, Farzaneh; Kløve, Bjørn

    2018-02-15

    Suspended sediment load (SSL) modelling is an important issue in integrated environmental and water resources management, as sediment affects water quality and aquatic habitats. Although classification and regression tree (CART) algorithms have been applied successfully to ecological and geomorphological modelling, their applicability to SSL estimation in rivers has not yet been investigated. In this study, we evaluated use of a CART model to estimate SSL based on hydro-meteorological data. We also compared the accuracy of the CART model with that of the four most commonly used models for time series modelling of SSL, i.e. adaptive neuro-fuzzy inference system (ANFIS), multi-layer perceptron (MLP) neural network and two kernels of support vector machines (RBF-SVM and P-SVM). The models were calibrated using river discharge, stage, rainfall and monthly SSL data for the Kareh-Sang River gauging station in the Haraz watershed in northern Iran, where sediment transport is a considerable issue. In addition, different combinations of input data with various time lags were explored to estimate SSL. The best input combination was identified through trial and error, percent bias (PBIAS), Taylor diagrams and violin plots for each model. For evaluating the capability of the models, different statistics such as Nash-Sutcliffe efficiency (NSE), Kling-Gupta efficiency (KGE) and percent bias (PBIAS) were used. The results showed that the CART model performed best in predicting SSL (NSE=0.77, KGE=0.8, PBIAS<±15), followed by RBF-SVM (NSE=0.68, KGE=0.72, PBIAS<±15). Thus the CART model can be a helpful tool in basins where hydro-meteorological data are readily available. Copyright © 2017 Elsevier B.V. All rights reserved.

  8. Forecasting Macroeconomic Variables using Neural Network Models and Three Automated Model Selection Techniques

    DEFF Research Database (Denmark)

    Kock, Anders Bredahl; Teräsvirta, Timo

    In this paper we consider the forecasting performance of a well-defined class of flexible models, the so-called single hidden-layer feedforward neural network models. A major aim of our study is to find out whether they, due to their flexibility, are as useful tools in economic forecasting as some...... previous studies have indicated. When forecasting with neural network models one faces several problems, all of which influence the accuracy of the forecasts. First, neural networks are often hard to estimate due to their highly nonlinear structure. In fact, their parameters are not even globally...... on the linearisation idea: the Marginal Bridge Estimator and Autometrics. Second, one must decide whether forecasting should be carried out recursively or directly. Comparisons of these two methodss exist for linear models and here these comparisons are extended to neural networks. Finally, a nonlinear model...

  9. A review of cutting mechanics and modeling techniques for biological materials.

    Science.gov (United States)

    Takabi, Behrouz; Tai, Bruce L

    2017-07-01

    This paper presents a comprehensive survey on the modeling of tissue cutting, including both soft tissue and bone cutting processes. In order to achieve higher accuracy in tissue cutting, as a critical process in surgical operations, the meticulous modeling of such processes is important in particular for surgical tool development and analysis. This review paper is focused on the mechanical concepts and modeling techniques utilized to simulate tissue cutting such as cutting forces and chip morphology. These models are presented in two major categories, namely soft tissue cutting and bone cutting. Fracture toughness is commonly used to describe tissue cutting while Johnson-Cook material model is often adopted for bone cutting in conjunction with finite element analysis (FEA). In each section, the most recent mathematical and computational models are summarized. The differences and similarities among these models, challenges, novel techniques, and recommendations for future work are discussed along with each section. This review is aimed to provide a broad and in-depth vision of the methods suitable for tissue and bone cutting simulations. Copyright © 2017 IPEM. Published by Elsevier Ltd. All rights reserved.

  10. Applications of soft computing in time series forecasting simulation and modeling techniques

    CERN Document Server

    Singh, Pritpal

    2016-01-01

    This book reports on an in-depth study of fuzzy time series (FTS) modeling. It reviews and summarizes previous research work in FTS modeling and also provides a brief introduction to other soft-computing techniques, such as artificial neural networks (ANNs), rough sets (RS) and evolutionary computing (EC), focusing on how these techniques can be integrated into different phases of the FTS modeling approach. In particular, the book describes novel methods resulting from the hybridization of FTS modeling approaches with neural networks and particle swarm optimization. It also demonstrates how a new ANN-based model can be successfully applied in the context of predicting Indian summer monsoon rainfall. Thanks to its easy-to-read style and the clear explanations of the models, the book can be used as a concise yet comprehensive reference guide to fuzzy time series modeling, and will be valuable not only for graduate students, but also for researchers and professionals working for academic, business and governmen...

  11. Composite use of numerical groundwater flow modeling and geoinformatics techniques for monitoring Indus Basin aquifer, Pakistan.

    Science.gov (United States)

    Ahmad, Zulfiqar; Ashraf, Arshad; Fryar, Alan; Akhter, Gulraiz

    2011-02-01

    The integration of the Geographic Information System (GIS) with groundwater modeling and satellite remote sensing capabilities has provided an efficient way of analyzing and monitoring groundwater behavior and its associated land conditions. A 3-dimensional finite element model (Feflow) has been used for regional groundwater flow modeling of Upper Chaj Doab in Indus Basin, Pakistan. The approach of using GIS techniques that partially fulfill the data requirements and define the parameters of existing hydrologic models was adopted. The numerical groundwater flow model is developed to configure the groundwater equipotential surface, hydraulic head gradient, and estimation of the groundwater budget of the aquifer. GIS is used for spatial database development, integration with a remote sensing, and numerical groundwater flow modeling capabilities. The thematic layers of soils, land use, hydrology, infrastructure, and climate were developed using GIS. The Arcview GIS software is used as additive tool to develop supportive data for numerical groundwater flow modeling and integration and presentation of image processing and modeling results. The groundwater flow model was calibrated to simulate future changes in piezometric heads from the period 2006 to 2020. Different scenarios were developed to study the impact of extreme climatic conditions (drought/flood) and variable groundwater abstraction on the regional groundwater system. The model results indicated a significant response in watertable due to external influential factors. The developed model provides an effective tool for evaluating better management options for monitoring future groundwater development in the study area.

  12. Testing and Measurement Potentials of Microcomputers for Cognitive Style Research and Individualized Instruction.

    Science.gov (United States)

    Burroway, Robert L.

    Microcomputer versions of three commonly used cognitive style measurement instruments are described and demonstrated: the Leveling-Sharpening House Test (LSHT) (Santostefano, 1964); Lowenfeld's Successive Impressions Test I (SIT-I) (1945); and the Group Embedded Figures Test (GEFT) (Witkin, Oltman, Ruskin, and Karp, 1971). At present, many…

  13. Huntington's Disease Research Roster Support with a Microcomputer Database Management System

    Science.gov (United States)

    Gersting, J. M.; Conneally, P. M.; Beidelman, K.

    1983-01-01

    This paper chronicles the MEGADATS (Medical Genetics Acquisition and DAta Transfer System) database development effort in collecting, storing, retrieving, and plotting human family pedigrees. The newest system, MEGADATS-3M, is detailed. Emphasis is on the microcomputer version of MEGADATS-3M and its use to support the Huntington's Disease research roster project. Examples of data input and pedigree plotting are included.

  14. JAX: a micro-computer based X-ray diffractometer controller

    International Nuclear Information System (INIS)

    Naval, P.C. Jr.

    1987-05-01

    This paper describes a micro-computer based X-ray diffractometer controller and explores its possibilities in simplifying acquisition and analysis of X-ray powder diffraction data. The interrupt-driven controller can operate in both present time and present count data acquisition modes and allows a data analysis program to execute concurrently with data collection. (Auth.). 16 figs.; 2 tabs

  15. Choosing the Machine; How Drexel University Selected a Microcomputer for Its Students.

    Science.gov (United States)

    Bulletin of the American Society for Information Science, 1983

    1983-01-01

    Describes process by which faculty member committee at Drexel University selected a microcomputer to be purchased by all freshmen beginning winter term, 1983-84. Considerations in decision (critical audience, computer power, degree of innovation, user interface, graphics, sound, growth capacity, cost to student and university, and source of…

  16. Library Micro-Computing, Vol. 2. Reprints from the Best of "ONLINE" [and]"DATABASE."

    Science.gov (United States)

    Online, Inc., Weston, CT.

    Reprints of 19 articles pertaining to library microcomputing appear in this collection, the second of two volumes on this topic in a series of volumes of reprints from "ONLINE" and "DATABASE" magazines. Edited for information professionals who use electronically distributed databases, these articles address such topics as: (1)…

  17. Automation by microcomputer of a geodetic distance measuring instrument: The distinvar

    International Nuclear Information System (INIS)

    Bain, G.; Bore, C.; Coosemans, W.; Dupont, J.; Fabre, J.P.; Gavaggio, R.; Peron, F.

    1985-01-01

    The distinvar is an instrument for the precision measurement of distances up to 50 m. This report describes the latest developments at CERN which have improved its resolution to 2 μm and increased its speed of use. Measurements are automated by incorporating a microcomputer programmed in BASIC. (orig.)

  18. A multi-channel data acquisition system with high resolution based on microcomputer

    International Nuclear Information System (INIS)

    An Qi; Wang Yanfang; Xing Tao

    1995-01-01

    The paper introduces the principle of a multi-channel data acquisition system with high resolution based on the microcomputer.The system consists of five parts.They are analog-to-digital converter, data buffer area, trigger logic circuit, control circuit, and digital-to-analog converter

  19. Development of a microcomputer data base of manufacturing, installation, and operating experience for the NSSS designer

    International Nuclear Information System (INIS)

    Borchers, W.A.; Markowski, E.S.

    1986-01-01

    Future nuclear steam supply systems (NSSSs) will be designed in an environment of powerful micro hardware and software and these systems will be linked by local area networks (LAN). With such systems, individual NSSS designers and design groups will establish and maintain local data bases to replace existing manual files and data sources. One such effort of this type in Combustion Engineering's (C-E's) NSSS engineering organization is the establishment of a data base of historical manufacturing, installation, and operating experience to provide designers with information to improve on current designs and practices. In contrast to large mainframe or minicomputer data bases, which compile industry-wide data, the data base described here is implemented on a microcomputer, is design specific, and contains a level of detail that is of interest to system and component designers. DBASE III, a popular microcomputer data base management software package, is used. In addition to the immediate benefits provided by the data base, the development itself provided a vehicle for identifying procedural and control aspects that need to be addressed in the environment of local microcomputer data bases. This paper describes the data base and provides some observations on the development, use, and control of local microcomputer data bases in a design organization

  20. The Ant and the Grasshopper: A Program for Adopting Microcomputer Graphics in the Arts.

    Science.gov (United States)

    Youngblood, Michael S.

    1988-01-01

    Discusses the major reasons why microcomputers have not yet become commonplace in the arts and describes how the College of Communications and Fine Arts at Southern Illinois University (Carbondale) researched, planned, and implemented a college-wide computer technology program. (GEA)

  1. Methods: a comparative analysis of radiography, microcomputed tomography, and histology for bone tissue engineering.

    NARCIS (Netherlands)

    Hedberg, E.L.; Kroese-Deutman, H.C.; Shih, C.K.; Lemoine, J.J.; Liebschner, M.A.; Miller, M.J.; Yasko, A.W.; Crowther, R.S.; Carney, D.H.; Mikos, A.G.; Jansen, J.A.

    2005-01-01

    This study focused on the assessment of radiography, microcomputed tomography, and histology for the evaluation of bone formation in a 15.0-mm defect in the rabbit radius after the implantation of a tissue-engineered construct. Radiography was found to be useful as a noninvasive method for obtaining

  2. Microdot: The Microcomputer as a Learning Aid in Sixth-Form Mathematics.

    Science.gov (United States)

    Higgo, J. R.

    1984-01-01

    Demonstrates how microcomputers can be used in teaching differential calculus, iteration, integral calculus, graphs, and statistics. Several ideas for putting this information into practice are outlined. Sample computer programs are included for the discussions on differential calculus, integral calculus, and iteration. (JN)

  3. Organizing the Program. The Logistics Involved in Introducing Microcomputing at Drexel University.

    Science.gov (United States)

    Drexel Univ., Philadelphia, PA. Microcomputing Program.

    The introduction of Drexel University's microcomputing program, which provides students access to this technology for coursework, is described, with attention to the groups who played a major role in preparing the university. Faculty helped set the initial direction of the program and have been central to its operation. Other faculty roles…

  4. Out of the Shoebox: Publishing a Regional Periodical Index Using a Microcomputer.

    Science.gov (United States)

    Schmidt, Jean M.; Wilson, John S.

    1988-01-01

    Describes production of the Bayou State Periodical Index, an index of Louisiana periodicals designed to provide access to information on local topics, and offers guidelines for producing such an index in a timely fashion with limited funds, using a microcomputer. Planning, procedures, financing, and publicity and marketing are covered. (seven…

  5. Classroom Use of Microcomputer Graphics and Probabilistic Sensitivity Analysis to Demonstrate Petroleum Engineering Concepts.

    Science.gov (United States)

    Whitman, David L.; Terry, R. E.

    1984-01-01

    A computer program which allows the solution of a Monte Carlo simulation (probabilistic sensitivity analysis) has been developed for the Vic-20 microcomputer. Theory of Monte Carlo simulation, program capabilities and operation, and sample calculations are discussed. Student comments on the program are included. (JN)

  6. Direct microcomputer controlled determination of zinc in human serum by flow injection atomic absorption spectrometry

    DEFF Research Database (Denmark)

    Simonsen, Kirsten Wiese; Nielsen, Bent; Jensen, Arne

    1986-01-01

    A procedure is described for the direct determination of zinc in human serum by fully automated, microcomputer controlled flow injection atomic absorption spectrometry (Fl-AAS). The Fl system is pumpless, using the negative pressure created by the nebuliser. It only consists of a three-way valve...

  7. Interfacing the Microcomputer and the Mainframe to Support Visit-Revisit Comparisons of Medical Records

    Science.gov (United States)

    Saltsman, Tom; Truax, Terry

    1982-01-01

    The microcomputer is finding an increasing role on the hospital patient floor to aid the collection, management and display of the patient's medical data. The data collected by the nursing staff on the microcomputer is stored locally on floppy disks and forms the initial basis for a computerized nursing assessment. The data, however, is permanently maintained at a central site using a mainframe database management system. This paper reports the development of methods which allow the nursing staff (using the same microcomputer used for local data collection) to interface with the mainframe database and send or retrieve components of the patient's record. The microcomputer interface allows the nursing staff to retrieve and view the patient's previous visit records using the same screen format used to input and display the current assessment. As data from the current visit is entered, the current data is compared with the relevant data from the previous visits. Significant deviations are detected and the nurse is alerted. The data can then be displayed for comparisons.

  8. The Effectiveness of Microcomputer Exercises in Teaching Marketing Planning and Control.

    Science.gov (United States)

    Calantone, Roger J.; di Benedetto, C. Anthony

    1989-01-01

    Student teams in a senior capstone course in marketing completed microcomputer simulation in which they developed six-month marketing plan for a new product. A procedure similar to Kelly's Role Repertory Test was applied to define the role that various information sources played in their decision-making process, a means of determining the amount…

  9. Numerical Time-Domain Modeling of Lamb Wave Propagation Using Elastodynamic Finite Integration Technique

    Directory of Open Access Journals (Sweden)

    Hussein Rappel

    2014-01-01

    integration technique (EFIT as well as its validation with analytical results. Lamb wave method is a long range inspection technique which is considered to have unique future in the field of structural health monitoring. One of the main problems facing the lamb wave method is how to choose the most appropriate frequency to generate the waves for adequate transmission capable of properly propagating in the material, interfering with defects/damages, and being received in good conditions. Modern simulation tools based on numerical methods such as finite integration technique (FIT, finite element method (FEM, and boundary element method (BEM may be used for modeling. In this paper, two sets of simulation are performed. In the first set, group velocities of lamb wave in a steel plate are obtained numerically. Results are then compared with analytical results to validate the simulation. In the second set, EFIT is employed to study fundamental symmetric mode interaction with a surface braking defect.

  10. Efficiency assessment of runoff harvesting techniques using a 3D coupled surface-subsurface hydrological model

    International Nuclear Information System (INIS)

    Verbist, K.; Cronelis, W. M.; McLaren, R.; Gabriels, D.; Soto, G.

    2009-01-01

    In arid and semi-arid zones runoff harvesting techniques are often applied to increase the water retention and infiltration on steep slopes. Additionally, they act as an erosion control measure to reduce land degradation hazards. Both in literature and in the field, a large variety of runoff collecting systems are found, as well as large variations in design and dimensions. Therefore, detailed measurements were performed on a semi-arid slope in central Chile to allow identification of the effect of a simple water harvesting technique on soil water availability. For this purpose, twenty two TDR-probes were installed and were monitored continuously during and after a simulated rainfall event. These data were used to calibrate the 3D distributed flow model HydroGeoSphere, to assess the runoff components and soil water retention as influenced by the water harvesting technique, both under simulated and natural rainfall conditions. (Author) 6 refs.

  11. Proposal of a congestion control technique in LAN networks using an econometric model ARIMA

    Directory of Open Access Journals (Sweden)

    Joaquín F Sánchez

    2017-01-01

    Full Text Available Hasty software development can produce immediate implementations with source code unnecessarily complex and hardly readable. These small kinds of software decay generate a technical debt that could be big enough to seriously affect future maintenance activities. This work presents an analysis technique for identifying architectural technical debt related to non-uniformity of naming patterns; the technique is based on term frequency over package hierarchies. The proposal has been evaluated on projects of two popular organizations, Apache and Eclipse. The results have shown that most of the projects have frequent occurrences of the proposed naming patterns, and using a graph model and aggregated data could enable the elaboration of simple queries for debt identification. The technique has features that favor its applicability on emergent architectures and agile software development.

  12. Analytical Modelling of the Effects of Different Gas Turbine Cooling Techniques on Engine Performance =

    Science.gov (United States)

    Uysal, Selcuk Can

    In this research, MATLAB SimulinkRTM was used to develop a cooled engine model for industrial gas turbines and aero-engines. The model consists of uncooled on-design, mean-line turbomachinery design and a cooled off-design analysis in order to evaluate the engine performance parameters by using operating conditions, polytropic efficiencies, material information and cooling system details. The cooling analysis algorithm involves a 2nd law analysis to calculate losses from the cooling technique applied. The model is used in a sensitivity analysis that evaluates the impacts of variations in metal Biot number, thermal barrier coating Biot number, film cooling effectiveness, internal cooling effectiveness and maximum allowable blade temperature on main engine performance parameters of aero and industrial gas turbine engines. The model is subsequently used to analyze the relative performance impact of employing Anti-Vortex Film Cooling holes (AVH) by means of data obtained for these holes by Detached Eddy Simulation-CFD Techniques that are valid for engine-like turbulence intensity conditions. Cooled blade configurations with AVH and other different external cooling techniques were used in a performance comparison study. (Abstract shortened by ProQuest.).

  13. Numerical modelling of radon-222 entry into houses: An outline of techniques and results

    DEFF Research Database (Denmark)

    Andersen, C.E.

    2001-01-01

    Numerical modelling is a powerful tool for studies of soil gas and radon-222 entry into houses. It is the purpose of this paper to review some main techniques and results. In the past, modelling has focused on Darcy flow of soil gas (driven by indoor–outdoor pressure differences) and combined......, fractures, moisture, non-uniform soil temperature, non-Darcy flow of gas, and flow caused by changes in the atmospheric pressure. Numerical models can be used to estimate the importance of specific factors for radon entry. Models are also helpful when results obtained in special laboratory or test structure...... experiments need to be extrapolated to more general situations (e.g. to real houses or even to other soil–gas pollutants). Finally, models provide a cost-effective test bench for improved designs of radon prevention systems. The paper includes a summary of transport equations and boundary conditions...

  14. A stochastic delay model for pricing debt and equity: Numerical techniques and applications

    Science.gov (United States)

    Tambue, Antoine; Kemajou Brown, Elisabeth; Mohammed, Salah

    2015-01-01

    Delayed nonlinear models for pricing corporate liabilities and European options were recently developed. Using self-financed strategy and duplication we were able to derive a Random Partial Differential Equation (RPDE) whose solutions describe the evolution of debt and equity values of a corporate in the last delay period interval in the accompanied paper (Kemajou et al., 2012) [14]. In this paper, we provide robust numerical techniques to solve the delayed nonlinear model for the corporate value, along with the corresponding RPDEs modeling the debt and equity values of the corporate. Using financial data from some firms, we forecast and compare numerical solutions from both the nonlinear delayed model and classical Merton model with the real corporate data. From this comparison, it comes up that in corporate finance the past dependence of the firm value process may be an important feature and therefore should not be ignored.

  15. New Diagnostic, Launch and Model Control Techniques in the NASA Ames HFFAF Ballistic Range

    Science.gov (United States)

    Bogdanoff, David W.

    2012-01-01

    This report presents new diagnostic, launch and model control techniques used in the NASA Ames HFFAF ballistic range. High speed movies were used to view the sabot separation process and the passage of the model through the model splap paper. Cavities in the rear of the sabot, to catch the muzzle blast of the gun, were used to control sabot finger separation angles and distances. Inserts were installed in the powder chamber to greatly reduce the ullage volume (empty space) in the chamber. This resulted in much more complete and repeatable combustion of the powder and hence, in much more repeatable muzzle velocities. Sheets of paper or cardstock, impacting one half of the model, were used to control the amplitudes of the model pitch oscillations.

  16. Forecasting performance of three automated modelling techniques during the economic crisis 2007-2009

    DEFF Research Database (Denmark)

    Kock, Anders Bredahl; Teräsvirta, Timo

    In this work we consider forecasting macroeconomic variables during an economic crisis. The focus is on a speci…c class of models, the so-called single hidden-layer feedforward autoregressive neural network models. What makes these models interesting in the present context is that they form a class...... during the economic crisis 2007–2009. Forecast accuracy is measured by the root mean square forecast error. Hypothesis testing is also used to compare the performance of the different techniques with each other....... of universal approximators and may be expected to work well during exceptional periods such as major economic crises. These models are often difficult to estimate, and we follow the idea of White (2006) to transform the speci…cation and nonlinear estimation problem into a linear model selection and estimation...

  17. Limitations in paleomagnetic data and modelling techniques and their impact on Holocene geomagnetic field models

    DEFF Research Database (Denmark)

    Panovska, S.; Korte, M.; Finlay, Chris

    2015-01-01

    in individual archeomagnetic data so that these data or models derived from them can be used for reliable initial relative paleointensity calibration and declination orientation in sediments. More work will be needed to assess whether co-estimation or an iterative approach to inversion is more efficient overall......Characterization of geomagnetic field behaviour on timescales of centuries to millennia is necessary to understand the mechanisms that sustain the geodynamo and drive its evolution. As Holocene paleomagnetic and archeomagnetic data have become more abundant, strategies for regularized inversion...... of modern field data have been adapted to produce numerous timevarying global field models. We evaluate the effectiveness of several approaches to inversion and data handling, by assessing both global and regional properties of the resulting models. Global Holocene field models cannot resolve Southern...

  18. Development of pathological anthropomorphic models using 3D modelling techniques for numerical dosimetry

    International Nuclear Information System (INIS)

    Costa, Kleber Souza Silva; Barbosa, Antonio Konrado de Santana; Vieira, Jose Wilson; Lima, Fernando Roberto de Andrade

    2011-01-01

    Computational exposure models can be used to estimate human body absorbed dose in a series of situations such as X-Ray exams for diagnosis, accidents and medical treatments. These models are fundamentally composed of an anthropomorphic simulator (phantom), an algorithm that simulates a radioactive source and a Monte Carlo Code. The accuracy of data obtained in the simulation is strongly connected to the adequacy of such simulation to the real situation. The phantoms are one of the key factors for the researcher manipulation. They are generally developed in supine position and its anatomy is patronized by compiled data from international institutions such as ICRP or ICRU. Several pathologies modify the structure of organs and body tissues. In order to measure how significant these alterations are, an anthropomorphic model was developed for this study: patient mastectomies. This model was developed using voxel phantom FASH and then coupled with EGSnrc Monte Carlo code

  19. Evaluation of 3-dimensional superimposition techniques on various skeletal structures of the head using surface models.

    Directory of Open Access Journals (Sweden)

    Nikolaos Gkantidis

    Full Text Available To test the applicability, accuracy, precision, and reproducibility of various 3D superimposition techniques for radiographic data, transformed to triangulated surface data.Five superimposition techniques (3P: three-point registration; AC: anterior cranial base; AC + F: anterior cranial base + foramen magnum; BZ: both zygomatic arches; 1Z: one zygomatic arch were tested using eight pairs of pre-existing CT data (pre- and post-treatment. These were obtained from non-growing orthodontic patients treated with rapid maxillary expansion. All datasets were superimposed by three operators independently, who repeated the whole procedure one month later. Accuracy was assessed by the distance (D between superimposed datasets on three form-stable anatomical areas, located on the anterior cranial base and the foramen magnum. Precision and reproducibility were assessed using the distances between models at four specific landmarks. Non parametric multivariate models and Bland-Altman difference plots were used for analyses.There was no difference among operators or between time points on the accuracy of each superimposition technique (p>0.05. The AC + F technique was the most accurate (D0.05, the detected structural changes differed significantly between different techniques (p<0.05. Bland-Altman difference plots showed that BZ superimposition was comparable to AC, though it presented slightly higher random error.Superimposition of 3D datasets using surface models created from voxel data can provide accurate, precise, and reproducible results, offering also high efficiency and increased post-processing capabilities. In the present study population, the BZ superimposition was comparable to AC, with the added advantage of being applicable to scans with a smaller field of view.

  20. Evaluation of 3-dimensional superimposition techniques on various skeletal structures of the head using surface models.

    Science.gov (United States)

    Gkantidis, Nikolaos; Schauseil, Michael; Pazera, Pawel; Zorkun, Berna; Katsaros, Christos; Ludwig, Björn

    2015-01-01

    To test the applicability, accuracy, precision, and reproducibility of various 3D superimposition techniques for radiographic data, transformed to triangulated surface data. Five superimposition techniques (3P: three-point registration; AC: anterior cranial base; AC + F: anterior cranial base + foramen magnum; BZ: both zygomatic arches; 1Z: one zygomatic arch) were tested using eight pairs of pre-existing CT data (pre- and post-treatment). These were obtained from non-growing orthodontic patients treated with rapid maxillary expansion. All datasets were superimposed by three operators independently, who repeated the whole procedure one month later. Accuracy was assessed by the distance (D) between superimposed datasets on three form-stable anatomical areas, located on the anterior cranial base and the foramen magnum. Precision and reproducibility were assessed using the distances between models at four specific landmarks. Non parametric multivariate models and Bland-Altman difference plots were used for analyses. There was no difference among operators or between time points on the accuracy of each superimposition technique (p>0.05). The AC + F technique was the most accurate (D0.05), the detected structural changes differed significantly between different techniques (p<0.05). Bland-Altman difference plots showed that BZ superimposition was comparable to AC, though it presented slightly higher random error. Superimposition of 3D datasets using surface models created from voxel data can provide accurate, precise, and reproducible results, offering also high efficiency and increased post-processing capabilities. In the present study population, the BZ superimposition was comparable to AC, with the added advantage of being applicable to scans with a smaller field of view.