Sample records for ranking computer program

  1. University rankings in computer science

    DEFF Research Database (Denmark)

    Ehret, Philip; Zuccala, Alesia Ann; Gipp, Bela


    This is a research-in-progress paper concerning two types of institutional rankings, the Leiden and QS World ranking, and their relationship to a list of universities’ ‘geo-based’ impact scores, and Computing Research and Education Conference (CORE) participation scores in the field of computer...... science. A ‘geo-based’ impact measure examines the geographical distribution of incoming citations to a particular university’s journal articles for a specific period of time. It takes into account both the number of citations and the geographical variability in these citations. The CORE participation...... score is calculated on the basis of the number of weighted proceedings papers that a university has contributed to either an A*, A, B, or C conference as ranked by the Computing Research and Education Association of Australasia. In addition to calculating the correlations between the distinct university...

  2. Let Us Rank Journalism Programs (United States)

    Weber, Joseph


    Unlike law, business, and medical schools, as well as universities in general, journalism schools and journalism programs have rarely been ranked. Publishers such as "U.S. News & World Report," "Forbes," "Bloomberg Businessweek," and "Washington Monthly" do not pay them much mind. What is the best…

  3. Ranking Forestry Investments With Parametric Linear Programming (United States)

    Paul A. Murphy


    Parametric linear programming is introduced as a technique for ranking forestry investments under multiple constraints; it combines the advantages of simple tanking and linear programming as capital budgeting tools.

  4. Algebraic and computational aspects of real tensor ranks

    CERN Document Server

    Sakata, Toshio; Miyazaki, Mitsuhiro


    This book provides comprehensive summaries of theoretical (algebraic) and computational aspects of tensor ranks, maximal ranks, and typical ranks, over the real number field. Although tensor ranks have been often argued in the complex number field, it should be emphasized that this book treats real tensor ranks, which have direct applications in statistics. The book provides several interesting ideas, including determinant polynomials, determinantal ideals, absolutely nonsingular tensors, absolutely full column rank tensors, and their connection to bilinear maps and Hurwitz-Radon numbers. In addition to reviews of methods to determine real tensor ranks in details, global theories such as the Jacobian method are also reviewed in details. The book includes as well an accessible and comprehensive introduction of mathematical backgrounds, with basics of positive polynomials and calculations by using the Groebner basis. Furthermore, this book provides insights into numerical methods of finding tensor ranks through...

  5. Monte Carlo methods of PageRank computation

    NARCIS (Netherlands)

    Litvak, Nelli


    We describe and analyze an on-line Monte Carlo method of PageRank computation. The PageRank is being estimated basing on results of a large number of short independent simulation runs initiated from each page that contains outgoing hyperlinks. The method does not require any storage of the hyperlink

  6. Effect of Doximity Residency Rankings on Residency Applicants’ Program Choices

    Directory of Open Access Journals (Sweden)

    Aimee M. Rolston


    Full Text Available Introduction: Choosing a residency program is a stressful and important decision. Doximity released residency program rankings by specialty in September 2014. This study sought to investigate the impact of those rankings on residency application choices made by fourth year medical students. Methods: A 12-item survey was administered in October 2014 to fourth year medical students at three schools. Students indicated their specialty, awareness of and perceived accuracy of the rankings, and the rankings’ impact on the programs to which they chose to apply. Descriptive statistics were reported for all students and those applying to Emergency Medicine (EM. Results: A total of 461 (75.8% students responded, with 425 applying in one of the 20 Doximity ranked specialties. Of the 425, 247 (58% were aware of the rankings and 177 looked at them. On a 1-100 scale (100=very accurate, students reported a mean ranking accuracy rating of 56.7 (SD 20.3. Forty-five percent of students who looked at the rankings modified the number of programs to which they applied. The majority added programs. Of the 47 students applying to EM, 18 looked at the rankings and 33% changed their application list with most adding programs. Conclusion: The Doximity rankings had real effects on students applying to residencies as almost half of students who looked at the rankings modified their program list. Additionally, students found the rankings to be moderately accurate. Graduating students might benefit from emphasis on more objective characterization of programs to assess in light of their own interests and personal/career goals

  7. An Adaptive Reordered Method for Computing PageRank

    Directory of Open Access Journals (Sweden)

    Yi-Ming Bu


    Full Text Available We propose an adaptive reordered method to deal with the PageRank problem. It has been shown that one can reorder the hyperlink matrix of PageRank problem to calculate a reduced system and get the full PageRank vector through forward substitutions. This method can provide a speedup for calculating the PageRank vector. We observe that in the existing reordered method, the cost of the recursively reordering procedure could offset the computational reduction brought by minimizing the dimension of linear system. With this observation, we introduce an adaptive reordered method to accelerate the total calculation, in which we terminate the reordering procedure appropriately instead of reordering to the end. Numerical experiments show the effectiveness of this adaptive reordered method.

  8. Low-rank quadratic semidefinite programming

    KAUST Repository

    Yuan, Ganzhao


    Low rank matrix approximation is an attractive model in large scale machine learning problems, because it can not only reduce the memory and runtime complexity, but also provide a natural way to regularize parameters while preserving learning accuracy. In this paper, we address a special class of nonconvex quadratic matrix optimization problems, which require a low rank positive semidefinite solution. Despite their non-convexity, we exploit the structure of these problems to derive an efficient solver that converges to their local optima. Furthermore, we show that the proposed solution is capable of dramatically enhancing the efficiency and scalability of a variety of concrete problems, which are of significant interest to the machine learning community. These problems include the Top-k Eigenvalue problem, Distance learning and Kernel learning. Extensive experiments on UCI benchmarks have shown the effectiveness and efficiency of our proposed method. © 2012.

  9. Ranking the dermatology programs based on measurements of academic achievement. (United States)

    Wu, Jashin J; Ramirez, Claudia C; Alonso, Carol A; Berman, Brian; Tyring, Stephen K


    The only dermatology rankings in the past were based on National Institutes of Health (NIH) funding and journal citations. To determine the highest ranking academic dermatology programs based on 5 outcome measures and on an overall ranking scale. To the best of our knowledge, this is the first report to rank the dermatology programs on 4 of the following outcome measures of academic achievement and with an overall ranking. We collected extensive 2001 to 2004 data ranging from total publications to grant funding on 107 U.S. dermatology programs and their full-time faculty. Data from part-time and volunteer faculty were not used. Publications in 2001 to 2004; NIH funding in 2004; Dermatology Foundation grants in 2001 to 2004; faculty lectures in 2004 delivered at national conferences; number of full-time faculty members who were on the editorial boards of the top 3 U.S. dermatology journals and the top 4 subspecialty journals We used the 5 outcome measures to tabulate the highest ranking programs in each category. Using a weighted ranking system, we also tabulated the overall top 30 dermatology programs based on these 5 outcome measures. We were not able to determine the total amount of NIH funding in dollars of the dermatology divisions. The impact factors of the journal in which these publications appeared was not factored into our calculations. Since faculty members may collaborate on the same publication, some publications may have been double-counted. In descending order, the 5 highest ranked academic programs are the University of Pennsylvania; University of California, San Francisco; Yale-New Haven Medical Center; New York University; and University of Michigan. This ranking system may allow residents and faculty to improve the academic achievements at their respective programs.

  10. Methods for evaluating and ranking transportation energy conservation programs. Final report

    Energy Technology Data Exchange (ETDEWEB)


    Methods for comparative evaluations of the Office of Transportation programs designed to help achieve significant reductions in the consumption of petroleum by different forms of transportation while maintaining public, commercial, and industrial mobility are described. Assessments of the programs in terms of petroleum savings, incremental costs to consumers of the technologies and activities, probability of technical and market success, and external impacts due to environmental, economic, and social factors are inputs to the evaluation methodologies presented. The methods described for evaluating the programs on a comparative basis are three ranking functions and a policy matrix listing important attributes of the programs and the technologies and activities with which they are concerned and include the traditional net present value measure which computes the present worth of petroleum savings less the present worth of costs. This is modified by dividing by the present value of DOE funding to obtain a net present value per program dollar, which is the second ranking function. The third ranking function is broader in that it takes external impacts into account and is known as the comprehensive ranking function. Procedures are described for making computations of the ranking functions and the attributes that require computation. Computations are made for the electric vehicle, Stirling engine, gas turbine, and MPG mileage guide program. (MCW)

  11. Computer Programs. (United States)

    Anderson, Tiffoni

    This module provides information on development and use of a Material Safety Data Sheet (MSDS) software program that seeks to link literacy skills education, safety training, and human-centered design. Section 1 discusses the development of the software program that helps workers understand the MSDSs that accompany the chemicals with which they…

  12. Computer programming and computer systems

    CERN Document Server

    Hassitt, Anthony


    Computer Programming and Computer Systems imparts a "reading knowledge? of computer systems.This book describes the aspects of machine-language programming, monitor systems, computer hardware, and advanced programming that every thorough programmer should be acquainted with. This text discusses the automatic electronic digital computers, symbolic language, Reverse Polish Notation, and Fortran into assembly language. The routine for reading blocked tapes, dimension statements in subroutines, general-purpose input routine, and efficient use of memory are also elaborated.This publication is inten

  13. Combinatorial conditions for low rank solutions in semidefinite programming

    NARCIS (Netherlands)

    A. Varvitsiotis (Antonios)


    htmlabstractIn this thesis we investigate combinatorial conditions that guarantee the existence of low-rank optimal solutions to semidefinite programs. Results of this type are important for approximation algorithms and for the study of geometric representations of graphs. The structure of the

  14. Combinatorial conditions for low rank solutions in semidefinite programming

    NARCIS (Netherlands)

    Varvitsiotis, A.


    In this thesis we investigate combinatorial conditions that guarantee the existence of low-rank optimal solutions to semidefinite programs. Results of this type are important for approximation algorithms and for the study of geometric representations of graphs. The structure of the thesis is as

  15. Monte Carlo methods in PageRank computation: When one iteration is sufficient

    NARCIS (Netherlands)

    Avrachenkov, K.; Litvak, Nelli; Nemirovsky, D.; Osipova, N.


    PageRank is one of the principle criteria according to which Google ranks Web pages. PageRank can be interpreted as a frequency of visiting a Web page by a random surfer and thus it reflects the popularity of a Web page. Google computes the PageRank using the power iteration method which requires

  16. Monte Carlo methods in PageRank computation: When one iteration is sufficient

    NARCIS (Netherlands)

    Avrachenkov, K.; Litvak, Nelli; Nemirovsky, D.; Osipova, N.

    PageRank is one of the principle criteria according to which Google ranks Web pages. PageRank can be interpreted as a frequency of visiting a Web page by a random surfer, and thus it reflects the popularity of a Web page. Google computes the PageRank using the power iteration method, which requires

  17. Low-rank regularization for learning gene expression programs. (United States)

    Ye, Guibo; Tang, Mengfan; Cai, Jian-Feng; Nie, Qing; Xie, Xiaohui


    Learning gene expression programs directly from a set of observations is challenging due to the complexity of gene regulation, high noise of experimental measurements, and insufficient number of experimental measurements. Imposing additional constraints with strong and biologically motivated regularizations is critical in developing reliable and effective algorithms for inferring gene expression programs. Here we propose a new form of regulation that constrains the number of independent connectivity patterns between regulators and targets, motivated by the modular design of gene regulatory programs and the belief that the total number of independent regulatory modules should be small. We formulate a multi-target linear regression framework to incorporate this type of regulation, in which the number of independent connectivity patterns is expressed as the rank of the connectivity matrix between regulators and targets. We then generalize the linear framework to nonlinear cases, and prove that the generalized low-rank regularization model is still convex. Efficient algorithms are derived to solve both the linear and nonlinear low-rank regularized problems. Finally, we test the algorithms on three gene expression datasets, and show that the low-rank regularization improves the accuracy of gene expression prediction in these three datasets.

  18. ICASE Computer Science Program (United States)


    The Institute for Computer Applications in Science and Engineering computer science program is discussed in outline form. Information is given on such topics as problem decomposition, algorithm development, programming languages, and parallel architectures.

  19. Risk-Assessment Computer Program (United States)

    Dias, William C.; Mittman, David S.


    RISK D/C is prototype computer program assisting in attempts to do program risk modeling for Space Exploration Initiative (SEI) architectures proposed in Synthesis Group Report. Risk assessment performed with respect to risk events, probabilities, and severities of potential results. Enables ranking, with respect to effectiveness, of risk-mitigation strategies proposed for exploration program architecture. Allows for fact that risk assessment in early phases of planning subjective. Although specific to SEI in present form, also used as software framework for development of risk-assessment programs for other specific uses. Developed for Macintosh(TM) series computer. Requires HyperCard(TM) 2.0 or later, as well as 2 Mb of random-access memory and System 6.0.8 or later.

  20. Rank diversity of languages: generic behavior in computational linguistics. (United States)

    Cocho, Germinal; Flores, Jorge; Gershenson, Carlos; Pineda, Carlos; Sánchez, Sergio


    Statistical studies of languages have focused on the rank-frequency distribution of words. Instead, we introduce here a measure of how word ranks change in time and call this distribution rank diversity. We calculate this diversity for books published in six European languages since 1800, and find that it follows a universal lognormal distribution. Based on the mean and standard deviation associated with the lognormal distribution, we define three different word regimes of languages: "heads" consist of words which almost do not change their rank in time, "bodies" are words of general use, while "tails" are comprised by context-specific words and vary their rank considerably in time. The heads and bodies reflect the size of language cores identified by linguists for basic communication. We propose a Gaussian random walk model which reproduces the rank variation of words in time and thus the diversity. Rank diversity of words can be understood as the result of random variations in rank, where the size of the variation depends on the rank itself. We find that the core size is similar for all languages studied.

  1. Programming in biomolecular computation

    DEFF Research Database (Denmark)

    Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue


    Our goal is to provide a top-down approach to biomolecular computation. In spite of widespread discussion about connections between biology and computation, one question seems notable by its absence: Where are the programs? We identify a number of common features in programming that seem...... conspicuously absent from the literature on biomolecular computing; to partially redress this absence, we introduce a model of computation that is evidently programmable, by programs reminiscent of low-level computer machine code; and at the same time biologically plausible: its functioning is defined...... by a single and relatively small set of chemical-like reaction rules. Further properties: the model is stored-program: programs are the same as data, so programs are not only executable, but are also compilable and interpretable. It is universal: all computable functions can be computed (in natural ways...

  2. Programming in biomolecular computation

    DEFF Research Database (Denmark)

    Hartmann, Lars Røeboe; Jones, Neil; Simonsen, Jakob Grue


    Our goal is to provide a top-down approach to biomolecular computation. In spite of widespread discussion about connections between biology and computation, one question seems notable by its absence: Where are the programs? We introduce a model of computation that is evidently programmable......, by programs reminiscent of low-level computer machine code; and at the same time biologically plausible: its functioning is defined by a single and relatively small set of chemical-like reaction rules. Further properties: the model is stored-program: programs are the same as data, so programs are not only...... in a strong sense: a universal algorithm exists, that is able to execute any program, and is not asymptotically inefficient. A prototype model has been implemented (for now in silico on a conventional computer). This work opens new perspectives on just how computation may be specified at the biological level....

  3. A computer-aided diagnosis system of nuclear cataract via ranking. (United States)

    Huang, Wei; Li, Huiqi; Chan, Kap Luk; Lim, Joo Hwee; Liu, Jiang; Wong, Tien Yin


    A novel computer-aided diagnosis system of nuclear cataract via ranking is firstly proposed in this paper. The grade of nuclear cataract in a slit-lamp image is predicted based on its neighboring labeled images in a ranked images list, which is achieved using an optimal ranking function. A new ranking evaluation measure is proposed for learning the optimal ranking function via direct optimization. Our system has been tested by a large dataset composed of 1000 slit-lamp images from 1000 different cases. Both experimental results and comparison with several state-of-the-art methods indicate the superiority of our system.

  4. Flexible Animation Computer Program (United States)

    Stallcup, Scott S.


    FLEXAN (Flexible Animation), computer program animating structural dynamics on Evans and Sutherland PS300-series graphics workstation with VAX/VMS host computer. Typical application is animation of spacecraft undergoing structural stresses caused by thermal and vibrational effects. Displays distortions in shape of spacecraft. Program displays single natural mode of vibration, mode history, or any general deformation of flexible structure. Written in FORTRAN 77.

  5. Computer Security Assistance Program (United States)


    Information COMPUTER SECURITY ASSISTANCE PROGRAM OPR: HQ AFCA/SYS (CMSgt Hogan) Certified by: HQ USAF/SCXX (Lt Col Francis X. McGovern) Pages: 5...Distribution: F This instruction implements Air Force Policy Directive (AFPD) 33-2, Information Protection, establishes the Air Force Computer Security Assistance...Force single point of contact for reporting and handling computer security incidents and vulnerabilities including AFCERT advisories and Defense

  6. Ranking of CMIP5-based global climate models for India using compromise programming (United States)

    Srinivasa Raju, K.; Sonali, P.; Nagesh Kumar, D.


    Thirty-six Coupled Model Intercomparison Project-5-based global climate models (GCMs) are explored to evaluate the performance of maximum ( T max) and minimum ( T min) temperature simulations for India covering 40 grid points. Three performance indicators used for evaluating GCMs are correlation coefficient (CC), normalised root mean square error (NRMSE) and skill score (SS). Entropy method is applied to compute the weights of the three indicators employed. However, equal weights are also considered as part of sensitivity analysis studies. Compromise programming (CP), a distance-based decision-making technique, is employed to rank the GCMs. Group decision-making approach is used to aggregate the ranking patterns obtained for individual grid points. A simple but effective ensemble approach is also suggested.

  7. The Extrapolation-Accelerated Multilevel Aggregation Method in PageRank Computation

    Directory of Open Access Journals (Sweden)

    Bing-Yuan Pu


    Full Text Available An accelerated multilevel aggregation method is presented for calculating the stationary probability vector of an irreducible stochastic matrix in PageRank computation, where the vector extrapolation method is its accelerator. We show how to periodically combine the extrapolation method together with the multilevel aggregation method on the finest level for speeding up the PageRank computation. Detailed numerical results are given to illustrate the behavior of this method, and comparisons with the typical methods are also made.

  8. Methods of computing vocabulary size for the two-parameter rank distribution (United States)

    Edmundson, H. P.; Fostel, G.; Tung, I.; Underwood, W.


    A summation method is described for computing the vocabulary size for given parameter values in the 1- and 2-parameter rank distributions. Two methods of determining the asymptotes for the family of 2-parameter rank-distribution curves are also described. Tables are computed and graphs are drawn relating paris of parameter values to the vocabulary size. The partial product formula for the Riemann zeta function is investigated as an approximation to the partial sum formula for the Riemann zeta function. An error bound is established that indicates that the partial product should not be used to approximate the partial sum in calculating the vocabulary size for the 2-parameter rank distribution.

  9. Appendix A: Rankings for All Programs in "NCTQ Teacher Prep Review 2014." [2014 Teacher Prep Review (United States)

    Greenberg, Julie; Walsh, Kate; McKee, Arthur


    The "NCTQ Teacher Prep Review" evaluates the quality of programs that provide preservice preparation of public school teachers. The rankings of all 2,400 elementary, secondary, and special education programs included in "NCTQ Teacher Prep Review 2014" are listed in this appendix as: (1) a numeric national ranking; (2)…

  10. Tensor Rank


    Erdtman, Elias; Jönsson, Carl


    This master's thesis addresses numerical methods of computing the typical ranks of tensors over the real numbers and explores some properties of tensors over finite fields. We present three numerical methods to compute typical tensor rank. Two of these have already been published and can be used to calculate the lowest typical ranks of tensors and an approximate percentage of how many tensors have the lowest typical ranks (for some tensor formats), respectively. The third method was developed...

  11. Computing Principal Eigenvectors of Large Web Graphs: Algorithms and Accelerations Related to PageRank and HITS (United States)

    Nagasinghe, Iranga


    This thesis investigates and develops a few acceleration techniques for the search engine algorithms used in PageRank and HITS computations. PageRank and HITS methods are two highly successful applications of modern Linear Algebra in computer science and engineering. They constitute the essential technologies accounted for the immense growth and…

  12. Identification and ranking of the risk factors of cloud computing in State-Owned organizations

    Directory of Open Access Journals (Sweden)

    Noor Mohammad Yaghoubi


    Full Text Available Rapid development of processing and storage technologies and the success of the Internet have made computing resources cheaper, more powerful and more available than before. This technological trend has enabled the realization of a new computing model called cloud computing. Recently, the State-Owned organizations have begun to utilize cloud computing architectures, platforms, and applications to deliver services and meet constituents’ needs. Despite all of the advantages and opportunities of cloud computing technology, there are so many risks that State-Owned organizations need to know about before their migration to cloud environment. The purpose of this study is to identify and rank the risks factors of cloud computing in State-Owned organizations by making use of IT experts’ opinion. Firstly, by reviewing key articles, a comprehensive list of risks factors were extracted and classified into two categories: tangible and intangible. Then, six experts were interviewed about these risks and their classifications, and 10 risks were identified. After that, process of ranking the risks was done by seeking help from 52 experts and by fuzzy analytic hierarchy process. The results show that experts have identified intangible risks as the most important risks in cloud computing usage by State-Owned organizations. As the results indicate, "data confidentiality" risk has the highest place among the other risks.

  13. Computational Methods for Large Spatio-temporal Datasets and Functional Data Ranking

    KAUST Repository

    Huang, Huang


    This thesis focuses on two topics, computational methods for large spatial datasets and functional data ranking. Both are tackling the challenges of big and high-dimensional data. The first topic is motivated by the prohibitive computational burden in fitting Gaussian process models to large and irregularly spaced spatial datasets. Various approximation methods have been introduced to reduce the computational cost, but many rely on unrealistic assumptions about the process and retaining statistical efficiency remains an issue. We propose a new scheme to approximate the maximum likelihood estimator and the kriging predictor when the exact computation is infeasible. The proposed method provides different types of hierarchical low-rank approximations that are both computationally and statistically efficient. We explore the improvement of the approximation theoretically and investigate the performance by simulations. For real applications, we analyze a soil moisture dataset with 2 million measurements with the hierarchical low-rank approximation and apply the proposed fast kriging to fill gaps for satellite images. The second topic is motivated by rank-based outlier detection methods for functional data. Compared to magnitude outliers, it is more challenging to detect shape outliers as they are often masked among samples. We develop a new notion of functional data depth by taking the integration of a univariate depth function. Having a form of the integrated depth, it shares many desirable features. Furthermore, the novel formation leads to a useful decomposition for detecting both shape and magnitude outliers. Our simulation studies show the proposed outlier detection procedure outperforms competitors in various outlier models. We also illustrate our methodology using real datasets of curves, images, and video frames. Finally, we introduce the functional data ranking technique to spatio-temporal statistics for visualizing and assessing covariance properties, such as

  14. Computer Assisted Parallel Program Generation

    CERN Document Server

    Kawata, Shigeo


    Parallel computation is widely employed in scientific researches, engineering activities and product development. Parallel program writing itself is not always a simple task depending on problems solved. Large-scale scientific computing, huge data analyses and precise visualizations, for example, would require parallel computations, and the parallel computing needs the parallelization techniques. In this Chapter a parallel program generation support is discussed, and a computer-assisted parallel program generation system P-NCAS is introduced. Computer assisted problem solving is one of key methods to promote innovations in science and engineering, and contributes to enrich our society and our life toward a programming-free environment in computing science. Problem solving environments (PSE) research activities had started to enhance the programming power in 1970's. The P-NCAS is one of the PSEs; The PSE concept provides an integrated human-friendly computational software and hardware system to solve a target ...

  15. A result-driven minimum blocking method for PageRank parallel computing (United States)

    Tao, Wan; Liu, Tao; Yu, Wei; Huang, Gan


    Matrix blocking is a common method for improving computational efficiency of PageRank, but the blocking rules are hard to be determined, and the following calculation is complicated. In tackling these problems, we propose a minimum blocking method driven by result needs to accomplish a parallel implementation of PageRank algorithm. The minimum blocking just stores the element which is necessary for the result matrix. In return, the following calculation becomes simple and the consumption of the I/O transmission is cut down. We do experiments on several matrixes of different data size and different sparsity degree. The results show that the proposed method has better computational efficiency than traditional blocking methods.

  16. Identify and rank key factors influencing the adoption of cloud computing for a healthy Electronics

    Directory of Open Access Journals (Sweden)

    Javad Shukuhy


    Full Text Available Cloud computing as a new technology with Internet infrastructure and new approaches can be significant benefits in providing medical services electronically. Aplying this technology in E-Health requires consideration of various factors. The main objective of this study is to identify and rank the factors influencing the adoption of e-health cloud. Based on the Technology-Organization-Environment (TOE framework and Human-Organization-Technology fit (HOT-fit model, 16 sub-factors were identified in four major factors. With survey of 60 experts, academics and experts in health information technology and with the help of fuzzy analytic hierarchy process had ranked these sub-factors and factors. In the literature, considering newness this study, no internal or external study, have not alluded these number of criteria. The results show that when deciding to adopt cloud computing in E-Health, respectively, must be considered technological, human, organizational and environmental factors.

  17. Computer programs as accounting object

    Directory of Open Access Journals (Sweden)

    I.V. Perviy


    Full Text Available Existing approaches to the regulation of accounting software as one of the types of intangible assets have been considered. The features and current state of the legal protection of computer programs have been analyzed. The reasons for the need to use patent law as a means of legal protection of individual elements of computer programs have been discovered. The influence of the legal aspects of the use of computer programs for national legislation to their accounting reflection has been analyzed. The possible options for the transfer of rights from computer programs copyright owners have been analyzed that should be considered during creation of software accounting system at the enterprise. Identified and analyzed the characteristics of computer software as an intangible asset under the current law. General economic characteristics of computer programs as one of the types of intangible assets have been grounded. The main distinguishing features of software compared to other types of intellectual property have been all ocated

  18. A programming approach to computability

    CERN Document Server

    Kfoury, A J; Arbib, Michael A


    Computability theory is at the heart of theoretical computer science. Yet, ironically, many of its basic results were discovered by mathematical logicians prior to the development of the first stored-program computer. As a result, many texts on computability theory strike today's computer science students as far removed from their concerns. To remedy this, we base our approach to computability on the language of while-programs, a lean subset of PASCAL, and postpone consideration of such classic models as Turing machines, string-rewriting systems, and p. -recursive functions till the final chapter. Moreover, we balance the presentation of un solvability results such as the unsolvability of the Halting Problem with a presentation of the positive results of modern programming methodology, including the use of proof rules, and the denotational semantics of programs. Computer science seeks to provide a scientific basis for the study of information processing, the solution of problems by algorithms, and the design ...

  19. Fuzzy bi-objective linear programming for portfolio selection problem with magnitude ranking function (United States)

    Kusumawati, Rosita; Subekti, Retno


    Fuzzy bi-objective linear programming (FBOLP) model is bi-objective linear programming model in fuzzy number set where the coefficients of the equations are fuzzy number. This model is proposed to solve portfolio selection problem which generate an asset portfolio with the lowest risk and the highest expected return. FBOLP model with normal fuzzy numbers for risk and expected return of stocks is transformed into linear programming (LP) model using magnitude ranking function.

  20. Computer Program Newsletter No. 7

    Energy Technology Data Exchange (ETDEWEB)

    Magnuson, W.G. Jr.


    This issue of the Computer Program Newsletter updates an earlier newsletter (Number 2, September 1979) and focuses on electrical network analysis computer programs. In particular, five network analysis programs (SCEPTRE, SPICE2, NET2, CALAHAN, and EMTP) will be described. The objective of this newsletter will be to provide a very brief description of the input syntax and semantics for each program, highlight their strong and weak points, illustrate how the programs are run at Lawrence Livermore National Laboratory using the Octopus computer network, and present examples of input for each of the programs to illustrate some of the features of each program. In a sense, this newsletter can be used as a quick reference guide to the programs.

  1. NASA's computer science research program (United States)

    Larsen, R. L.


    Following a major assessment of NASA's computing technology needs, a new program of computer science research has been initiated by the Agency. The program includes work in concurrent processing, management of large scale scientific databases, software engineering, reliable computing, and artificial intelligence. The program is driven by applications requirements in computational fluid dynamics, image processing, sensor data management, real-time mission control and autonomous systems. It consists of university research, in-house NASA research, and NASA's Research Institute for Advanced Computer Science (RIACS) and Institute for Computer Applications in Science and Engineering (ICASE). The overall goal is to provide the technical foundation within NASA to exploit advancing computing technology in aerospace applications.

  2. Low-rank coal study. Volume 5. RD and D program evaluation

    Energy Technology Data Exchange (ETDEWEB)


    A national program is recommended for research, development, and demonstration (RD and D) of improved technologies for the enviromentally acceptable use of low-rank coals. RD and D project recommendations are outlined in all applicable technology areas, including extraction, transportation, preparation, handling and storage, conventional combustion and environmental control technology, fluidized bed combustion, gasification, liquefaction, and pyrolysis. Basic research topics are identified separately, as well as a series of crosscutting research activities addressing environmental, economic, and regulatory issues. The recommended RD and D activities are classified into Priority I and Priority II categories, reflecting their relative urgency and potential impact on the advancement of low-rank coal development. Summaries of ongoing research projects on low-rank coals in the US are presented in an Appendix, and the relationships of these ongoing efforts to the recommended RD and D program are discussed.

  3. A Multiobjective Programming Method for Ranking All Units Based on Compensatory DEA Model

    Directory of Open Access Journals (Sweden)

    Haifang Cheng


    Full Text Available In order to rank all decision making units (DMUs on the same basis, this paper proposes a multiobjective programming (MOP model based on a compensatory data envelopment analysis (DEA model to derive a common set of weights that can be used for the full ranking of all DMUs. We first revisit a compensatory DEA model for ranking all units, point out the existing problem for solving the model, and present an improved algorithm for which an approximate global optimal solution of the model can be obtained by solving a sequence of linear programming. Then, we applied the key idea of the compensatory DEA model to develop the MOP model in which the objectives are to simultaneously maximize all common weights under constraints that the sum of efficiency values of all DMUs is equal to unity and the sum of all common weights is also equal to unity. In order to solve the MOP model, we transform it into a single objective programming (SOP model using a fuzzy programming method and solve the SOP model using the proposed approximation algorithm. To illustrate the ranking method using the proposed method, two numerical examples are solved.

  4. Individual Differences in Learning Computer Programming: A Social Cognitive Approach (United States)

    Akar, Sacide Guzin Mazman; Altun, Arif


    The purpose of this study is to investigate and conceptualize the ranks of importance of social cognitive variables on university students' computer programming performances. Spatial ability, working memory, self-efficacy, gender, prior knowledge and the universities students attend were taken as variables to be analyzed. The study has been…

  5. Designing computer programs

    CERN Document Server

    Haigh, Jim


    This is a book for students at every level who are learning to program for the first time - and for the considerable number who learned how to program but were never taught to structure their programs. The author presents a simple set of guidelines that show the programmer how to design in a manageable structure from the outset. The method is suitable for most languages, and is based on the widely used 'JSP' method, to which the student may easily progress if it is needed at a later stage.Most language specific texts contain very little if any information on design, whilst books on des

  6. Computer Program NIKE

    DEFF Research Database (Denmark)

    Spanget-Larsen, Jens


    FORTRAN source code for program NIKE (PC version of QCPE 343). Sample input and output for two model chemical reactions are appended: I. Three consecutive monomolecular reactions, II. A simple chain mechanism...

  7. Relations between Prestige Rankings of Clinical Psychology Doctoral Programs and Scores on the Examination for Professional Practice in Psychology (EPPP) (United States)

    Townsend, James M.; Ryan, Joseph J.


    We assessed the relationship between "U.S. News and World Report" 2008 rankings of clinical psychology doctoral programs and scores earned by graduates on the Examination for Professional Practice in Psychology (EPPP). For the top 25 programs, relationship between ranking and EPPP scores was not significant, r[subscript s] = -0.28. EPPP scores…

  8. Line-Editor Computer Program (United States)

    Scott, Peter J.


    ZED editing program for DEC VAX computer simple, powerful line editor for text, program source code, and nonbinary data. Excels in processing of text by use of procedure files. Also features versatile search qualifiers, global changes, conditionals, online help, hexadecimal mode, space compression, looping, logical combinations of search strings, journaling, visible control characters, and automatic detabbing. Users of Cambridge implementation devised such ZED procedures as chess games, calculators, and programs for evaluating pi. Written entirely in C.

  9. The psychology of computer programming

    CERN Document Server

    Weinberg, Gerald Marvin


    This landmark 1971 classic is reprinted with a new preface, chapter-by-chapter commentary, and straight-from-the-heart observations on topics that affect the professional life of programmers. Long regarded as one of the first books to pioneer a people-oriented approach to computing, The Psychology of Computer Programming endures as a penetrating analysis of the intelligence, skill, teamwork, and problem-solving power of the computer programmer. Finding the chapters strikingly relevant to today's issues in programming, Gerald M. Weinberg adds new insights and highlights the similarities and differences between now and then. Using a conversational style that invites the reader to join him, Weinberg reunites with some of his most insightful writings on the human side of software engineering. Topics include egoless programming, intelligence, psychological measurement, personality factors, motivation, training, social problems on large projects, problem-solving ability, programming language design, team formati...

  10. An Efficient Ranking Technique for Intuitionistic Fuzzy Numbers with Its Application in Chance Constrained Bilevel Programming

    Directory of Open Access Journals (Sweden)

    Animesh Biswas


    Full Text Available The aim of this paper is to develop a new ranking technique for intuitionistic fuzzy numbers using the method of defuzzification based on probability density function of the corresponding membership function, as well as the complement of nonmembership function. Using the proposed ranking technique a methodology for solving linear bilevel fuzzy stochastic programming problem involving normal intuitionistic fuzzy numbers is developed. In the solution process each objective is solved independently to set the individual goal value of the objectives of the decision makers and thereby constructing fuzzy membership goal of the objectives of each decision maker. Finally, a fuzzy goal programming approach is considered to achieve the highest membership degree to the extent possible of each of the membership goals of the decision makers in the decision making context. Illustrative numerical examples are provided to demonstrate the applicability of the proposed methodology and the achieved results are compared with existing techniques.

  11. A C++11 implementation of arbitrary-rank tensors for high-performance computing (United States)

    Aragón, Alejandro M.


    This article discusses an efficient implementation of tensors of arbitrary rank by using some of the idioms introduced by the recently published C++ ISO Standard (C++11). With the aims at providing a basic building block for high-performance computing, a single Array class template is carefully crafted, from which vectors, matrices, and even higher-order tensors can be created. An expression template facility is also built around the array class template to provide convenient mathematical syntax. As a result, by using templates, an extra high-level layer is added to the C++ language when dealing with algebraic objects and their operations, without compromising performance. The implementation is tested running on both CPU and GPU.

  12. Functional Programming in Computer Science

    Energy Technology Data Exchange (ETDEWEB)

    Anderson, Loren James [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Davis, Marion Kei [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)


    We explore functional programming through a 16-week internship at Los Alamos National Laboratory. Functional programming is a branch of computer science that has exploded in popularity over the past decade due to its high-level syntax, ease of parallelization, and abundant applications. First, we summarize functional programming by listing the advantages of functional programming languages over the usual imperative languages, and we introduce the concept of parsing. Second, we discuss the importance of lambda calculus in the theory of functional programming. Lambda calculus was invented by Alonzo Church in the 1930s to formalize the concept of effective computability, and every functional language is essentially some implementation of lambda calculus. Finally, we display the lasting products of the internship: additions to a compiler and runtime system for the pure functional language STG, including both a set of tests that indicate the validity of updates to the compiler and a compiler pass that checks for illegal instances of duplicate names.

  13. Sonic boom research. [computer program (United States)

    Zakkay, V.; Ting, L.


    A computer program for CDC 6600 is developed for the nonlinear sonic boom analysis including the asymmetric effect of lift near the vertical plane of symmetry. The program is written in FORTRAN 4 language. This program carries out the numerical integration of the nonlinear governing equations from the input data at a finite distance from the airplane configuration at a flight altitude to yield the pressure signitude at ground. The required input data and the format for the output are described. A complete program listing and a sample calculation are given.

  14. Elliptical Orbit Performance Computer Program (United States)

    Myler, T.


    Elliptical Orbit Performance (ELOPE) computer program for analyzing orbital performance of space boosters uses orbit insertion data obtained from trajectory simulation to generate parametric data on apogee and perigee altitudes as function of payload data. Data used to generate presentation plots that display elliptical orbit performance capability of space booster.

  15. Theoretical issues in quantum computing: Graph isomorphism, PageRank, and Hamiltonian determination (United States)

    Rudinger, Kenneth Michael

    This thesis explores several theoretical questions pertaining to quantum computing. First we examine several questions regarding multi-particle quantum random walk-based algorithms for the graph isomorphism problem. We find that there exists a non-trivial difference between continuous-time walks of one and two non-interacting particles as compared to non-interacting walks of three or more particles, in that the latter are able to distinguish many strongly regular graphs (SRGs), a class of graphs with many graph pairs that are difficult to distinguish. We demonstrate analytically where this distinguishing power comes from, and we show numerically that three-particle and four-particle non-interacting continuous-time walks can distinguish many pairs of strongly regular graphs. We additionally show that this distinguishing power, while it grows with particle number, is bounded, so that no continuous-time non-interacting walk of fixed particle number can distinguish all strongly regular graphs. We then investigate the relationship between continuous-time and discrete-time walks, in the context of the graph isomorphism problem. While it has been previously demonstrated numerically that discrete-time walks of non-interacting particles can distinguish some SRGs, we demonstrate where this distinguishing power comes from. We also show that while no continuous-time non-interacting walk of fixed particle number can distinguish SRGs, it remains a possibility that such a discrete-time walk could, leaving open the possibility of a non-trivial difference between discrete-time and continuous-time walks. The last piece of our work on graph isomorphism examines limitations on certain kinds of continuous-time walk-based algorithms for distinguishing graphs. We show that a very general class of continuous-time walk algorithms, with a broad class of allowable interactions, cannot distinguish all graphs. We next consider a previously-proposed quantum adiabatic algorithm for computing the


    Directory of Open Access Journals (Sweden)

    Monalisha Pattnaik


    Full Text Available Background: This paper explores the solutions to the fuzzy optimization linear program problems (FOLPP where some parameters are fuzzy numbers. In practice, there are many problems in which all decision parameters are fuzzy numbers, and such problems are usually solved by either probabilistic programming or multi-objective programming methods. Methods: In this paper, using the concept of comparison of fuzzy numbers, a very effective method is introduced for solving these problems. This paper extends linear programming based problem in fuzzy environment. With the problem assumptions, the optimal solution can still be theoretically solved using the two phase simplex based method in fuzzy environment. To handle the fuzzy decision variables can be initially generated and then solved and improved sequentially using the fuzzy decision approach by introducing robust ranking technique. Results and conclusions: The model is illustrated with an application and a post optimal analysis approach is obtained. The proposed procedure was programmed with MATLAB (R2009a version software for plotting the four dimensional slice diagram to the application. Finally, numerical example is presented to illustrate the effectiveness of the theoretical results, and to gain additional managerial insights. 

  17. Computer Education in Dental Laboratory Technology Programs. (United States)

    Rogers, William A.; Hawkins, Robert Ross


    A 1990 survey of 37 dental technology programs investigated 3 areas of computer use: current and anticipated general computer education courses; incorporation of computer applications into technology and management courses; and faculty use of the computer. Most programs are beginning to expand use of technology. (MSE)

  18. Computer technology and computer programming research and strategies

    CERN Document Server

    Antonakos, James L


    Covering a broad range of new topics in computer technology and programming, this volume discusses encryption techniques, SQL generation, Web 2.0 technologies, and visual sensor networks. It also examines reconfigurable computing, video streaming, animation techniques, and more. Readers will learn about an educational tool and game to help students learn computer programming. The book also explores a new medical technology paradigm centered on wireless technology and cloud computing designed to overcome the problems of increasing health technology costs.

  19. Use of the computer program in a cloud computing

    Directory of Open Access Journals (Sweden)

    Radovanović Sanja


    Full Text Available Cloud computing represents a specific networking, in which a computer program simulates the operation of one or more server computers. In terms of copyright, all technological processes that take place within the cloud computing are covered by the notion of copying computer programs, and exclusive right of reproduction. However, this right suffers some limitations in order to allow normal use of computer program by users. Based on the fact that the cloud computing is virtualized network, the issue of normal use of the computer program requires to put all aspects of the permitted copying into the context of a specific computing environment and specific processes within the cloud. In this sense, the paper pointed out that the user of a computer program in cloud computing, needs to obtain the consent of the right holder for any act which he undertakes using the program. In other words, the copyright in the cloud computing is a full scale, and thus the freedom of contract (in the case of this particular restriction as well.

  20. Predicted binding site information improves model ranking in protein docking using experimental and computer-generated target structures. (United States)

    Maheshwari, Surabhi; Brylinski, Michal


    Protein-protein interactions (PPIs) mediate the vast majority of biological processes, therefore, significant efforts have been directed to investigate PPIs to fully comprehend cellular functions. Predicting complex structures is critical to reveal molecular mechanisms by which proteins operate. Despite recent advances in the development of new methods to model macromolecular assemblies, most current methodologies are designed to work with experimentally determined protein structures. However, because only computer-generated models are available for a large number of proteins in a given genome, computational tools should tolerate structural inaccuracies in order to perform the genome-wide modeling of PPIs. To address this problem, we developed eRank(PPI), an algorithm for the identification of near-native conformations generated by protein docking using experimental structures as well as protein models. The scoring function implemented in eRank(PPI) employs multiple features including interface probability estimates calculated by eFindSite(PPI) and a novel contact-based symmetry score. In comparative benchmarks using representative datasets of homo- and hetero-complexes, we show that eRank(PPI) consistently outperforms state-of-the-art algorithms improving the success rate by ~10 %. eRank(PPI) was designed to bridge the gap between the volume of sequence data, the evidence of binary interactions, and the atomic details of pharmacologically relevant protein complexes. Tolerating structure imperfections in computer-generated models opens up a possibility to conduct the exhaustive structure-based reconstruction of PPI networks across proteomes. The methods and datasets used in this study are available at

  1. Exact fast computation of band depth for large functional datasets: How quickly can one million curves be ranked?

    KAUST Repository

    Sun, Ying


    © 2012 John Wiley & Sons, Ltd. Band depth is an important nonparametric measure that generalizes order statistics and makes univariate methods based on order statistics possible for functional data. However, the computational burden of band depth limits its applicability when large functional or image datasets are considered. This paper proposes an exact fast method to speed up the band depth computation when bands are defined by two curves. Remarkable computational gains are demonstrated through simulation studies comparing our proposal with the original computation and one existing approximate method. For example, we report an experiment where our method can rank one million curves, evaluated at fifty time points each, in 12.4 seconds with Matlab.

  2. Personal Computer Transport Analysis Program (United States)

    DiStefano, Frank, III; Wobick, Craig; Chapman, Kirt; McCloud, Peter


    The Personal Computer Transport Analysis Program (PCTAP) is C++ software used for analysis of thermal fluid systems. The program predicts thermal fluid system and component transients. The output consists of temperatures, flow rates, pressures, delta pressures, tank quantities, and gas quantities in the air, along with air scrubbing component performance. PCTAP s solution process assumes that the tubes in the system are well insulated so that only the heat transfer between fluid and tube wall and between adjacent tubes is modeled. The system described in the model file is broken down into its individual components; i.e., tubes, cold plates, heat exchangers, etc. A solution vector is built from the components and a flow is then simulated with fluid being transferred from one component to the next. The solution vector of components in the model file is built at the initiation of the run. This solution vector is simply a list of components in the order of their inlet dependency on other components. The component parameters are updated in the order in which they appear in the list at every time step. Once the solution vectors have been determined, PCTAP cycles through the components in the solution vector, executing their outlet function for each time-step increment.

  3. Debugging a high performance computing program (United States)

    Gooding, Thomas M.


    Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

  4. Computer Assistance for Writing Interactive Programs: TICS (United States)

    Kaplow, Ray; And Others


    A description of an on-line and interactive programing system (TICS - Teacher-Interactive-Computer-System), which is aimed at facilitating the authoring of interactive, instructional computer programs by persons who are experts on the subject matter being addressed, but not necessarily programers. (Author)

  5. The NASA computer science research program plan (United States)


    A taxonomy of computer science is included, one state of the art of each of the major computer science categories is summarized. A functional breakdown of NASA programs under Aeronautics R and D, space R and T, and institutional support is also included. These areas were assessed against the computer science categories. Concurrent processing, highly reliable computing, and information management are identified.

  6. High performance computing and communications program (United States)

    Holcomb, Lee


    A review of the High Performance Computing and Communications (HPCC) program is provided in vugraph format. The goals and objectives of this federal program are as follows: extend U.S. leadership in high performance computing and computer communications; disseminate the technologies to speed innovation and to serve national goals; and spur gains in industrial competitiveness by making high performance computing integral to design and production.

  7. A heuristic ranking approach on capacity benefit margin determination using Pareto-based evolutionary programming technique. (United States)

    Othman, Muhammad Murtadha; Abd Rahman, Nurulazmi; Musirin, Ismail; Fotuhi-Firuzabad, Mahmud; Rajabi-Ghahnavieh, Abbas


    This paper introduces a novel multiobjective approach for capacity benefit margin (CBM) assessment taking into account tie-line reliability of interconnected systems. CBM is the imperative information utilized as a reference by the load-serving entities (LSE) to estimate a certain margin of transfer capability so that a reliable access to generation through interconnected system could be attained. A new Pareto-based evolutionary programming (EP) technique is used to perform a simultaneous determination of CBM for all areas of the interconnected system. The selection of CBM at the Pareto optimal front is proposed to be performed by referring to a heuristic ranking index that takes into account system loss of load expectation (LOLE) in various conditions. Eventually, the power transfer based available transfer capability (ATC) is determined by considering the firm and nonfirm transfers of CBM. A comprehensive set of numerical studies are conducted on the modified IEEE-RTS79 and the performance of the proposed method is numerically investigated in detail. The main advantage of the proposed technique is in terms of flexibility offered to an independent system operator in selecting an appropriate solution of CBM simultaneously for all areas.

  8. Ranking Journalism and Mass Communications Programs: Administrators and Faculty Approve of the Idea and Assess Potential Criteria (United States)

    Weber, Joseph


    Rankings of universities and colleges are common and controversial. However, few rankers produce useful lists that assess and compare journalism and mass communications programs. The few currently available involve superficial reputational surveys or are less than transparent about their methodology. To determine potential criteria for a useful…

  9. What do reversible programs compute?

    DEFF Research Database (Denmark)

    Axelsen, Holger Bock; Glück, Robert


    are not strictly classically universal, but that they support another notion of universality; we call this RTM-universality. Thus, even though the RTMs are sub-universal in the classical sense, they are powerful enough as to include a self-interpreter. Lifting this to other computation models, we propose r...... be the starting point of a computational theory of reversible computing. We provide a novel semantics-based approach to such a theory, using reversible Turing machines (RTMs) as the underlying computation model. We show that the RTMs can compute exactly all injective, computable functions. We find that the RTMs...

  10. Integer programming theory, applications, and computations

    CERN Document Server

    Taha, Hamdy A


    Integer Programming: Theory, Applications, and Computations provides information pertinent to the theory, applications, and computations of integer programming. This book presents the computational advantages of the various techniques of integer programming.Organized into eight chapters, this book begins with an overview of the general categorization of integer applications and explains the three fundamental techniques of integer programming. This text then explores the concept of implicit enumeration, which is general in a sense that it is applicable to any well-defined binary program. Other

  11. Mastering cloud computing foundations and applications programming

    CERN Document Server

    Buyya, Rajkumar; Selvi, SThamarai


    Mastering Cloud Computing is designed for undergraduate students learning to develop cloud computing applications. Tomorrow's applications won't live on a single computer but will be deployed from and reside on a virtual server, accessible anywhere, any time. Tomorrow's application developers need to understand the requirements of building apps for these virtual systems, including concurrent programming, high-performance computing, and data-intensive systems. The book introduces the principles of distributed and parallel computing underlying cloud architectures and specifical

  12. Evaluation and Ranking of Geothermal Resources for Electrical Generation or Electrical Offset in Idaho, Montana, Oregon and Washington. Volume II.

    Energy Technology Data Exchange (ETDEWEB)

    Bloomquist, R. Gordon


    This volume contains appendices on: (1) resource assessment - electrical generation computer results; (2) resource assessment summary - direct use computer results; (3) electrical generation (high temperature) resource assessment computer program listing; (4) direct utilization (low temperature) resource assessment computer program listing; (5) electrical generation computer program CENTPLANT and related documentation; (6) electrical generation computer program WELLHEAD and related documentation; (7) direct utilization computer program HEATPLAN and related documentation; (8) electrical generation ranking computer program GEORANK and related documentation; (9) direct utilization ranking computer program GEORANK and related documentation; and (10) life cycle cost analysis computer program and related documentation. (ACR)

  13. On Verified Numerical Computations in Convex Programming


    Jansson, Christian


    This survey contains recent developments for computing verified results of convex constrained optimization problems, with emphasis on applications. Especially, we consider the computation of verified error bounds for non-smooth convex conic optimization in the framework of functional analysis, for linear programming, and for semidefinite programming. A discussion of important problem transformations to special types of convex problems and convex relaxations is included...

  14. An Efficient Monte Carlo Approach to Compute PageRank for Large Graphs on a Single PC

    Directory of Open Access Journals (Sweden)

    Sonobe Tomohiro


    Full Text Available This paper describes a novel Monte Carlo based random walk to compute PageRanks of nodes in a large graph on a single PC. The target graphs of this paper are ones whose size is larger than the physical memory. In such an environment, memory management is a difficult task for simulating the random walk among the nodes. We propose a novel method that partitions the graph into subgraphs in order to make them fit into the physical memory, and conducts the random walk for each subgraph. By evaluating the walks lazily, we can conduct the walks only in a subgraph and approximate the random walk by rotating the subgraphs. In computational experiments, the proposed method exhibits good performance for existing large graphs with several passes of the graph data.

  15. Structured Parallel Programming Patterns for Efficient Computation

    CERN Document Server

    McCool, Michael; Robison, Arch


    Programming is now parallel programming. Much as structured programming revolutionized traditional serial programming decades ago, a new kind of structured programming, based on patterns, is relevant to parallel programming today. Parallel computing experts and industry insiders Michael McCool, Arch Robison, and James Reinders describe how to design and implement maintainable and efficient parallel algorithms using a pattern-based approach. They present both theory and practice, and give detailed concrete examples using multiple programming models. Examples are primarily given using two of th

  16. A Similarity-Ranking Method on Semantic Computing for Providing Information-Services in Station-Concierge System

    Directory of Open Access Journals (Sweden)

    Motoki Yokoyama


    Full Text Available The prevalence of smartphones and wireless broadband networks have been progressing as a new Railway infomration environment. According to the spread of such devices and information technology, various types of information can be obtained from databases connected to the Internet. One scenario of obtaining such a wide variety of information resources is in the phase of user’s transportation. This paper proposes an information provision system, named the Station Concierge System that matches the situation and intention of passengers. The purpose of this system is to estimate the needs of passengers like station staff or hotel concierge and to provide information resources that satisfy user’s expectations dynamically. The most important module of the system is constructed based on a new information ranking method for passenger intention prediction and service recommendation. This method has three main features, which are (1 projecting a user to semantic vector space by using her current context, (2 predicting the intention of a user based on selecting a semantic vector subspace, and (3 ranking the services by a descending order of relevant scores to the user’ intention. By comparing the predicted results of our method with those of two straightforward computation methods, the experimental studies show the effectiveness and efficiency of the proposed method. Using this system, users can obtain transit information and service map that dynamically matches their context.

  17. A Semi-Preemptive Computational Service System with Limited Resources and Dynamic Resource Ranking

    Directory of Open Access Journals (Sweden)

    Fang-Yie Leu


    Full Text Available In this paper, we integrate a grid system and a wireless network to present a convenient computational service system, called the Semi-Preemptive Computational Service system (SePCS for short, which provides users with a wireless access environment and through which a user can share his/her resources with others. In the SePCS, each node is dynamically given a score based on its CPU level, available memory size, current length of waiting queue, CPU utilization and bandwidth. With the scores, resource nodes are classified into three levels. User requests based on their time constraints are also classified into three types. Resources of higher levels are allocated to more tightly constrained requests so as to increase the total performance of the system. To achieve this, a resource broker with the Semi-Preemptive Algorithm (SPA is also proposed. When the resource broker cannot find suitable resources for the requests of higher type, it preempts the resource that is now executing a lower type request so that the request of higher type can be executed immediately. The SePCS can be applied to a Vehicular Ad Hoc Network (VANET, users of which can then exploit the convenient mobile network services and the wireless distributed computing. As a result, the performance of the system is higher than that of the tested schemes.

  18. Computer Programs in Marine Science (United States)


    between two locations. Requires subroutines COS, SIN, ARCOS . Author - Ralph Johnson. Oceanographic Services Branch Copy on file at XODC National...STEREOGRAPHIC PROJECTICN 65 FORTRAN CDC 3800 PIE SCATTERINC COMPUTATIONS 41 FURTRAN COC 36UO PLCTS TRACK $AD DATA PROFILE .RACK 47 FORTRAN COC 3.00 PLCTS

  19. FORTRAN computer program for seismic risk analysis (United States)

    McGuire, Robin K.


    A program for seismic risk analysis is described which combines generality of application, efficiency and accuracy of operation, and the advantage of small storage requirements. The theoretical basis for the program is first reviewed, and the computational algorithms used to apply this theory are described. The information required for running the program is listed. Published attenuation functions describing the variation with earthquake magnitude and distance of expected values for various ground motion parameters are summarized for reference by the program user. Finally, suggestions for use of the program are made, an example problem is described (along with example problem input and output) and the program is listed.

  20. Computer programming and architecture the VAX

    CERN Document Server

    Levy, Henry


    Takes a unique systems approach to programming and architecture of the VAXUsing the VAX as a detailed example, the first half of this book offers a complete course in assembly language programming. The second describes higher-level systems issues in computer architecture. Highlights include the VAX assembler and debugger, other modern architectures such as RISCs, multiprocessing and parallel computing, microprogramming, caches and translation buffers, and an appendix on the Berkeley UNIX assembler.

  1. The Dynamic Geometrisation of Computer Programming (United States)

    Sinclair, Nathalie; Patterson, Margaret


    The goal of this paper is to explore dynamic geometry environments (DGE) as a type of computer programming language. Using projects created by secondary students in one particular DGE, we analyse the extent to which the various aspects of computational thinking--including both ways of doing things and particular concepts--were evident in their…

  2. 43 Computer Assisted Programmed Instruction and Cognitive ...

    African Journals Online (AJOL)


    Computer Assisted Programmed Instruction and Cognitive Preference Style as. Determinant of Achievement of Secondary School Physics Students. Sotayo, M. A. O.. Federal College of Education, Osiele, Abeokuta, Nigeria. Abstract. The study probes into the effect of Computer Assisted Instruction and Cognitive preference.

  3. NASA High-End Computing Program Website (United States)

    Cohen, Jarrett S.


    If you are a NASA-sponsored scientist or engineer. computing time is available to you at the High-End Computing (HEC) Program's NASA Advanced Supercomputing (NAS) Facility and NASA Center for Computational Sciences (NCCS). The Science Mission Directorate will select from requests NCCS Portals submitted to the e-Books online system for awards beginning on May 1. Current projects set to explore on April 30 must have a request in e-Books to be considered for renewal

  4. Computer-Aided Corrosion Program Management (United States)

    MacDowell, Louis


    This viewgraph presentation reviews Computer-Aided Corrosion Program Management at John F. Kennedy Space Center. The contents include: 1) Corrosion at the Kennedy Space Center (KSC); 2) Requirements and Objectives; 3) Program Description, Background and History; 4) Approach and Implementation; 5) Challenges; 6) Lessons Learned; 7) Successes and Benefits; and 8) Summary and Conclusions.

  5. Developing computer training programs for blood bankers. (United States)

    Eisenbrey, L


    Two surveys were conducted in July 1991 to gather information about computer training currently performed within American Red Cross Blood Services Regions. One survey was completed by computer trainers from software developer-vendors and regional centers. The second survey was directed to the trainees, to determine their perception of the computer training. The surveys identified the major concepts, length of training, evaluations, and methods of instruction used. Strengths and weaknesses of training programs were highlighted by trainee respondents. Using the survey information and other sources, recommendations (including those concerning which computer skills and tasks should be covered) are made that can be used as guidelines for developing comprehensive computer training programs at any blood bank or blood center.

  6. Estimation of the P(CS) with a Computer Program. Revision (United States)


    ESTIMATIONI OF THlE P(CS) WITH A COMPUTER PROGRAM4 by Edward J. Dudewlcz* .BTIMATICE OF THE P(cs) WITHI A COMUTER PROGRAM by IEdward J. Dudewicz* . t...J. Dudevicz was aI Visiting Scholar, Department of Mathematics and Statistics, University of Nebraska, Lincoln. Fdward J. Dudewicz is Professor...1954’. A single-sample multiple decision procedure for ranking means of normal populations with known variances. Annals of Mathematical Statistics

  7. The Computational Physics Program of the national MFE Computer Center

    Energy Technology Data Exchange (ETDEWEB)

    Mirin, A.A.


    Since June 1974, the MFE Computer Center has been engaged in a significant computational physics effort. The principal objective of the Computational Physics Group is to develop advanced numerical models for the investigation of plasma phenomena and the simulation of present and future magnetic confinement devices. Another major objective of the group is to develop efficient algorithms and programming techniques for current and future generations of supercomputers. The Computational Physics Group has been involved in several areas of fusion research. One main area is the application of Fokker-Planck/quasilinear codes to tokamaks. Another major area is the investigation of resistive magnetohydrodynamics in three dimensions, with applications to tokamaks and compact toroids. A third area is the investigation of kinetic instabilities using a 3-D particle code; this work is often coupled with the task of numerically generating equilibria which model experimental devices. Ways to apply statistical closure approximations to study tokamak-edge plasma turbulence have been under examination, with the hope of being able to explain anomalous transport. Also, we are collaborating in an international effort to evaluate fully three-dimensional linear stability of toroidal devices. In addition to these computational physics studies, the group has developed a number of linear systems solvers for general classes of physics problems and has been making a major effort at ascertaining how to efficiently utilize multiprocessor computers. A summary of these programs are included in this paper. 6 tabs.

  8. An introduction to Python and computer programming

    CERN Document Server

    Zhang, Yue


    This book introduces Python programming language and fundamental concepts in algorithms and computing. Its target audience includes students and engineers with little or no background in programming, who need to master a practical programming language and learn the basic thinking in computer science/programming. The main contents come from lecture notes for engineering students from all disciplines, and has received high ratings. Its materials and ordering have been adjusted repeatedly according to classroom reception. Compared to alternative textbooks in the market, this book introduces the underlying Python implementation of number, string, list, tuple, dict, function, class, instance and module objects in a consistent and easy-to-understand way, making assignment, function definition, function call, mutability and binding environments understandable inside-out. By giving the abstraction of implementation mechanisms, this book builds a solid understanding of the Python programming language.

  9. Rank Dynamics (United States)

    Gershenson, Carlos

    Studies of rank distributions have been popular for decades, especially since the work of Zipf. For example, if we rank words of a given language by use frequency (most used word in English is 'the', rank 1; second most common word is 'of', rank 2), the distribution can be approximated roughly with a power law. The same applies for cities (most populated city in a country ranks first), earthquakes, metabolism, the Internet, and dozens of other phenomena. We recently proposed ``rank diversity'' to measure how ranks change in time, using the Google Books Ngram dataset. Studying six languages between 1800 and 2009, we found that the rank diversity curves of languages are universal, adjusted with a sigmoid on log-normal scale. We are studying several other datasets (sports, economies, social systems, urban systems, earthquakes, artificial life). Rank diversity seems to be universal, independently of the shape of the rank distribution. I will present our work in progress towards a general description of the features of rank change in time, along with simple models which reproduce it

  10. Role of logic programming in computer studies

    Directory of Open Access Journals (Sweden)

    Nicolae PELIN


    Full Text Available The paper contains the analysis of the opinions of a number of scholars and specialists on the importance and the role in logic programming methodology of studying computer science, philosophy about the logic programs and interpreter, concerning the burden of which is opposite to the programmer if there is logic interpreter. The presented material is meant, according to the author, to help the reader to understand more easily the analyzed multilateral problem.

  11. Program computes turbine steam rates and properties

    Energy Technology Data Exchange (ETDEWEB)

    Ganapathy, V. (ABCO Industries, Inc., Abilene, TX (US))


    BASIC computer program quickly evaluates steam properties and rates during expansion in a steam turbine. Engineers involved in cogeneration projects and power plant studies often need to calculate the steam properties during expansion in a steam turbine to evaluate the theoretical and actual steam rates and hence, the electrical power output. With the help of this program written in BASIC, one can quickly evaluate all the pertinent data. Correlations used for steam property evaluation are also presented.

  12. A Computer-Aided Bibliometrics System for Journal Citation Analysis and Departmental Core Journal Ranking List Generation

    Directory of Open Access Journals (Sweden)

    Yih-Chearng Shiue


    Full Text Available Due to the tremendous increase and variation in serial publications, faculties in department of university are finding it difficult to generate and update their departmental core journal list regularly and accurately, and libraries are finding it difficult to maintain their current serial collection for different departments. Therefore, the evaluation of a departmental core journal list is an important task for departmental faculties and librarians. A departmental core journal list not only helps departments understand research performances of faculties and students, but also helps librarians make decisions about which journals to retain and which to cancel. In this study, a Computer-Aided Bibliometrics System was implemented and two methodologies (JCDF and LibJF were proposed in order to generate a departmental core journal ranking list and make the journal citation analysis. Six departments were taken as examples, with MIS as the major one. One journal citation pattern was found and the ratio of Turning point-to-No. journal was always around 0.07 among the 10 journals and 6 departments. After comparing with four methodologies via overlapping rate and standard deviation distances, the two proposed methodologies were shown to be better than questionnaire and library subscription method.

  13. Academic Productivity of US Neurosurgery Residents as Measured by H-Index: Program Ranking with Correlation to Faculty Productivity. (United States)

    Sarkiss, Christopher A; Riley, Kyle J; Hernandez, Christopher M; Oermann, Eric K; Ladner, Travis R; Bederson, Joshua B; Shrivastava, Raj K


    Engagement in research and academic productivity are crucial components in the training of a neurosurgeon. This process typically begins in residency training. In this study, we analyzed individual resident productivity as it correlated to publications across all Accreditation Council for Graduate Medical Education (ACGME)-accredited neurosurgery training programs in an attempt to identify how programs have developed and fostered a research culture and environment. We obtained a list of current neurosurgery residents in ACGME-accredited programs from the American Association of Neurological Surgeons database. An expanded PubMed and Scopus search was conducted for each resident through the present time. We tabulated all articles attributed to each resident. We then categorized the publications based on each neurosurgical subspecialty while in residency. A spreadsheet-based statistical analysis was performed. This formulated the average number of resident articles, h-indices, and most common subspecialty categories by training program. We analyzed 1352 current neurosurgery residents in 105 programs. There were a total of 10 645 publications, of which 3985 were resident first-author publications during the period of study. The most common subspecialties among all resident publications were vascular (24.9%), spine (16.9%), oncology (16.1%), pediatric (5.6%), functional (4.9%), and trauma (3.8%). The average resident published 2.9 first-author papers with average of 38.0 first-author publications by total residents at each program (range 0-241). The average h-index per resident is 2.47 ± 3.25. When comparing previously published faculty h-index program rankings against our resident h-index rankings, there is a strong correlation between the 2 datasets with a clear delineation between Top-20 productivity and that of other programs (average h-index 4.2 vs 1.7, respectively, P productivity on both the resident and faculty level (average h-index 1.6, 1.9, 3.9 for 1, 2, and

  14. A CAD (Classroom Assessment Design) of a Computer Programming Course (United States)

    Hawi, Nazir S.


    This paper presents a CAD (classroom assessment design) of an entry-level undergraduate computer programming course "Computer Programming I". CAD has been the product of a long experience in teaching computer programming courses including teaching "Computer Programming I" 22 times. Each semester, CAD is evaluated and modified…

  15. Introduction to programming multiple-processor computers

    Energy Technology Data Exchange (ETDEWEB)

    Hicks, H.R.; Lynch, V.E.


    FORTRAN applications programs can be executed on multiprocessor computers in either a unitasking (traditional) or multitasking form. The latter allows a single job to use more than one processor simultaneously, with a consequent reduction in wall-clock time and, perhaps, the cost of the calculation. An introduction to programming in this environment is presented. The concepts of synchronization and data sharing using EVENTS and LOCKS are illustrated with examples. The strategy of strong synchronization and the use of synchronization templates are proposed. We emphasize that incorrect multitasking programs can produce irreproducible results, which makes debugging more difficult.

  16. Computer Assisted Programmed Instruction and Cognitive ...

    African Journals Online (AJOL)

    The achievement of students of application learning mode was also significantly higher than those of recall and principle respectively. There was no significant interaction effect between Cognitive Preference Style and Computer Assisted Programmed Instruction. The implications of the result to the stakeholder were ...

  17. Computer program package for PIXE spectra evaluation

    Energy Technology Data Exchange (ETDEWEB)

    Kajfosz, J. [Institute of Nuclear Physics, Cracow (Poland)


    The computer programs described here were developed for calculating the concentrations of elements in samples analysed by the PIXE (Proton Induced X-ray Emission) method from the X-ray spectra obtained in those analyses. (author). 10 refs, 2 figs.

  18. Computer programming students head to Tokyo


    Crumbley, Liz


    "The Milk's Gone Bad," a team of three undergraduate students from the Virginia Tech College of Engineering, will compete in the World Finals of the Association of Computing Machinery International Collegiate Programming Contest (ACM-ICPC) March 12-16 in Tokyo, Japan.

  19. Computer Program Re-layers Engineering Drawings (United States)

    Crosby, Dewey C., III


    RULCHK computer program aids in structuring layers of information pertaining to part or assembly designed with software described in article "Software for Drawing Design Details Concurrently" (MFS-28444). Checks and optionally updates structure of layers for part. Enables designer to construct model and annotate its documentation without burden of manually layering part to conform to standards at design time.

  20. Data systems and computer science programs: Overview (United States)

    Smith, Paul H.; Hunter, Paul


    An external review of the Integrated Technology Plan for the Civil Space Program is presented. The topics are presented in viewgraph form and include the following: onboard memory and storage technology; advanced flight computers; special purpose flight processors; onboard networking and testbeds; information archive, access, and retrieval; visualization; neural networks; software engineering; and flight control and operations.

  1. Employing subgoals in computer programming education (United States)

    Margulieux, Lauren E.; Catrambone, Richard; Guzdial, Mark


    The rapid integration of technology into our professional and personal lives has left many education systems ill-equipped to deal with the influx of people seeking computing education. To improve computing education, we are applying techniques that have been developed for other procedural fields. The present study applied such a technique, subgoal labeled worked examples, to explore whether it would improve programming instruction. The first two experiments, conducted in a laboratory, suggest that the intervention improves undergraduate learners' problem-solving performance and affects how learners approach problem-solving. The third experiment demonstrates that the intervention has similar, and perhaps stronger, effects in an online learning environment with in-service K-12 teachers who want to become qualified to teach computing courses. By implementing this subgoal intervention as a tool for educators to teach themselves and their students, education systems could improve computing education and better prepare learners for an increasingly technical world.

  2. Computer Presentation Programs and Teaching Research Methodologies

    Directory of Open Access Journals (Sweden)

    Vahid Motamedi


    Full Text Available Supplementing traditional chalk and board instruction with computer delivery has been viewed positively by students who have reported increased understanding and more interaction with the instructor when computer presentations are used in the classroom. Some problems contributing to student errors while taking class notes might be transcription of numbers to the board, and handwriting of the instructor can be resolved in careful construction of computer presentations. The use of computer presentation programs promises to increase the effectiveness of learning by making content more readily available, by reducing the cost and effort of producing quality content, and by allowing content to be more easily shared. This paper describes how problems can be overcome by using presentation packages for instruction.

  3. Foresters' Metric Conversions program (version 1.0). [Computer program (United States)

    Jefferson A. Palmer


    The conversion of scientific measurements has become commonplace in the fields of - engineering, research, and forestry. Foresters? Metric Conversions is a Windows-based computer program that quickly converts user-defined measurements from English to metric and from metric to English. Foresters? Metric Conversions was derived from the publication "Metric...

  4. GAP: A computer program for gene assembly

    Energy Technology Data Exchange (ETDEWEB)

    Eisnstein, J.R.; Uberbacher, E.C.; Guan, X.; Mural, R.J.; Mann, R.C.


    A computer program, GAP (Gene Assembly Program), has been written to assemble and score hypothetical genes, given a DNA sequence containing the gene, and the outputs of several other programs which analyze the sequence. These programs include the codign-recognition and splice-junction-recognition modules developed in this laboratory. GAP is a prototype of a planned system in which it will be integrated with an expert system and rule base. Initial tests of GAP have been carried out with four sequences, the exons of which have been determined by biochemcial methods. The highest-scoring hypothetical genes for each of the four sequences had percent correct splice junctions ranging from 50 to 100% (average 81%) and percent correct bases ranging from 92 to 100% (average 96%). 9 refs., 1 tab.

  5. Heat Exchanger Ranking Program Using Genetic Algorithm and ε-NTU Method for Optimal Design

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Soon Ho; Kim, Minsung; Ha, Man Yeong; Park, Sang-Hu; Min, June Kee [Pusan National University, Busan (Korea, Republic of)


    Today, computational fluid dynamics (CFD) is widely used in industry because of the availability of high performance computers. However, full-scale analysis poses problems owing to the limited resources and time. In this study, the performance and optimal size of a heat exchanger were calculated using the effectiveness-number of transfer units (ε-NTU) method and a database of characteristics heat exchanger. Information about the geometry and performance of various heat exchangers is collected, and the performance of the heat exchanger is calculated under the given operating conditions. To determine the optimal size of the heat exchanger, a Genetic Algorithm (GA) is used, and MATLAB and REFPROP are used for the calculation.

  6. The computational physics program of the National MFE Computer Center

    Energy Technology Data Exchange (ETDEWEB)

    Mirin, A.A.


    The principal objective of the Computational Physics Group is to develop advanced numerical models for the investigation of plasma phenomena and the simulation of present and future magnetic confinement devices. Another major objective of the group is to develop efficient algorithms and programming techniques for current and future generation of supercomputers. The computational physics group is involved in several areas of fusion research. One main area is the application of Fokker-Planck/quasilinear codes to tokamaks. Another major area is the investigation of resistive magnetohydrodynamics in three dimensions, with applications to compact toroids. Another major area is the investigation of kinetic instabilities using a 3-D particle code. This work is often coupled with the task of numerically generating equilibria which model experimental devices. Ways to apply statistical closure approximations to study tokamak-edge plasma turbulence are being examined. In addition to these computational physics studies, the group has developed a number of linear systems solvers for general classes of physics problems and has been making a major effort at ascertaining how to efficiently utilize multiprocessor computers.

  7. PageRank (II): Mathematics

    African Journals Online (AJOL)


    INTRODUCTION. PageRank is Google's system for ranking web pages. A page with a higher PageRank is deemed more important and is more likely to be listed above a ... Felix U. Ogban, Department of Mathematics/Statistics and Computer Science, Faculty of Science, University of ..... probability, 2004, 41, (3): 721-734.


    Directory of Open Access Journals (Sweden)

    José Manuel Cabrera Delgado


    Full Text Available This article aims to get an overview of the process of including the computer programming and the robotics in the educational curriculum of the basic education in several European countries, including Spain. For this purpose, the cases of Estonia and France are briefly analyzed, two countries in the European Union, which can be considered pioneers in implementing such teaching. Also, in relation to Spain, it is analyzed some of the current initiatives implemented by some Autonomous Communities in this sense.





    As most of the global economic activity takes place in the form of projects/programs, their effective management and governance is becoming more and more critical to the competitive position of organizations and societies. Project auditing and risk management, elaborate on methodologies that could be used for analyzing project progress, by identifying potential risks and liabilities, and finally recommending corrective and preventive actions. In relation to these fields, this paper proposes a...

  10. Scientific Computing in the CH Programming Language

    Directory of Open Access Journals (Sweden)

    Harry H. Cheng


    Full Text Available We have developed a general-purpose block-structured interpretive programming Ianguage. The syntax and semantics of this language called CH are similar to C. CH retains most features of C from the scientific computing point of view. In this paper, the extension of C to CH for numerical computation of real numbers will be described. Metanumbers of −0.0, 0.0, Inf, −Inf, and NaN are introduced in CH. Through these metanumbers, the power of the IEEE 754 arithmetic standard is easily available to the programmer. These metanumbers are extended to commonly used mathematical functions in the spirit of the IEEE 754 standard and ANSI C. The definitions for manipulation of these metanumbers in I/O; arithmetic, relational, and logic operations; and built-in polymorphic mathematical functions are defined. The capabilities of bitwise, assignment, address and indirection, increment and decrement, as well as type conversion operations in ANSI C are extended in CH. In this paper, mainly new linguistic features of CH in comparison to C will be described. Example programs programmed in CH with metanumbers and polymorphic mathematical functions will demonstrate capabilities of CH in scientific computing.

  11. NASA High Performance Computing and Communications program (United States)

    Holcomb, Lee; Smith, Paul; Hunter, Paul


    The National Aeronautics and Space Administration's HPCC program is part of a new Presidential initiative aimed at producing a 1000-fold increase in supercomputing speed and a 1(X)-fold improvement in available communications capability by 1997. As more advanced technologies are developed under the HPCC program, they will be used to solve NASA's 'Grand Challenge' problems, which include improving the design and simulation of advanced aerospace vehicles, allowing people at remote locations to communicate more effectively and share information, increasing scientists' abilities to model the Earth's climate and forecast global environmental trends, and improving the development of advanced spacecraft. NASA's HPCC program is organized into three projects which are unique to the agency's mission: the Computational Aerosciences (CAS) project, the Earth and Space Sciences (ESS) project, and the Remote Exploration and Experimentation (REE) project. An additional project, the Basic Research and Human Resources (BRHR) project, exists to promote long term research in computer science and engineering and to increase the pool of trained personnel in a variety of scientific disciplines. This document presents an overview of the objectives and organization of these projects, as well as summaries of early accomplishments and the significance, status, and plans for individual research and development programs within each project. Areas of emphasis include benchmarking, testbeds, software and simulation methods.

  12. User's manual for computer program BASEPLOT (United States)

    Sanders, Curtis L.


    The checking and reviewing of daily records of streamflow within the U.S. Geological Survey is traditionally accomplished by hand-plotting and mentally collating tables of data. The process is time consuming, difficult to standardize, and subject to errors in computation, data entry, and logic. In addition, the presentation of flow data on the internet requires more timely and accurate computation of daily flow records. BASEPLOT was developed for checking and review of primary streamflow records within the U.S. Geological Survey. Use of BASEPLOT enables users to (1) provide efficiencies during the record checking and review process, (2) improve quality control, (3) achieve uniformity of checking and review techniques of simple stage-discharge relations, and (4) provide a tool for teaching streamflow computation techniques. The BASEPLOT program produces tables of quality control checks and produces plots of rating curves and discharge measurements; variable shift (V-shift) diagrams; and V-shifts converted to stage-discharge plots, using data stored in the U.S. Geological Survey Automatic Data Processing System database. In addition, the program plots unit-value hydrographs that show unit-value stages, shifts, and datum corrections; input shifts, datum corrections, and effective dates; discharge measurements; effective dates for rating tables; and numeric quality control checks. Checklist/tutorial forms are provided for reviewers to ensure completeness of review and standardize the review process. The program was written for the U.S. Geological Survey SUN computer using the Statistical Analysis System (SAS) software produced by SAS Institute, Incorporated.

  13. First sale doctrine on computer programs in the European Union. reiteration of the precedent or jurisprudential change?

    Directory of Open Access Journals (Sweden)

    Sergio Raúl Cardoso González


    Full Text Available Comparative analysis of rulings C-128/11 (UsedSoft of July 3, 2012 and C-166/15 (Ranks and Vasiļevičs of October 12, 2016 by the Court of Justice of the European Union, concerning first sale doctrine on computer programs. Reiteration of the precedent or jurisprudential change?

  14. A research program in empirical computer science (United States)

    Knight, J. C.


    During the grant reporting period our primary activities have been to begin preparation for the establishment of a research program in experimental computer science. The focus of research in this program will be safety-critical systems. Many questions that arise in the effort to improve software dependability can only be addressed empirically. For example, there is no way to predict the performance of the various proposed approaches to building fault-tolerant software. Performance models, though valuable, are parameterized and cannot be used to make quantitative predictions without experimental determination of underlying distributions. In the past, experimentation has been able to shed some light on the practical benefits and limitations of software fault tolerance. It is common, also, for experimentation to reveal new questions or new aspects of problems that were previously unknown. A good example is the Consistent Comparison Problem that was revealed by experimentation and subsequently studied in depth. The result was a clear understanding of a previously unknown problem with software fault tolerance. The purpose of a research program in empirical computer science is to perform controlled experiments in the area of real-time, embedded control systems. The goal of the various experiments will be to determine better approaches to the construction of the software for computing systems that have to be relied upon. As such it will validate research concepts from other sources, provide new research results, and facilitate the transition of research results from concepts to practical procedures that can be applied with low risk to NASA flight projects. The target of experimentation will be the production software development activities undertaken by any organization prepared to contribute to the research program. Experimental goals, procedures, data analysis and result reporting will be performed for the most part by the University of Virginia.

  15. Gender Differences in the Use of Computers, Programming, and Peer Interactions in Computer Science Classrooms (United States)

    Stoilescu, Dorian; Egodawatte, Gunawardena


    Research shows that female and male students in undergraduate computer science programs view computer culture differently. Female students are interested more in the use of computers than in doing programming, whereas male students see computer science mainly as a programming activity. The overall purpose of our research was not to find new…

  16. STEW A Nonlinear Data Modeling Computer Program

    CERN Document Server

    Chen, H


    A nonlinear data modeling computer program, STEW, employing the Levenberg-Marquardt algorithm, has been developed to model the experimental sup 2 sup 3 sup 9 Pu(n,f) and sup 2 sup 3 sup 5 U(n,f) cross sections. This report presents results of the modeling of the sup 2 sup 3 sup 9 Pu(n,f) and sup 2 sup 3 sup 5 U(n,f) cross-section data. The calculation of the fission transmission coefficient is based on the double-humped-fission-barrier model of Bjornholm and Lynn. Incident neutron energies of up to 5 MeV are considered.

  17. Computer programming: Science, art, or both? (United States)

    Gum, Sandra Trent

    The purpose of this study was to determine if spatial intelligence contributes to a student's success in a computer science major or if mathematical-logical intelligence is sufficient data on which to base a prediction of success. The study was performed at a small university. The sample consisted of 15 computer science (CS) majors, enrolled in a computer science class, and 15 non-CS-majors, enrolled in a statistics class. Seven of the CS-majors were considered advanced and seven were considered less advanced. The independent measures were: the mathematics and the English scores from the ACT/SAT (CS-majors); a questionnaire to obtain personal information; the major area of study which compared CS-majors to all other majors; and the number of completed computer science classes (CS-majors) to determine advanced and less advanced CS-majors. The dependent measures were: a multiple intelligence inventory for adults to determine perception of intelligences; the GEFT to determine field independence independence; the Card Rotations Test to determine spatial orientation ability; the Maze Tracing Speed Test to determine spatial scanning ability; and the Surface Development test to determine visualization ability. The visualization measure correlated positively and significantly with the GEFT. The year in college correlated positively and significantly with the GEFT and visualization measure for CS-majors and correlated negatively for non-CS-majors. Although non-CS-majors scored higher on the spatial orientation measure, CS-majors scored significantly higher on the spatial scanning measure. The year in college correlated negatively with many of the measures and perceptions of intelligences among both groups; however, there were more significant negative correlations among non-CS-majors. Results indicated that experience in computer programming may increase field independence, visualization ability, and spatial scanning ability while decreasing spatial orientation ability. The

  18. H Scan/AHP decision process planning for evaluating and ranking of proposals submitted to the DOE hydrogen program

    Energy Technology Data Exchange (ETDEWEB)

    Szoka de Valladares, M.R. [National Renwable Energy Lab., Washington, DC (United States); Mack, S. [Energtics, Washington, DC (United States)


    The DOE Hydrogen Program needs to develop criteria as part of a systematic evaluation process for proposal identification, evaluation and selection. The H Scan component of this process provides a framework in which a project proposer can fully describe their candidate technology system and its components. The H Scan complements traditional methods of capturing cost and technical information. It consists of a special set of survey forms designed to elicit information so expert reviewers can assess the proposal relative to DOE specified selection criteria. The Analytic Hierarchy Process (AHP) component of the decision process assembles the management defined evaluation and selection criteria into a coherent multi-level decision construct by which projects can be evaluated in pair-wise comparisons. The AHP model will reflect management`s objectives and it will assist in the ranking of individual projects based on the extent to which each contributes to management`s objectives. This paper contains a detailed description of the products and activities associated with the planning and evaluation process: The objectives or criteria; the H Scan; and The Analytic Hierarchy Process (AHP).

  19. Can College Rankings Be Believed?

    Directory of Open Access Journals (Sweden)

    Meredith Davis

    Full Text Available The article summarizes literature on college and university rankings worldwide and the strategies used by various ranking organizations, including those of government and popular media. It traces the history of national and global rankings, indicators used by ranking systems, and the effect of rankings on academic programs and their institutions. Although ranking systems employ diverse criteria and most weight certain indicators over others, there is considerable skepticism that most actually measure educational quality. At the same time, students and their families increasingly consult these evaluations when making college decisions, and sponsors of faculty research consider reputation when forming academic partnerships. While there are serious concerns regarding the validity of ranking institutions when so little data can support differences between one institution and another, college rankings appear to be here to stay.

  20. Computing Low-Rank Approximation of a Dense Matrix on Multicore CPUs with a GPU and Its Application to Solving a Hierarchically Semiseparable Linear System of Equations

    Directory of Open Access Journals (Sweden)

    Ichitaro Yamazaki


    of their low-rank properties. To compute a low-rank approximation of a dense matrix, in this paper, we study the performance of QR factorization with column pivoting or with restricted pivoting on multicore CPUs with a GPU. We first propose several techniques to reduce the postprocessing time, which is required for restricted pivoting, on a modern CPU. We then examine the potential of using a GPU to accelerate the factorization process with both column and restricted pivoting. Our performance results on two eight-core Intel Sandy Bridge CPUs with one NVIDIA Kepler GPU demonstrate that using the GPU, the factorization time can be reduced by a factor of more than two. In addition, to study the performance of our implementations in practice, we integrate them into a recently developed software StruMF which algebraically exploits such low-rank structures for solving a general sparse linear system of equations. Our performance results for solving Poisson's equations demonstrate that the proposed techniques can significantly reduce the preconditioner construction time of StruMF on the CPUs, and the construction time can be further reduced by 10%–50% using the GPU.

  1. Human operator identification model and related computer programs (United States)

    Kessler, K. M.; Mohr, J. N.


    Four computer programs which provide computational assistance in the analysis of man/machine systems are reported. The programs are: (1) Modified Transfer Function Program (TF); (2) Time Varying Response Program (TVSR); (3) Optimal Simulation Program (TVOPT); and (4) Linear Identification Program (SCIDNT). The TV program converts the time domain state variable system representative to frequency domain transfer function system representation. The TVSR program computes time histories of the input/output responses of the human operator model. The TVOPT program is an optimal simulation program and is similar to TVSR in that it produces time histories of system states associated with an operator in the loop system. The differences between the two programs are presented. The SCIDNT program is an open loop identification code which operates on the simulated data from TVOPT (or TVSR) or real operator data from motion simulators.

  2. Experiences with Efficient Methodologies for Teaching Computer Programming to Geoscientists (United States)

    Jacobs, Christian T.; Gorman, Gerard J.; Rees, Huw E.; Craig, Lorraine E.


    Computer programming was once thought of as a skill required only by professional software developers. But today, given the ubiquitous nature of computation and data science it is quickly becoming necessary for all scientists and engineers to have at least a basic knowledge of how to program. Teaching how to program, particularly to those students…

  3. Computer Programs for Characteristic Modes of Bodies of Revolution (United States)

    Computer programs are given for calculating the characteristic currents and characteristic gain patterns of conducting bodies of revolution. Also...given are computer programs for using these characteristic currents in aperture radiation and plane-wave scattering problems. Plot programs for use with

  4. 01010000 01001100 01000001 01011001: Play Elements in Computer Programming (United States)

    Breslin, Samantha


    This article explores the role of play in human interaction with computers in the context of computer programming. The author considers many facets of programming including the literary practice of coding, the abstract design of programs, and more mundane activities such as testing, debugging, and hacking. She discusses how these incorporate the…

  5. ROUTES: a computer program for preliminary route location. (United States)

    S.E. Reutebuch


    An analytical description of the ROUTES computer program is presented. ROUTES is part of the integrated preliminary harvest- and transportation-planning software package, PLANS. The ROUTES computer program is useful where grade and sideslope limitations are important in determining routes for vehicular travel. With the program, planners can rapidly identify route...

  6. From rankings to mission. (United States)

    Kirch, Darrell G; Prescott, John E


    Since the 1980s, school ranking systems have been a topic of discussion among leaders of higher education. Various ranking systems are based on inadequate data that fail to illustrate the complex nature and special contributions of the institutions they purport to rank, including U.S. medical schools, each of which contributes uniquely to meeting national health care needs. A study by Tancredi and colleagues in this issue of Academic Medicine illustrates the limitations of rankings specific to primary care training programs. This commentary discusses, first, how each school's mission and strengths, as well as the impact it has on the community it serves, are distinct, and, second, how these schools, which are each unique, are poorly represented by overly subjective ranking methodologies. Because academic leaders need data that are more objective to guide institutional development, the Association of American Medical Colleges (AAMC) has been developing tools to provide valid data that are applicable to each medical school. Specifically, the AAMC's Medical School Admissions Requirements and its Missions Management Tool each provide a comprehensive assessment of medical schools that leaders are using to drive institutional capacity building. This commentary affirms the importance of mission while challenging the leaders of medical schools, teaching hospitals, and universities to use reliable data to continually improve the quality of their training programs to improve the health of all.

  7. Programs=data=first-class citizens in a computational world

    DEFF Research Database (Denmark)

    Jones, Neil; Simonsen, Jakob Grue


    From a programming perspective, Alan Turing's epochal 1936 paper on computable functions introduced several new concepts, including what is today known as self-interpreters and programs as data, and invented a great many now-common programming techniques. We begin by reviewing Turing's contribution...... from a programming perspective; and then systematize and mention some of the many ways that later developments in models of computation (MOCs) have interacted with computability theory and programming language research. Next, we describe the ‘blob’ MOC: a recent stored-program computational model...... without pointers. In the blob model, programs are truly first-class citizens, capable of being automatically compiled, or interpreted, or executed directly. Further, the blob model appears closer to being physically realizable than earlier computation models. In part, this is due to strong finiteness...

  8. Dynamic Matrix Rank

    DEFF Research Database (Denmark)

    Frandsen, Gudmund Skovbjerg; Frandsen, Peter Frands


    We consider maintaining information about the rank of a matrix under changes of the entries. For n×n matrices, we show an upper bound of O(n1.575) arithmetic operations and a lower bound of Ω(n) arithmetic operations per element change. The upper bound is valid when changing up to O(n0.575) entries...... in a single column of the matrix. We also give an algorithm that maintains the rank using O(n2) arithmetic operations per rank one update. These bounds appear to be the first nontrivial bounds for the problem. The upper bounds are valid for arbitrary fields, whereas the lower bound is valid for algebraically...... closed fields. The upper bound for element updates uses fast rectangular matrix multiplication, and the lower bound involves further development of an earlier technique for proving lower bounds for dynamic computation of rational functions....

  9. Quantum probability ranking principle for ligand-based virtual screening. (United States)

    Al-Dabbagh, Mohammed Mumtaz; Salim, Naomie; Himmat, Mubarak; Ahmed, Ali; Saeed, Faisal


    Chemical libraries contain thousands of compounds that need screening, which increases the need for computational methods that can rank or prioritize compounds. The tools of virtual screening are widely exploited to enhance the cost effectiveness of lead drug discovery programs by ranking chemical compounds databases in decreasing probability of biological activity based upon probability ranking principle (PRP). In this paper, we developed a novel ranking approach for molecular compounds inspired by quantum mechanics, called quantum probability ranking principle (QPRP). The QPRP ranking criteria would make an attempt to draw an analogy between the physical experiment and molecular structure ranking process for 2D fingerprints in ligand based virtual screening (LBVS). The development of QPRP criteria in LBVS has employed the concepts of quantum at three different levels, firstly at representation level, this model makes an effort to develop a new framework of molecular representation by connecting the molecular compounds with mathematical quantum space. Secondly, estimate the similarity between chemical libraries and references based on quantum-based similarity searching method. Finally, rank the molecules using QPRP approach. Simulated virtual screening experiments with MDL drug data report (MDDR) data sets showed that QPRP outperformed the classical ranking principle (PRP) for molecular chemical compounds.

  10. Quantum probability ranking principle for ligand-based virtual screening (United States)

    Al-Dabbagh, Mohammed Mumtaz; Salim, Naomie; Himmat, Mubarak; Ahmed, Ali; Saeed, Faisal


    Chemical libraries contain thousands of compounds that need screening, which increases the need for computational methods that can rank or prioritize compounds. The tools of virtual screening are widely exploited to enhance the cost effectiveness of lead drug discovery programs by ranking chemical compounds databases in decreasing probability of biological activity based upon probability ranking principle (PRP). In this paper, we developed a novel ranking approach for molecular compounds inspired by quantum mechanics, called quantum probability ranking principle (QPRP). The QPRP ranking criteria would make an attempt to draw an analogy between the physical experiment and molecular structure ranking process for 2D fingerprints in ligand based virtual screening (LBVS). The development of QPRP criteria in LBVS has employed the concepts of quantum at three different levels, firstly at representation level, this model makes an effort to develop a new framework of molecular representation by connecting the molecular compounds with mathematical quantum space. Secondly, estimate the similarity between chemical libraries and references based on quantum-based similarity searching method. Finally, rank the molecules using QPRP approach. Simulated virtual screening experiments with MDL drug data report (MDDR) data sets showed that QPRP outperformed the classical ranking principle (PRP) for molecular chemical compounds.

  11. Stochastic linear programming models, theory, and computation

    CERN Document Server

    Kall, Peter


    This new edition of Stochastic Linear Programming: Models, Theory and Computation has been brought completely up to date, either dealing with or at least referring to new material on models and methods, including DEA with stochastic outputs modeled via constraints on special risk functions (generalizing chance constraints, ICC’s and CVaR constraints), material on Sharpe-ratio, and Asset Liability Management models involving CVaR in a multi-stage setup. To facilitate use as a text, exercises are included throughout the book, and web access is provided to a student version of the authors’ SLP-IOR software. Additionally, the authors have updated the Guide to Available Software, and they have included newer algorithms and modeling systems for SLP. The book is thus suitable as a text for advanced courses in stochastic optimization, and as a reference to the field. From Reviews of the First Edition: "The book presents a comprehensive study of stochastic linear optimization problems and their applications. … T...

  12. Macroevolution simulated with autonomously replicating computer programs. (United States)

    Yedid, Gabriel; Bell, Graham

    The process of adaptation occurs on two timescales. In the short term, natural selection merely sorts the variation already present in a population, whereas in the longer term genotypes quite different from any that were initially present evolve through the cumulation of new mutations. The first process is described by the mathematical theory of population genetics. However, this theory begins by defining a fixed set of genotypes and cannot provide a satisfactory analysis of the second process because it does not permit any genuinely new type to arise. The evolutionary outcome of selection acting on novel variation arising over long periods is therefore difficult to predict. The classical problem of this kind is whether 'replaying the tape of life' would invariably lead to the familiar organisms of the modern biota. Here we study the long-term behaviour of populations of autonomously replicating computer programs and find that the same type, introduced into the same simple environment, evolves on any given occasion along a unique trajectory towards one of many well-adapted end points.

  13. Basis And Application Of The CARES/LIFE Computer Program (United States)

    Nemeth, Noel N.; Janosik, Lesley A.; Gyekenyesi, John P.; Powers, Lynn M.


    Report discusses physical and mathematical basis of Ceramics Analysis and Reliability Evaluation of Structures LIFE prediction (CARES/LIFE) computer program, described in "Program for Evaluation of Reliability of Ceramic Parts" (LEW-16018).

  14. Positioning Continuing Education Computer Programs for the Corporate Market. (United States)

    Tilney, Ceil


    Summarizes the findings of the market assessment phase of Bellevue Community College's evaluation of its continuing education computer training program. Indicates that marketing efforts must stress program quality and software training to help overcome strong antiacademic client sentiment. (MGB)

  15. Educational ranking of schools providing nutrition program in Iran: A strategy to identify the strengths and weaknesses

    Directory of Open Access Journals (Sweden)

    R Mojtahedzadeh


    Full Text Available Introduction: After two decades of expansion of universities of medical sciences and increasing the number of students as one of the most attainable solutions to overcome sub-standard status of Iranian workforce in health section, quality of medical education has recently taken priority in national health plans. The purpose of this study was designing a benchmark tool for determining differences in the field of education among schools offering nutrition programs in Iran by stratifying their educational services. Methods: To measure the educational performance of schools providing nutrition programs nationwide, a benchmarking tool consisting of more than 50 indicators was developed by a special committee via brainstorming and nominal group. The weight given to each indicator was defined through consensus developing methods. Each school was asked to introduce a representative who would complete a questionnaire designed to collect schools’ information. Next, all the divisions were visited by one of the project's members and the school’s representative. Then, data retrieval was performed and verified at the project’s office. The average score of each school for each indicator was calculated. Finally, a special computer software was applied to perform the final analysis. Results: A total of 50 indicators were designed in the form of tree diagram and were weighted via consensus agreement. The medical universities offering nutritional programs were compared using the designed model. Conclusion: This study defined the educational strengths and weaknesses of nutrition program in Iranian schools that could be used as a measure for authorities to devise practical strategies for qualitative and quantitative improvement in medical education.

  16. Tool for Ranking Research Options (United States)

    Ortiz, James N.; Scott, Kelly; Smith, Harold


    Tool for Research Enhancement Decision Support (TREDS) is a computer program developed to assist managers in ranking options for research aboard the International Space Station (ISS). It could likely also be adapted to perform similar decision-support functions in industrial and academic settings. TREDS provides a ranking of the options, based on a quantifiable assessment of all the relevant programmatic decision factors of benefit, cost, and risk. The computation of the benefit for each option is based on a figure of merit (FOM) for ISS research capacity that incorporates both quantitative and qualitative inputs. Qualitative inputs are gathered and partly quantified by use of the time-tested analytical hierarchical process and used to set weighting factors in the FOM corresponding to priorities determined by the cognizant decision maker(s). Then by use of algorithms developed specifically for this application, TREDS adjusts the projected benefit for each option on the basis of levels of technical implementation, cost, and schedule risk. Based partly on Excel spreadsheets, TREDS provides screens for entering cost, benefit, and risk information. Drop-down boxes are provided for entry of qualitative information. TREDS produces graphical output in multiple formats that can be tailored by users.

  17. The Outlook for Computer Professions: 1985 Rewrites the Program. (United States)

    Drake, Larry


    The author states that graduates of junior college programs who learn COBOL will continue to find jobs, but employers will increasingly seek college graduates when filling positions for computer programers and systems analysts. Areas of growth for computer applications (services, military, data communications, and artificial intelligence) are…

  18. Basic BASIC; An Introduction to Computer Programming in BASIC Language. (United States)

    Coan, James S.

    With the increasing availability of computer access through remote terminals and time sharing, more and more schools and colleges are able to introduce programing to substantial numbers of students. This book is an attempt to incorporate computer programming, using BASIC language, and the teaching of mathematics. The general approach of the book…

  19. A Computer Program for Short Circuit Analysis of Electric Power ...

    African Journals Online (AJOL)

    This paper described the mathematical basis and computational framework of a computer program developed for short circuit studies of electric power systems. The Short Circuit Analysis Program (SCAP) is to be used to assess the composite effects of unbalanced and balanced faults on the overall reliability of electric ...

  20. Software survey: VOSviewer, a computer program for bibliometric mapping

    NARCIS (Netherlands)

    N.J.P. van Eck (Nees Jan); L. Waltman (Ludo)


    textabstractWe present VOSviewer, a freely available computer program that we have developed for constructing and viewing bibliometric maps. Unlike most computer programs that are used for bibliometric mapping, VOSviewer pays special attention to the graphical representation of bibliometric maps.

  1. Near-Surface Seismic Velocity Data: A Computer Program For ...

    African Journals Online (AJOL)

    A computer program (NESURVELANA) has been developed in Visual Basic Computer programming language to carry out a near surface velocity analysis. The method of analysis used includes: Algorithms design and Visual Basic codes generation for plotting arrival time (ms) against geophone depth (m) employing the ...

  2. Case Studies of Liberal Arts Computer Science Programs (United States)

    Baldwin, D.; Brady, A.; Danyluk, A.; Adams, J.; Lawrence, A.


    Many undergraduate liberal arts institutions offer computer science majors. This article illustrates how quality computer science programs can be realized in a wide variety of liberal arts settings by describing and contrasting the actual programs at five liberal arts colleges: Williams College, Kalamazoo College, the State University of New York…

  3. Programming language for computations in the Interkosmos program (United States)

    Schmidt, K.


    The programming system for Intercosmos data processing, based on the structural programming theory, which considers a program as an ordered set of standardized elementary parts, from which the user programs are automatically generated, is described. The programs are comprised of several modules, which are briefly summarized. The general structure of the programming system is presented in a block diagram. A programming control language developed to formulate the problem quickly and completely is presented along with basic symbols which are characteristic of the Intercosmos programming system.

  4. Method and computer program product for maintenance and modernization backlogging (United States)

    Mattimore, Bernard G; Reynolds, Paul E; Farrell, Jill M


    According to one embodiment, a computer program product for determining future facility conditions includes a computer readable medium having computer readable program code stored therein. The computer readable program code includes computer readable program code for calculating a time period specific maintenance cost, for calculating a time period specific modernization factor, and for calculating a time period specific backlog factor. Future facility conditions equal the time period specific maintenance cost plus the time period specific modernization factor plus the time period specific backlog factor. In another embodiment, a computer-implemented method for calculating future facility conditions includes calculating a time period specific maintenance cost, calculating a time period specific modernization factor, and calculating a time period specific backlog factor. Future facility conditions equal the time period specific maintenance cost plus the time period specific modernization factor plus the time period specific backlog factor. Other embodiments are also presented.

  5. An Experimental Investigation of Computer Program Development Approaches and Computer Programming Metrics. (United States)


    Approved for plbi elaSO; 81 1ibutio u0lim t 03 /1 Technical Report TR-853 December 1979 An Experimental Investigation of Computer Program Development...this Report) Approved for public release; distribution unlimited. 17. DISTRIBUTION STATEMENT (of rhe entled i f [31-k 20, ! dil(enot fo, ).R...55t -3 O 2 a. tL loco o.uaZ 4 1 0 0 - .. a .CO acca 4 * ao -- c- - .an za-e~a- 2.5 aa34z98ba a- CHAPTER VII coipletely differentiated outcome is

  6. Designing Educational Games for Computer Programming: A Holistic Framework (United States)

    Malliarakis, Christos; Satratzemi, Maya; Xinogalos, Stelios


    Computer science is continuously evolving during the past decades. This has also brought forth new knowledge that should be incorporated and new learning strategies must be adopted for the successful teaching of all sub-domains. For example, computer programming is a vital knowledge area within computer science with constantly changing curriculum…

  7. Seventy Years of Computing in the Nuclear Weapons Program

    Energy Technology Data Exchange (ETDEWEB)

    Archer, Billy Joe [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)


    Los Alamos has continuously been on the forefront of scientific computing since it helped found the field. This talk will explore the rich history of computing in the Los Alamos weapons program. The current status of computing will be discussed, as will the expectations for the near future.

  8. A Methodology for Teaching Computer Programming: first year students’ perspective


    Bassey Isong


    The teaching of computer programming is one of the greatest challenges that have remained for years in Computer Science Education. A particular case is computer programming course for the beginners. While the traditional objectivist lecture-based approaches do not actively engage students to achieve their learning outcome, we believe that integrating some cutting-edge processes and practices like agile method into the teaching approaches will be leverage. Agile software development has gained...

  9. Attitude, Gender and Achievement in Computer Programming (United States)

    Baser, Mustafa


    The aim of this research was to explore the relationship among students' attitudes toward programming, gender and academic achievement in programming. The scale used for measuring students' attitudes toward programming was developed by the researcher and consisted of 35 five-point Likert type items in four subscales. The scale was administered to…

  10. Efficient Top-k Search for PageRank

    National Research Council Canada - National Science Library

    Fujiwara, Yasuhiro; Nakatsuji, Makoto; Shiokawa, Hiroaki; Mishima, Takeshi; Onizuka, Makoto


      In AI communities, many applications utilize PageRank. To obtain high PageRank score nodes, the original approach iteratively computes the PageRank score of each node until convergence from the whole graph...

  11. Generic Assessment Rubrics for Computer Programming Courses (United States)

    Mustapha, Aida; Samsudin, Noor Azah; Arbaiy, Nurieze; Mohammed, Rozlini; Hamid, Isredza Rahmi


    In programming, one problem can usually be solved using different logics and constructs but still producing the same output. Sometimes students get marked down inappropriately if their solutions do not follow the answer scheme. In addition, lab exercises and programming assignments are not necessary graded by the instructors but most of the time…

  12. Instructional Uses of the Computer: Program Force (United States)

    Ostrander, P.


    Describes a program which simulates motion in two dimensions of a point mass subject to a force which is a function of position, velocity, or time. Sample applications are noted and a source of a complete list of applications and programs is given. (GH)

  13. Introduction of handheld computing to a family practice residency program. (United States)

    Rao, Goutham


    Handheld computers are valuable practice tools. It is important for residency programs to introduce their trainees and faculty to this technology. This article describes a formal strategy to introduce handheld computing to a family practice residency program. Objectives were selected for the handheld computer training program that reflected skills physicians would find useful in practice. TRGpro handheld computers preloaded with a suite of medical reference programs, a medical calculator, and a database program were supplied to participants. Training consisted of four 1-hour modules each with a written evaluation quiz. Participants completed a self-assessment questionnaire after the program to determine their ability to meet each objective. Sixty of the 62 participants successfully completed the training program. The mean composite score on quizzes was 36 of 40 (90%), with no significant differences by level of residency training. The mean self-ratings of participants across all objectives was 3.31 of 4.00. Third-year residents had higher mean self-ratings than others (mean of group, 3.62). Participants were very comfortable with practical skills, such as using drug reference software, and less comfortable with theory, such as knowing the different types of handheld computers available. Structured training is a successful strategy for introducing handheld computing to a residency program.

  14. Teacher Training Programs for Computer Education and Computer Assisted Education in Turkey (United States)

    Usun, Salih


    The aim of this descriptive study is to review the applications and problems on the teacher training programs for computer education and computer assisted education (CAE) in Turkey. The study, firstly, introduces some applications and major problems on using instructional media and computers in developing countries and instructional technology…

  15. A computer program for planimetric analysis of digitized images

    DEFF Research Database (Denmark)

    Lynnerup, N; Lynnerup, O; Homøe, P


    bones as seen on X-rays. By placing the X-rays on a digitizer tablet and tracing the outline of the cell system, the area was calculated by the program. The calculated data and traced images could be stored and printed. The program is written in BASIC; necessary hardware is an IBM-compatible personal......Planimetrical measurements are made to calculate the area of an entity. By digitizing the entity the planimetrical measurements may be done by computer. This computer program was developed in conjunction with a research project involving measurement of the pneumatized cell system of the temporal...... computer, a digitizer tablet and a printer....

  16. Ranking Support Vector Machine with Kernel Approximation. (United States)

    Chen, Kai; Li, Rongchun; Dou, Yong; Liang, Zhengfa; Lv, Qi


    Learning to rank algorithm has become important in recent years due to its successful application in information retrieval, recommender system, and computational biology, and so forth. Ranking support vector machine (RankSVM) is one of the state-of-art ranking models and has been favorably used. Nonlinear RankSVM (RankSVM with nonlinear kernels) can give higher accuracy than linear RankSVM (RankSVM with a linear kernel) for complex nonlinear ranking problem. However, the learning methods for nonlinear RankSVM are still time-consuming because of the calculation of kernel matrix. In this paper, we propose a fast ranking algorithm based on kernel approximation to avoid computing the kernel matrix. We explore two types of kernel approximation methods, namely, the Nyström method and random Fourier features. Primal truncated Newton method is used to optimize the pairwise L2-loss (squared Hinge-loss) objective function of the ranking model after the nonlinear kernel approximation. Experimental results demonstrate that our proposed method gets a much faster training speed than kernel RankSVM and achieves comparable or better performance over state-of-the-art ranking algorithms.

  17. Intelligent physical blocks for introducing computer programming in developing countries

    CSIR Research Space (South Africa)

    Smith, Adrew C


    Full Text Available This paper reports on the evaluation of a novel affordable system that incorporates intelligent physical blocks to introduce illiterate children in developing countries to the logical thinking process required in computer programming. Both...

  18. Advanced wellbore thermal simulator GEOTEMP2. Appendix. Computer program listing

    Energy Technology Data Exchange (ETDEWEB)

    Mitchell, R.F.


    This appendix gives the program listing of GEOTEMP2 with comments and discussion to make the program organization more understandable. This appendix is divided into an introduction and four main blocks of code: main program, program initiation, wellbore flow, and wellbore heat transfer. The purpose and use of each subprogram is discussed and the program listing is given. Flowcharts will be included to clarify code organization when needed. GEOTEMP2 was written in FORTRAN IV. Efforts have been made to keep the programing as conventional as possible so that GEOTEMP2 will run without modification on most computers.

  19. Architecture and Programming Models for High Performance Intensive Computation (United States)


    AFRL-AFOSR-VA-TR-2016-0230 Architecture and Programming Models for High Performance Intensive Computation XiaoMing Li UNIVERSITY OF DELAWARE Final...TITLE AND SUBTITLE Architecture and Programming Models for High Performance Intensive Computation 5a. CONTRACT NUMBER 5b. GRANT NUMBER FA9550-13-1-0213...developing an efficient system architecture and software tools for building and running Dynamic Data Driven Application Systems (DDDAS). The foremost

  20. Computing the Line Index of Balance Using Integer Programming Optimisation


    Aref, Samin; Andrew J. Mason; Wilson, Mark C.


    An important measure of a signed graph is the line index of balance which has several applications in many fields. However, this graph-theoretic measure was underused for decades because of the inherent complexity in its computation which is closely related to solving NP-hard graph optimisation problems like MAXCUT. We develop new quadratic and linear programming models to compute the line index of balance exactly. Using the Gurobi integer programming optimisation solver, we evaluate the line...

  1. Software survey: VOSviewer, a computer program for bibliometric mapping


    van Eck, Nees Jan; Waltman, Ludo


    textabstractWe present VOSviewer, a freely available computer program that we have developed for constructing and viewing bibliometric maps. Unlike most computer programs that are used for bibliometric mapping, VOSviewer pays special attention to the graphical representation of bibliometric maps. The functionality of VOSviewer is especially useful for displaying large bibliometric maps in an easy-to-interpret way. The paper consists of three parts. In the first part, an overview of VOSviewer'...

  2. Newnes circuit calculations pocket book with computer programs

    CERN Document Server

    Davies, Thomas J


    Newnes Circuit Calculations Pocket Book: With Computer Programs presents equations, examples, and problems in circuit calculations. The text includes 300 computer programs that help solve the problems presented. The book is comprised of 20 chapters that tackle different aspects of circuit calculation. The coverage of the text includes dc voltage, dc circuits, and network theorems. The book also covers oscillators, phasors, and transformers. The text will be useful to electrical engineers and other professionals whose work involves electronic circuitry.

  3. Computer program compatible with a laser nephelometer (United States)

    Paroskie, R. M.; Blau, H. H., Jr.; Blinn, J. C., III


    The laser nephelometer data system was updated to provide magnetic tape recording of data, and real time or near real time processing of data to provide particle size distribution and liquid water content. Digital circuits were provided to interface the laser nephelometer to a Data General Nova 1200 minicomputer. Communications are via a teletypewriter. A dual Linc Magnetic Tape System is used for program storage and data recording. Operational programs utilize the Data General Real-Time Operating System (RTOS) and the ERT AIRMAP Real-Time Operating System (ARTS). The programs provide for acquiring data from the laser nephelometer, acquiring data from auxiliary sources, keeping time, performing real time calculations, recording data and communicating with the teletypewriter.

  4. Contributions to computational stereology and parallel programming

    DEFF Research Database (Denmark)

    Rasmusson, Allan

    Stereology is the science of interpreting 3D structures from 2D sections planes and it is used in a multitude of disciplines including bioscience, material science and more. At its core is the use of random systematic sampling and geometrical probes which allow valid statistical inference...... between computer science and stereology, we try to overcome these problems by developing new virtual stereological probes and virtual tissue sections. A concrete result is the development of a new virtual 3D probe, the spatial rotator, which was found to have lower variance than the widely used planar...... simulator and a memory efficient, GPU implementation of for connected components labeling. This was furthermore extended to produce signed distance fields and Voronoi diagrams, all with real-time performance. It has during the course of the project been realized that many disciplines within computer science...

  5. The engineering design integration (EDIN) system. [digital computer program complex (United States)

    Glatt, C. R.; Hirsch, G. N.; Alford, G. E.; Colquitt, W. N.; Reiners, S. J.


    A digital computer program complex for the evaluation of aerospace vehicle preliminary designs is described. The system consists of a Univac 1100 series computer and peripherals using the Exec 8 operating system, a set of demand access terminals of the alphanumeric and graphics types, and a library of independent computer programs. Modification of the partial run streams, data base maintenance and construction, and control of program sequencing are provided by a data manipulation program called the DLG processor. The executive control of library program execution is performed by the Univac Exec 8 operating system through a user established run stream. A combination of demand and batch operations is employed in the evaluation of preliminary designs. Applications accomplished with the EDIN system are described.

  6. 77 FR 38610 - Privacy Act of 1974; Computer Matching Program (United States)


    ... education at the time of the parent or guardian's death. Beginning July 1, 2010, students who are otherwise... From the Federal Register Online via the Government Publishing Office DEPARTMENT OF EDUCATION Privacy Act of 1974; Computer Matching Program AGENCY: Department of Education. ACTION: Notice--Computer...

  7. Computing, Information, and Communications Technology (CICT) Program Overview (United States)

    VanDalsem, William R.


    The Computing, Information and Communications Technology (CICT) Program's goal is to enable NASA's Scientific Research, Space Exploration, and Aerospace Technology Missions with greater mission assurance, for less cost, with increased science return through the development and use of advanced computing, information and communication technologies

  8. 78 FR 1275 - Privacy Act of 1974; Computer Matching Program (United States)


    ...: Notice--computer matching between the Office of Personnel Management and the Social Security... matching program with the Social Security Administration (SSA). DATES: OPM will file a report of the..., as amended, regulates the use of computer matching by Federal agencies when records in a system of...

  9. 77 FR 74518 - Privacy Act of 1974; Computer Matching Program (United States)


    ...: Notice--computer matching between the Office of Personnel Management and the Social Security... Personnel Management (OPM) is publishing notice of its new computer matching program with the Social... matching by Federal agencies when records in a system of records are matched with other Federal, State, or...

  10. A computer program for analysis of fuelwood harvesting costs (United States)

    George B. Harpole; Giuseppe Rensi


    The fuelwood harvesting computer program (FHP) is written in FORTRAN 60 and designed to select a collection of harvest units and systems from among alternatives to satisfy specified energy requirements at a lowest cost per million Btu's as recovered in a boiler, or thousand pounds of H2O evaporative capacity kiln drying. Computed energy costs are used as a...

  11. Gender Digital Divide and Challenges in Undergraduate Computer Science Programs (United States)

    Stoilescu, Dorian; McDougall, Douglas


    Previous research revealed a reduced number of female students registered in computer science studies. In addition, the female students feel isolated, have reduced confidence, and underperform. This article explores differences between female and male students in undergraduate computer science programs in a mid-size university in Ontario. Based on…

  12. Computer Programming with Infants and Juniors. (United States)

    Hind, Jim


    The article argues that even extremely young children can be taught to program microcomputers from their very first contact. A teaching strategy is proposed, having more in common with the teaching of language than with the more traditional didactic-reinforcement cycle commonly employed in the text books. (Author/CL)

  13. Programming Languages for Distributed Computing Systems

    NARCIS (Netherlands)

    Bal, H.E.; Steiner, J.G.; Tanenbaum, A.S.


    When distributed systems first appeared, they were programmed in traditional sequential languages, usually with the addition of a few library procedures for sending and receiving messages. As distributed applications became more commonplace and more sophisticated, this ad hoc approach became less

  14. EZLP: An Interactive Computer Program for Solving Linear Programming Problems. Final Report. (United States)

    Jarvis, John J.; And Others

    Designed for student use in solving linear programming problems, the interactive computer program described (EZLP) permits the student to input the linear programming model in exactly the same manner in which it would be written on paper. This report includes a brief review of the development of EZLP; narrative descriptions of program features,…

  15. From Clover to computer. Towards programmed anaesthesia? (United States)

    Mapleson, W W


    The control of depth of anaesthesia has been viewed as a control-system problem the solution of which can involve both feedback and feedforward techniques. The nature of the problem in Clover's day and the solutions he found have been examined. A similar analysis has been made in respect of the modern anaesthetist. Finally, the way in which computers may aid the anaesthetist in his task has been illustrated by reference to various attempts reported from around the world and, in particular, by describing the development in Cardiff of a system which should produce, in the brain of the patient, any tension of an inhaled anaesthetic which the anaesthetist chooses to specify.

  16. MULGRES: a computer program for stepwise multiple regression analysis (United States)

    A. Jeff Martin


    MULGRES is a computer program source deck that is designed for multiple regression analysis employing the technique of stepwise deletion in the search for most significant variables. The features of the program, along with inputs and outputs, are briefly described, with a note on machine compatibility.

  17. A Successful Course of Study in Computer Programming (United States)

    Seeger, David H.


    Three keys to the successful development of the program of the computer programming department of the Technical Institute of Oklahoma State University are discussed: Community involvement, faculty/administration commitment to the basic principles of technical career education, and availability of appropriate equipment for student use. (HD)

  18. A Research Program in Computer Technology (United States)


    14 (7), 1971, 453-360. 5. Donzeau-Gouge, V., G. Kahn, and B. Lang , A Complete Machine-Checked Definition of a Simple Programming Language Using...Denotational Semantics, IRIA Laborla, Technical Report 330, October 1978. 6. Donzeau-Gouge, V., G. Kahn, and B. Lang , Formal Definition of Ada, Honeywell...May 1976. r S.-..-. . . . . . . . 12. ARPANET TENEX SERVICE T’fhttiral Staff Marion McKinley, Jr. William H. Moore Robert Hines Serge Poievitzky Edward

  19. Programs=data=first-class citizens in a computational world. (United States)

    Jones, Neil D; Simonsen, Jakob Grue


    From a programming perspective, Alan Turing's epochal 1936 paper on computable functions introduced several new concepts, including what is today known as self-interpreters and programs as data, and invented a great many now-common programming techniques. We begin by reviewing Turing's contribution from a programming perspective; and then systematize and mention some of the many ways that later developments in models of computation (MOCs) have interacted with computability theory and programming language research. Next, we describe the 'blob' MOC: a recent stored-program computational model without pointers. In the blob model, programs are truly first-class citizens, capable of being automatically compiled, or interpreted, or executed directly. Further, the blob model appears closer to being physically realizable than earlier computation models. In part, this is due to strong finiteness owing to early binding in the program; and a strong adjacency property: the active instruction is always adjacent to the piece of data on which it operates. The model is Turing complete in a strong sense: a universal interpretation algorithm exists that is able to run any program in a natural way and without arcane data encodings. Next, some of the best known among the numerous existing MOCs are described, and we develop a list of traits an 'ideal' MOC should possess from our perspective. We make no attempt to consider all models put forth since Turing's 1936 paper, and the selection of models covered concerns only models with discrete, atomic computation steps. The next step is to classify the selected models by qualitative rather than quantitative features. Finally, we describe how the blob model differs from an 'ideal' MOC, and identify some natural next steps to achieve such a model.

  20. Integrating computer programs for engineering analysis and design (United States)

    Wilhite, A. W.; Crisp, V. K.; Johnson, S. C.


    The design of a third-generation system for integrating computer programs for engineering and design has been developed for the Aerospace Vehicle Interactive Design (AVID) system. This system consists of an engineering data management system, program interface software, a user interface, and a geometry system. A relational information system (ARIS) was developed specifically for the computer-aided engineering system. It is used for a repository of design data that are communicated between analysis programs, for a dictionary that describes these design data, for a directory that describes the analysis programs, and for other system functions. A method is described for interfacing independent analysis programs into a loosely-coupled design system. This method emphasizes an interactive extension of analysis techniques and manipulation of design data. Also, integrity mechanisms exist to maintain database correctness for multidisciplinary design tasks by an individual or a team of specialists. Finally, a prototype user interface program has been developed to aid in system utilization.

  1. A Tangible Programming Tool for Children to Cultivate Computational Thinking (United States)

    Wang, Danli; Liu, Zhen


    Game and creation are activities which have good potential for computational thinking skills. In this paper we present T-Maze, an economical tangible programming tool for children aged 5–9 to build computer programs in maze games by placing wooden blocks. Through the use of computer vision technology, T-Maze provides a live programming interface with real-time graphical and voice feedback. We conducted a user study with 7 children using T-Maze to play two levels of maze-escape games and create their own mazes. The results show that T-Maze is not only easy to use, but also has the potential to help children cultivate computational thinking like abstraction, problem decomposition, and creativity. PMID:24719575

  2. On Computational Power of Quantum Read-Once Branching Programs

    Directory of Open Access Journals (Sweden)

    Farid Ablayev


    Full Text Available In this paper we review our current results concerning the computational power of quantum read-once branching programs. First of all, based on the circuit presentation of quantum branching programs and our variant of quantum fingerprinting technique, we show that any Boolean function with linear polynomial presentation can be computed by a quantum read-once branching program using a relatively small (usually logarithmic in the size of input number of qubits. Then we show that the described class of Boolean functions is closed under the polynomial projections.

  3. Sync-rank: Robust Ranking, Constrained Ranking and Rank Aggregation via Eigenvector and SDP Synchronization (United States)


    eigenvector of the associated Laplacian matrix (i.e., the Fiedler vector) matches that of the variables. In other words, this approach (reminiscent of...S1), i.e., Dii = ∑n j=1Gi,j is the degree of node i in the measurement graph G. 3: Compute the Fiedler vector of S (eigenvector corresponding to the...smallest nonzero eigenvalue of LS). 4: Output the ranking induced by sorting the Fiedler vector of S, with the global ordering (increasing or decreasing

  4. Injecting Artificial Memory Errors Into a Running Computer Program (United States)

    Bornstein, Benjamin J.; Granat, Robert A.; Wagstaff, Kiri L.


    Single-event upsets (SEUs) or bitflips are computer memory errors caused by radiation. BITFLIPS (Basic Instrumentation Tool for Fault Localized Injection of Probabilistic SEUs) is a computer program that deliberately injects SEUs into another computer program, while the latter is running, for the purpose of evaluating the fault tolerance of that program. BITFLIPS was written as a plug-in extension of the open-source Valgrind debugging and profiling software. BITFLIPS can inject SEUs into any program that can be run on the Linux operating system, without needing to modify the program s source code. Further, if access to the original program source code is available, BITFLIPS offers fine-grained control over exactly when and which areas of memory (as specified via program variables) will be subjected to SEUs. The rate of injection of SEUs is controlled by specifying either a fault probability or a fault rate based on memory size and radiation exposure time, in units of SEUs per byte per second. BITFLIPS can also log each SEU that it injects and, if program source code is available, report the magnitude of effect of the SEU on a floating-point value or other program variable.

  5. BraX-Ray: an X-ray of the Brazilian computer science graduate programs. (United States)

    Digiampietri, Luciano A; Mena-Chalco, Jesús P; Vaz de Melo, Pedro O S; Malheiro, Ana P R; Meira, Dânia N O; Franco, Laryssa F; Oliveira, Leonardo B


    Research productivity assessment is increasingly relevant for allocation of research funds. On one hand, this assessment is challenging because it involves both qualitative and quantitative analysis of several characteristics, most of them subjective in nature. On the other hand, current tools and academic social networks make bibliometric data web-available to everyone for free. Those tools, especially when combined with other data, are able to create a rich environment from which information on research productivity can be extracted. In this context, our work aims at characterizing the Brazilian Computer Science graduate programs and the relationship among themselves. We (i) present views of the programs from different perspectives, (ii) rank the programs according to each perspective and a combination of them, (iii) show correlation between assessment metrics, (iv) discuss how programs relate to another, and (v) infer aspects that boost programs' research productivity. The results indicate that programs with a higher insertion in the coauthorship network topology also possess a higher research productivity between 2004 and 2009.

  6. Ranking library materials


    Lewandowski, Dirk


    Purpose: This paper discusses ranking factors suitable for library materials and shows that ranking in general is a complex process and that ranking for library materials requires a variety of techniques. Design/methodology/approach: The relevant literature is reviewed to provide a systematic overview of suitable ranking factors. The discussion is based on an overview of ranking factors used in Web search engines. Findings: While there are a wide variety of ranking factors appl...

  7. Material Programming: a Design Practice for Computational Composites

    DEFF Research Database (Denmark)

    Vallgårda, Anna; Boer, Laurens; Tsaknaki, Vasiliki


    In this paper we propose the notion of material programming as a future design practice for computational composites. Material programming would be a way for the interaction designer to better explore the dynamic potential of computational materials at hand and through that familiarity be able...... to compose more sophisticated and complex temporal forms in their designs. The contribution of the paper is an analysis of qualities that we find a material programming practice would and should support: designs grounded in material properties and experiences, embodied programming practice, real-time on......-site explorations, and finally a reasonable level of complexity in couplings between input and output. We propose material programming knowing that the technology and materials are not entirely ready to support this practice yet, however, we are certain they will be and that the interaction design community...

  8. Ranking Baltic States Researchers

    Directory of Open Access Journals (Sweden)

    Gyula Mester


    Full Text Available In this article, using the h-index and the total number of citations, the best 10 Lithuanian, Latvian and Estonian researchers from several disciplines are ranked. The list may be formed based on the h-index and the total number of citations, given in Web of Science, Scopus, Publish or Perish Program and Google Scholar database. Data for the first 10 researchers are presented. Google Scholar is the most complete. Therefore, to define a single indicator, h-index calculated by Google Scholar may be a good and simple one. The author chooses the Google Scholar database as it is the broadest one.

  9. PageRank of integers (United States)

    Frahm, K. M.; Chepelianskii, A. D.; Shepelyansky, D. L.


    We up a directed network tracing links from a given integer to its divisors and analyze the properties of the Google matrix of this network. The PageRank vector of this matrix is computed numerically and it is shown that its probability is approximately inversely proportional to the PageRank index thus being similar to the Zipf law and the dependence established for the World Wide Web. The spectrum of the Google matrix of integers is characterized by a large gap and a relatively small number of nonzero eigenvalues. A simple semi-analytical expression for the PageRank of integers is derived that allows us to find this vector for matrices of billion size. This network provides a new PageRank order of integers.

  10. Ranking Fuzzy Numbers and Its Application to Products Attributes Preferences


    Abdullah, Lazim; Fauzee, Nor Nashrah Ahmad


    Ranking is one of the widely used methods in fuzzy decision making environment. The recent ranking fuzzy numbers proposed by Wang and Li is claimed to be the improved version in ranking. However, the method was never been simplified and tested in real life application. This paper presents a four-step computation of ranking fuzzy numbers and its application in ranking attributes of selected chocolate products. The four steps algorithm was formulated to rank fuzzy numbers and followed by a tes...

  11. Web Program for Development of GUIs for Cluster Computers (United States)

    Czikmantory, Akos; Cwik, Thomas; Klimeck, Gerhard; Hua, Hook; Oyafuso, Fabiano; Vinyard, Edward


    WIGLAF (a Web Interface Generator and Legacy Application Facade) is a computer program that provides a Web-based, distributed, graphical-user-interface (GUI) framework that can be adapted to any of a broad range of application programs, written in any programming language, that are executed remotely on any cluster computer system. WIGLAF enables the rapid development of a GUI for controlling and monitoring a specific application program running on the cluster and for transferring data to and from the application program. The only prerequisite for the execution of WIGLAF is a Web-browser program on a user's personal computer connected with the cluster via the Internet. WIGLAF has a client/server architecture: The server component is executed on the cluster system, where it controls the application program and serves data to the client component. The client component is an applet that runs in the Web browser. WIGLAF utilizes the Extensible Markup Language to hold all data associated with the application software, Java to enable platform-independent execution on the cluster system and the display of a GUI generator through the browser, and the Java Remote Method Invocation software package to provide simple, effective client/server networking.

  12. An introduction to NASA's advanced computing program: Integrated computing systems in advanced multichip modules (United States)

    Fang, Wai-Chi; Alkalai, Leon


    Recent changes within NASA's space exploration program favor the design, implementation, and operation of low cost, lightweight, small and micro spacecraft with multiple launches per year. In order to meet the future needs of these missions with regard to the use of spacecraft microelectronics, NASA's advanced flight computing (AFC) program is currently considering industrial cooperation and advanced packaging architectures. In relation to this, the AFC program is reviewed, considering the design and implementation of NASA's AFC multichip module.

  13. A full ranking for decision making units using ideal and anti-ideal points in DEA. (United States)

    Barzegarinegad, A; Jahanshahloo, G; Rostamy-Malkhalifeh, M


    We propose a procedure for ranking decision making units in data envelopment analysis, based on ideal and anti-ideal points in the production possibility set. Moreover, a model has been introduced to compute the performance of a decision making unit for these two points through using common set of weights. One of the best privileges of this method is that we can make ranking for all decision making units by solving only three programs, and also solving these programs is not related to numbers of decision making units. One of the other advantages of this procedure is to rank all the extreme and nonextreme efficient decision making units. In other words, the suggested ranking method tends to seek a set of common weights for all units to make them fully ranked. Finally, it was applied for different sets holding real data, and then it can be compared with other procedures.

  14. CRECTJ: a computer program for compilation of evaluated nuclear data

    Energy Technology Data Exchange (ETDEWEB)

    Nakagawa, Tsuneo [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment


    In order to compile evaluated nuclear data in the ENDF format, the computer program CRECTJ has been developed. CRECTJ has two versions; CRECTJ5 treats the data in the ENDF/B-IV and ENDF/B-V format, and CRECTJ6 the data in the ENDF-6 format. These programs have been frequently used to make Japanese Evaluated Nuclear Data Library (JENDL). This report describes input data and examples of CRECTJ. (author)

  15. Department of Energy Mathematical, Information, and Computational Sciences Division: High Performance Computing and Communications Program

    Energy Technology Data Exchange (ETDEWEB)



    This document is intended to serve two purposes. Its first purpose is that of a program status report of the considerable progress that the Department of Energy (DOE) has made since 1993, the time of the last such report (DOE/ER-0536, The DOE Program in HPCC), toward achieving the goals of the High Performance Computing and Communications (HPCC) Program. The second purpose is that of a summary report of the many research programs administered by the Mathematical, Information, and Computational Sciences (MICS) Division of the Office of Energy Research under the auspices of the HPCC Program and to provide, wherever relevant, easy access to pertinent information about MICS-Division activities via universal resource locators (URLs) on the World Wide Web (WWW).

  16. Department of Energy: MICS (Mathematical Information, and Computational Sciences Division). High performance computing and communications program

    Energy Technology Data Exchange (ETDEWEB)



    This document is intended to serve two purposes. Its first purpose is that of a program status report of the considerable progress that the Department of Energy (DOE) has made since 1993, the time of the last such report (DOE/ER-0536, {open_quotes}The DOE Program in HPCC{close_quotes}), toward achieving the goals of the High Performance Computing and Communications (HPCC) Program. The second purpose is that of a summary report of the many research programs administered by the Mathematical, Information, and Computational Sciences (MICS) Division of the Office of Energy Research under the auspices of the HPCC Program and to provide, wherever relevant, easy access to pertinent information about MICS-Division activities via universal resource locators (URLs) on the World Wide Web (WWW). The information pointed to by the URL is updated frequently, and the interested reader is urged to access the WWW for the latest information.

  17. Computer Program Plagiarism Detection: The Limits of the Halstead Metric. (United States)

    Berghel, H. L.; Sallach, David L.


    Discusses two alternative metrics to detect computer software plagiarism: the Halstead metric drawn from the software science discipline and an ad hoc method drawn from program grading experience and identified by factor analysis. Possible explanations as to why the ad hoc method is more useful in identical-task environments are considered.…

  18. Computer Programming with Early Elementary Students with Down Syndrome (United States)

    Taylor, Matthew S.; Vasquez, Eleazar; Donehower, Claire


    Students of all ages and abilities must be given the opportunity to learn academic skills that can shape future opportunities and careers. Researchers in the mid-1970s and 1980s began teaching young students the processes of computer programming using basic coding skills and limited technology. As technology became more personalized and easily…

  19. 77 FR 56824 - Privacy Act of 1974; Computer Matching Program (United States)


    ... information contained in the USCIS database is referred to as the Verification Information System (VIS), which... records entitled ``Verification Information System Records Notice (DHS-2007-0010).'' Where there is a... Information: Privacy Act of 1974; Computer Matching Program between the U.S. Department of Education and the...

  20. Computers for All Children: A Handbook for Program Design. (United States)

    Sharp, Pamela; Crist-Whitzel, Janet

    One of three publications of the Research on Equitable Access to Technology (REAT) project, this practitioner's handbook is designed to assist educators in the design and implementation of computer instruction programs for underserved groups of students, including low-income, minority, low-achieving, limited-English speaking, female, and rural…

  1. Tpetra, and the Use of Generic Programming in Scientific Computing

    Directory of Open Access Journals (Sweden)

    C.G. Baker


    Full Text Available We present Tpetra, a Trilinos package for parallel linear algebra primitives implementing the Petra object model. We describe Tpetra's design, based on generic programming via C++ templated types and template metaprogramming. We discuss some benefits of this approach in the context of scientific computing, with illustrations consisting of code and notable empirical results.

  2. What's New in Software? Integrated Computer Programs and Daily Living. (United States)

    Hedley, Carolyn N.


    Various kinds of electronic information media can now be integrated to plan educational programs, through use of computer videodiscs, hypercards, and hypertexts. Discussed are the components of integrative technology, including audio technology, video technology, and electronic text and graphics, and possibilities for interfacing the various…

  3. Intellectual Property Law and the Protection of Computer Programs. (United States)

    Lomio, J. Paul


    Briefly reviews the laws pertaining to copyrights, patents, and trade secrets, and discusses how each of these may be applied to the protection of computer programs. The comparative merits and limitations of each category of law are discussed and recent court decisions are summarized. (CLB)

  4. Learning Computer Programming: Implementing a Fractal in a Turing Machine (United States)

    Pereira, Hernane B. de B.; Zebende, Gilney F.; Moret, Marcelo A.


    It is common to start a course on computer programming logic by teaching the algorithm concept from the point of view of natural languages, but in a schematic way. In this sense we note that the students have difficulties in understanding and implementation of the problems proposed by the teacher. The main idea of this paper is to show that the…

  5. A Domain-Specific Programming Language for Secure Multiparty Computation

    DEFF Research Database (Denmark)

    Nielsen, Janus Dam; Schwartzbach, Michael Ignatieff


    We present a domain-specific programming language for Secure Multiparty Computation (SMC). Information is a resource of vital importance and considerable economic value to individuals, public administration, and private companies. This means that the confidentiality of information is crucial, but...... application development. The language is implemented in a prototype compiler that generates Java code exploiting a distributed cryptographic runtime....

  6. A Research Program in Computer Technology. 1987 Annual Technical Report (United States)


    mathematical approach to computational network design," in E. E. Swartzlander (ed.), Systolic Signal Processing Systems, chapter 1, Marcel Dekker, 1987...Intention-Based Diagnosis of Novice Programming Errors, Morgan Kaufmann, Los Altos, California, 1986. 32. Johnson, W. L., and E. Soloway, " PROUST

  7. An Analysis on Distance Education Computer Programming Students' Attitudes Regarding Programming and Their Self-Efficacy for Programming (United States)

    Ozyurt, Ozcan


    This study aims to analyze the attitudes of students studying computer programming through the distance education regarding programming, and their self-efficacy for programming and the relation between these two factors. The study is conducted with 104 students being thought with distance education in a university in the north region of Turkey in…

  8. Computing Programs for Determining Traffic Flows from Roundabouts (United States)

    Boroiu, A. A.; Tabacu, I.; Ene, A.; Neagu, E.; Boroiu, A.


    For modelling road traffic at the level of a road network it is necessary to specify the flows of all traffic currents at each intersection. These data can be obtained by direct measurements at the traffic light intersections, but in the case of a roundabout this is not possible directly and the literature as well as the traffic modelling software doesn’t offer ways to solve this issue. Two sets of formulas are proposed by which all traffic flows from the roundabouts with 3 or 4 arms are calculated based on the streams that can be measured. The objective of this paper is to develop computational programs to operate with these formulas. For each of the two sets of analytical relations, a computational program was developed in the Java operating language. The obtained results fully confirm the applicability of the calculation programs. The final stage for capitalizing these programs will be to make them web pages in HTML format, so that they can be accessed and used on the Internet. The achievements presented in this paper are an important step to provide a necessary tool for traffic modelling because these computational programs can be easily integrated into specialized software.

  9. Reduced Rank Regression

    DEFF Research Database (Denmark)

    Johansen, Søren


    The reduced rank regression model is a multivariate regression model with a coefficient matrix with reduced rank. The reduced rank regression algorithm is an estimation procedure, which estimates the reduced rank regression model. It is related to canonical correlations and involves calculating e...... eigenvalues and eigenvectors. We give a number of different applications to regression and time series analysis, and show how the reduced rank regression estimator can be derived as a Gaussian maximum likelihood estimator. We briefly mention asymptotic results......The reduced rank regression model is a multivariate regression model with a coefficient matrix with reduced rank. The reduced rank regression algorithm is an estimation procedure, which estimates the reduced rank regression model. It is related to canonical correlations and involves calculating...

  10. Computer programs for eddy-current defect studies

    Energy Technology Data Exchange (ETDEWEB)

    Pate, J. R.; Dodd, C. V. [Oak Ridge National Lab., TN (USA)


    Several computer programs to aid in the design of eddy-current tests and probes have been written. The programs, written in Fortran, deal in various ways with the response to defects exhibited by four types of probes: the pancake probe, the reflection probe, the circumferential boreside probe, and the circumferential encircling probe. Programs are included which calculate the impedance or voltage change in a coil due to a defect, which calculate and plot the defect sensitivity factor of a coil, and which invert calculated or experimental readings to obtain the size of a defect. The theory upon which the programs are based is the Burrows point defect theory, and thus the calculations of the programs will be more accurate for small defects. 6 refs., 21 figs.

  11. Introductory Computer Programming Course Teaching Improvement Using Immersion Language, Extreme Programming, and Education Theories (United States)

    Velez-Rubio, Miguel


    Teaching computer programming to freshmen students in Computer Sciences and other Information Technology areas has been identified as a complex activity. Different approaches have been studied looking for the best one that could help to improve this teaching process. A proposed approach was implemented which is based in the language immersion…

  12. Modeling of rolling element bearing mechanics. Computer program user's manual (United States)

    Greenhill, Lyn M.; Merchant, David H.


    This report provides the user's manual for the Rolling Element Bearing Analysis System (REBANS) analysis code which determines the quasistatic response to external loads or displacement of three types of high-speed rolling element bearings: angular contact ball bearings, duplex angular contact ball bearings, and cylindrical roller bearings. The model includes the defects of bearing ring and support structure flexibility. It is comprised of two main programs: the Preprocessor for Bearing Analysis (PREBAN) which creates the input files for the main analysis program, and Flexibility Enhanced Rolling Element Bearing Analysis (FEREBA), the main analysis program. This report addresses input instructions for and features of the computer codes. A companion report addresses the theoretical basis for the computer codes. REBANS extends the capabilities of the SHABERTH (Shaft and Bearing Thermal Analysis) code to include race and housing flexibility, including such effects as dead band and preload springs.

  13. Portable Computer Technology (PCT) Research and Development Program Phase 2 (United States)

    Castillo, Michael; McGuire, Kenyon; Sorgi, Alan


    The subject of this project report, focused on: (1) Design and development of two Advanced Portable Workstation 2 (APW 2) units. These units incorporate advanced technology features such as a low power Pentium processor, a high resolution color display, National Television Standards Committee (NTSC) video handling capabilities, a Personal Computer Memory Card International Association (PCMCIA) interface, and Small Computer System Interface (SCSI) and ethernet interfaces. (2) Use these units to integrate and demonstrate advanced wireless network and portable video capabilities. (3) Qualification of the APW 2 systems for use in specific experiments aboard the Mir Space Station. A major objective of the PCT Phase 2 program was to help guide future choices in computing platforms and techniques for meeting National Aeronautics and Space Administration (NASA) mission objectives. The focus being on the development of optimal configurations of computing hardware, software applications, and network technologies for use on NASA missions.

  14. New Mexico district work-effort analysis computer program (United States)

    Hiss, W.L.; Trantolo, A.P.; Sparks, J.L.


    The computer program (CAN 2) described in this report is one of several related programs used in the New Mexico District cost-analysis system. The work-effort information used in these programs is accumulated and entered to the nearest hour on forms completed by each employee. Tabulating cards are punched directly from these forms after visual examinations for errors are made. Reports containing detailed work-effort data itemized by employee within each project and account and by account and project for each employee are prepared for both current-month and year-to-date periods by the CAN 2 computer program. An option allowing preparation of reports for a specified 3-month period is provided. The total number of hours worked on each account and project and a grand total of hours worked in the New Mexico District is computed and presented in a summary report for each period. Work effort not chargeable directly to individual projects or accounts is considered as overhead and can be apportioned to the individual accounts and projects on the basis of the ratio of the total hours of work effort for the individual accounts or projects to the total New Mexico District work effort at the option of the user. The hours of work performed by a particular section, such as General Investigations or Surface Water, are prorated and charged to the projects or accounts within the particular section. A number of surveillance or buffer accounts are employed to account for the hours worked on special events or on those parts of large projects or accounts that require a more detailed analysis. Any part of the New Mexico District operation can be separated and analyzed in detail by establishing an appropriate buffer account. With the exception of statements associated with word size, the computer program is written in FORTRAN IV in a relatively low and standard language level to facilitate its use on different digital computers. The program has been run only on a Control Data Corporation

  15. Solutions manual and computer programs for physical and computational aspects of convective heat transfer

    CERN Document Server

    Cebeci, Tuncer


    This book is designed to accompany Physical and Computational Aspects of Convective Heat Transfer by T Cebeci and P Bradshaw and contains solutions to the exercises and computer programs for the numerical methods contained in that book Physical and Computational Aspects of Convective Heat Transfer begins with a thorough discussion of the physical aspects of convective heat transfer and presents in some detail the partial differential equations governing the transport of thermal energy in various types of flows The book is intended for senior undergraduate and graduate students of aeronautical, chemical, civil and mechanical engineering It can also serve as a reference for the practitioner

  16. My Program Is Ok--Am I? Computing Freshmen's Experiences of Doing Programming Assignments (United States)

    Kinnunen, Paivi; Simon, Beth


    This article provides insight into how computing majors experience the process of doing programming assignments in their first programming course. This grounded theory study sheds light on the various processes and contexts through which students constantly assess their self-efficacy as a programmer. The data consists of a series of four…

  17. Motivating Programming: Using Storytelling to Make Computer Programming Attractive to Middle School Girls (United States)


    In Generic Alice, there are two common default Chapter 5: Developing the Storytelling Gallery 114 positions for characters and, consequently , two...balance or generating the nth Fibonacci number. Often, students write programs in introductory computer science using professional programming...handle without fundamentally changing the common control structures found in general -purpose languages. Consequently , when a student moves from one of

  18. Computer Programming Games and Gender Oriented Cultural Forms (United States)

    AlSulaiman, Sarah Abdulmalik

    I present the design and evaluation of two games designed to help elementary and middle school students learn computer programming concepts. The first game was designed to be "gender neutral", aligning with might be described as a consensus opinion on best practices for computational learning environments. The second game, based on the cultural form of dress up dolls was deliberately designed to appeal to females. I recruited 70 participants in an international two-phase study to investigate the relationship between games, gender, attitudes towards computer programming, and learning. My findings suggest that while the two games were equally effective in terms of learning outcomes, I saw differences in motivation between players of the two games. Specifically, participants who reported a preference for female- oriented games were more motivated to learn about computer programming when they played a game that they perceived as designed for females. In addition, I describe how the two games seemed to encourage different types of social activity between players in a classroom setting. Based on these results, I reflect on the strategy of exclusively designing games and activities as "gender neutral", and suggest that employing cultural forms, including gendered ones, may help create a more productive experience for learners.

  19. Computer-Guided Deep Brain Stimulation Programming for Parkinson's Disease. (United States)

    Heldman, Dustin A; Pulliam, Christopher L; Urrea Mendoza, Enrique; Gartner, Maureen; Giuffrida, Joseph P; Montgomery, Erwin B; Espay, Alberto J; Revilla, Fredy J


    Pilot study to evaluate computer-guided deep brain stimulation (DBS) programming designed to optimize stimulation settings using objective motion sensor-based motor assessments. Seven subjects (five males; 54-71 years) with Parkinson's disease (PD) and recently implanted DBS systems participated in this pilot study. Within two months of lead implantation, the subject returned to the clinic to undergo computer-guided programming and parameter selection. A motion sensor was placed on the index finger of the more affected hand. Software guided a monopolar survey during which monopolar stimulation on each contact was iteratively increased followed by an automated assessment of tremor and bradykinesia. After completing assessments at each setting, a software algorithm determined stimulation settings designed to minimize symptom severities, side effects, and battery usage. Optimal DBS settings were chosen based on average severity of motor symptoms measured by the motion sensor. Settings chosen by the software algorithm identified a therapeutic window and improved tremor and bradykinesia by an average of 35.7% compared with baseline in the "off" state (p computer-guided DBS programming identified stimulation parameters that significantly improved tremor and bradykinesia with minimal clinician involvement. Automated motion sensor-based mapping is worthy of further investigation and may one day serve to extend programming to populations without access to specialized DBS centers. © 2015 International Neuromodulation Society.

  20. Ranking Operations Management conferences

    NARCIS (Netherlands)

    Steenhuis, H.J.; de Bruijn, E.J.; Gupta, Sushil; Laptaned, U


    Several publications have appeared in the field of Operations Management which rank Operations Management related journals. Several ranking systems exist for journals based on , for example, perceived relevance and quality, citation, and author affiliation. Many academics also publish at conferences

  1. An intertwined method for making low-rank, sum-of-product basis functions that makes it possible to compute vibrational spectra of molecules with more than 10 atoms (United States)

    Thomas, Phillip S.; Carrington, Tucker


    We propose a method for solving the vibrational Schrödinger equation with which one can compute spectra for molecules with more than ten atoms. It uses sum-of-product (SOP) basis functions stored in a canonical polyadic tensor format and generated by evaluating matrix-vector products. By doing a sequence of partial optimizations, in each of which the factors in a SOP basis function for a single coordinate are optimized, the rank of the basis functions is reduced as matrix-vector products are computed. This is better than using an alternating least squares method to reduce the rank, as is done in the reduced-rank block power method. Partial optimization is better because it speeds up the calculation by about an order of magnitude and allows one to significantly reduce the memory cost. We demonstrate the effectiveness of the new method by computing vibrational spectra of two molecules, ethylene oxide (C2H4O ) and cyclopentadiene (C5H6 ) , with 7 and 11 atoms, respectively.

  2. Engineering and programming manual: Two-dimensional kinetic reference computer program (TDK) (United States)

    Nickerson, G. R.; Dang, L. D.; Coats, D. E.


    The Two Dimensional Kinetics (TDK) computer program is a primary tool in applying the JANNAF liquid rocket thrust chamber performance prediction methodology. The development of a methodology that includes all aspects of rocket engine performance from analytical calculation to test measurements, that is physically accurate and consistent, and that serves as an industry and government reference is presented. Recent interest in rocket engines that operate at high expansion ratio, such as most Orbit Transfer Vehicle (OTV) engine designs, has required an extension of the analytical methods used by the TDK computer program. Thus, the version of TDK that is described in this manual is in many respects different from the 1973 version of the program. This new material reflects the new capabilities of the TDK computer program, the most important of which are described.

  3. Center for Programming Models for Scalable Parallel Computing

    Energy Technology Data Exchange (ETDEWEB)

    John Mellor-Crummey


    Rice University's achievements as part of the Center for Programming Models for Scalable Parallel Computing include: (1) design and implemention of cafc, the first multi-platform CAF compiler for distributed and shared-memory machines, (2) performance studies of the efficiency of programs written using the CAF and UPC programming models, (3) a novel technique to analyze explicitly-parallel SPMD programs that facilitates optimization, (4) design, implementation, and evaluation of new language features for CAF, including communication topologies, multi-version variables, and distributed multithreading to simplify development of high-performance codes in CAF, and (5) a synchronization strength reduction transformation for automatically replacing barrier-based synchronization with more efficient point-to-point synchronization. The prototype Co-array Fortran compiler cafc developed in this project is available as open source software from

  4. Techniques for Engaging Students in an Online Computer Programming Course

    Directory of Open Access Journals (Sweden)

    Eman M. El-Sheikh


    Full Text Available Many institutions of higher education are significantly expanding their online program and course offerings to deal with the rapidly increasing demand for flexible educational alternatives. One of the main challenges that faculty who teach online courses face is determining how to engage students in an online environment. Teaching computer programming effectively requires demonstration of programming techniques, examples, and environments, and interaction with the students, making online delivery even more challenging. This paper describes efforts to engage students in an online introductory programming course at our institution. The tools and methods used to promote student engagement in the course are described, in addition to the lessons learned from the design and delivery of the online course and opportunities for future work.

  5. A Fast Algorithm for Generating Permutation Distribution of Ranks in ...

    African Journals Online (AJOL)

    ... function of the distribution of the ranks. This further gives insight into the permutation distribution of a rank statistics. The algorithm is implemented with the aid of the computer algebra system Mathematica. Key words: Combinatorics, generating function, permutation distribution, rank statistics, partitions, computer algebra.

  6. Final Report: Center for Programming Models for Scalable Parallel Computing

    Energy Technology Data Exchange (ETDEWEB)

    Mellor-Crummey, John [William Marsh Rice University


    As part of the Center for Programming Models for Scalable Parallel Computing, Rice University collaborated with project partners in the design, development and deployment of language, compiler, and runtime support for parallel programming models to support application development for the “leadership-class” computer systems at DOE national laboratories. Work over the course of this project has focused on the design, implementation, and evaluation of a second-generation version of Coarray Fortran. Research and development efforts of the project have focused on the CAF 2.0 language, compiler, runtime system, and supporting infrastructure. This has involved working with the teams that provide infrastructure for CAF that we rely on, implementing new language and runtime features, producing an open source compiler that enabled us to evaluate our ideas, and evaluating our design and implementation through the use of benchmarks. The report details the research, development, findings, and conclusions from this work.

  7. Qualification of a computer program for drill string dynamics

    Energy Technology Data Exchange (ETDEWEB)

    Stone, C.M.; Carne, T.G.; Caskey, B.C.


    A four point plan for the qualification of the GEODYN drill string dynamics computer program is described. The qualification plan investigates both modal response and transient response of a short drill string subjected to simulated cutting loads applied through a polycrystalline diamond compact (PDC) bit. The experimentally based qualification shows that the analytical techniques included in Phase 1 GEODYN correctly simulate the dynamic response of the bit-drill string system. 6 refs., 8 figs.

  8. Introduction to ''Interactive models of computation and program behaviour"


    Curien, Pierre-Louis


    Since the mid-eighties of the last century, a fruitful interplay between computer scientists and mathematicians has led to much progress in the understanding of programming languages, and has given new impulse to areas of mathematics such as proof theory or category theory. The volume of which this text is an introduction contains three contributions: Categorical semantics of linear logic, by P.-A. Melliès, Realizability in classical logic, by J.-L. Krivien, Abstract machines for dialogue gam...

  9. PET computer programs for use with the 88-inch cyclotron

    Energy Technology Data Exchange (ETDEWEB)

    Gough, R.A.; Chlosta, L.


    This report describes in detail several offline programs written for the PET computer which provide an efficient data management system to assist with the operation of the 88-Inch Cyclotron. This function includes the capability to predict settings for all cyclotron and beam line parameters for all beams within the present operating domain of the facility. The establishment of a data base for operational records is also described from which various aspects of the operating history can be projected.

  10. Designing, programming, and optimizing a (small) quantum computer (United States)

    Svore, Krysta

    In 1982, Richard Feynman proposed to use a computer founded on the laws of quantum physics to simulate physical systems. In the more than thirty years since, quantum computers have shown promise to solve problems in number theory, chemistry, and materials science that would otherwise take longer than the lifetime of the universe to solve on an exascale classical machine. The practical realization of a quantum computer requires understanding and manipulating subtle quantum states while experimentally controlling quantum interference. It also requires an end-to-end software architecture for programming, optimizing, and implementing a quantum algorithm on the quantum device hardware. In this talk, we will introduce recent advances in connecting abstract theory to present-day real-world applications through software. We will highlight recent advancement of quantum algorithms and the challenges in ultimately performing a scalable solution on a quantum device.

  11. Computationally intensive econometrics using a distributed matrix-programming language. (United States)

    Doornik, Jurgen A; Hendry, David F; Shephard, Neil


    This paper reviews the need for powerful computing facilities in econometrics, focusing on concrete problems which arise in financial economics and in macroeconomics. We argue that the profession is being held back by the lack of easy-to-use generic software which is able to exploit the availability of cheap clusters of distributed computers. Our response is to extend, in a number of directions, the well-known matrix-programming interpreted language Ox developed by the first author. We note three possible levels of extensions: (i) Ox with parallelization explicit in the Ox code; (ii) Ox with a parallelized run-time library; and (iii) Ox with a parallelized interpreter. This paper studies and implements the first case, emphasizing the need for deterministic computing in science. We give examples in the context of financial economics and time-series modelling.

  12. The ACP (Advanced Computer Program) multiprocessor system at Fermilab

    Energy Technology Data Exchange (ETDEWEB)

    Nash, T.; Areti, H.; Atac, R.; Biel, J.; Case, G.; Cook, A.; Fischler, M.; Gaines, I.; Hance, R.; Husby, D.


    The Advanced Computer Program at Fermilab has developed a multiprocessor system which is easy to use and uniquely cost effective for many high energy physics problems. The system is based on single board computers which cost under $2000 each to build including 2 Mbytes of on board memory. These standard VME modules each run experiment reconstruction code in Fortran at speeds approaching that of a VAX 11/780. Two versions have been developed: one uses Motorola's 68020 32 bit microprocessor, the other runs with AT and T's 32100. both include the corresponding floating point coprocessor chip. The first system, when fully configured, uses 70 each of the two types of processors. A 53 processor system has been operated for several months with essentially no down time by computer operators in the Fermilab Computer Center, performing at nearly the capacity of 6 CDC Cyber 175 mainframe computers. The VME crates in which the processing ''nodes'' sit are connected via a high speed ''Branch Bus'' to one or more MicroVAX computers which act as hosts handling system resource management and all I/O in offline applications. An interface from Fastbus to the Branch Bus has been developed for online use which has been tested error free at 20 Mbytes/sec for 48 hours. ACP hardware modules are now available commercially. A major package of software, including a simulator that runs on any VAX, has been developed. It allows easy migration of existing programs to this multiprocessor environment. This paper describes the ACP Multiprocessor System and early experience with it at Fermilab and elsewhere.

  13. Communities in Large Networks: Identification and Ranking

    DEFF Research Database (Denmark)

    Olsen, Martin


    We study the problem of identifying and ranking the members of a community in a very large network with link analysis only, given a set of representatives of the community. We define the concept of a community justified by a formal analysis of a simple model of the evolution of a directed graph. ...... and its immediate surroundings. The members are ranked with a “local” variant of the PageRank algorithm. Results are reported from successful experiments on identifying and ranking Danish Computer Science sites and Danish Chess pages using only a few representatives....

  14. Sparse structure regularized ranking

    KAUST Repository

    Wang, Jim Jing-Yan


    Learning ranking scores is critical for the multimedia database retrieval problem. In this paper, we propose a novel ranking score learning algorithm by exploring the sparse structure and using it to regularize ranking scores. To explore the sparse structure, we assume that each multimedia object could be represented as a sparse linear combination of all other objects, and combination coefficients are regarded as a similarity measure between objects and used to regularize their ranking scores. Moreover, we propose to learn the sparse combination coefficients and the ranking scores simultaneously. A unified objective function is constructed with regard to both the combination coefficients and the ranking scores, and is optimized by an iterative algorithm. Experiments on two multimedia database retrieval data sets demonstrate the significant improvements of the propose algorithm over state-of-the-art ranking score learning algorithms.

  15. Natural Resources Research Program: Catalog of Computer Programs for Project Management. (United States)

    project management . These include programs developed for use on a microcomputer, as well as those which run on a host computer but are accessed by a terminal in a field office. A one-page description of each program contains the title; preparing agency; abstract; a summary of the data inputs and outputs; equipment, disk, and memory requirements; operating system and programming language; and a contact for further information. The programs described in this publication are not limited to those available within the Corps, but also include those available from other

  16. Towards a Serious Game to Help Students Learn Computer Programming

    Directory of Open Access Journals (Sweden)

    Mathieu Muratet


    Full Text Available Video games are part of our culture like TV, movies, and books. We believe that this kind of software can be used to increase students' interest in computer science. Video games with other goals than entertainment, serious games, are present, today, in several fields such as education, government, health, defence, industry, civil security, and science. This paper presents a study around a serious game dedicated to strengthening programming skills. Real-Time Strategy, which is a popular game genre, seems to be the most suitable kind of game to support such a serious game. From programming teaching features to video game characteristics, we define a teaching organisation to experiment if a serious game can be adapted to learn programming.

  17. HAL/SM language specification. [programming languages and computer programming for space shuttles (United States)

    Williams, G. P. W., Jr.; Ross, C.


    A programming language is presented for the flight software of the NASA Space Shuttle program. It is intended to satisfy virtually all of the flight software requirements of the space shuttle. To achieve this, it incorporates a wide range of features, including applications-oriented data types and organizations, real time control mechanisms, and constructs for systems programming tasks. It is a higher order language designed to allow programmers, analysts, and engineers to communicate with the computer in a form approximating natural mathematical expression. Parts of the English language are combined with standard notation to provide a tool that readily encourages programming without demanding computer hardware expertise. Block diagrams and flow charts are included. The semantics of the language is discussed.

  18. [An algol program for the computation of empiric regressions]. (United States)

    Peil, J; Schmerling, S


    An explanation is given about the meaning of empirical regression and on the domain of application of this biomathematical-statistical procedure. It may be helpful in data handling after the measurements and in a first stage of data processing especially if there is a large amount of datas. An empirical regression can provide the basis for a functional relationship analysis by giving hints for the choice of empirical mathematical functions. This will be useful and necessary in such cases where the measured values have a greater dispersion and one wants to get an analytical expression for the course of measured points. In the appendix a program listing of the ALGOL-program for empirical regression is presented. Detailed remarks are made in the text concerning the program structure, the data input and output resp. the program control parameters to enable the biological or medical user to adapt the program to their special problems without the help by a mathematician, and neither with deeper knowledge of mathematics nor with detailed insight to computer technical aspects of data processing.

  19. A Computer Program for Assessing Nuclear Safety Culture Impact

    Energy Technology Data Exchange (ETDEWEB)

    Han, Kiyoon; Jae, Moosung [Hanyang Univ., Seoul (Korea, Republic of)


    Through several accidents of NPP including the Fukushima Daiichi in 2011 and Chernobyl accidents in 1986, a lack of safety culture was pointed out as one of the root cause of these accidents. Due to its latent influences on safety performance, safety culture has become an important issue in safety researches. Most of the researches describe how to evaluate the state of the safety culture of the organization. However, they did not include a possibility that the accident occurs due to the lack of safety culture. Because of that, a methodology for evaluating the impact of the safety culture on NPP's safety is required. In this study, the methodology for assessing safety culture impact is suggested and a computer program is developed for its application. SCII model which is the new methodology for assessing safety culture impact quantitatively by using PSA model. The computer program is developed for its application. This program visualizes the SCIs and the SCIIs. It might contribute to comparing the level of the safety culture among NPPs as well as improving the management safety of NPP.

  20. Frequency Domain Computer Programs for Prediction and Analysis of Rail Vehicle Dynamics : Volume 1. Technical Report (United States)


    Frequency domain computer programs developed or acquired by TSC for the analysis of rail vehicle dynamics are described in two volumes. Volume I defines the general analytical capabilities required for computer programs applicable to single rail vehi...

  1. Ethics of prevention: an interactive computer-tailored program. (United States)

    Van Hooren, Rob H; Van den Borne, Bart W; Curfs, Leopold M G; Widdershoven, Guy A M


    This article describes the contents of an interactive computer-tailored program. The program is based on previous studies of the practice of care for persons with Prader-Willi syndrome. This genetic condition is associated with a constant overeating behaviour with the risk of obesity. The aim of the program is to start a process of awareness, reflection, and discussion by caregivers who are confronted with the moral dilemma of respect for autonomy versus restricting overeating behaviour. The program focuses on values (such as health and well-being) that are relevant to caregivers in daily practice. Furthermore, the focus is on various ways of interaction with the client. Caregivers were expected to focus mainly on health, and on both paternalistic and interpretive/deliberative forms of interaction. Sixteen professionals and 12 parents pilot-tested the program contents. With a pre-test, responses on one central case were collected for tailored feedback; with a post-test, the effects of the program were measured. Significant correlations were found between the values of autonomy and consultation and between autonomy and well-being. In contrast to our expectations respondents valued all categories (autonomy, consultation, health, well-being, and liveability for others) as equally important in the pre-test. No significant changes in scores were found between pre- and post-test. The open answers and remarks of participants support the program contents. Participants' responses support previous research findings, advocating a concept of autonomy in terms of positive freedom, through support by others. The promotion of the client's self-understanding and self-development is central in this concept.

  2. High Performance Computing - Power Application Programming Interface Specification.

    Energy Technology Data Exchange (ETDEWEB)

    Laros, James H.,; Kelly, Suzanne M.; Pedretti, Kevin; Grant, Ryan; Olivier, Stephen Lecler; Levenhagen, Michael J.; DeBonis, David


    Measuring and controlling the power and energy consumption of high performance computing systems by various components in the software stack is an active research area [13, 3, 5, 10, 4, 21, 19, 16, 7, 17, 20, 18, 11, 1, 6, 14, 12]. Implementations in lower level software layers are beginning to emerge in some production systems, which is very welcome. To be most effective, a portable interface to measurement and control features would significantly facilitate participation by all levels of the software stack. We present a proposal for a standard power Application Programming Interface (API) that endeavors to cover the entire software space, from generic hardware interfaces to the input from the computer facility manager.

  3. Maximum Waring ranks of monomials


    Holmes, Erik; Plummer, Paul; Siegert, Jeremy; Teitler, Zach


    We show that monomials and sums of pairwise coprime monomials in four or more variables have Waring rank less than the generic rank, with a short list of exceptions. We asymptotically compare their ranks with the generic rank.

  4. Building Computer-Based Experiments in Psychology without Programming Skills. (United States)

    Ruisoto, Pablo; Bellido, Alberto; Ruiz, Javier; Juanes, Juan A


    Research in Psychology usually requires to build and run experiments. However, although this task has required scripting, recent computer tools based on graphical interfaces offer new opportunities in this field for researchers with non-programming skills. The purpose of this study is to illustrate and provide a comparative overview of two of the main free open source "point and click" software packages for building and running experiments in Psychology: PsychoPy and OpenSesame. Recommendations for their potential use are further discussed.

  5. Structural mode significance using INCA. [Interactive Controls Analysis computer program (United States)

    Bauer, Frank H.; Downing, John P.; Thorpe, Christopher J.


    Structural finite element models are often too large to be used in the design and analysis of control systems. Model reduction techniques must be applied to reduce the structural model to manageable size. In the past, engineers either performed the model order reduction by hand or used distinct computer programs to retrieve the data, to perform the significance analysis and to reduce the order of the model. To expedite this process, the latest version of INCA has been expanded to include an interactive graphical structural mode significance and model order reduction capability.

  6. Computational Hydrodynamics: How Portable and Scalable Are Heterogeneous Programming Paradigms?

    DEFF Research Database (Denmark)

    Pawlak, Wojciech; Glimberg, Stefan Lemvig; Engsig-Karup, Allan Peter

    New many-core era applications at the interface of mathematics and computer science adopt modern parallel programming paradigms and expose parallelism through proper algorithms. We present new performance results for a novel massively parallel free surface wave model suitable for advanced......-device system sizes from desktops to large HPC systems such as superclusters and in the cloud utilizing heterogeneous devices like multi-core CPUs, GPUs, and Xeon Phi coprocessors. The numerical efficiency is evaluated on heterogeneous devices like multi-core CPUs, GPUs and Xeon Phi coprocessors to test...

  7. Activity computer program for calculating ion irradiation activation (United States)

    Palmer, Ben; Connolly, Brian; Read, Mark


    A computer program, Activity, was developed to predict the activity and gamma lines of materials irradiated with an ion beam. It uses the TENDL (Koning and Rochman, 2012) [1] proton reaction cross section database, the Stopping and Range of Ions in Matter (SRIM) (Biersack et al., 2010) code, a Nuclear Data Services (NDS) radioactive decay database (Sonzogni, 2006) [2] and an ENDF gamma decay database (Herman and Chadwick, 2006) [3]. An extended version of Bateman's equation is used to calculate the activity at time t, and this equation is solved analytically, with the option to also solve by numeric inverse Laplace Transform as a failsafe. The program outputs the expected activity and gamma lines of the activated material.

  8. Programming a massively parallel, computation universal system: static behavior

    Energy Technology Data Exchange (ETDEWEB)

    Lapedes, A.; Farber, R.


    In previous work by the authors, the ''optimum finding'' properties of Hopfield neural nets were applied to the nets themselves to create a ''neural compiler.'' This was done in such a way that the problem of programming the attractors of one neural net (called the Slave net) was expressed as an optimization problem that was in turn solved by a second neural net (the Master net). In this series of papers that approach is extended to programming nets that contain interneurons (sometimes called ''hidden neurons''), and thus deals with nets capable of universal computation. 22 refs.

  9. Accelerated Strategic Computing Initiative (ASCI) Program Plan [FY2000

    Energy Technology Data Exchange (ETDEWEB)



    In August 1995, the United States took a significant step to reduce the nuclear danger. The decision to pursue a zero- yield Comprehensive Test Ban Treaty will allow greater control over the proliferation of nuclear weapons and will halt the growth of new nuclear systems. This step is only possible because of the Stockpile Stewardship Program, which provides an alternative means of ensuring the safety, performance, and reliability of the United States' enduring stockpile. At the heart of the Stockpile Stewardship Program is ASCI, which will create the high-confidence simulation capabilities needed to integrate fundamental science, experiments, and archival data into the stewardship of the actual weapons in the stockpile. ASCI will also serve to drive the development of simulation as a national resource by working closely with the computer industry and with universities.

  10. A Computer Program for a Canonical Problem in Underwater Shock

    Directory of Open Access Journals (Sweden)

    Thomas L. Geers


    Full Text Available Finite-element/boundary-element codes are widely used to analyze the response of marine structures to underwater explosions. An important step in verifying the correctness and accuracy of such codes is the comparison of code-generated results for canonical problems with corresponding analytical or semianalytical results. At the present time, such comparisons rely on hardcopy results presented in technical journals and reports. This article describes a computer program available from SAVIAC that produces user-selected numerical results for a step-wave-excited spherical shell submerged in and (optionally filled with an acoustic fluid. The method of solution employed in the program is based on classical expansion of the field quantities in generalized Fourier series in the meridional coordinate. Convergence of the series is enhanced by judicious application of modified Cesàro summation and partial closed-form solution.

  11. The Effects of a Robot Game Environment on Computer Programming Education for Elementary School Students (United States)

    Shim, Jaekwoun; Kwon, Daiyoung; Lee, Wongyu


    In the past, computer programming was perceived as a task only carried out by computer scientists; in the 21st century, however, computer programming is viewed as a critical and necessary skill that everyone should learn. In order to improve teaching of problem-solving abilities in a computing environment, extensive research is being done on…

  12. Easy-to-use application programs for decay heat and delayed neutron calculations on personal computers

    Energy Technology Data Exchange (ETDEWEB)

    Oyamatsu, Kazuhiro [Nagoya Univ. (Japan)


    Application programs for personal computers are developed to calculate the decay heat power and delayed neutron activity from fission products. The main programs can be used in any computers from personal computers to main frames because their sources are written in Fortran. These programs have user friendly interfaces to be used easily not only for research activities but also for educational purposes. (author)

  13. Modified Hazard Ranking System/Hazard Ranking System for sites with mixed radioactive and hazardous wastes: Software documentation

    Energy Technology Data Exchange (ETDEWEB)

    Stenner, R.D.; Peloquin, R.A.; Hawley, K.A.


    The mHRS/HRS software package was developed by the Pacific Northwest Laboratory (PNL) under contract with the Department of Energy (DOE) to provide a uniform method for DOE facilities to use in performing their Conservation Environmental Response Compensation and Liability Act (CERCLA) Phase I Modified Hazard Ranking System or Hazard Ranking System evaluations. The program is designed to remove the tedium and potential for error associated with the performing of hand calculations and the interpreting of information on tables and in reference books when performing an evaluation. The software package is designed to operate on a microcomputer (IBM PC, PC/XT, or PC/AT, or a compatible system) using either a dual floppy disk drive or a hard disk storage system. It is written in the dBASE III language and operates using the dBASE III system. Although the mHRS/HRS software package was developed for use at DOE facilities, it has direct applicability to the performing of CERCLA Phase I evaluations for any facility contaminated by hazardous waste. The software can perform evaluations using either the modified hazard ranking system methodology developed by DOE/PNL, the hazard ranking system methodology developed by EPA/MITRE Corp., or a combination of the two. This document is a companion manual to the mHRS/HRS user manual. It is intended for the programmer who must maintain the software package and for those interested in the computer implementation. This manual documents the system logic, computer programs, and data files that comprise the package. Hardware and software implementation requirements are discussed. In addition, hand calculations of three sample situations (problems) with associated computer runs used for the verification of program calculations are included.

  14. Computer Program Recognizes Patterns in Time-Series Data (United States)

    Hand, Charles


    A computer program recognizes selected patterns in time-series data like digitized samples of seismic or electrophysiological signals. The program implements an artificial neural network (ANN) and a set of N clocks for the purpose of determining whether N or more instances of a certain waveform, W, occur within a given time interval, T. The ANN must be trained to recognize W in the incoming stream of data. The first time the ANN recognizes W, it sets clock 1 to count down from T to zero; the second time it recognizes W, it sets clock 2 to count down from T to zero, and so forth through the Nth instance. On the N + 1st instance, the cycle is repeated, starting with clock 1. If any clock has not reached zero when it is reset, then N instances of W have been detected within time T, and the program so indicates. The program can readily be encoded in a field-programmable gate array or an application-specific integrated circuit that could be used, for example, to detect electroencephalographic or electrocardiographic waveforms indicative of epileptic seizures or heart attacks, respectively.

  15. Program Predicts Time Courses of Human/Computer Interactions (United States)

    Vera, Alonso; Howes, Andrew


    CPM X is a computer program that predicts sequences of, and amounts of time taken by, routine actions performed by a skilled person performing a task. Unlike programs that simulate the interaction of the person with the task environment, CPM X predicts the time course of events as consequences of encoded constraints on human behavior. The constraints determine which cognitive and environmental processes can occur simultaneously and which have sequential dependencies. The input to CPM X comprises (1) a description of a task and strategy in a hierarchical description language and (2) a description of architectural constraints in the form of rules governing interactions of fundamental cognitive, perceptual, and motor operations. The output of CPM X is a Program Evaluation Review Technique (PERT) chart that presents a schedule of predicted cognitive, motor, and perceptual operators interacting with a task environment. The CPM X program allows direct, a priori prediction of skilled user performance on complex human-machine systems, providing a way to assess critical interfaces before they are deployed in mission contexts.

  16. How to Rank Journals. (United States)

    Bradshaw, Corey J A; Brook, Barry W


    There are now many methods available to assess the relative citation performance of peer-reviewed journals. Regardless of their individual faults and advantages, citation-based metrics are used by researchers to maximize the citation potential of their articles, and by employers to rank academic track records. The absolute value of any particular index is arguably meaningless unless compared to other journals, and different metrics result in divergent rankings. To provide a simple yet more objective way to rank journals within and among disciplines, we developed a κ-resampled composite journal rank incorporating five popular citation indices: Impact Factor, Immediacy Index, Source-Normalized Impact Per Paper, SCImago Journal Rank and Google 5-year h-index; this approach provides an index of relative rank uncertainty. We applied the approach to six sample sets of scientific journals from Ecology (n = 100 journals), Medicine (n = 100), Multidisciplinary (n = 50); Ecology + Multidisciplinary (n = 25), Obstetrics & Gynaecology (n = 25) and Marine Biology & Fisheries (n = 25). We then cross-compared the κ-resampled ranking for the Ecology + Multidisciplinary journal set to the results of a survey of 188 publishing ecologists who were asked to rank the same journals, and found a 0.68-0.84 Spearman's ρ correlation between the two rankings datasets. Our composite index approach therefore approximates relative journal reputation, at least for that discipline. Agglomerative and divisive clustering and multi-dimensional scaling techniques applied to the Ecology + Multidisciplinary journal set identified specific clusters of similarly ranked journals, with only Nature & Science separating out from the others. When comparing a selection of journals within or among disciplines, we recommend collecting multiple citation-based metrics for a sample of relevant and realistic journals to calculate the composite rankings and their relative uncertainty windows.

  17. Computational Ranking of Yerba Mate Small Molecules Based on Their Predicted Contribution to Antibacterial Activity against Methicillin-Resistant Staphylococcus aureus.

    Directory of Open Access Journals (Sweden)

    Caroline S Rempe

    Full Text Available The aqueous extract of yerba mate, a South American tea beverage made from Ilex paraguariensis leaves, has demonstrated bactericidal and inhibitory activity against bacterial pathogens, including methicillin-resistant Staphylococcus aureus (MRSA. The gas chromatography-mass spectrometry (GC-MS analysis of two unique fractions of yerba mate aqueous extract revealed 8 identifiable small molecules in those fractions with antimicrobial activity. For a more comprehensive analysis, a data analysis pipeline was assembled to prioritize compounds for antimicrobial testing against both MRSA and methicillin-sensitive S. aureus using forty-two unique fractions of the tea extract that were generated in duplicate, assayed for activity, and analyzed with GC-MS. As validation of our automated analysis, we checked our predicted active compounds for activity in literature references and used authentic standards to test for antimicrobial activity. 3,4-dihydroxybenzaldehyde showed the most antibacterial activity against MRSA at low concentrations in our bioassays. In addition, quinic acid and quercetin were identified using random forests analysis and 5-hydroxy pipecolic acid was identified using linear discriminant analysis. We also generated a ranked list of unidentified compounds that may contribute to the antimicrobial activity of yerba mate against MRSA. Here we utilized GC-MS data to implement an automated analysis that resulted in a ranked list of compounds that likely contribute to the antimicrobial activity of aqueous yerba mate extract against MRSA.

  18. Computer programs for forward and inverse modeling of acoustic and electromagnetic data (United States)

    Ellefsen, Karl J.


    A suite of computer programs was developed by U.S. Geological Survey personnel for forward and inverse modeling of acoustic and electromagnetic data. This report describes the computer resources that are needed to execute the programs, the installation of the programs, the program designs, some tests of their accuracy, and some suggested improvements.

  19. Academic rankings: an approach to a Portuguese ranking


    Bernardino, Pedro; Marques,Rui


    The academic rankings are a controversial subject in higher education. However, despite all the criticism, academic rankings are here to stay and more and more different stakeholders use rankings to obtain information about the institutions’ performance. The two most well-known rankings, The Times and the Shanghai Jiao Tong University rankings have different methodologies. The Times ranking is based on peer review, whereas the Shanghai ranking has only quantitative indicators and is mainly ba...

  20. Threshold evaluation data revision and computer program enhancement. Final report

    Energy Technology Data Exchange (ETDEWEB)


    The Threshold Evaluation System was developed to assist the Division of Buildings and Community Systems of the Department of Energy in performing preliminary evaluation of projects being considered for funding. In addition, the evaluation has been applied to on-going projects, because information obtained through RD and D may alter the expected benefits and costs of a project, making it necessary to reevaluate project funding. The system evaluates each project according to its expected energy savings and costs. A number of public and private sector criteria are calculated, upon which comparisons between projects may be based. A summary of the methodology is given in Appendix B. The purpose of this task is to upgrade both the quality of the data used for input to the system and the usefulness and efficiency of the computer program used to perform the analysis. The modifications required to produce a better, more consistent set of data are described in Section 2. Program changes that have had a significant impact on the methodology are discussed in Section 3, while those that affected only the computer code are presented as a system flow diagram and program listing in Appendix C. These improvements in the project evaluation methodology and data will provide BCS with a more efficient and comprehensive management tool. The direction of future work will be toward integrating this system with a large scale (at ORNL) so that information used by both systems may be stored in a common data base. A discussion of this, and other unresolved problems is given in Section 4.

  1. Frequency Domain Computer Programs for Prediction and Analysis of Rail Vehicle Dynamics : Volume 2. Appendixes (United States)


    Frequency domain computer programs developed or acquired by TSC for the analysis of rail vehicle dynamics are described in two volumes. Volume 2 contains program listings including subroutines for the four TSC frequency domain programs described in V...

  2. Second Annual AEC Scientific Computer Information Exhange Meeting. Proceedings of the technical program theme: computer graphics

    Energy Technology Data Exchange (ETDEWEB)

    Peskin,A.M.; Shimamoto, Y.


    The topic of computer graphics serves well to illustrate that AEC affiliated scientific computing installations are well represented in the forefront of computing science activities. The participant response to the technical program was overwhelming--both in number of contributions and quality of the work described. Session I, entitled Advanced Systems, contains presentations describing systems that contain features not generally found in graphics facilities. These features can be roughly classified as extensions of standard two-dimensional monochromatic imaging to higher dimensions including color and time as well as multidimensional metrics. Session II presents seven diverse applications ranging from high energy physics to medicine. Session III describes a number of important developments in establishing facilities, techniques and enhancements in the computer graphics area. Although an attempt was made to schedule as many of these worthwhile presentations as possible, it appeared impossible to do so given the scheduling constraints of the meeting. A number of prospective presenters 'came to the rescue' by graciously withdrawing from the sessions. Some of their abstracts have been included in the Proceedings.

  3. Rankings matter: nurse graduates from higher-ranked institutions have higher productivity. (United States)

    Yakusheva, Olga; Weiss, Marianne


    Increasing demand for baccalaureate-prepared nurses has led to rapid growth in the number of baccalaureate-granting programs, and to concerns about educational quality and potential effects on productivity of the graduating nursing workforce. We examined the association of individual productivity of a baccalaureate-prepared nurse with the ranking of the degree-granting institution. For a sample of 691 nurses from general medical-surgical units at a large magnet urban hospital between 6/1/2011-12/31/2011, we conducted multivariate regression analysis of nurse productivity on the ranking of the degree-granting institution, adjusted for age, hospital tenure, gender, and unit-specific effects. Nurse productivity was coded as "top"/"average"/"bottom" based on a computation of individual nurse value-added to patient outcomes. Ranking of the baccalaureate-granting institution was derived from the US News and World Report Best Colleges Rankings' categorization of the nurse's institution as the "first tier" or the "second tier", with diploma or associate degree as the reference category. Relative to diploma or associate degree nurses, nurses who had attended first-tier universities had three-times the odds of being in the top productivity category (OR = 3.18, p productivity (OR = 1.73, p = 0.11). Being in the bottom productivity category was not associated with having a baccalaureate degree or the quality tier. The productivity boost from a nursing baccalaureate degree depends on the quality of the educational institution. Recognizing differences in educational outcomes, initiatives to build a baccalaureate-educated nursing workforce should be accompanied by improved access to high-quality educational institutions.

  4. Multi-Language Programming Environments for High Performance Java Computing

    Directory of Open Access Journals (Sweden)

    Vladimir Getov


    Full Text Available Recent developments in processor capabilities, software tools, programming languages and programming paradigms have brought about new approaches to high performance computing. A steadfast component of this dynamic evolution has been the scientific community’s reliance on established scientific packages. As a consequence, programmers of high‐performance applications are reluctant to embrace evolving languages such as Java. This paper describes the Java‐to‐C Interface (JCI tool which provides application programmers wishing to use Java with immediate accessibility to existing scientific packages. The JCI tool also facilitates rapid development and reuse of existing code. These benefits are provided at minimal cost to the programmer. While beneficial to the programmer, the additional advantages of mixed‐language programming in terms of application performance and portability are addressed in detail within the context of this paper. In addition, we discuss how the JCI tool is complementing other ongoing projects such as IBM’s High‐Performance Compiler for Java (HPCJ and IceT’s metacomputing environment.

  5. Ranking Fuzzy Numbers and Its Application to Products Attributes Preferences


    Lazim Abdullah; Nor Nashrah Ahmad Fauzee


    Ranking is one of the widely used methods in fuzzy decision making environment. The recent ranking fuzzy numbers proposed by Wang and Li is claimed to be the improved version in ranking. However, the method was never been simplified and tested in real life application. This paper presents a four-step computation of ranking fuzzy numbers and its application in ranking attributes of selected chocolate products.  The four steps algorithm was formulated to rank fuzzy numbers and followed by a tes...

  6. Ranking Economic History Journals

    DEFF Research Database (Denmark)

    Di Vaio, Gianfranco; Weisdorf, Jacob Louis

    This study ranks - for the first time - 12 international academic journals that have economic history as their main topic. The ranking is based on data collected for the year 2007. Journals are ranked using standard citation analysis where we adjust for age, size and self-citation of journals. We...... also compare the leading economic history journals with the leading journals in economics in order to measure the influence on economics of economic history, and vice versa. With a few exceptions, our results confirm the general idea about what economic history journals are the most influential...

  7. Ranking economic history journals

    DEFF Research Database (Denmark)

    Di Vaio, Gianfranco; Weisdorf, Jacob Louis


    This study ranks-for the first time-12 international academic journals that have economic history as their main topic. The ranking is based on data collected for the year 2007. Journals are ranked using standard citation analysis where we adjust for age, size and self-citation of journals. We also...... compare the leading economic history journals with the leading journals in economics in order to measure the influence on economics of economic history, and vice versa. With a few exceptions, our results confirm the general idea about what economic history journals are the most influential for economic...

  8. Recurrent fuzzy ranking methods (United States)

    Hajjari, Tayebeh


    With the increasing development of fuzzy set theory in various scientific fields and the need to compare fuzzy numbers in different areas. Therefore, Ranking of fuzzy numbers plays a very important role in linguistic decision-making, engineering, business and some other fuzzy application systems. Several strategies have been proposed for ranking of fuzzy numbers. Each of these techniques has been shown to produce non-intuitive results in certain case. In this paper, we reviewed some recent ranking methods, which will be useful for the researchers who are interested in this area.


    Hui, J.


    KNET facilitates distributed computing between a UNIX compatible local host and a remote host which may or may not be UNIX compatible. It is capable of automatic remote login. That is, it performs on the user's behalf the chore of handling host selection, user name, and password to the designated host. Once the login has been successfully completed, the user may interactively communicate with the remote host. Data output from the remote host may be directed to the local screen, to a local file, and/or to a local process. Conversely, data input from the keyboard, a local file, or a local process may be directed to the remote host. KNET takes advantage of the multitasking and terminal mode control features of the UNIX operating system. A parent process is used as the upper layer for interfacing with the local user. A child process is used for a lower layer for interfacing with the remote host computer, and optionally one or more child processes can be used for the remote data output. Output may be directed to the screen and/or to the local processes under the control of a data pipe switch. In order for KNET to operate, the local and remote hosts must observe a common communications protocol. KNET is written in ANSI standard C-language for computers running UNIX. It has been successfully implemented on several Sun series computers and a DECstation 3100 and used to run programs remotely on VAX VMS and UNIX based computers. It requires 100K of RAM under SunOS and 120K of RAM under DEC RISC ULTRIX. An electronic copy of the documentation is provided on the distribution medium. The standard distribution medium for KNET is a .25 inch streaming magnetic tape cartridge in UNIX tar format. It is also available on a 3.5 inch diskette in UNIX tar format. KNET was developed in 1991 and is a copyrighted work with all copyright vested in NASA. UNIX is a registered trademark of AT&T Bell Laboratories. Sun and SunOS are trademarks of Sun Microsystems, Inc. DECstation, VAX, VMS, and

  10. PageRank as a method to rank biomedical literature by importance. (United States)

    Yates, Elliot J; Dixon, Louise C


    Optimal ranking of literature importance is vital in overcoming article overload. Existing ranking methods are typically based on raw citation counts, giving a sum of 'inbound' links with no consideration of citation importance. PageRank, an algorithm originally developed for ranking webpages at the search engine, Google, could potentially be adapted to bibliometrics to quantify the relative importance weightings of a citation network. This article seeks to validate such an approach on the freely available, PubMed Central open access subset (PMC-OAS) of biomedical literature. On-demand cloud computing infrastructure was used to extract a citation network from over 600,000 full-text PMC-OAS articles. PageRanks and citation counts were calculated for each node in this network. PageRank is highly correlated with citation count (R = 0.905, P PageRank can be trivially computed on commodity cluster hardware and is linearly correlated with citation count. Given its putative benefits in quantifying relative importance, we suggest it may enrich the citation network, thereby overcoming the existing inadequacy of citation counts alone. We thus suggest PageRank as a feasible supplement to, or replacement of, existing bibliometric ranking methods.

  11. Aether: Leveraging Linear Programming For Optimal Cloud Computing In Genomics. (United States)

    Luber, Jacob M; Tierney, Braden T; Cofer, Evan M; Patel, Chirag J; Kostic, Aleksandar D


    Across biology we are seeing rapid developments in scale of data production without a corresponding increase in data analysis capabilities. Here, we present Aether (, an intuitive, easy-to-use, cost-effective, and scalable framework that uses linear programming (LP) to optimally bid on and deploy combinations of underutilized cloud computing resources. Our approach simultaneously minimizes the cost of data analysis and provides an easy transition from users' existing HPC pipelines. Data utilized are available at and with EBI SRA accession ERP005989. Source code is available at ( Examples, documentation, and a tutorial are available at ( and Supplementary data are available at Bioinformatics online.

  12. Interactive, Computer-Based Training Program for Radiological Workers

    Energy Technology Data Exchange (ETDEWEB)

    Trinoskey, P.A.; Camacho, P.I.; Wells, L.


    Lawrence Livermore National Laboratory (LLNL) is redesigning its Computer-Based Training (CBT) program for radiological workers. The redesign represents a major effort to produce a single, highly interactive and flexible CBT program that will meet the training needs of a wide range of radiological workers--from researchers and x-ray operators to individuals working in tritium, uranium, plutonium, and accelerator facilities. The new CBT program addresses the broad diversity of backgrounds found at a national laboratory. When a training audience is homogeneous in terms of education level and type of work performed, it is difficult to duplicate the effectiveness of a flexible, technically competent instructor who can tailor a course to the express needs and concerns of a course's participants. Unfortunately, such homogeneity is rare. At LLNL, they have a diverse workforce engaged in a wide range of radiological activities, from the fairly common to the quite exotic. As a result, the Laboratory must offer a wide variety of radiological worker courses. These include a general contamination-control course in addition to radioactive-material-handling courses for both low-level laboratory (i.e., bench-top) activities as well as high-level work in tritium, uranium, and plutonium facilities. They also offer training courses for employees who work with radiation-generating devices--x-ray, accelerator, and E-beam operators, for instance. However, even with the number and variety of courses the Laboratory offers, they are constrained by the diversity of backgrounds (i.e., knowledge and experience) of those to be trained. Moreover, time constraints often preclude in-depth coverage of site- and/or task-specific details. In response to this situation, several years ago LLNL began moving toward computer-based training for radiological workers. Today, that CBT effort includes a general radiological safety course developed by the Department of Energy's Hanford facility and

  13. The Julia programming language: the future of scientific computing (United States)

    Gibson, John


    Julia is an innovative new open-source programming language for high-level, high-performance numerical computing. Julia combines the general-purpose breadth and extensibility of Python, the ease-of-use and numeric focus of Matlab, the speed of C and Fortran, and the metaprogramming power of Lisp. Julia uses type inference and just-in-time compilation to compile high-level user code to machine code on the fly. A rich set of numeric types and extensive numerical libraries are built-in. As a result, Julia is competitive with Matlab for interactive graphical exploration and with C and Fortran for high-performance computing. This talk interactively demonstrates Julia's numerical features and benchmarks Julia against C, C++, Fortran, Matlab, and Python on a spectral time-stepping algorithm for a 1d nonlinear partial differential equation. The Julia code is nearly as compact as Matlab and nearly as fast as Fortran. This material is based upon work supported by the National Science Foundation under Grant No. 1554149.

  14. A program for reading DNA sequence gels using a small computer equipped with a graphics tablet. (United States)

    Lautenberger, J A


    A program has been written in BASIC that allows DNA sequence gels to be read by a Tektronix model 4052 computer equipped with a graphics tablet. Sequences from each gel are stored on tape for later transfer to a larger computer where they are melded into a complete overall sequence. The program should be adaptable to other small computers. PMID:7063401

  15. A Study of the Programming Languages Used in Information Systems and in Computer Science Curricula (United States)

    Russell, Jack; Russell, Barbara; Pollacia, Lissa F.; Tastle, William J.


    This paper researches the computer languages taught in the first, second and third programming courses in Computer Information Systems (CIS), Management Information Systems (MIS or IS) curricula as well as in Computer Science (CS) and Information Technology (IT) curricula. Instructors teaching the first course in programming within a four year…

  16. TRECII: a computer program for transportation risk assessment

    Energy Technology Data Exchange (ETDEWEB)

    Franklin, A.L.


    A risk-based fault tree analysis method has been developed at the Pacific Northwest Laboratory (PNL) for analysis of nuclear fuel cycle operations. This methodology was developed for the Department of Energy (DOE) as a risk analysis tool for evaluating high level waste management systems. A computer package consisting of three programs was written at that time to assist in the performance of risk assessment: ACORN (draws fault trees), MFAULT (analyzes fault trees), and RAFT (calculates risk). This methodology evaluates release consequences and estimates the frequency of occurrence of these consequences. This document describes an additional risk calculating code which can be used in conjunction with two of the three codes for transportation risk assessment. TRECII modifies the definition of risk used in RAFT (prob. x release) to accommodate release consequences in terms of fatalities. Throughout this report risk shall be defined as probability times consequences (fatalities are one possible health effect consequence). This methodology has been applied to a variety of energy material transportation systems. Typically the material shipped has been radioactive, although some adaptation to fossil fuels has occurred. The approach is normally applied to truck or train transport systems with some adaptation to pipelines and aircraft. TRECII is designed to be used primarily in conjunction with MFAULT; however, with a moderate amount of effort by the user, it can be implemented independent of the risk analysis package developed at PNL. Code description and user instructions necessary for the implementation of the TRECII program are provided.

  17. Execution Models for Mapping Programs onto Distributed Memory Parallel Computers (United States)


    DISTRIBUTED MEMORY PARALLEL COMPUTERS Alan Sussman Contract No. NAS1-18605 March 1992 Institute for Computer Applications in Science and Engineering NASA...MEMORY PARALLEL COMPUTERS Alan Sussman 1 Institute for Computer Applications in Science and Engineering NASA Langley Research Center Hampton, VA 23665...Computation onto Distributed Mem- ory Parallel Computers . PhD thesis, Carnegie Mellon University, September 1991. Also available as Technical Report

  18. Asset ranking manager (ranking index of components)

    Energy Technology Data Exchange (ETDEWEB)

    Maloney, S.M.; Engle, A.M.; Morgan, T.A. [Applied Reliability, Maracor Software and Engineering (United States)


    The Ranking Index of Components (RIC) is an Asset Reliability Manager (ARM), which itself is a Web Enabled front end where plant database information fields from several disparate databases are combined. That information is used to create a specific weighted number (Ranking Index) relating to that components health and risk to the site. The higher the number, the higher priority that any work associated with that component receives. ARM provides site Engineering, Maintenance and Work Control personnel with a composite real time - (current condition) look at the components 'risk of not working' to the plant. Information is extracted from the existing Computerized Maintenance management System (CMMS) and specific site applications and processed nightly. ARM helps to ensure that the most important work is placed into the workweeks and the non value added work is either deferred, frequency changed or deleted. This information is on the web, updated each night, and available for all employees to use. This effort assists the work management specialist when allocating limited resources to the most important work. The use of this tool has maximized resource usage, performing the most critical work with available resources. The ARM numbers are valued inputs into work scoping for the workweek managers. System and Component Engineers are using ARM to identify the components that are at 'risk of failure' and therefore should be placed into the appropriate work week schedule.

  19. Multithreaded transactions in scientific computing: New versions of a computer program for kinematical calculations of RHEED intensity oscillations (United States)

    Brzuszek, Marcin; Daniluk, Andrzej


    Writing a concurrent program can be more difficult than writing a sequential program. Programmer needs to think about synchronisation, race conditions and shared variables. Transactions help reduce the inconvenience of using threads. A transaction is an abstraction, which allows programmers to group a sequence of actions on the program into a logical, higher-level computation unit. This paper presents multithreaded versions of the GROWTH program, which allow to calculate the layer coverages during the growth of thin epitaxial films and the corresponding RHEED intensities according to the kinematical approximation. The presented programs also contain graphical user interfaces, which enable displaying program data at run-time. New version program summaryTitles of programs:GROWTHGr, GROWTH06 Catalogue identifier:ADVL_v2_0 Program summary URL: Program obtainable from:CPC Program Library, Queen's University of Belfast, N. Ireland Catalogue identifier of previous version:ADVL Does the new version supersede the original program:No Computer for which the new version is designed and others on which it has been tested: Pentium-based PC Operating systems or monitors under which the new version has been tested: Windows 9x, XP, NT Programming language used:Object Pascal Memory required to execute with typical data:More than 1 MB Number of bits in a word:64 bits Number of processors used:1 No. of lines in distributed program, including test data, etc.:20 931 Number of bytes in distributed program, including test data, etc.: 1 311 268 Distribution format:tar.gz Nature of physical problem: The programs compute the RHEED intensities during the growth of thin epitaxial structures prepared using the molecular beam epitaxy (MBE). The computations are based on the use of kinematical diffraction theory [P.I. Cohen, G.S. Petrich, P.R. Pukite, G.J. Whaley, A.S. Arrott, Surf. Sci. 216 (1989) 222. [1

  20. Middle School Teachers' Perceptions of Computer-Assisted Reading Intervention Programs (United States)

    Bippert, Kelli; Harmon, Janis


    Middle schools often turn to computer-assisted reading intervention programs to improve student reading. The questions guiding this study are (a) in what ways are computer-assisted reading intervention programs utilized, and (b) what are teachers' perceptions about these intervention programs? Nineteen secondary reading teachers were interviewed…

  1. Design and Curriculum Considerations for a Computer Graphics Program in the Arts. (United States)

    Leeman, Ruedy W.

    This history and state-of-the-art review of computer graphics describes computer graphics programs and proposed programs at Sheridan College (Canada), the Rhode Island School of Design, the University of Oregon, Northern Illinois University, and Ohio State University. These programs are discussed in terms of their philosophy, curriculum, student…

  2. Case Study: Creation of a Degree Program in Computer Security. White Paper. (United States)

    Belon, Barbara; Wright, Marie

    This paper reports on research into the field of computer security, and undergraduate degrees offered in that field. Research described in the paper reveals only one computer security program at the associate's degree level in the entire country. That program, at Texas State Technical College in Waco, is a 71-credit-hour program leading to an…

  3. The DOE Program in HPCC: High-Performance Computing and Communications. (United States)

    Department of Energy, Washington, DC. Office of Energy Research.

    This document reports to Congress on the progress that the Department of Energy has made in 1992 toward achieving the goals of the High Performance Computing and Communications (HPCC) program. Its second purpose is to provide a picture of the many programs administered by the Office of Scientific Computing under the auspices of the HPCC program.…

  4. Children Learning Computer Programming: Experiments with Languages, Curricula and Programmable Devices. Technical Report No. 250. (United States)

    Weyer, S. A.; Cannara, A. B.

    An experiment was conducted to study how children, aged 10-15 years, learn concepts relevant to computer programing and how they learn modern programing languages. The implicit educational goal was to teach thinking strategies through the medium of programing concepts and their applications. The computer languages Simper and Logo were chosen…

  5. Combined Reduced-Rank Transform

    Directory of Open Access Journals (Sweden)

    Anatoli Torokhti


    Full Text Available We propose and justify a new approach to constructing optimal nonlinear transforms of random vectors. We show that the proposed transform improves such characteristics of {rank-reduced} transforms as compression ratio, accuracy of decompression and reduces required computational work. The proposed transform ${mathcal T}_p$ is presented in the form of a sum with $p$ terms where each term is interpreted as a particular rank-reduced transform. Moreover, terms in ${mathcal T}_p$ are represented as a combination of three operations ${mathcal F}_k$, ${mathcal Q}_k$ and ${oldsymbol{varphi}}_k$ with $k=1,ldots,p$. The prime idea is to determine ${mathcal F}_k$ separately, for each $k=1,ldots,p$, from an associated rank-constrained minimization problem similar to that used in the Karhunen--Lo`{e}ve transform. The operations ${mathcal Q}_k$ and ${oldsymbol{varphi}}_k$ are auxiliary for f/inding ${mathcal F}_k$. The contribution of each term in ${mathcal T}_p$ improves the entire transform performance. A corresponding unconstrained nonlinear optimal transform is also considered. Such a transform is important in its own right because it is treated as an optimal filter without signal compression. A rigorous analysis of errors associated with the proposed transforms is given.

  6. Multiplex PageRank.

    Directory of Open Access Journals (Sweden)

    Arda Halu

    Full Text Available Many complex systems can be described as multiplex networks in which the same nodes can interact with one another in different layers, thus forming a set of interacting and co-evolving networks. Examples of such multiplex systems are social networks where people are involved in different types of relationships and interact through various forms of communication media. The ranking of nodes in multiplex networks is one of the most pressing and challenging tasks that research on complex networks is currently facing. When pairs of nodes can be connected through multiple links and in multiple layers, the ranking of nodes should necessarily reflect the importance of nodes in one layer as well as their importance in other interdependent layers. In this paper, we draw on the idea of biased random walks to define the Multiplex PageRank centrality measure in which the effects of the interplay between networks on the centrality of nodes are directly taken into account. In particular, depending on the intensity of the interaction between layers, we define the Additive, Multiplicative, Combined, and Neutral versions of Multiplex PageRank, and show how each version reflects the extent to which the importance of a node in one layer affects the importance the node can gain in another layer. We discuss these measures and apply them to an online multiplex social network. Findings indicate that taking the multiplex nature of the network into account helps uncover the emergence of rankings of nodes that differ from the rankings obtained from one single layer. Results provide support in favor of the salience of multiplex centrality measures, like Multiplex PageRank, for assessing the prominence of nodes embedded in multiple interacting networks, and for shedding a new light on structural properties that would otherwise remain undetected if each of the interacting networks were analyzed in isolation.

  7. Multiplex PageRank. (United States)

    Halu, Arda; Mondragón, Raúl J; Panzarasa, Pietro; Bianconi, Ginestra


    Many complex systems can be described as multiplex networks in which the same nodes can interact with one another in different layers, thus forming a set of interacting and co-evolving networks. Examples of such multiplex systems are social networks where people are involved in different types of relationships and interact through various forms of communication media. The ranking of nodes in multiplex networks is one of the most pressing and challenging tasks that research on complex networks is currently facing. When pairs of nodes can be connected through multiple links and in multiple layers, the ranking of nodes should necessarily reflect the importance of nodes in one layer as well as their importance in other interdependent layers. In this paper, we draw on the idea of biased random walks to define the Multiplex PageRank centrality measure in which the effects of the interplay between networks on the centrality of nodes are directly taken into account. In particular, depending on the intensity of the interaction between layers, we define the Additive, Multiplicative, Combined, and Neutral versions of Multiplex PageRank, and show how each version reflects the extent to which the importance of a node in one layer affects the importance the node can gain in another layer. We discuss these measures and apply them to an online multiplex social network. Findings indicate that taking the multiplex nature of the network into account helps uncover the emergence of rankings of nodes that differ from the rankings obtained from one single layer. Results provide support in favor of the salience of multiplex centrality measures, like Multiplex PageRank, for assessing the prominence of nodes embedded in multiple interacting networks, and for shedding a new light on structural properties that would otherwise remain undetected if each of the interacting networks were analyzed in isolation.

  8. Ranking of Rankings: Benchmarking Twenty-Five Higher Education Ranking Systems in Europe (United States)

    Stolz, Ingo; Hendel, Darwin D.; Horn, Aaron S.


    The purpose of this study is to evaluate the ranking practices of 25 European higher education ranking systems (HERSs). Ranking practices were assessed with 14 quantitative measures derived from the Berlin Principles on Ranking of Higher Education Institutions (BPs). HERSs were then ranked according to their degree of congruence with the BPs.…

  9. A tilting approach to ranking influence

    KAUST Repository

    Genton, Marc G.


    We suggest a new approach, which is applicable for general statistics computed from random samples of univariate or vector-valued or functional data, to assessing the influence that individual data have on the value of a statistic, and to ranking the data in terms of that influence. Our method is based on, first, perturbing the value of the statistic by ‘tilting’, or reweighting, each data value, where the total amount of tilt is constrained to be the least possible, subject to achieving a given small perturbation of the statistic, and, then, taking the ranking of the influence of data values to be that which corresponds to ranking the changes in data weights. It is shown, both theoretically and numerically, that this ranking does not depend on the size of the perturbation, provided that the perturbation is sufficiently small. That simple result leads directly to an elegant geometric interpretation of the ranks; they are the ranks of the lengths of projections of the weights onto a ‘line’ determined by the first empirical principal component function in a generalized measure of covariance. To illustrate the generality of the method we introduce and explore it in the case of functional data, where (for example) it leads to generalized boxplots. The method has the advantage of providing an interpretable ranking that depends on the statistic under consideration. For example, the ranking of data, in terms of their influence on the value of a statistic, is different for a measure of location and for a measure of scale. This is as it should be; a ranking of data in terms of their influence should depend on the manner in which the data are used. Additionally, the ranking recognizes, rather than ignores, sign, and in particular can identify left- and right-hand ‘tails’ of the distribution of a random function or vector.

  10. Debugging the Program. Computer Equity Strategies for the Classroom Teacher. (United States)

    Wolfe, Leslie R.; And Others

    Designed to provide classroom teachers with activities to enhance computer equity for female students, this kit is divided into four sections which present excerpts from four other publications: (1) "The Neuter Computer: Computers for Boys and Girls" (Jo Schuchat Sanders and Antonia Stone for the Computer Equity Training Project, Women's…

  11. A Financial Technology Entrepreneurship Program for Computer Science Students (United States)

    Lawler, James P.; Joseph, Anthony


    Education in entrepreneurship is becoming a critical area of curricula for computer science students. Few schools of computer science have a concentration in entrepreneurship in the computing curricula. The paper presents Technology Entrepreneurship in the curricula at a leading school of computer science and information systems, in which students…

  12. A new kind of science - or a not-so-new kind of computer program?

    CERN Multimedia

    Naiditch, D


    Critical discussion of Stephen Wolfram's theory that "scientists give up their unwieldy equations and instead employ the types of computational rules used in cellular automata (CA) and related computer programs" (2 pages).

  13. Computation of vibration mode elastic-rigid and effective weight coefficients from finite-element computer program output (United States)

    Levy, R.


    Post-processing algorithms are given to compute the vibratory elastic-rigid coupling matrices and the modal contributions to the rigid-body mass matrices and to the effective modal inertias and masses. Recomputation of the elastic-rigid coupling matrices for a change in origin is also described. A computational example is included. The algorithms can all be executed by using standard finite-element program eigenvalue analysis output with no changes to existing code or source programs.

  14. Possible means of legal protection for computer programs and perspectives of future development


    Toufar, Pavel


    This theses focuses on the possible legal protection of a computer program as well as on the legal nature of computer program as an intangible asset. Both copyright protection (as a standard and worldwide accepted means of protection) and also the other possibilities, i.e. patent protection and protection based on provisions regulating an unfair competition are discussed. Each means of protection is assessed based on its usability in relation with the computer program taking the overall impac...

  15. Playable Serious Games for Studying and Programming Computational STEM and Informatics Applications of Distributed and Parallel Computer Architectures (United States)

    Amenyo, John-Thones


    Carefully engineered playable games can serve as vehicles for students and practitioners to learn and explore the programming of advanced computer architectures to execute applications, such as high performance computing (HPC) and complex, inter-networked, distributed systems. The article presents families of playable games that are grounded in…

  16. Proceedings of seventh symposium on sharing of computer programs and technology in nuclear medicine, computer assisted data processing

    Energy Technology Data Exchange (ETDEWEB)

    Howard, B.Y.; McClain, W.J.; Landay, M. (comps.)


    The Council on Computers (CC) of the Society of Nuclear Medicine (SNM) annually publishes the Proceedings of its Symposium on the Sharing of Computer Programs and Technology in Nuclear Medicine. This is the seventh such volume and has been organized by topic, with the exception of the invited papers and the discussion following them. An index arranged by author and by subject is included.

  17. When technology became language: the origins of the linguistic conception of computer programming, 1950-1960. (United States)

    Nofre, David; Priestley, Mark; Alberts, Gerard


    Language is one of the central metaphors around which the discipline of computer science has been built. The language metaphor entered modern computing as part of a cybernetic discourse, but during the second half of the 1950s acquired a more abstract meaning, closely related to the formal languages of logic and linguistics. The article argues that this transformation was related to the appearance of the commercial computer in the mid-1950s. Managers of computing installations and specialists on computer programming in academic computer centers, confronted with an increasing variety of machines, called for the creation of "common" or "universal languages" to enable the migration of computer code from machine to machine. Finally, the article shows how the idea of a universal language was a decisive step in the emergence of programming languages, in the recognition of computer programming as a proper field of knowledge, and eventually in the way we think of the computer.

  18. Thinking processes used by high-performing students in a computer programming task

    Directory of Open Access Journals (Sweden)

    Marietjie Havenga


    Full Text Available Computer programmers must be able to understand programming source code and write programs that execute complex tasks to solve real-world problems. This article is a trans- disciplinary study at the intersection of computer programming, education and psychology. It outlines the role of mental processes in the process of programming and indicates how successful thinking processes can support computer science students in writing correct and well-defined programs. A mixed methods approach was used to better understand the thinking activities and programming processes of participating students. Data collection involved both computer programs and students’ reflective thinking processes recorded in their journals. This enabled analysis of psychological dimensions of participants’ thinking processes and their problem-solving activities as they considered a programming problem. Findings indicate that the cognitive, reflective and psychological processes used by high-performing programmers contributed to their success in solving a complex programming problem. Based on the thinking processes of high performers, we propose a model of integrated thinking processes, which can support computer programming students. Keywords: Computer programming, education, mixed methods research, thinking processes.  Disciplines: Computer programming, education, psychology

  19. Global Low-Rank Image Restoration With Gaussian Mixture Model. (United States)

    Zhang, Sibo; Jiao, Licheng; Liu, Fang; Wang, Shuang


    Low-rank restoration has recently attracted a lot of attention in the research of computer vision. Empirical studies show that exploring the low-rank property of the patch groups can lead to superior restoration performance, however, there is limited achievement on the global low-rank restoration because the rank minimization at image level is too strong for the natural images which seldom match the low-rank condition. In this paper, we describe a flexible global low-rank restoration model which introduces the local statistical properties into the rank minimization. The proposed model can effectively recover the latent global low-rank structure via nuclear norm, as well as the fine details via Gaussian mixture model. An alternating scheme is developed to estimate the Gaussian parameters and the restored image, and it shows excellent convergence and stability. Besides, experiments on image and video sequence datasets show the effectiveness of the proposed method in image inpainting problems.

  20. Automation and schema acquisition in learning elementary computer programming : implications for the design of practice

    NARCIS (Netherlands)

    van Merrienboer, Jeroen J.G.; van Merrienboer, J.J.G.; Paas, Fred G.W.C.


    Two complementary processes may be distinguished in learning a complex cognitive skill such as computer programming. First, automation offers task-specific procedures that may directly control programming behavior, second, schema acquisition offers cognitive structures that provide analogies in new

  1. Who's bigger? where historical figures really rank

    CERN Document Server

    Skiena, Steven


    Is Hitler bigger than Napoleon? Washington bigger than Lincoln? Picasso bigger than Einstein? Quantitative analysts are rapidly finding homes in social and cultural domains, from finance to politics. What about history? In this fascinating book, Steve Skiena and Charles Ward bring quantitative analysis to bear on ranking and comparing historical reputations. They evaluate each person by aggregating the traces of millions of opinions, just as Google ranks webpages. The book includes a technical discussion for readers interested in the details of the methods, but no mathematical or computational background is necessary to understand the rankings or conclusions. Along the way, the authors present the rankings of more than one thousand of history's most significant people in science, politics, entertainment, and all areas of human endeavor. Anyone interested in history or biography can see where their favorite figures place in the grand scheme of things.

  2. The Application of Visual Basic Computer Programming Language to Simulate Numerical Iterations

    Directory of Open Access Journals (Sweden)

    Abdulkadir Baba HASSAN


    Full Text Available This paper examines the application of Visual Basic Computer Programming Language to Simulate Numerical Iterations, the merit of Visual Basic as a Programming Language and the difficulties faced when solving numerical iterations analytically, this research paper encourage the uses of Computer Programming methods for the execution of numerical iterations and finally fashion out and develop a reliable solution using Visual Basic package to write a program for some selected iteration problems.

  3. Federal High Performance Computing and Communications Program. The Department of Energy Component. (United States)

    Department of Energy, Washington, DC. Office of Energy Research.

    This report, profusely illustrated with color photographs and other graphics, elaborates on the Department of Energy (DOE) research program in High Performance Computing and Communications (HPCC). The DOE is one of seven agency programs within the Federal Research and Development Program working on HPCC. The DOE HPCC program emphasizes research in…

  4. Two Computer Programs for Equipment Cost Estimation and Economic Evaluation of Chemical Processes. (United States)

    Kuri, Carlos J.; Corripio, Armando B.


    Describes two computer programs for use in process design courses: an easy-to-use equipment cost estimation program based on latest cost correlations available and an economic evaluation program which calculates two profitability indices. Comparisons between programed and hand-calculated results are included. (JM)

  5. Partial Kernelization for Rank Aggregation: Theory and Experiments (United States)

    Betzler, Nadja; Bredereck, Robert; Niedermeier, Rolf

    Rank Aggregation is important in many areas ranging from web search over databases to bioinformatics. The underlying decision problem Kemeny Score is NP-complete even in case of four input rankings to be aggregated into a "median ranking". We study efficient polynomial-time data reduction rules that allow us to find optimal median rankings. On the theoretical side, we improve a result for a "partial problem kernel" from quadratic to linear size. On the practical side, we provide encouraging experimental results with data based on web search and sport competitions, e.g., computing optimal median rankings for real-world instances with more than 100 candidates within milliseconds.

  6. Diversifying customer review rankings. (United States)

    Krestel, Ralf; Dokoohaki, Nima


    E-commerce Web sites owe much of their popularity to consumer reviews accompanying product descriptions. On-line customers spend hours and hours going through heaps of textual reviews to decide which products to buy. At the same time, each popular product has thousands of user-generated reviews, making it impossible for a buyer to read everything. Current approaches to display reviews to users or recommend an individual review for a product are based on the recency or helpfulness of each review. In this paper, we present a framework to rank product reviews by optimizing the coverage of the ranking with respect to sentiment or aspects, or by summarizing all reviews with the top-K reviews in the ranking. To accomplish this, we make use of the assigned star rating for a product as an indicator for a review's sentiment polarity and compare bag-of-words (language model) with topic models (latent Dirichlet allocation) as a mean to represent aspects. Our evaluation on manually annotated review data from a commercial review Web site demonstrates the effectiveness of our approach, outperforming plain recency ranking by 30% and obtaining best results by combining language and topic model representations. Copyright © 2015 Elsevier Ltd. All rights reserved.

  7. OutRank

    DEFF Research Database (Denmark)

    Müller, Emmanuel; Assent, Ira; Steinhausen, Uwe


    Outlier detection is an important data mining task for consistency checks, fraud detection, etc. Binary decision making on whether or not an object is an outlier is not appropriate in many applications and moreover hard to parametrize. Thus, recently, methods for outlier ranking have been proposed...

  8. Improving Ranking Using Quantum Probability


    Melucci, Massimo


    The paper shows that ranking information units by quantum probability differs from ranking them by classical probability provided the same data used for parameter estimation. As probability of detection (also known as recall or power) and probability of false alarm (also known as fallout or size) measure the quality of ranking, we point out and show that ranking by quantum probability yields higher probability of detection than ranking by classical probability provided a given probability of ...

  9. DEP : a computer program for evaluating lumber drying costs and investments (United States)

    Stewart Holmes; George B. Harpole; Edward Bilek


    The DEP computer program is a modified discounted cash flow computer program designed for analysis of problems involving economic analysis of wood drying processes. Wood drying processes are different from other processes because of the large amounts of working capital required to finance inventories, and because of relatively large shares of costs charged to inventory...

  10. Computer Literacy of Turkish Preservice Teachers in Different Teacher Training Programs (United States)

    Ozsevgec, Tuncay


    This paper reports on an investigation into the sophomore and senior preservice teachers' computer literacy in different teacher training programs and to determine relationship between grades and the teacher training programs in terms of their computer literacy. The study used case study research methodology, and the sample consisted of 276…

  11. Computer program for distance learning of pesticide application technology

    Directory of Open Access Journals (Sweden)

    Bruno Maia


    Full Text Available Distance learning presents great potential for mitigating field problems on pesticide application technology. Thus, due to the lack of teaching material about pesticide spraying technology in the Portuguese language and the increasing availability of distance learning, this study developed and evaluated a computer program for distance learning about the theory of pesticide spraying technology using the tools of information technology. The modules comprising the course, named Pulverizar, were: (1 Basic concepts, (2 Factors that affect application, (3 Equipments, (4 Spraying nozzles, (5 Sprayer calibration, (6 Aerial application, (7 Chemigation, (8 Physical-chemical properties, (9 Formulations, (10 Adjuvants, (11 Water quality, and (12 Adequate use of pesticides. The program was made available to the public on July 1st, 2008, hosted at the web site, and was simple, robust and practical on the complementation of traditional teaching for the education of professionals in Agricultural Sciences. Mastering pesticide spraying technology by people involved in agricultural production can be facilitated by the program Pulverizar, which was well accepted in its initial evaluation.O ensino à distância apresenta grande potencial para minorar os problemas ocorridos no campo na área de tecnologia de aplicação de agroquímicos. Dessa forma, diante da escassez de material instrucional na área de tecnologia de aplicação de agroquímicos em Português e do crescimento elevado da educação à distância, o objetivo deste trabalho foi desenvolver e avaliar um programa computacional para o ensino à distância da parte teórica de tecnologia de aplicação de agroquímicos, utilizando as ferramentas de tecnologia da informação. Os módulos que compuseram o curso, intitulado Pulverizar, foram: (1 Conceitos básicos, (2 Fatores que afetam a aplicação, (3 Equipamentos, (4 Pontas de pulverização, (5 Calibração de pulverizadores

  12. Artificial intelligence programming languages for computer aided manufacturing (United States)

    Rieger, C.; Samet, H.; Rosenberg, J.


    Eight Artificial Intelligence programming languages (SAIL, LISP, MICROPLANNER, CONNIVER, MLISP, POP-2, AL, and QLISP) are presented and surveyed, with examples of their use in an automated shop environment. Control structures are compared, and distinctive features of each language are highlighted. A simple programming task is used to illustrate programs in SAIL, LISP, MICROPLANNER, and CONNIVER. The report assumes reader knowledge of programming concepts, but not necessarily of the languages surveyed.

  13. Ranking Workplace Competencies: Student and Graduate Perceptions. (United States)

    Rainsbury, Elizabeth; Hodges, Dave; Burchell, Noel; Lay, Mark


    New Zealand business students and graduates made similar rankings of the five most important workplace competencies: computer literacy, customer service orientation, teamwork and cooperation, self-confidence, and willingness to learn. Graduates placed greater importance on most of the 24 competencies, resulting in a statistically significant…

  14. An algorithm for ranking assignments using reoptimization

    DEFF Research Database (Denmark)

    Pedersen, Christian Roed; Nielsen, Lars Relund; Andersen, Kim Allan


    We consider the problem of ranking assignments according to cost in the classical linear assignment problem. An algorithm partitioning the set of possible assignments, as suggested by Murty, is presented where, for each partition, the optimal assignment is calculated using a new reoptimization...... technique. Computational results for the new algorithm are presented...

  15. Fractional cointegration rank estimation

    DEFF Research Database (Denmark)

    Lasak, Katarzyna; Velasco, Carlos

    We consider cointegration rank estimation for a p-dimensional Fractional Vector Error Correction Model. We propose a new two-step procedure which allows testing for further long-run equilibrium relations with possibly different persistence levels. The fi…rst step consists in estimating the parame......We consider cointegration rank estimation for a p-dimensional Fractional Vector Error Correction Model. We propose a new two-step procedure which allows testing for further long-run equilibrium relations with possibly different persistence levels. The fi…rst step consists in estimating...... to control for stochastic trend estimation effects from the first step. The critical values of the tests proposed depend only on the number of common trends under the null, p - r, and on the interval of the cointegration degrees b allowed, but not on the true cointegration degree b0. Hence, no additional...

  16. HISD Magnet Evaluation: Science, Math, and Computer Enrichment Programs, 1990-91. (United States)

    Kirkpatrick, Nanda D.; And Others

    Twenty-one magnet programs in the Houston Independent School District in Texas feature an enriched curriculum in science, math, and/or computers (science/math). Of these, 12 are elementary programs, 4 are middle school programs, and 5 are high school programs. In these programs, a total of 9,574 students were served during the 1990-91 school year:…

  17. The NASA Computational Fluid Dynamics (CFD) program - Building technology to solve future challenges (United States)

    Richardson, Pamela F.; Dwoyer, Douglas L.; Kutler, Paul; Povinelli, Louis A.


    This paper presents the NASA Computational Fluid Dynamics program in terms of a strategic vision and goals as well as NASA's financial commitment and personnel levels. The paper also identifies the CFD program customers and the support to those customers. In addition, the paper discusses technical emphasis and direction of the program and some recent achievements. NASA's Ames, Langley, and Lewis Research Centers are the research hubs of the CFD program while the NASA Headquarters Office of Aeronautics represents and advocates the program.

  18. Factors Influencing Women's Attitudes towards Computers in a Computer Literacy Training Program (United States)

    Chang, Sung-Lu; Shieh, Ruey S.; Liu, Eric Zhi-Feng; Yu, Pao-Ta


    In the "Digital Divide" research, adult women have generally been found to be the weakest group when compared with others. There is thus a need to provide this particular group with computer literacy training, and to give them opportunities to learn about using computers. In such training, women not only need to learn computer skills,…

  19. Tile Low Rank Cholesky Factorization for Climate/Weather Modeling Applications on Manycore Architectures

    KAUST Repository

    Akbudak, Kadir


    Covariance matrices are ubiquitous in computational science and engineering. In particular, large covariance matrices arise from multivariate spatial data sets, for instance, in climate/weather modeling applications to improve prediction using statistical methods and spatial data. One of the most time-consuming computational steps consists in calculating the Cholesky factorization of the symmetric, positive-definite covariance matrix problem. The structure of such covariance matrices is also often data-sparse, in other words, effectively of low rank, though formally dense. While not typically globally of low rank, covariance matrices in which correlation decays with distance are nearly always hierarchically of low rank. While symmetry and positive definiteness should be, and nearly always are, exploited for performance purposes, exploiting low rank character in this context is very recent, and will be a key to solving these challenging problems at large-scale dimensions. The authors design a new and flexible tile row rank Cholesky factorization and propose a high performance implementation using OpenMP task-based programming model on various leading-edge manycore architectures. Performance comparisons and memory footprint saving on up to 200K×200K covariance matrix size show a gain of more than an order of magnitude for both metrics, against state-of-the-art open-source and vendor optimized numerical libraries, while preserving the numerical accuracy fidelity of the original model. This research represents an important milestone in enabling large-scale simulations for covariance-based scientific applications.

  20. Addition of flexible body option to the TOLA computer program. Part 2: User and programmer documentation (United States)

    Dick, J. W.; Benda, B. J.


    User and programmer oriented documentation for the flexible body option of the Takeoff and Landing Analysis (TOLA) computer program are provided. The user information provides sufficient knowledge of the development and use of the option to enable the engineering user to successfully operate the modified program and understand the results. The programmer's information describes the option structure and logic enabling a programmer to make major revisions to this part of the TOLA computer program.

  1. 78 FR 11815 - Creation of a New Computer Matching Program That Will Expire on August 13, 2014 (United States)


    ... Creation of a New Computer Matching Program That Will Expire on August 13, 2014 AGENCY: National Finance... publishing notice of a computer matching program (CMP) that NFC will conduct on behalf of itself and the... description of the matching program. Notice of Computer Matching Program NFC will match applicant records with...

  2. Programming language, natural language? Supporting the diverse computational activities of novice programmers


    Good, Judith; Howland, Kate


    Given the current focus on teaching computational concepts to all from an early age, combined with the growing trend to empower end users to become producers of technology rather than mere consumers, we consider the issue of “computational notation”. Specifically, where the goal is to help individuals develop their understanding of computation and/or use computation in real world settings, we question whether natural language might be a preferred notation to traditional programming languages,...

  3. FREQFIT: Computer program which performs numerical regression and statistical chi-squared goodness of fit analysis

    Energy Technology Data Exchange (ETDEWEB)

    Hofland, G.S.; Barton, C.C.


    The computer program FREQFIT is designed to perform regression and statistical chi-squared goodness of fit analysis on one-dimensional or two-dimensional data. The program features an interactive user dialogue, numerous help messages, an option for screen or line printer output, and the flexibility to use practically any commercially available graphics package to create plots of the program`s results. FREQFIT is written in Microsoft QuickBASIC, for IBM-PC compatible computers. A listing of the QuickBASIC source code for the FREQFIT program, a user manual, and sample input data, output, and plots are included. 6 refs., 1 fig.

  4. [Computer program for medical records of patients with diabetic retinopathy]. (United States)

    Davidescu, Livia; Ignat, Florica; Popescu, G L


    The purpose of this study was to point out the advantages or the computerised evidence in patients with diabetic retinopathy examined in "Diabetic Eye Department" from Ophthalmology Clinic Craiova. This program represents a new national performance regarding the evidence of the diabetic patients clinic examination. The main characteristics of our program are the accessibility and the coherence; the utility of this program has been demonstrated by long term use and the big number of patents recorded.

  5. Data systems and computer science: Software Engineering Program (United States)

    Zygielbaum, Arthur I.


    An external review of the Integrated Technology Plan for the Civil Space Program is presented. This review is specifically concerned with the Software Engineering Program. The goals of the Software Engineering Program are as follows: (1) improve NASA's ability to manage development, operation, and maintenance of complex software systems; (2) decrease NASA's cost and risk in engineering complex software systems; and (3) provide technology to assure safety and reliability of software in mission critical applications.

  6. Programmer's guide for the GNAT computer program (numerical analysis of stratification in supercritical oxygen) (United States)

    Heinmiller, J. P.


    This document is the programmer's guide for the GNAT computer program developed under MSC/TRW Task 705-2, Apollo cryogenic storage system analysis, subtask 2, is reported. Detailed logic flow charts and compiled program listings are provided for all program elements.

  7. Generic Mathematical Programming Formulation and Solution for Computer-Aided Molecular Design

    DEFF Research Database (Denmark)

    Zhang, Lei; Cignitti, Stefano; Gani, Rafiqul


    This short communication presents a generic mathematical programming formulation for Computer-Aided Molecular Design (CAMD). A given CAMD problem, based on target properties, is formulated as a Mixed Integer Linear/Non-Linear Program (MILP/MINLP). The mathematical programming model presented here...

  8. Regression Computer Programs for Setwise Regression and Three Related Analysis of Variance Techniques. (United States)

    Williams, John D.; Lindem, Alfred C.

    Four computer programs using the general purpose multiple linear regression program have been developed. Setwise regression analysis is a stepwise procedure for sets of variables; there will be as many steps as there are sets. Covarmlt allows a solution to the analysis of covariance design with multiple covariates. A third program has three…

  9. Issue Management Risk Ranking Systems

    Energy Technology Data Exchange (ETDEWEB)

    Novack, Steven David; Marshall, Frances Mc Clellan; Stromberg, Howard Merion; Grant, Gary Michael


    Thousands of safety issues have been collected on-line at the Idaho National Engineering and Environmental Laboratory (INEEL) as part of the Issue Management Plan. However, there has been no established approach to prioritize collected and future issues. The authors developed a methodology, based on hazards assessment, to identify and risk rank over 5000 safety issues collected at INEEL. This approach required that it was easily applied and understandable for site adaptation and commensurate with the Integrated Safety Plan. High-risk issues were investigated and mitigative/preventive measures were suggested and ranked based on a cost-benefit scheme to provide risk-informed safety measures. This methodology was consistent with other integrated safety management goals and tasks providing a site-wide risk informed decision tool to reduce hazardous conditions and focus resources on high-risk safety issues. As part of the issue management plan, this methodology was incorporated at the issue collection level and training was provided to management to better familiarize decision-makers with concepts of safety and risk. This prioritization methodology and issue dissemination procedure will be discussed. Results of issue prioritization and training efforts will be summarized. Difficulties and advantages of the process will be reported. Development and incorporation of this process into INEELs lessons learned reporting and the site-wide integrated safety management program will be shown with an emphasis on establishing self reliance and ownership of safety issues.

  10. Comparison of Image reformation Using Personal Computer with Dentascan Program

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Eun Kyung [Dept of OMF Radiology, School of Dentistry, Dankook University, Yongin (Korea, Republic of)


    This study was performed to demonstrate the method of image reformation of dental implants, using a personal computer with inexpensive software and to compare the images reformatted using the above method with those using Dentascan software. CT axial slices of 5 mandibles of 5 volunteers from GE Highspeed Advantage (GE Medical systems, U.S.A.) were used, Personal computer used for image reformation was PowerWave 604/120 (Power Computing Co, U.S.A.) and software used were Osiris (Univ. Hospital Of Geneva, Switzerland) and Import ACCESS V1.1(Designed Access Co., U.S.A.) for importing CT image and NIH Image 1.58 (NIH, U.S.A.) for image processing. Seven images were selected among the serial reconstructed cross-sectional images produced by Dentascan (DS group). Seven resliced cross-sectional images at the same position were obtained ned at the personal computer (PC group). Regression analysis of the measurements of PC group was done against those of DS group. Measurements of the bone height and width at the reformed cross-sectional images using Mac-compatible computer were highly correlated with those using workstation with Dentascan software (height : r2=0.999, p<0.001, width : p=0.991, p<0.001). So, it is considered that we can use a personal computer with inexpensive software for the dental implant planning, instead of the expensive software and workstation.

  11. Quantum computer simulation using the CUDA programming model (United States)

    Gutiérrez, Eladio; Romero, Sergio; Trenas, María A.; Zapata, Emilio L.


    Quantum computing emerges as a field that captures a great theoretical interest. Its simulation represents a problem with high memory and computational requirements which makes advisable the use of parallel platforms. In this work we deal with the simulation of an ideal quantum computer on the Compute Unified Device Architecture (CUDA), as such a problem can benefit from the high computational capacities of Graphics Processing Units (GPU). After all, modern GPUs are becoming very powerful computational architectures which is causing a growing interest in their application for general purpose. CUDA provides an execution model oriented towards a more general exploitation of the GPU allowing to use it as a massively parallel SIMT (Single-Instruction Multiple-Thread) multiprocessor. A simulator that takes into account memory reference locality issues is proposed, showing that the challenge of achieving a high performance depends strongly on the explicit exploitation of memory hierarchy. Several strategies have been experimentally evaluated obtaining good performance results in comparison with conventional platforms.


    CERN Multimedia

    M. Kasemann

    Overview In autumn the main focus was to process and handle CRAFT data and to perform the Summer08 MC production. The operational aspects were well covered by regular Computing Shifts, experts on duty and Computing Run Coordination. At the Computing Resource Board (CRB) in October a model to account for service work at Tier 2s was approved. The computing resources for 2009 were reviewed for presentation at the C-RRB. The quarterly resource monitoring is continuing. Facilities/Infrastructure operations Operations during CRAFT data taking ran fine. This proved to be a very valuable experience for T0 workflows and operations. The transfers of custodial data to most T1s went smoothly. A first round of reprocessing started at the Tier-1 centers end of November; it will take about two weeks. The Computing Shifts procedure was tested full scale during this period and proved to be very efficient: 30 Computing Shifts Persons (CSP) and 10 Computing Resources Coordinators (CRC). The shift program for the shut down w...

  13. DORCA 2 computer program. Volume 2: Programmer's manual (United States)

    Gold, B. J.


    A guide for coding the Dynamic Operational Requirements and Cost Analysis Program (DORCA 2) is presented. The manual provides a detailed operation of every subroutine, the layout in core of the major matrices and arrays, and the meaning of all program values. Flow charts are included.

  14. General purpose computer program for interacting supersonic configurations: Programmer's manual (United States)

    Crill, W.; Dale, B.


    The program ISCON (Interacting Supersonic Configuration) is described. The program is in support of the problem to generate a numerical procedure for determining the unsteady dynamic forces on interacting wings and tails in supersonic flow. Subroutines are presented along with the complete FORTRAN source listing.

  15. SSME structural computer program development: BOPACE programmer manual (United States)


    The programming manual for the two-dimensional Boeing Plastic Analysis Capability for Engines (BOPACE) program written in FORTRAN IV is presented. The BOPACE flow logic is summarized along with subroutines and input/output files. Emphasis is placed on linear equation solver routines written as an independent package.

  16. 76 FR 11435 - Privacy Act of 1974; Computer Matching Program (United States)


    ... Federal student financial assistance, and the ED personal information number (PIN) Registration System of Records (18-11-12), which contains the applicant's information to receive an ED PIN, will be matched... Dates of the Matching Program This matching program must be approved by the Data Integrity Board of each...

  17. Review of the Fusion Theory and Computing Program. Fusion Energy Sciences Advisory Committee (FESAC)

    Energy Technology Data Exchange (ETDEWEB)

    Antonsen, Thomas M. [Univ. of Maryland, College Park, MD (United States); Berry, Lee A. [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Brown, Michael R. [Swarthmore College, PA (United States); Dahlburg, Jill P. [General Atomics, San Diego, CA (United States); Davidson, Ronald C. [Princeton Plasma Physics Lab. (PPPL), Princeton, NJ (United States); Greenwald, Martin [Massachusetts Inst. of Technology (MIT), Cambridge, MA (United States); Hegna, Chris C. [Univ. of Wisconsin, Madison, WI (United States); McCurdy, William [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Newman, David E. [Univ. of Alaska, Fairbanks, AK (United States); Pellegrini, Claudio [Univ. of California, Los Angeles, CA (United States); Phillips, Cynthia K. [Princeton Plasma Physics Lab. (PPPL), Princeton, NJ (United States); Post, Douglass E. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States); Rosenbluth, Marshall N. [Univ. of California, San Diego, CA (United States); Sheffield, John [Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States); Simonen, Thomas C. [Munising, MI (United States); Van Dam, James [Univ. of Texas, Austin, TX (United States)


    At the November 14-15, 2000, meeting of the Fusion Energy Sciences Advisory Committee, a Panel was set up to address questions about the Theory and Computing program, posed in a charge from the Office of Fusion Energy Sciences (see Appendix A). This area was of theory and computing/simulations had been considered in the FESAC Knoxville meeting of 1999 and in the deliberations of the Integrated Program Planning Activity (IPPA) in 2000. A National Research Council committee provided a detailed review of the scientific quality of the fusion energy sciences program, including theory and computing, in 2000.

  18. Method for Statically Checking an Object-oriented Computer Program Module (United States)

    Bierhoff, Kevin M. (Inventor); Aldrich, Jonathan (Inventor)


    A method for statically checking an object-oriented computer program module includes the step of identifying objects within a computer program module, at least one of the objects having a plurality of references thereto, possibly from multiple clients. A discipline of permissions is imposed on the objects identified within the computer program module. The permissions enable tracking, from among a discrete set of changeable states, a subset of states each object might be in. A determination is made regarding whether the imposed permissions are violated by a potential reference to any of the identified objects. The results of the determination are output to a user.

  19. Liquid rocket combustion computer model with distributed energy release. DER computer program documentation and user's guide, volume 1 (United States)

    Combs, L. P.


    A computer program for analyzing rocket engine performance was developed. The program is concerned with the formation, distribution, flow, and combustion of liquid sprays and combustion product gases in conventional rocket combustion chambers. The capabilities of the program to determine the combustion characteristics of the rocket engine are described. Sample data code sheets show the correct sequence and formats for variable values and include notes concerning options to bypass the input of certain data. A seperate list defines the variables and indicates their required dimensions.

  20. Computer Programming: An Activity as Compelling as Game Play

    Directory of Open Access Journals (Sweden)

    Tom Goulding


    Full Text Available Game motif programming exercises (GM-Games were developed to help novices develop complex client server game systems within their freshman year. GM-Games foster a strong work ethic in as much as they reproduce the challenges and excitement associated with game play; yet their purpose is the development of advanced programming skills. We have found that young people are just as interested in mastering programming skills as they are in mastering the shooting, racing or strategy skills required in many entertainment games. We describe in this paper how GM-Games imitate many of the aspects of game play.

  1. Rankings from Fuzzy Pairwise Comparisons

    NARCIS (Netherlands)

    van den Broek, P.M.; Noppen, J.A.R.; Mohammadian, M.


    We propose a new method for deriving rankings from fuzzy pairwise comparisons. It is based on the observation that quantification of the uncertainty of the pairwise comparisons should be used to obtain a better crisp ranking, instead of a fuzzified version of the ranking obtained from crisp pairwise

  2. University Rankings and Social Science (United States)

    Marginson, Simon


    University rankings widely affect the behaviours of prospective students and their families, university executive leaders, academic faculty, governments and investors in higher education. Yet the social science foundations of global rankings receive little scrutiny. Rankings that simply recycle reputation without any necessary connection to real…

  3. Sequential rank agreement methods for comparison of ranked lists

    DEFF Research Database (Denmark)

    Ekstrøm, Claus Thorn; Gerds, Thomas Alexander; Jensen, Andreas Kryger


    The comparison of alternative rankings of a set of items is a general and prominent task in applied statistics. Predictor variables are ranked according to magnitude of association with an outcome, prediction models rank subjects according to the personalized risk of an event, and genetic studies...... are illustrated using gene rankings, and using data from two Danish ovarian cancer studies where we assess the within and between agreement of different statistical classification methods.......The comparison of alternative rankings of a set of items is a general and prominent task in applied statistics. Predictor variables are ranked according to magnitude of association with an outcome, prediction models rank subjects according to the personalized risk of an event, and genetic studies...

  4. Low Rank Approximation Algorithms, Implementation, Applications

    CERN Document Server

    Markovsky, Ivan


    Matrix low-rank approximation is intimately related to data modelling; a problem that arises frequently in many different fields. Low Rank Approximation: Algorithms, Implementation, Applications is a comprehensive exposition of the theory, algorithms, and applications of structured low-rank approximation. Local optimization methods and effective suboptimal convex relaxations for Toeplitz, Hankel, and Sylvester structured problems are presented. A major part of the text is devoted to application of the theory. Applications described include: system and control theory: approximate realization, model reduction, output error, and errors-in-variables identification; signal processing: harmonic retrieval, sum-of-damped exponentials, finite impulse response modeling, and array processing; machine learning: multidimensional scaling and recommender system; computer vision: algebraic curve fitting and fundamental matrix estimation; bioinformatics for microarray data analysis; chemometrics for multivariate calibration; ...

  5. Adaptive distributional extensions to DFR ranking

    DEFF Research Database (Denmark)

    Petersen, Casper; Simonsen, Jakob Grue; Järvelin, Kalervo


    Divergence From Randomness (DFR) ranking models assume that informative terms are distributed in a corpus differently than non-informative terms. Different statistical models (e.g. Poisson, geometric) are used to model the distribution of non-informative terms, producing different DFR models....... An informative term is then detected by measuring the divergence of its distribution from the distribution of non-informative terms. However, there is little empirical evidence that the distributions of non-informative terms used in DFR actually fit current datasets. Practically this risks providing a poor...... separation between informative and non-informative terms, thus compromising the discriminative power of the ranking model. We present a novel extension to DFR, which first detects the best-fitting distribution of non-informative terms in a collection, and then adapts the ranking computation to this best...

  6. Estimating Independent Locally Shifted Random Utility Models for Ranking Data (United States)

    Lam, Kar Yin; Koning, Alex J.; Franses, Philip Hans


    We consider the estimation of probabilistic ranking models in the context of conjoint experiments. By using approximate rather than exact ranking probabilities, we avoided the computation of high-dimensional integrals. We extended the approximation technique proposed by Henery (1981) in the context of the Thurstone-Mosteller-Daniels model to any…

  7. Rank range test for equality of dispersion | Odiase | Journal of ...

    African Journals Online (AJOL)

    This paper exploits the computational simplicity of the range of a set of data to formulate a twosample scale test called the Rank Range test. The performance of the test statistic is compared with other tests of scale. The exact distribution of the Rank Range test statistic is generated empirically through the unconditional ...

  8. Computer Science and Perl Programming Best of Perl Journal

    CERN Document Server


    With more than a million dedicated programmers, Perl has proven to be the best computing language for the latest trends in computing and business. While other languages have stagnated, Perl remains fresh, thanks to its community-based development model, which encourages the sharing of information among users. This tradition of knowledge-sharing allows developers to find answers to almost any Perl question they can dream up.And you can find many of those answers right here in Perl Hacks. Like all books in O'Reilly's Hacks Series, Perl Hacks appeals to a variety of programmers, whether you're a

  9. Computer program for machine design of Cassegrain feed systems (United States)

    Potter, P. D.


    Program designs the feed system geometry and the subreflector surface, with the main reflector configuration and frequency of operation as input data. Although the feedhorn is not designed, its required gain, beamwidth, and approximate radiation pattern are specified.

  10. GEODYN programmer's guide, volume 2, part 2. [computer program for estimation of orbit and geodetic parameters (United States)

    Mullins, N. E.; Dao, N. C.; Martin, T. V.; Goad, C. C.; Boulware, N. L.; Chin, M. M.


    A computer program for executive control routine for orbit integration of artificial satellites is presented. At the beginning of each arc, the program initiates required constants as well as the variational partials at epoch. If epoch needs to be reset to a previous time, the program negates the stepsize, and calls for integration backward to the desired time. After backward integration is completed, the program resets the stepsize to the proper positive quantity.

  11. Conceptual Understanding of Computer Program Execution: Application to C++


    Al-Fedaghi, Sabah


    A visual programming language uses pictorial tools such as diagrams to represent its structural units and control stream. It is useful for enhancing understanding, maintenance, verification, testing, and parallelism. This paper proposes a diagrammatic methodology that produces a conceptual representation of instructions for programming source codes. Without loss of generality in the potential for using the methodology in a wider range of applications, this paper focuses on using these diagram...

  12. Verifying the Absence of Common Runtime Errors in Computer Programs (United States)


    the langauge designer another way to test a design, in addition to the usual ways based on experience with other languages and difficulty of...simplified verification conditions is a special skill that one must learn in order to use the verifier. In theI process of analyzing a VC, une notes...Discovery of Linear Restraints Among Variables of a Program, Proceedings of the Fifth ACM Symposium on Principles of Programming Languages, January

  13. Computation of Program Source Code Similarity by Composition of Parse Tree and Call Graph

    Directory of Open Access Journals (Sweden)

    Hyun-Je Song


    Full Text Available This paper proposes a novel method to compute how similar two program source codes are. Since a program source code is represented as a structural form, the proposed method adopts convolution kernel functions as a similarity measure. Actually, a program source code has two kinds of structural information. One is syntactic information and the other is the dependencies of function calls lying on the program. Since the syntactic information of a program is expressed as its parse tree, the syntactic similarity between two programs is computed by a parse tree kernel. The function calls within a program provide a global structure of a program and can be represented as a graph. Therefore, the similarity of function calls is computed with a graph kernel. Then, both structural similarities are reflected simultaneously into comparing program source codes by composing the parse tree and the graph kernels based on a cyclomatic complexity. According to the experimental results on a real data set for program plagiarism detection, the proposed method is proved to be effective in capturing the similarity between programs. The experiments show that the plagiarized pairs of programs are found correctly and thoroughly by the proposed method.

  14. Encouraging more women into computer science: Initiating a single-sex intervention program in Sweden (United States)

    Brandell, Gerd; Carlsson, Svante; Ekblom, Håkan; Nord, Ann-Charlotte


    The process of starting a new program in computer science and engineering, heavily based on applied mathematics and only open to women, is described in this paper. The program was introduced into an educational system without any tradition in single-sex education. Important observations made during the process included the considerable interest in mathematics and curiosity about computer science found among female students at the secondary school level, and the acceptance of the single-sex program by the staff, administration, and management of the university as well as among male and female students. The process described highlights the importance of preparing the environment for a totally new type of educational program.

  15. DB90: A Fortran Callable Relational Database Routine for Scientific and Engineering Computer Programs (United States)

    Wrenn, Gregory A.


    This report describes a database routine called DB90 which is intended for use with scientific and engineering computer programs. The software is written in the Fortran 90/95 programming language standard with file input and output routines written in the C programming language. These routines should be completely portable to any computing platform and operating system that has Fortran 90/95 and C compilers. DB90 allows a program to supply relation names and up to 5 integer key values to uniquely identify each record of each relation. This permits the user to select records or retrieve data in any desired order.

  16. The Design of an Undergraduate Degree Program in Computer & Digital Forensics

    Directory of Open Access Journals (Sweden)

    Gary C. Kessler


    Full Text Available Champlain College formally started an undergraduate degree program in Computer & Digital Forensics in 2003. The underlying goals were that the program be multidisciplinary, bringing together the law, computer technology, and the basics of digital investigations; would be available as on online and on-campus offering; and would have a process-oriented focus. Success of this program has largely been due to working closely with practitioners, maintaining activity in events related to both industry and academia, and flexibility to respond to ever-changing needs. This paper provides an overview of how this program was conceived, developed, and implemented; its evolution over time; and current and planned initiatives.

  17. Computer Programming in the UK Undergraduate Mathematics Curriculum (United States)

    Sangwin, Christopher J.; O'Toole, Claire


    This paper reports a study which investigated the extent to which undergraduate mathematics students in the United Kingdom are currently taught to programme a computer as a core part of their mathematics degree programme. We undertook an online survey, with significant follow-up correspondence, to gather data on current curricula and received…

  18. development of a computer program for the design of auger

    African Journals Online (AJOL)


    Computer Science Department, U.N.N.. E- mail: ABSTRACT. The screw or auger conveyor is an important conveyor in agricultural and food processing industries. The classical approach to its analysis and design involve the determination of the type and dimensions of the pitch, capacity, screw ...

  19. A Research Program in Computer Technology. Volume 1 (United States)


    Kahn, and Bernard Lang . Formal Definition of Ada, Honeywell, Inc. and CII-Honeywell Bull, 1980. 2. Ichbiah, Jean D., et al., Reference Manual for the...14. COMPUTER RESEARCH SUPPORT Technical Staff: Support Staff: Dan Lynch John Metzger Walt Edmison Ray Bates Bill Moore Chloe HoIg Dale Chase Serge

  20. High Performance Computing Modernization Program Kerberos Throughput Test Report (United States)


    ORGANIZATION REPORT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) 10. SPONSOR / MONITOR’S ACRONYM(S)9. SPONSORING / MONITORING AGENCY...the high computing power of the main supercomputer. Each supercomputer is different in node architecture as well as hardware specifications. 2

  1. Computer Adaptive Testing for Small Scale Programs and Instructional Systems (United States)

    Rudner, Lawrence M.; Guo, Fanmin


    This study investigates measurement decision theory (MDT) as an underlying model for computer adaptive testing when the goal is to classify examinees into one of a finite number of groups. The first analysis compares MDT with a popular item response theory model and finds little difference in terms of the percentage of correct classifications. The…

  2. A Cognitive Model for Aggregating People's Rankings (United States)

    Lee, Michael D.; Steyvers, Mark; Miller, Brent


    We develop a cognitive modeling approach, motivated by classic theories of knowledge representation and judgment from psychology, for combining people's rankings of items. The model makes simple assumptions about how individual differences in knowledge lead to observed ranking data in behavioral tasks. We implement the cognitive model as a Bayesian graphical model, and use computational sampling to infer an aggregate ranking and measures of the individual expertise. Applications of the model to 23 data sets, dealing with general knowledge and prediction tasks, show that the model performs well in producing an aggregate ranking that is often close to the ground truth and, as in the “wisdom of the crowd” effect, usually performs better than most of individuals. We also present some evidence that the model outperforms the traditional statistical Borda count method, and that the model is able to infer people's relative expertise surprisingly well without knowing the ground truth. We discuss the advantages of the cognitive modeling approach to combining ranking data, and in wisdom of the crowd research generally, as well as highlighting a number of potential directions for future model development. PMID:24816733

  3. Workshop on programming languages for high performance computing (HPCWPL): final report.

    Energy Technology Data Exchange (ETDEWEB)

    Murphy, Richard C.


    This report summarizes the deliberations and conclusions of the Workshop on Programming Languages for High Performance Computing (HPCWPL) held at the Sandia CSRI facility in Albuquerque, NM on December 12-13, 2006.

  4. Intraobserver and intermethod reliability for using two different computer programs in preoperative lower limb alignment analysis

    Directory of Open Access Journals (Sweden)

    Mohamed Kenawey


    Conclusion: Computer assisted lower limb alignment analysis is reliable whether using graphics editing program or specialized planning software. However slight higher variability for angles away from the knee joint can be expected.

  5. Gender and stereotypes in motivation to study computer programming for careers in multimedia (United States)

    Doubé, Wendy; Lang, Catherine


    A multimedia university programme with relatively equal numbers of male and female students in elective programming subjects provided a rare opportunity to investigate female motivation to study and pursue computer programming in a career. The MSLQ was used to survey 85 participants. In common with research into deterrence of females from STEM domains, females displayed significantly lower self-efficacy and expectancy for success. In contrast to research into deterrence of females from STEM domains, both genders placed similar high values on computer programming and shared high extrinsic and intrinsic goal orientation. The authors propose that the stereotype associated with a creative multimedia career could attract female participation in computer programming whereas the stereotype associated with computer science could be a deterrent.

  6. A Teaching Exercise for the Identification of Bacteria Using An Interactive Computer Program. (United States)

    Bryant, Trevor N.; Smith, John E.


    Describes an interactive Fortran computer program which provides an exercise in the identification of bacteria. Provides a way of enhancing a student's approach to systematic bacteriology and numerical identification procedures. (Author/MA)


    Directory of Open Access Journals (Sweden)

    R. Sagan


    Full Text Available This article considers different aspects which allow defining correctness of choosing sorting algorithms. Also some algorithms, needed for computational experiments for certain class of programs, are compared.

  8. The Use of Computer-Based Programming Environments as Computer Modelling Tools in Early Science Education: The Cases of Textual and Graphical Program Languages (United States)

    Louca, Loucas T.; Zacharia, Zacharia C.


    This is an interpretive case study seeking to develop detailed and comparative descriptions of how two groups of fifth-grade students used two different Computer-based Programming Environments (CPEs) (namely Microworlds Logo and Stagecast Creator) during scientific modelling. The primary sources of data that were used in this 4-month study include…



    Aries Alfian Prasetyo; Setiadi Cahyono Putro; I Made Wirawan


    This study aims to (1) describe supporting and inhibiting factors in blended learning implementation for the students of computer and network engineering expertise program and (2) describe the accomplishment level of the implementation. This study is designed as a descriptive study with quantitative approach. The research object is the blended learning implementation in computer and network engineering expertise program in SMK N 1 Baureno Bojonegoro. The research subjects consist of teachers,...

  10. Computer-based Programs in Speech Therapy of Dyslalia and Dyslexia- Dysgraphia

    Directory of Open Access Journals (Sweden)

    Mirela Danubianu


    Full Text Available During the last years, the researchers and therapists in speech therapy have been more and more concerned with the elaboration and use of computer programs in speech disorders therapy. The main objective of this study was to evaluate the therapeutic effectiveness of computer-based programs for the Romanian language in speech therapy. Along the study, we will present the experimental research through assessing the effectiveness of computer programs in the speech therapy for speech disorders: dyslalia, dyslexia and dysgraphia. Methodologically, the use of the computer in the therapeutic phases was carried out with the help of some computer-based programs (Logomon, Dislex-Test etc. that we elaborated and we experimented with during several years of therapeutic activity. The sample used in our experiments was composed of 120 subjects; two groups of 60 children with speech disorders were selected for both speech disorders: 30 for the experimental ('computer-based' group and 30 for the control ('classical method' group. The study hypotheses verified whether the results, obtained by the subjects within the experimental group, improved significantly after using the computer-based program, compared to the subjects within the control group, who did not use this program but got a classical therapy. The hypotheses were confirmed for the speech disorders included in this research; the conclusions of the study confirm the advantages of using computer-based programs within speech therapy by correcting these disorders, as well as due to the positive influence these programs have on the development of children’s personality.

  11. VNAP2: a computer program for computation of two-dimensional, time-dependent, compressible, turbulent flow

    Energy Technology Data Exchange (ETDEWEB)

    Cline, M.C.


    VNAP2 is a computer program for calculating turbulent (as well as laminar and inviscid), steady, and unsteady flow. VNAP2 solves the two-dimensional, time-dependent, compressible Navier-Stokes equations. The turbulence is modeled with either an algebraic mixing-length model, a one-equation model, or the Jones-Launder two-equation model. The geometry may be a single- or a dual-flowing stream. The interior grid points are computed using the unsplit MacCormack scheme. Two options to speed up the calculations for high Reynolds number flows are included. The boundary grid points are computed using a reference-plane-characteristic scheme with the viscous terms treated as source functions. An explicit artificial viscosity is included for shock computations. The fluid is assumed to be a perfect gas. The flow boundaries may be arbitrary curved solid walls, inflow/outflow boundaries, or free-jet envelopes. Typical problems that can be solved concern nozzles, inlets, jet-powered afterbodies, airfoils, and free-jet expansions. The accuracy and efficiency of the program are shown by calculations of several inviscid and turbulent flows. The program and its use are described completely, and six sample cases and a code listing are included.

  12. A two-dimensional graphing program for the Tektronix 4050-series graphics computers (United States)

    Kipp, K.L.


    A refined, two-dimensional graph-plotting program was developed for use on Tektronix 4050-series graphics computers. Important features of this program include: any combination of logarithmic and linear axes, optional automatic scaling and numbering of the axes, multiple-curve plots, character or drawn symbol-point plotting, optional cartridge-tape data input and plot-format storage, optional spline fitting for smooth curves, and built-in data-editing options. The program is run while the Tektronix is not connected to any large auxiliary computer, although data from files on an auxiliary computer easily can be transferred to data-cartridge for later plotting. The user is led through the plot-construction process by a series of questions and requests for data input. Five example plots are presented to illustrate program capability and the sequence of program operation. (USGS)

  13. CMX: The Effects of an Educational MMORPG on Learning and Teaching Computer Programming (United States)

    Malliarakis, Christos; Satratzemi, Maya; Xinogalos, Stelios


    Computer programming has for decades posed several difficulties for students of all educational levels. A number of teaching approaches have been proposed over the years but none seems to fulfil the needs of students nowadays. Students use computers mainly for playing games and the Internet and as quite a few researchers state this aspect of…

  14. Computer Literacy and Online Learning Attitude toward GSOE Students in Distance Education Programs (United States)

    Li, Lung-Yu; Lee, Long-Yuan


    The purpose of this study was to explore graduate students' competencies in computer use and their attitudes toward online learning in asynchronous online courses of distance learning programs in a Graduate School of Education (GSOE) in Taiwan. The research examined the relationship between computer literacy and the online learning attitudes of…

  15. Abstracts of computer programs and data libraries pertaining to photon production data

    Energy Technology Data Exchange (ETDEWEB)

    White, J.E.; Manneschmidt, J.B.; Finch, S.Y.; Dickens, J.K.


    Abstracts, or descriptions, of computer programs and data libraries pertaining to Photon Production Data (Measurements, Evaluations and Calculations) maintained in the collections of the Radiation Safety Information Computational Center, Oak Ridge, Tennessee USA and at the OECD/NEA Data Bank, Paris, are collected in this document.

  16. General design methodology applied to the research domain of physical programming for computer illiterate

    CSIR Research Space (South Africa)

    Smith, Andrew C


    Full Text Available The authors discuss the application of the 'general design methodology‘ in the context of a physical computing project. The aim of the project was to design and develop physical objects that could serve as metaphors for computer programming elements...

  17. Enhancing Student Writing and Computer Programming with LATEX and MATLAB in Multivariable Calculus (United States)

    Sullivan, Eric; Melvin, Timothy


    Written communication and computer programming are foundational components of an undergraduate degree in the mathematical sciences. All lower-division mathematics courses at our institution are paired with computer-based writing, coding, and problem-solving activities. In multivariable calculus we utilize MATLAB and LATEX to have students explore…

  18. Community Needs Assessment for an Electronics and Computer Engineering Technology Program at Maui, Molokai, and Lanai. (United States)

    Pezzoli, Jean A.

    In June 1992, Maui Community College (MCC), in Hawaii, conducted a survey of the communities of Maui, Molokai, Lanai, and Hana to determine perceived needs for an associate degree and certificate program in electronics and computer engineering. Questionnaires were mailed to 500 firms utilizing electronic or computer services, seeking information…

  19. Why zombies cannot write significant source code: The Knowledge Game and the art of computer programming

    NARCIS (Netherlands)

    Gobbo, F.; Benini, M.


    This article analyses the knowledge needed to understand a computer program within the philosophy of information. L. Floridi's method of levels of abstraction is applied to the relation between an ideal programmer and a modern computer seen together as an informational organism. The results obtained

  20. Permitted decompilation of a computer program in order to protect the general interest

    Directory of Open Access Journals (Sweden)

    Radovanović Sanja M.


    Full Text Available Computer program is an intellectual creation protected by copyright. However, unlike other items with equivalent legal protection, a computer program has a strong technical functionality, which is, in nowadays' society, an indispensable factor in everyday business activities, exchange of information, entertainment or achieving other similar purposes. Precisely because of this feature, computer program can rarely be seen in isolation from the hardware and software environment. In other words, the functionality of a computer program reaches its full scope only in interaction with other computer program or device. Bearing in mind the fact that this intellectual creation is in the focus of technological, and thus social, development, legislators are trying to provide a legal framework in which these interactions take place unhindered. In fact, considering that each aspect of the use of a computer program presents the exclusive right of the author, relying on his or her consent to undertake certain perpetration which would provide the necessary connectivity of the various components, could put in risk further technological development. Therefore, the lawmakers provide that, in certain cases and under certain conditions, the author's exclusive right could be restricted or excluded. This paper aims to analyze a normative contribution in achieving, technical and technological needed, and therefore, in terms of general interest justified, interactions.

  1. School Meal Programs: Few Outbreaks of Foodborne Illness Reported. Report to the Ranking Minority Member, Committee on Agriculture, Nutrition, and Forestry, U.S. Senate. (United States)

    Robertson, Robert E.

    Twenty outbreaks of foodborne illness in schools were reported to the Centers for Disease Control and Prevention (CDC) during 1997; however, only 8 cases were associated with food served in the school meal programs. Preliminary findings identified nine outbreaks in 1998, affecting an estimated 1,609 individuals. CDC notes that such outbreaks are…

  2. Demand side management scheme in smart grid with cloud computing approach using stochastic dynamic programming

    Directory of Open Access Journals (Sweden)

    S. Sofana Reka


    Full Text Available This paper proposes a cloud computing framework in smart grid environment by creating small integrated energy hub supporting real time computing for handling huge storage of data. A stochastic programming approach model is developed with cloud computing scheme for effective demand side management (DSM in smart grid. Simulation results are obtained using GUI interface and Gurobi optimizer in Matlab in order to reduce the electricity demand by creating energy networks in a smart hub approach.

  3. Certainty in Stockpile Computing: Recommending a Verification and Validation Program for Scientific Software

    Energy Technology Data Exchange (ETDEWEB)

    Lee, J.R.


    As computing assumes a more central role in managing the nuclear stockpile, the consequences of an erroneous computer simulation could be severe. Computational failures are common in other endeavors and have caused project failures, significant economic loss, and loss of life. This report examines the causes of software failure and proposes steps to mitigate them. A formal verification and validation program for scientific software is recommended and described.

  4. Econophysics of a ranked demand and supply resource allocation problem (United States)

    Priel, Avner; Tamir, Boaz


    We present a two sided resource allocation problem, between demands and supplies, where both parties are ranked. For example, in Big Data problems where a set of different computational tasks is divided between a set of computers each with its own resources, or between employees and employers where both parties are ranked, the employees by their fitness and the employers by their package benefits. The allocation process can be viewed as a repeated game where in each iteration the strategy is decided by a meta-rule, based on the ranks of both parties and the results of the previous games. We show the existence of a phase transition between an absorbing state, where all demands are satisfied, and an active one where part of the demands are always left unsatisfied. The phase transition is governed by the ratio between supplies and demand. In a job allocation problem we find positive correlation between the rank of the workers and the rank of the factories; higher rank workers are usually allocated to higher ranked factories. These all suggest global emergent properties stemming from local variables. To demonstrate the global versus local relations, we introduce a local inertial force that increases the rank of employees in proportion to their persistence time in the same factory. We show that such a local force induces non trivial global effects, mostly to benefit the lower ranked employees.

  5. 78 FR 21713 - Privacy Act of 1974: Computer Matching Program (United States)


    ..., and the Office of Management and Budget (OMB) Guidelines on the Conduct of Matching Programs, notice... cost of medical care. These veterans supply household income information that includes their spouses and dependents at the time of application for VA health care benefits. DATES: Effective Date: This...

  6. 76 FR 47299 - Privacy Act of 1974: Computer Matching Program (United States)


    ..., and the Office of Management and Budget (OMB) Guidelines on the Conduct of Matching Programs, notice... cost of medical care. These veterans supply household income information that includes their spouses and dependents at the time of application for VA health care benefits. DATES: Effective Date: This...

  7. 78 FR 50146 - Privacy Act of 1974: Computer Matching Program (United States)


    ..., and the Office of Management and Budget (OMB) Guidelines on the Conduct of Matching Programs, notice... care. These veterans supply household income information that includes their spouses and dependents at the time of application for VA health care benefits. DATES: Effective Date: The effective (start) date...

  8. 75 FR 54966 - Privacy Act of 1974: Computer Matching Program (United States)


    ..., and the Office of Management and Budget (OMB) Guidelines on the Conduct of Matching Programs, notice... cost of medical care. These veterans supply household income information that includes their spouses and dependents at the time of application for VA health care benefits. DATES: Effective Date: This...

  9. Evaluation of SHABERTH: A bearing simulation computer program (United States)


    To investigate lubrication effects on bearing thermal performance, an investigation was performed to determine the feasibility of using the SKF program SHABERTH for simulating the performance of cryogenically lubricated ball bearings. As a part of this study, the particular application chosen for SHABERTH was to simulate the performance of the Space Shuttle main engine turbo-pump and pre-burner bearing system.

  10. 75 FR 69988 - Privacy Act of 1974; Computer Matching Program (United States)


    ... section 5301 of the Anti-Drug Abuse Act of 1988, and which was amended and redesignated as section 421 of... Abuse Act of 1988, Pub. L. 100-690, 21 U.S.C. 853a, which was amended and redesignated as section 421 of... court orders issued pursuant to the Denial of Federal Benefits Program. 3. Authority for Conducting the...

  11. 75 FR 12226 - Privacy Act of 1974; Computer Matching Program (United States)


    ...; the Academic Competitiveness Grant Program; the National Science and Mathematics Access to Retain... (FRS), toll free, at 1-800-877-8339. Individuals with disabilities can obtain this document in an... person listed in the preceding paragraph. Electronic Access to This Document You can view this document...

  12. A computer program for the estimation of time of death

    DEFF Research Database (Denmark)

    Lynnerup, N


    and that the temperature at death is known. Also, Marshall and Hoare's formula expresses the temperature as a function of time, and not vice versa, the latter being the problem most often encountered by forensic scientists. A simple BASIC program that enables solving of Marshall and Hoare's equation for the postmortem...

  13. INTEL: Intel based systems move up in supercomputing ranks

    CERN Document Server


    "The TOP500 supercomputer rankings released today at the Supercomputing 2002 conference show a dramatic increase in the number of Intel-based systems being deployed in high-performance computing (HPC) or supercomputing areas" (1/2 page).

  14. Do PageRank-based author rankings outperform simple citation counts?

    CERN Document Server

    Fiala, Dalibor; Žitnik, Slavko; Bajec, Marko


    The basic indicators of a researcher's productivity and impact are still the number of publications and their citation counts. These metrics are clear, straightforward, and easy to obtain. When a ranking of scholars is needed, for instance in grant, award, or promotion procedures, their use is the fastest and cheapest way of prioritizing some scientists over others. However, due to their nature, there is a danger of oversimplifying scientific achievements. Therefore, many other indicators have been proposed including the usage of the PageRank algorithm known for the ranking of webpages and its modifications suited to citation networks. Nevertheless, this recursive method is computationally expensive and even if it has the advantage of favouring prestige over popularity, its application should be well justified, particularly when compared to the standard citation counts. In this study, we analyze three large datasets of computer science papers in the categories of artificial intelligence, software engineering,...

  15. Improving the Incoherence of a Learned Dictionary via Rank Shrinkage. (United States)

    Ubaru, Shashanka; Seghouane, Abd-Krim; Saad, Yousef


    This letter considers the problem of dictionary learning for sparse signal representation whose atoms have low mutual coherence. To learn such dictionaries, at each step, we first update the dictionary using the method of optimal directions (MOD) and then apply a dictionary rank shrinkage step to decrease its mutual coherence. In the rank shrinkage step, we first compute a rank 1 decomposition of the column-normalized least squares estimate of the dictionary obtained from the MOD step. We then shrink the rank of this learned dictionary by transforming the problem of reducing the rank to a nonnegative garrotte estimation problem and solving it using a path-wise coordinate descent approach. We establish theoretical results that show that the rank shrinkage step included will reduce the coherence of the dictionary, which is further validated by experimental results. Numerical experiments illustrating the performance of the proposed algorithm in comparison to various other well-known dictionary learning algorithms are also presented.

  16. LogDet Rank Minimization with Application to Subspace Clustering. (United States)

    Kang, Zhao; Peng, Chong; Cheng, Jie; Cheng, Qiang


    Low-rank matrix is desired in many machine learning and computer vision problems. Most of the recent studies use the nuclear norm as a convex surrogate of the rank operator. However, all singular values are simply added together by the nuclear norm, and thus the rank may not be well approximated in practical problems. In this paper, we propose using a log-determinant (LogDet) function as a smooth and closer, though nonconvex, approximation to rank for obtaining a low-rank representation in subspace clustering. Augmented Lagrange multipliers strategy is applied to iteratively optimize the LogDet-based nonconvex objective function on potentially large-scale data. By making use of the angular information of principal directions of the resultant low-rank representation, an affinity graph matrix is constructed for spectral clustering. Experimental results on motion segmentation and face clustering data demonstrate that the proposed method often outperforms state-of-the-art subspace clustering algorithms.

  17. Ranking nodes in growing networks: When PageRank fails. (United States)

    Mariani, Manuel Sebastian; Medo, Matúš; Zhang, Yi-Cheng


    PageRank is arguably the most popular ranking algorithm which is being applied in real systems ranging from information to biological and infrastructure networks. Despite its outstanding popularity and broad use in different areas of science, the relation between the algorithm's efficacy and properties of the network on which it acts has not yet been fully understood. We study here PageRank's performance on a network model supported by real data, and show that realistic temporal effects make PageRank fail in individuating the most valuable nodes for a broad range of model parameters. Results on real data are in qualitative agreement with our model-based findings. This failure of PageRank reveals that the static approach to information filtering is inappropriate for a broad class of growing systems, and suggest that time-dependent algorithms that are based on the temporal linking patterns of these systems are needed to better rank the nodes.

  18. Neophilia Ranking of Scientific Journals (United States)

    Packalen, Mikko; Bhattacharya, Jay


    The ranking of scientific journals is important because of the signal it sends to scientists about what is considered most vital for scientific progress. Existing ranking systems focus on measuring the influence of a scientific paper (citations)—these rankings do not reward journals for publishing innovative work that builds on new ideas. We propose an alternative ranking based on the proclivity of journals to publish papers that build on new ideas, and we implement this ranking via a text-based analysis of all published biomedical papers dating back to 1946. In addition, we compare our neophilia ranking to citation-based (impact factor) rankings; this comparison shows that the two ranking approaches are distinct. Prior theoretical work suggests an active role for our neophilia index in science policy. Absent an explicit incentive to pursue novel science, scientists underinvest in innovative work because of a coordination problem: for work on a new idea to flourish, many scientists must decide to adopt it in their work. Rankings that are based purely on influence thus do not provide sufficient incentives for publishing innovative work. By contrast, adoption of the neophilia index as part of journal-ranking procedures by funding agencies and university administrators would provide an explicit incentive for journals to publish innovative work and thus help solve the coordination problem by increasing scientists' incentives to pursue innovative work. PMID:28713181

  19. A Research Program in Computer Technology. 1983 Annual Technical Report (United States)


    has been made in porting the MOSIS geometry software to the VAX 11/750. MOSIS is on the verge of generating MEBES files on the VAX 11/750. In recent...pilot tests, the MEBES files generated vn a VAX 11/750 were identical to those generated on a KL-2060. Computing on VAX 11/750s has several benefits...used primarily for geometry crunching. preparing MEBES files, and writing MEBES tapes. The MOSIS software provides for placement and routing of parts

  20. Pre-university training in computer science and office programming

    Directory of Open Access Journals (Sweden)

    Александр Иванович Громов


    Full Text Available Some features of the teaching of ICT in terms of building plans and learning objectives, to be the forefront in preparing students for the preparatory phase (in secondary and higher schools which brings mass teaching of computer science at an early stage to a qualitatively new level in accordance with modern requirements both in terms of national and international aspects are discussed in the article. The materials of the article are the basis for further development of modern EMC in accordance with the trends and the spirit of the time.

  1. Algebraic Functions, Computer Programming, and the Challenge of Transfer


    Schanzer, Emmanuel Tanenbaum


    Students' struggles with algebra are well documented. Prior to the introduction of functions, mathematics is typically focused on applying a set of arithmetic operations to compute an answer. The introduction of functions, however, marks the point at which mathematics begins to focus on building up abstractions as a way to solve complex problems. A common refrain about word problems is that “the equations are easy to solve - the hard part is setting them up!” A student of algebra is asked to ...

  2. A program to compute the area of an irregular polygon on a spheroidal surface

    Digital Repository Service at National Institute of Oceanography (India)

    Sivakholundu, K.M.; Prabaharan, N.

    stream_size 10201 stream_content_type text/plain stream_name Comput_Geosci_24_823.pdf.txt stream_source_info Comput_Geosci_24_823.pdf.txt Content-Encoding UTF-8 Content-Type text/plain; charset=UTF-8 SHORT NOTE A PROGRAM... FTS.IAMG.ORG PROGRAM DESCRIPTION The input polygon coordinates are converted to seconds and the latitudical origin is shifted by add- ing 90 36000 to avoid a negative latitude value (as in the situation of the southern hemisphere). Further computations...

  3. REEFER: a digital computer program for the simulation of high energy electron tubes. [Reefer

    Energy Technology Data Exchange (ETDEWEB)

    Boers, J.E.


    A digital computer program for the simulation of very high-energy electron and ion beams is described. The program includes space-charge effects through the solution of Poisson's equation and magnetic effects (both induced and applied) through the relativistic trajectory equations. Relaxation techniques are employed while alternately computing electric fields and trajectories. Execution time is generally less than 15 minutes on a CDC 6600 digital computer. Either space-charge-limited or field-emission sources may be simulated. The input data is described in detail and an example data set is included.

  4. Plane stress calculations with a two dimensional elastic-plastic computer program. [HEMP

    Energy Technology Data Exchange (ETDEWEB)

    Wilkins, M.L.; Guinan, M.W.


    In the study of ductile fracture it is useful to simulate fracture on the computer under plane stress conditions. In general, this is a three dimensional problem. Presented here is a method for adapting a two dimensional elastic-plastic computer program to calculate problems in plane stress as well as plane strain geometry. A simulation of a tension test of a flat aluminum plate pulled to failure is calculated with the modified two dimensional program. The results are compared with a fully three dimensional calculation. Finally a comparison is made with an experiment to demonstrate the effectiveness of the computational methods for studying fracture of work hardening materials.

  5. Using and Evaluating ISEE, a New Computer Program for Teaching Sampling and Statistical Inference. (United States)

    Dimitrova, Ganka; And Others


    Describes the background, development, and testing of a computer software program designed to teach principles of statistical sampling in social science courses. Compares results of a field test of the program, Introduction to Sampling Error Experiments (ISEE), with traditional instruction. (CFR)

  6. Problem-Solving Strategies and Group Processes in Small Groups Learning Computer Programming. (United States)

    Webb, Noreen M.; And Others


    Planning and debugging strategies and group processes predicting learning of computer programming were studied in 30 students aged 11 to 14. Students showed little advance planning. Factors associated with learning included high-level planning of program chunks, debugging of single statements, explaining, and verbalizing aloud while typing.…

  7. Incorporating a Computer Assisted Reading Program into an Adult Vocational Basic Skills Lab. (United States)

    Vescial, Ann; And Others

    A computer-assisted reading program was implemented in the VESL (Vocational English as a Second Language) Center at Hacienda La Puente Adult Schools (California), which provides support services to adult special needs vocational students. The purpose of the program was to improve the technical reading skills of the vocational students. The basic…

  8. Manned space station environmental control and life support system computer-aided technology assessment program (United States)

    Hall, J. B., Jr.; Pickett, S. J.; Sage, K. H.


    A computer program for assessing manned space station environmental control and life support systems technology is described. The methodology, mission model parameters, evaluation criteria, and data base for 17 candidate technologies for providing metabolic oxygen and water to the crew are discussed. Examples are presented which demonstrate the capability of the program to evaluate candidate technology options for evolving space station requirements.

  9. Computer Programs in Marine Science: Key to Oceanographic Records Documentation No. 5. (United States)

    Firestone, Mary A.

    Presented are abstracts of 700 computer programs in marine science. The programs listed are categorized under a wide range of headings which include physical oceanography, chemistry, coastal and estuarine processes, biology, pollution, air-sea interaction and heat budget, navigation and charting, curve fitting, and applied mathematics. The…

  10. Teaching Ethical Copyright Behavior: Assessing the Effects of a University-Sponsored Computing Ethics Program (United States)

    Siemens, Jennifer Christie; Kopp, Steven W.


    Universities have become sensitized to the potential for students' illegal downloading of copyrighted materials. Education has been advocated as one way to curb downloading of copyrighted digital content. This study investigates the effectiveness of a university-sponsored computing ethics education program. The program positively influenced…

  11. A multi-media computer program for training in basic professional counseling skills

    NARCIS (Netherlands)

    Adema, J.; Van Der Zee, K. I.


    This paper concerns the development of a self-instructional program for training in basic counseling skills. The product was a multimedia computer program, named GEVAT. The training under consideration was based on a traditional training in which students enhance these skills under supervision.

  12. 77 FR 43639 - Privacy Act of 1974, as Amended; Computer Matching Program (Social Security Administration (SSA... (United States)


    ...)/Department of Veterans Affairs (VA), Veterans Benefits Administration (VBA))--Match Number 1008 AGENCY: SSA... of an existing computer matching program that we are currently conducting with VA/VBA. DATES: We will... Benefits Administration (VBA) A. Participating Agencies SSA and VA/VBA. B. Purpose of the Matching Program...

  13. 77 FR 54943 - Privacy Act of 1974, as Amended; Computer Matching Program (Social Security Administration (SSA... (United States)


    ...)/Department of Veterans Affairs (VA), Veterans Benefits Administration (VBA))--Match Number 1309 AGENCY: SSA... of an existing computer matching program that we are currently conducting with VA/VBA. DATES: We will... Benefits Administration (VBA). A. Participating Agencies SSA and VA/VBA. B. Purpose of the Matching Program...

  14. Sets, Probability and Statistics: The Mathematics of Life Insurance. [Computer Program.] Second Edition. (United States)

    King, James M.; And Others

    The materials described here represent the conversion of a highly popular student workbook "Sets, Probability and Statistics: The Mathematics of Life Insurance" into a computer program. The program is designed to familiarize students with the concepts of sets, probability, and statistics, and to provide practice using real life examples. It also…

  15. A microcomputer-based tutorial program for computer-based training and patient care. (United States)

    Hayes, J R; Menke, J A


    This program provides the nonprogrammer a tool to develop easily modified computer assisted tutorials, tests, and simulations with a word processor. In addition, by using the branching and query features of this program, one can easily develop text files that provide rapid access to information related to patient care such as neonatal formularies and patient care algorithms.

  16. [The physician and the computer. 22. Programming in the Prolog language]. (United States)

    Feber, J; Kasal, P


    A modern tool of interaction with the computer is the programming language Prolog. The authors discuss its position among other languages, emphasize its different character (declarative programming). They outline basic structures of Prolog and its type of function. They also give examples of creating data bases of patients and the possibility to use this language in the creation of expert systems in medicine.

  17. Gender and Stereotypes in Motivation to Study Computer Programming for Careers in Multimedia (United States)

    Doube, Wendy; Lang, Catherine


    A multimedia university programme with relatively equal numbers of male and female students in elective programming subjects provided a rare opportunity to investigate female motivation to study and pursue computer programming in a career. The MSLQ was used to survey 85 participants. In common with research into deterrence of females from STEM…

  18. The Use of a Computer-Based Writing Program: Facilitation or Frustration? (United States)

    Chen, Chi-Fen Emily; Cheng, Wei-Yuan


    The invention of computer-based writing program has revolutionized the way of teaching second language writing. Embedded with artificial intelligence scoring engine, it can provide students with both immediate score and diagnostic feedback on their essays. In addition, some of such programs offer convenient writing and editing tools to facilitate…

  19. ADAM: A computer program to simulate selective-breeding schemes for animals

    DEFF Research Database (Denmark)

    Pedersen, L D; Sørensen, A C; Henryon, M


    ADAM is a computer program that models selective breeding schemes for animals using stochastic simulation. The program simulates a population of animals and traces the genetic changes in the population under different selective breeding scenarios. It caters to different population structures......, genetic models, selection strategies, and mating designs. ADAM can be used to evaluate breeding schemes and generate genetic data to test statistical tools...

  20. Computers-for-edu: An Advanced Business Application Programming (ABAP) Teaching Case (United States)

    Boyle, Todd A.


    The "Computers-for-edu" case is designed to provide students with hands-on exposure to creating Advanced Business Application Programming (ABAP) reports and dialogue programs, as well as navigating various mySAP Enterprise Resource Planning (ERP) transactions needed by ABAP developers. The case requires students to apply a wide variety…

  1. Factors Impacting Adult Learner Achievement in a Technology Certificate Program on Computer Networks (United States)

    Delialioglu, Omer; Cakir, Hasan; Bichelmeyer, Barbara A.; Dennis, Alan R.; Duffy, Thomas M.


    This study investigates the factors impacting the achievement of adult learners in a technology certificate program on computer networks. We studied 2442 participants in 256 institutions. The participants were older than age 18 and were enrolled in the Cisco Certified Network Associate (CCNA) technology training program as "non-degree" or…

  2. The Impact on Future Guidance Programs of Current Developments in Computer Science, Telecommunications, and Biotechnology. (United States)

    Mitchell, Lynda K.; Hardy, Philippe L.

    The purpose of this chapter is to envision how the era of technological revolution will affect the guidance, counseling, and student support programs of the future. Advances in computer science, telecommunications, and biotechnology are discussed. These advances have the potential to affect dramatically the services of guidance programs of the…

  3. Evaluation of a Pilot Program for Computer Literacy in Grades K-8. (United States)

    Still, Jacquelyn H.


    A curriculum guide and suggested instructional resource materials were developed for a school district program in computer literacy for kindergarten through grade 8. The program was piloted with 43 teachers in 9 elementary schools and 1 junior high school in 1984 in the Orange County (Florida) School District. A structured response questionnaire…




  5. DORMAN computer program (study 2.5). Volume 2: User's guide and programmer's guide. [development of data bank for computerized information storage of NASA programs (United States)

    Wray, S. T., Jr.


    The DORMAN program was developed to create and modify a data bank containing data decks which serve as input to the DORCA Computer Program. Via a remote terminal a user can access the bank, extract any data deck, modify that deck, output the modified deck to be input to the DORCA program, and save the modified deck in the data bank. This computer program is an assist in the utilization of the DORCA program. The program is dimensionless and operates almost entirely in integer mode. The program was developed on the CDC 6400/7600 complex for implementation on a UNIVAC 1108 computer.

  6. Algorithmic differentiation of pragma-defined parallel regions differentiating computer programs containing OpenMP

    CERN Document Server

    Förster, Michael


    Numerical programs often use parallel programming techniques such as OpenMP to compute the program's output values as efficient as possible. In addition, derivative values of these output values with respect to certain input values play a crucial role. To achieve code that computes not only the output values simultaneously but also the derivative values, this work introduces several source-to-source transformation rules. These rules are based on a technique called algorithmic differentiation. The main focus of this work lies on the important reverse mode of algorithmic differentiation. The inh

  7. A digital computer program for the dynamic interaction simulation of controls and structure (DISCOS), volume 1 (United States)

    Bodley, C. S.; Devers, A. D.; Park, A. C.; Frisch, H. P.


    A theoretical development and associated digital computer program system for the dynamic simulation and stability analysis of passive and actively controlled spacecraft are presented. The dynamic system (spacecraft) is modeled as an assembly of rigid and/or flexible bodies not necessarily in a topological tree configuration. The computer program system is used to investigate total system dynamic characteristics, including interaction effects between rigid and/or flexible bodies, control systems, and a wide range of environmental loadings. In addition, the program system is used for designing attitude control systems and for evaluating total dynamic system performance, including time domain response and frequency domain stability analyses.

  8. User’s Guide: Computer Program for Three-Dimensional Analysis of Building Systems (CTABS80). (United States)


    computer program TABS (1,2,3) was the direct " feedback " from the profession in the use of these programs. The first of these programs, FRMSTC, is a static...II. ~ gil a gilt eig gggg go gte *,0 glee 10 eeC ’MO * eel a 10 III. * il - C a iii - - egg’ a *. - let 4 0 a gill tel 4 C - - tie 10 lii

  9. Wikipedia ranking of world universities (United States)

    Lages, José; Patt, Antoine; Shepelyansky, Dima L.


    We use the directed networks between articles of 24 Wikipedia language editions for producing the wikipedia ranking of world Universities (WRWU) using PageRank, 2DRank and CheiRank algorithms. This approach allows to incorporate various cultural views on world universities using the mathematical statistical analysis independent of cultural preferences. The Wikipedia ranking of top 100 universities provides about 60% overlap with the Shanghai university ranking demonstrating the reliable features of this approach. At the same time WRWU incorporates all knowledge accumulated at 24 Wikipedia editions giving stronger highlights for historically important universities leading to a different estimation of efficiency of world countries in university education. The historical development of university ranking is analyzed during ten centuries of their history.

  10. Low-rank coal research

    Energy Technology Data Exchange (ETDEWEB)

    Weber, G. F.; Laudal, D. L.


    This work is a compilation of reports on ongoing research at the University of North Dakota. Topics include: Control Technology and Coal Preparation Research (SO{sub x}/NO{sub x} control, waste management), Advanced Research and Technology Development (turbine combustion phenomena, combustion inorganic transformation, coal/char reactivity, liquefaction reactivity of low-rank coals, gasification ash and slag characterization, fine particulate emissions), Combustion Research (fluidized bed combustion, beneficiation of low-rank coals, combustion characterization of low-rank coal fuels, diesel utilization of low-rank coals), Liquefaction Research (low-rank coal direct liquefaction), and Gasification Research (hydrogen production from low-rank coals, advanced wastewater treatment, mild gasification, color and residual COD removal from Synfuel wastewaters, Great Plains Gasification Plant, gasifier optimization).

  11. Dynamic Programming and Graph Algorithms in Computer Vision* (United States)

    Felzenszwalb, Pedro F.; Zabih, Ramin


    Optimization is a powerful paradigm for expressing and solving problems in a wide range of areas, and has been successfully applied to many vision problems. Discrete optimization techniques are especially interesting, since by carefully exploiting problem structure they often provide non-trivial guarantees concerning solution quality. In this paper we briefly review dynamic programming and graph algorithms, and discuss representative examples of how these discrete optimization techniques have been applied to some classical vision problems. We focus on the low-level vision problem of stereo; the mid-level problem of interactive object segmentation; and the high-level problem of model-based recognition. PMID:20660950

  12. Effectiveness of computer ergonomics interventions for an engineering company: a program evaluation. (United States)

    Goodman, Glenn; Landis, James; George, Christina; McGuire, Sheila; Shorter, Crystal; Sieminski, Michelle; Wilson, Tamika


    Ergonomic principles at the computer workstation may reduce the occurrence of work related injuries commonly associated with intensive computer use. A program implemented in 2001 by an occupational therapist and a physical therapist utilized these preventative measures with education about ergonomics, individualized evaluations of computer workstations, and recommendations for ergonomic and environmental changes. This study examined program outcomes and perceived effectiveness based on review of documents, interviews, and surveys of the employees and the plant manager. The program was deemed successful as shown by 59% of all therapist recommendations and 74% of ergonomic recommendations being implemented by the company, with an 85% satisfaction rate for the ergonomic interventions and an overall employee satisfaction rate of 70%. Eighty-one percent of the physical problems reported by employees were resolved to their satisfaction one year later. Successful implementation of ergonomics programs depend upon effective communication and education of the consumers, and the support, cooperation and collaboration of management and employees.

  13. The Implementation of Blended Learning Using Android-Based Tutorial Video in Computer Programming Course II (United States)

    Huda, C.; Hudha, M. N.; Ain, N.; Nandiyanto, A. B. D.; Abdullah, A. G.; Widiaty, I.


    Computer programming course is theoretical. Sufficient practice is necessary to facilitate conceptual understanding and encouraging creativity in designing computer programs/animation. The development of tutorial video in an Android-based blended learning is needed for students’ guide. Using Android-based instructional material, students can independently learn anywhere and anytime. The tutorial video can facilitate students’ understanding about concepts, materials, and procedures of programming/animation making in detail. This study employed a Research and Development method adapting Thiagarajan’s 4D model. The developed Android-based instructional material and tutorial video were validated by experts in instructional media and experts in physics education. The expert validation results showed that the Android-based material was comprehensive and very feasible. The tutorial video was deemed feasible as it received average score of 92.9%. It was also revealed that students’ conceptual understanding, skills, and creativity in designing computer program/animation improved significantly.

  14. High performance computing and communications grand challenges program

    Energy Technology Data Exchange (ETDEWEB)

    Solomon, J.E.; Barr, A.; Chandy, K.M.; Goddard, W.A., III; Kesselman, C.


    The so-called protein folding problem has numerous aspects, however it is principally concerned with the {ital de novo} prediction of three-dimensional (3D) structure from the protein primary amino acid sequence, and with the kinetics of the protein folding process. Our current project focuses on the 3D structure prediction problem which has proved to be an elusive goal of molecular biology and biochemistry. The number of local energy minima is exponential in the number of amino acids in the protein. All current methods of 3D structure prediction attempt to alleviate this problem by imposing various constraints that effectively limit the volume of conformational space which must be searched. Our Grand Challenge project consists of two elements: (1) a hierarchical methodology for 3D protein structure prediction; and (2) development of a parallel computing environment, the Protein Folding Workbench, for carrying out a variety of protein structure prediction/modeling computations. During the first three years of this project, we are focusing on the use of two proteins selected from the Brookhaven Protein Data Base (PDB) of known structure to provide validation of our prediction algorithms and their software implementation, both serial and parallel. Both proteins, protein L from {ital peptostreptococcus magnus}, and {ital streptococcal} protein G, are known to bind to IgG, and both have an {alpha} {plus} {beta} sandwich conformation. Although both proteins bind to IgG, they do so at different sites on the immunoglobin and it is of considerable biological interest to understand structurally why this is so. 12 refs., 1 fig.


    Directory of Open Access Journals (Sweden)

    Marek Horyński


    Full Text Available Currently, cloud computing models are provided with increasing a number of new functionalities. The paper describes a laboratory model of intelligent KNX system which makes it possible to test the practical use of cloud computing in components programming for this system. The cloud computing elements were used for this purpose. Innovatory features of services transferred to the cloud computing models consists in the integration of advanced IT techniques and Internet technologies based systems with the users services, in this case with the management of an intelligent building system. The software ETS5 installed on workstation is another important component of the station being discussed.

  16. Automatic Parallelization Tool: Classification of Program Code for Parallel Computing

    Directory of Open Access Journals (Sweden)

    Mustafa Basthikodi


    Full Text Available Performance growth of single-core processors has come to a halt in the past decade, but was re-enabled by the introduction of parallelism in processors. Multicore frameworks along with Graphical Processing Units empowered to enhance parallelism broadly. Couples of compilers are updated to developing challenges forsynchronization and threading issues. Appropriate program and algorithm classifications will have advantage to a great extent to the group of software engineers to get opportunities for effective parallelization. In present work we investigated current species for classification of algorithms, in that related work on classification is discussed along with the comparison of issues that challenges the classification. The set of algorithms are chosen which matches the structure with different issues and perform given task. We have tested these algorithms utilizing existing automatic species extraction toolsalong with Bones compiler. We have added functionalities to existing tool, providing a more detailed characterization. The contributions of our work include support for pointer arithmetic, conditional and incremental statements, user defined types, constants and mathematical functions. With this, we can retain significant data which is not captured by original speciesof algorithms. We executed new theories into the device, empowering automatic characterization of program code.

  17. Tangible computer programming: Exploring the use of emerging technology in classrooms and science museums (United States)

    Horn, Michael S.

    In considering ways to improve the use of digital technology in educational settings, it is helpful to look beyond desktop computers and conventional modes of interaction and consider the flood of emerging technologies that already play a prominent role in the everyday lives of children. In this dissertation, I will present a research project that builds on tangible user interface (TUI) technology to support computer programming and robotics activities in education settings. In particular, I will describe the design and implementation of a novel tangible computer programming language called Tern. I will also describe an evaluation of Tern's use in both formal and informal educational settings--as part of an interactive exhibit on robotics and computer programming called Robot Park on display at the Boston Museum of Science; and as part of a curriculum unit piloted in several kindergarten classrooms in the greater Boston area. In both cases, Tern allows children to create simple computer programs to control a robot. However, rather than using a keyboard or mouse to write programs on a computer screen, children instead use Tern to construct physical algorithmic structures using a collection of interlocking wooden blocks. The goal of this work is not to propose that tangible programming languages are general purpose tools that should replace existing graphical programming languages; rather, I will present evidence to support the argument that tangible programming begins to make sense when one considers the contexts and constraints of specific educational settings. Moreover, in these settings tangible languages can compensate for some of the shortcomings of graphical and text-based systems that have limited their use.

  18. Statistical methods for ranking data

    CERN Document Server

    Alvo, Mayer


    This book introduces advanced undergraduate, graduate students and practitioners to statistical methods for ranking data. An important aspect of nonparametric statistics is oriented towards the use of ranking data. Rank correlation is defined through the notion of distance functions and the notion of compatibility is introduced to deal with incomplete data. Ranking data are also modeled using a variety of modern tools such as CART, MCMC, EM algorithm and factor analysis. This book deals with statistical methods used for analyzing such data and provides a novel and unifying approach for hypotheses testing. The techniques described in the book are illustrated with examples and the statistical software is provided on the authors’ website.


    CERN Multimedia

    I. Fisk


    Introduction It has been a very active quarter in Computing with interesting progress in all areas. The activity level at the computing facilities, driven by both organised processing from data operations and user analysis, has been steadily increasing. The large-scale production of simulated events that has been progressing throughout the fall is wrapping-up and reprocessing with pile-up will continue. A large reprocessing of all the proton-proton data has just been released and another will follow shortly. The number of analysis jobs by users each day, that was already hitting the computing model expectations at the time of ICHEP, is now 33% higher. We are expecting a busy holiday break to ensure samples are ready in time for the winter conferences. Heavy Ion An activity that is still in progress is computing for the heavy-ion program. The heavy-ion events are collected without zero suppression, so the event size is much large at roughly 11 MB per event of RAW. The central collisions are more complex and...

  20. Near-term hybrid vehicle program, phase 1. Appendix B: Design trade-off studies report. Volume 3: Computer program listings (United States)


    A description and listing is presented of two computer programs: Hybrid Vehicle Design Program (HYVELD) and Hybrid Vehicle Simulation Program (HYVEC). Both of the programs are modifications and extensions of similar programs developed as part of the Electric and Hybrid Vehicle System Research and Development Project.

  1. Pulling Rank: Military Rank Affects Hormone Levels and Fairness in an Allocation Experiment

    Directory of Open Access Journals (Sweden)

    Benjamin Siart


    Full Text Available Status within social hierarchies has great effects on the lives of socially organized mammals. Its effects on human behavior and related physiology however is relatively little studied. The present study investigated the impact of military rank on fairness and behavior in relation to salivary cortisol (C and testosterone (T levels in male soldiers. For this purpose 180 members of the Austrian Armed Forces belonging to two distinct rank groups participated in two variations of a computer-based guard duty allocation experiment. The rank groups were 1 warrant officers (High Rank, HR and 2 enlisted men (Low Rank, LR. One soldier from each rank group participated in every experiment. At the beginning of the experiment, one participant was assigned to start standing guard and the other participant at rest. The participant who started at rest could choose if and when to relieve his fellow soldier and therefore had control over the experiment. In order to trigger perception of unfair behavior, an additional experiment was conducted which was manipulated by the experimenter. In the manipulated version both soldiers started in the standing guard position and were never relieved, believing that their opponent was at rest, not relieving them. Our aim was to test whether unfair behavior causes a physiological reaction. Saliva samples for hormone analysis were collected at regular intervals throughout the experiment.We found that in the un-manipulated setup high-ranking soldiers spent less time standing guard than lower ranking individuals. Rank was a significant predictor for C but not for T levels during the experiment. C levels in the HR group were higher than in LR group. C levels were also elevated in the manipulated experiment compared to the un-manipulated experiment, especially in LR. We assume that the elevated C levels in HR were caused by HR feeling their status challenged by the situation of having to negotiate with an individual of lower military

  2. Pulling Rank: Military Rank Affects Hormone Levels and Fairness in an Allocation Experiment. (United States)

    Siart, Benjamin; Pflüger, Lena S; Wallner, Bernard


    Status within social hierarchies has great effects on the lives of socially organized mammals. Its effects on human behavior and related physiology, however, is relatively little studied. The present study investigated the impact of military rank on fairness and behavior in relation to salivary cortisol (C) and testosterone (T) levels in male soldiers. For this purpose 180 members of the Austrian Armed Forces belonging to two distinct rank groups participated in two variations of a computer-based guard duty allocation experiment. The rank groups were (1) warrant officers (high rank, HR) and (2) enlisted men (low rank, LR). One soldier from each rank group participated in every experiment. At the beginning of the experiment, one participant was assigned to start standing guard and the other participant at rest. The participant who started at rest could choose if and when to relieve his fellow soldier and therefore had control over the experiment. In order to trigger perception of unfair behavior, an additional experiment was conducted which was manipulated by the experimenter. In the manipulated version both soldiers started in the standing guard position and were never relieved, believing that their opponent was at rest, not relieving them. Our aim was to test whether unfair behavior causes a physiological reaction. Saliva samples for hormone analysis were collected at regular intervals throughout the experiment. We found that in the un-manipulated setup high-ranking soldiers spent less time standing guard than lower ranking individuals. Rank was a significant predictor for C but not for T levels during the experiment. C levels in the HR group were higher than in the LR group. C levels were also elevated in the manipulated experiment compared to the un-manipulated experiment, especially in LR. We assume that the elevated C levels in HR were caused by HR feeling their status challenged by the situation of having to negotiate with an individual of lower military rank

  3. Space shuttle solid rocket booster recovery system definition. Volume 3: SRB water impact loads computer program, user's manual (United States)


    This user's manual describes the FORTRAN IV computer program developed to compute the total vertical load, normal concentrated pressure loads, and the center of pressure of typical SRB water impact slapdown pressure distributions specified in the baseline configuration. The program prepares the concentrated pressure load information in punched card format suitable for input to the STAGS computer program. In addition, the program prepares for STAGS input the inertia reacting loads to the slapdown pressure distributions.

  4. Low-rank and sparse modeling for visual analysis

    CERN Document Server

    Fu, Yun


    This book provides a view of low-rank and sparse computing, especially approximation, recovery, representation, scaling, coding, embedding and learning among unconstrained visual data. The book includes chapters covering multiple emerging topics in this new field. It links multiple popular research fields in Human-Centered Computing, Social Media, Image Classification, Pattern Recognition, Computer Vision, Big Data, and Human-Computer Interaction. Contains an overview of the low-rank and sparse modeling techniques for visual analysis by examining both theoretical analysis and real-world applic

  5. Ranking nodes in growing networks: When PageRank fails (United States)

    Mariani, Manuel Sebastian; Medo, Matúš; Zhang, Yi-Cheng


    PageRank is arguably the most popular ranking algorithm which is being applied in real systems ranging from information to biological and infrastructure networks. Despite its outstanding popularity and broad use in different areas of science, the relation between the algorithm’s efficacy and properties of the network on which it acts has not yet been fully understood. We study here PageRank’s performance on a network model supported by real data, and show that realistic temporal effects make PageRank fail in individuating the most valuable nodes for a broad range of model parameters. Results on real data are in qualitative agreement with our model-based findings. This failure of PageRank reveals that the static approach to information filtering is inappropriate for a broad class of growing systems, and suggest that time-dependent algorithms that are based on the temporal linking patterns of these systems are needed to better rank the nodes.

  6. Implementation of the Distributed Parallel Program for Geoid Heights Computation Using MPI and Openmp (United States)

    Lee, S.; Kim, J.; Jung, Y.; Choi, J.; Choi, C.


    Much research have been carried out using optimization algorithms for developing high-performance program, under the parallel computing environment with the evolution of the computer hardware technology such as dual-core processor and so on. Then, the studies by the parallel computing in geodesy and surveying fields are not so many. The present study aims to reduce running time for the geoid heights computation and carrying out least-squares collocation to improve its accuracy using distributed parallel technology. A distributed parallel program was developed in which a multi-core CPU-based PC cluster was adopted using MPI and OpenMP library. Geoid heights were calculated by the spherical harmonic analysis using the earth geopotential model of the National Geospatial-Intelligence Agency(2008). The geoid heights around the Korean Peninsula were calculated and tested in diskless-based PC cluster environment. As results, for the computing geoid heights by a earth geopotential model, the distributed parallel program was confirmed more effective to reduce the computational time compared to the sequential program.

  7. Ranking documents with a thesaurus. (United States)

    Rada, R; Bicknell, E


    This article reports on exploratory experiments in evaluating and improving a thesaurus through studying its effect on retrieval. A formula called DISTANCE was developed to measure the conceptual distance between queries and documents encoded as sets of thesaurus terms. DISTANCE references MeSH (Medical Subject Headings) and assesses the degree of match between a MeSH-encoded query and document. The performance of DISTANCE on MeSH is compared to the performance of people in the assessment of conceptual distance between queries and documents, and is found to simulate with surprising accuracy the human performance. The power of the computer simulation stems both from the tendency of people to rely heavily on broader-than (BT) relations in making decisions about conceptual distance and from the thousands of accurate BT relations in MeSH. One source for discrepancy between the algorithms' measurement of closeness between query and document and people's measurement of closeness between query and document is occasional inconsistency in the BT relations. Our experiments with adding non-BT relations to MeSH showed how these non-BT non-BT relations to MeSH showed how these non-BT relations could improve document ranking, if DISTANCE were also appropriately revised to treat these relations differently from BT relations.

  8. Multithreaded transactions in scientific computing. The Growth06_v2 program (United States)

    Daniluk, Andrzej


    Writing a concurrent program can be more difficult than writing a sequential program. Programmer needs to think about synchronization, race conditions and shared variables. Transactions help reduce the inconvenience of using threads. A transaction is an abstraction, which allows programmers to group a sequence of actions on the program into a logical, higher-level computation unit. This paper presents a new version of the GROWTHGr and GROWTH06 programs. New version program summaryProgram title: GROWTH06_v2 Catalogue identifier: ADVL_v2_1 Program summary URL: Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, No. of lines in distributed program, including test data, etc.: 65 255 No. of bytes in distributed program, including test data, etc.: 865 985 Distribution format: tar.gz Programming language: Object Pascal Computer: Pentium-based PC Operating system: Windows 9x, XP, NT, Vista RAM: more than 1 MB Classification: 4.3, 7.2, 6.2, 8, 14 Catalogue identifier of previous version: ADVL_v2_0 Journal reference of previous version: Comput. Phys. Comm. 175 (2006) 678 Does the new version supersede the previous version?: Yes Nature of problem: The programs compute the RHEED intensities during the growth of thin epitaxial structures prepared using the molecular beam epitaxy (MBE). The computations are based on the use of kinematical diffraction theory. Solution method: Epitaxial growth of thin films is modelled by a set of non-linear differential equations [1]. The Runge-Kutta method with adaptive stepsize control was used for solving initial value problem for non-linear differential equations [2]. Reasons for new version: According to the users' suggestions functionality of the program has been improved. Moreover, new use cases have been added which make the handling of the program easier and more

  9. Computer program to calculate three-dimensional boundary layer flows over wings with wall mass transfer (United States)

    Mclean, J. D.; Randall, J. L.


    A system of computer programs for calculating three dimensional transonic flow over wings, including details of the three dimensional viscous boundary layer flow, was developed. The flow is calculated in two overlapping regions: an outer potential flow region, and a boundary layer region in which the first order, three dimensional boundary layer equations are numerically solved. A consistent matching of the two solutions is achieved iteratively, thus taking into account viscous-inviscid interaction. For the inviscid outer flow calculations, the Jameson-Caughey transonic wing program FLO 27 is used, and the boundary layer calculations are performed by a finite difference boundary layer prediction program. Interface programs provide communication between the two basic flow analysis programs. Computed results are presented for the NASA F8 research wing, both with and without distributed surface suction.


    Directory of Open Access Journals (Sweden)

    D. G. Koliushko


    Full Text Available Purpose. Creating a computer program for interpreting the results of vertical sounding the soil in the form of multilayer model most typical for Ukraine. Methodology. The algorithm of the program is constructed on determination the soil structure with the help of the method of point source current, method of analogy and method of equivalent. The option of automatic interpretation based on Hook-Jeeves method. The program is implemented in the programming language Delphi. Results. The computer program «VEZ-4A» has a possibility of the interactive and automatic interpretation sounding results in the multi-layered geoelectrical model. Originality. In first time the computer program for analyzing and interpreting results of the soil sounding by Wenner configuration was created on the base of the analytical solution for field of current point source located in four-, three- or two-layer structure. In paper the review is presented and basic functions of our program are analyzed. Practical value. The program «VEZ-4A» is created and adapted for use in the electromagnetic diagnostics of grounding of existing power plants and substations.

  11. Teaching Scientific Computing: A Model-Centered Approach to Pipeline and Parallel Programming with C

    Directory of Open Access Journals (Sweden)

    Vladimiras Dolgopolovas


    Full Text Available The aim of this study is to present an approach to the introduction into pipeline and parallel computing, using a model of the multiphase queueing system. Pipeline computing, including software pipelines, is among the key concepts in modern computing and electronics engineering. The modern computer science and engineering education requires a comprehensive curriculum, so the introduction to pipeline and parallel computing is the essential topic to be included in the curriculum. At the same time, the topic is among the most motivating tasks due to the comprehensive multidisciplinary and technical requirements. To enhance the educational process, the paper proposes a novel model-centered framework and develops the relevant learning objects. It allows implementing an educational platform of constructivist learning process, thus enabling learners’ experimentation with the provided programming models, obtaining learners’ competences of the modern scientific research and computational thinking, and capturing the relevant technical knowledge. It also provides an integral platform that allows a simultaneous and comparative introduction to pipelining and parallel computing. The programming language C for developing programming models and message passing interface (MPI and OpenMP parallelization tools have been chosen for implementation.

  12. Kriging for Simulation Metamodeling: Experimental Design, Reduced Rank Kriging, and Omni-Rank Kriging (United States)

    Hosking, Michael Robert

    explanation through tests of reduced rank kriging's performance over different situations. In total, reduced rank kriging is a useful tool for simulation metamodeling. For the third contribution we will answer the question of how to find the best rank for reduced rank kriging. We do this by creating an alternative method which does not need to search for a particular rank. Instead it uses all potential ranks; we call this approach omnirank kriging. This modification realizes the potential gains from reduced rank kriging and provides a workable methodology for simulation metamodeling. Finally, we will demonstrate the use and value of these developments on two case studies, a clinic operation problem and a location problem. These cases will validate the value of this research. Simulation metamodeling always attempts to extract maximum information from limited data. Each one of these contributions will allow analysts to make better use of their constrained computational budgets.

  13. University Rankings in Critical Perspective (United States)

    Pusser, Brian; Marginson, Simon


    This article addresses global postsecondary ranking systems by using critical-theoretical perspectives on power. This research suggests rankings are at once a useful lens for studying power in higher education and an important instrument for the exercise of power in service of dominant norms in global higher education. (Contains 1 table and 1…

  14. University Ranking as Social Exclusion (United States)

    Amsler, Sarah S.; Bolsmann, Chris


    In this article we explore the dual role of global university rankings in the creation of a new, knowledge-identified, transnational capitalist class and in facilitating new forms of social exclusion. We examine how and why the practice of ranking universities has become widely defined by national and international organisations as an important…

  15. Computer program system for evaluation of FP nuclear data for JENDL. Smooth part

    Energy Technology Data Exchange (ETDEWEB)

    Nakagawa, Tsuneo [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment; Watanabe, Takashi; Iijima, Shungo


    This report describes computer programs used to evaluate nuclear data of fission product (FP) nuclides stored in an evaluated nuclear data library JENDL, especially in the smooth part above the resonance region. Many programs were used for determination of nuclear model parameters, calculation of nuclear data, handling of experimental and/or calculated data, and so on. Among them, reported here are programs for determination of level density parameters (ENSDFRET, LVLPLOT, LEVDES), for making sets of JCL and input data for the theoretical calculation program CASTHY (JOBSETTER, INDES/CASTHY), and for conversion of format of CASTHY output files to the ENDF format (CTOB2). (author). 51 refs.

  16. Uses of Computed Tomography in the NASA Materials Science Program (United States)

    Engel, H. Peter; Gillies, Donald C.; Curreri, Peter (Technical Monitor)


    Computed Tomography (CT) has proved to be of inestimable use in providing a rapid evaluation of a variety of samples from Mechanics of Granular Materials (MGM) to electronic materials (Ge-Si alloys) to space grown materials such as meteorites. The system at Kennedy Space Center (KSC), because of its convenient geographical location, is ideal for examining samples immediately after returning to Earth. It also has the advantage of the choice of fluxes, and in particular the use of a radioactive cobalt source, which is basically monochromatic. This permits a reasonable measurement of density to be made from which chemical composition can be determined. Due to the current dearth of long duration space grown materials, the CT instrument has been used to characterize materials in preparation for flight, to determine thermal expansion values, and to examine long duration space grown materials, i.e. meteorites. The work will first describe the establishment of the protocol for obtaining the optimum density readings for any material. This will include both the effects of the hardware or instrumental parameters that can be controlled, and the techniques used to process the CT data. Examples will be given of the compositional variation along single crystals of germanium-silicon alloys. Density variation with temperature has been measured in preparation for future materials science experiments; this involved the fabrication and installation of a single zone furnace incorporating a heat pipe to ensure of high temperature uniformity. At the time of writing the thermal expansion of lead has been measured from room temperature to 900 C. Three methods are available. Digital radiography enable length changes to be determined. Prior to melting the sample is small than the container and the diameter change can be measured. Most critical, however, is the density change in solid, through the melting region, and in the liquid state. These data are needed for engineering purposes to aid

  17. Strategies for casualty mitigation programs by using advanced tsunami computation (United States)

    IMAI, K.; Imamura, F.


    . Finally, compiling all this information on a tsunami hazard and risk map with the tsunami inundation animation, it is possible to create and propose strategies for casualty mitigation programs.

  18. Implications of the Turing machine model of computation for processor and programming language design (United States)

    Hunter, Geoffrey


    A computational process is classified according to the theoretical model that is capable of executing it; computational processes that require a non-predeterminable amount of intermediate storage for their execution are Turing-machine (TM) processes, while those whose storage are predeterminable are Finite Automation (FA) processes. Simple processes (such as traffic light controller) are executable by Finite Automation, whereas the most general kind of computation requires a Turing Machine for its execution. This implies that a TM process must have a non-predeterminable amount of memory allocated to it at intermediate instants of its execution; i.e. dynamic memory allocation. Many processes encountered in practice are TM processes. The implication for computational practice is that the hardware (CPU) architecture and its operating system must facilitate dynamic memory allocation, and that the programming language used to specify TM processes must have statements with the semantic attribute of dynamic memory allocation, for in Alan Turing"s thesis on computation (1936) the "standard description" of a process is invariant over the most general data that the process is designed to process; i.e. the program describing the process should never have to be modified to allow for differences in the data that is to be processed in different instantiations; i.e. data-invariant programming. Any non-trivial program is partitioned into sub-programs (procedures, subroutines, functions, modules, etc). Examination of the calls/returns between the subprograms reveals that they are nodes in a tree-structure; this tree-structure is independent of the programming language used to encode (define) the process. Each sub-program typically needs some memory for its own use (to store values intermediate between its received data and its computed results); this locally required memory is not needed before the subprogram commences execution, and it is not needed after its execution terminates

  19. Teaching programming to non-STEM novices: a didactical study of computational thinking and non-STEM computing education

    DEFF Research Database (Denmark)

    Spangsberg, Thomas Hvid

    The PhD project presented here is concerned with teaching introductory programming to non-STEM students. Particularly students in the field of Information Studies and HCI. The research is carried out through three Teaching and Learning Explorations (TLE - I-III) which are set in an inductive acti......, the translation- grammar method of a structural approach to natural language teaching shows promising results when actualized in a concrete exercise named code-labelling.......The PhD project presented here is concerned with teaching introductory programming to non-STEM students. Particularly students in the field of Information Studies and HCI. The research is carried out through three Teaching and Learning Explorations (TLE - I-III) which are set in an inductive action...... research approach. Computational thinking plays a significant role in computing education but it is still unclear how it should be interpreted to best serve its purpose. Constructionism and Computational Making seems to be promising frameworks to do this. In regards to specific teaching activities...

  20. PageRank tracker: from ranking to tracking. (United States)

    Gong, Chen; Fu, Keren; Loza, Artur; Wu, Qiang; Liu, Jia; Yang, Jie


    Video object tracking is widely used in many real-world applications, and it has been extensively studied for over two decades. However, tracking robustness is still an issue in most existing methods, due to the difficulties with adaptation to environmental or target changes. In order to improve adaptability, this paper formulates the tracking process as a ranking problem, and the PageRank algorithm, which is a well-known webpage ranking algorithm used by Google, is applied. Labeled and unlabeled samples in tracking application are analogous to query webpages and the webpages to be ranked, respectively. Therefore, determining the target is equivalent to finding the unlabeled sample that is the most associated with existing labeled set. We modify the conventional PageRank algorithm in three aspects for tracking application, including graph construction, PageRank vector acquisition and target filtering. Our simulations with the use of various challenging public-domain video sequences reveal that the proposed PageRank tracker outperforms mean-shift tracker, co-tracker, semiboosting and beyond semiboosting trackers in terms of accuracy, robustness and stability.

  1. Computer-based programs on acquisition of reading skills in schoolchildren (review of contemporary foreign investigations

    Directory of Open Access Journals (Sweden)

    Prikhoda N.A.


    Full Text Available The article presents a description of 17 computer-based programs, which were used over the last 5 years (2008—2013 in 15 studies of computer-assisted reading instruction and intervention of schoolchildren. The article includes a description of specificity of various terms used in the above-mentioned studies and the contents of training sessions. The article also carries out a brief analysis of main characteristics of computer-based techniques — language of instruction, age and basic characteristics of students, duration and frequency of training sessions, dependent variables of education. Special attention is paid to efficiency of acquisition of different reading skills through computer-based programs in comparison to traditional school instruction.


    Directory of Open Access Journals (Sweden)

    Vitalii M. Bazurin


    Full Text Available Research of physical processes on computer models is the one of ways of research approach in the study of general physics in pedagogical universities. The basic elements of software for research of physical processes on computer models are certain in the article: structure of program mean and feature of interface. The author offers his vision of structure of program facilities by means which the computer models of the physical phenomena and processes are realized: block of registration, block of entrance background check, block of design of physical process and block of results verification. A structure of this kinds of software is given, in opinion of author, is universal. The program facilities developed by an author are described in the article.

  3. Development of a Computer Program for the Integrated Control of the Fuel Homogeneity Measurement System

    Energy Technology Data Exchange (ETDEWEB)

    Shin, H. S.; Jang, J. W.; Lee, Y. H.; Oh, S. J.; Park, H. D.; Kim, C. K


    The computer program is developed based on Visual C++, which is equipped with a user-friendly interface of the input/output(I/O) and a display function for the measuring conditions. This program consists of three parts which are the port communication, PLC(Programmable Logic Controller) and the MCA(Multi Channel Analyzer) control parts. The communication type between the CPU of the PLC module box and the computer is selected as be the Rs-232 asynchronous type and the thread method is adapted in the development of the first part of the program. The PLC-related program has been developed so that the data communication between the PLC CPU and the computer could be harmonized with the unique commands which have already been defined in the PLC. The measuring space and time intervals, the start and end ROI(region of interest) values, and the allowable error limitation are input at each measurement in this program. Finally the controlling MCA program has been developed by using Canberra's programming library which contains several files including the head files in which the variable and the function of C++ are declared according to the MCA function. The performance test has been carried out through an application of the developed computer program to the homogeneity measurement system. The gamma counts at 28 measuring points of a fuel rod of 700 mm in length are measured for 50 sec at each point. It was revealed that the measurement results are better than the previous ones in respects of the measurement accuracy and a measurement time saving could be achieved. It was concluded that the gamma measurement system can be improved through equipping it with the developed control program.

  4. An Overview of the Integration of Problem Based Learning into an existing Computer Science Programming Module


    O'Kelly, Jackie; Mooney, Aidan; Bergin, Susan; Gaughran, Peter; Ghent, John


    In this paper we present an overview of the use of Problem Based Learning (PBL) in a first year Computer Science programming module.PBL was not employed in any of the programmong modules within the Department of Computer Science and assessment and learning for this module was on an individual student basis. We outline the problems that we encountered with our previous approach for teaching this module and our rationale for enhancing our approach through PBL.

  5. A Computer Program for the Computation of Running Gear Temperatures Using Green's Function (United States)

    Koshigoe, S.; Murdock, J. W.; Akin, L. S.; Townsend, D. P.


    A new technique has been developed to study two dimensional heat transfer problems in gears. This technique consists of transforming the heat equation into a line integral equation with the use of Green's theorem. The equation is then expressed in terms of eigenfunctions that satisfy the Helmholtz equation, and their corresponding eigenvalues for an arbitrarily shaped region of interest. The eigenfunction are obtalned by solving an intergral equation. Once the eigenfunctions are found, the temperature is expanded in terms of the eigenfunctions with unknown time dependent coefficients that can be solved by using Runge Kutta methods. The time integration is extremely efficient. Therefore, any changes in the time dependent coefficients or source terms in the boundary conditions do not impose a great computational burden on the user. The method is demonstrated by applying it to a sample gear tooth. Temperature histories at representative surface locatons are given.

  6. Using the TouchMath Program to Teach Mathematical Computation to At-Risk Students and Students with Disabilities (United States)

    Ellingsen, Ryleigh; Clinton, Elias


    This manuscript reviews the empirical literature of the TouchMath© instructional program. The TouchMath© program is a commercial mathematics series that uses a dot notation system to provide multisensory instruction of computation skills. Using the program, students are taught to solve computational tasks in a multisensory manner that does not…

  7. STARS: An Integrated, Multidisciplinary, Finite-Element, Structural, Fluids, Aeroelastic, and Aeroservoelastic Analysis Computer Program (United States)

    Gupta, K. K.


    A multidisciplinary, finite element-based, highly graphics-oriented, linear and nonlinear analysis capability that includes such disciplines as structures, heat transfer, linear aerodynamics, computational fluid dynamics, and controls engineering has been achieved by integrating several new modules in the original STARS (STructural Analysis RoutineS) computer program. Each individual analysis module is general-purpose in nature and is effectively integrated to yield aeroelastic and aeroservoelastic solutions of complex engineering problems. Examples of advanced NASA Dryden Flight Research Center projects analyzed by the code in recent years include the X-29A, F-18 High Alpha Research Vehicle/Thrust Vectoring Control System, B-52/Pegasus Generic Hypersonics, National AeroSpace Plane (NASP), SR-71/Hypersonic Launch Vehicle, and High Speed Civil Transport (HSCT) projects. Extensive graphics capabilities exist for convenient model development and postprocessing of analysis results. The program is written in modular form in standard FORTRAN language to run on a variety of computers, such as the IBM RISC/6000, SGI, DEC, Cray, and personal computer; associated graphics codes use OpenGL and IBM/graPHIGS language for color depiction. This program is available from COSMIC, the NASA agency for distribution of computer programs.

  8. Beyond Zipf's Law: The Lavalette Rank Function and its Properties

    CERN Document Server

    Fontanelli, Oscar; Yang, Yaning; Cocho, Germinal; Li, Wentian


    Although Zipf's law is widespread in natural and social data, one often encounters situations where one or both ends of the ranked data deviate from the power-law function. Previously we proposed the Beta rank function to improve the fitting of data which does not follow a perfect Zipf's law. Here we show that when the two parameters in the Beta rank function have the same value, the Lavalette rank function, the probability density function can be derived analytically. We also show both computationally and analytically that Lavalette distribution is approximately equal, though not identical, to the lognormal distribution. We illustrate the utility of Lavalette rank function in several datasets. We also address three analysis issues on the statistical testing of Lavalette fitting function, comparison between Zipf's law and lognormal distribution through Lavalette function, and comparison between lognormal distribution and Lavalette distribution.

  9. A program of computer-aided coordination analysis for an undergraduate course in protective relaying

    Energy Technology Data Exchange (ETDEWEB)

    Chu, W.C.; Liu, M.C. (Tatung Inst. of Tech., Taipei (Taiwan, Province of China))


    This paper introduces a program used as a practicing tool for the protective relaying course of undergraduate level. This progress allows students to practice protective devices coordination in power systems on a personal computer by representing the time-current curve with dynamic computer graphics. The program is also provided with a debugging system that is designed to inform the user of any problem found in his protection scheme. In order to make students fully understand the dynamic performance of protective devices coordination, the program allows students to assign the location of a fault and it will display the operating time of the protective devices according to the tripping sequence. The program also has the function of automatic coordination analysis, the results of which can be compared with the user's scheme and can also help the user to have a better understanding of what are actually used in the industry today.

  10. Design and evaluation of the computer-based training program Calcularis for enhancing numerical cognition

    Directory of Open Access Journals (Sweden)

    Tanja eKäser


    Full Text Available This article presents the design and a first pilot evaluation of the computer-based training program Calcularis for children with developmental dyscalculia (DD or difficulties in learning mathematics. The program has been designed according to insights on the typical and atypical development of mathematical abilities. The learning process is supported through multimodal cues, which encode different properties of numbers. To offer optimal learning conditions, a user model completes the program and allows flexible adaptation to a child’s individual learning and knowledge profile. 32 children with difficulties in learning mathematics completed the 6 to 12-weeks computer training. The children played the game for 20 minutes per day for 5 days a week. The training effects were evaluated using neuropsychological tests. Generally, children benefited significantly from the training regarding number representation and arithmetic operations. Furthermore, children liked to play with the program and reported that the training improved their mathematical abilities.

  11. Noise studies of communication systems using the SYSTID computer aided analysis program (United States)

    Tranter, W. H.; Dawson, C. T.


    SYSTID computer aided design is a simple program for simulating data systems and communication links. A trial of the efficiency of the method was carried out by simulating a linear analog communication system to determine its noise performance and by comparing the SYSTID result with the result arrived at by theoretical calculation. It is shown that the SYSTID program is readily applicable to the analysis of these types of systems.

  12. A new computer program for QSAR-analysis: ARTE-QSAR. (United States)

    Van Damme, Sofie; Bultinck, Patrick


    A new computer program has been designed to build and analyze quantitative-structure activity relationship (QSAR) models through regression analysis. The user is provided with a range of regression and validation techniques. The emphasis of the program lies mainly in the validation of QSAR models in chemical applications. ARTE-QSAR produces an easy interpretable output from which the user can conclude if the obtained model is suitable for prediction and analysis.

  13. Universal scaling in sports ranking (United States)

    Deng, Weibing; Li, Wei; Cai, Xu; Bulou, Alain; Wang, Qiuping A.


    Ranking is a ubiquitous phenomenon in human society. On the web pages of Forbes, one may find all kinds of rankings, such as the world's most powerful people, the world's richest people, the highest-earning tennis players, and so on and so forth. Herewith, we study a specific kind—sports ranking systems in which players' scores and/or prize money are accrued based on their performances in different matches. By investigating 40 data samples which span 12 different sports, we find that the distributions of scores and/or prize money follow universal power laws, with exponents nearly identical for most sports. In order to understand the origin of this universal scaling we focus on the tennis ranking systems. By checking the data we find that, for any pair of players, the probability that the higher-ranked player tops the lower-ranked opponent is proportional to the rank difference between the pair. Such a dependence can be well fitted to a sigmoidal function. By using this feature, we propose a simple toy model which can simulate the competition of players in different matches. The simulations yield results consistent with the empirical findings. Extensive simulation studies indicate that the model is quite robust with respect to the modifications of some parameters.

  14. Universal scaling in sports ranking

    CERN Document Server

    Deng, Weibing; Cai, Xu; Bulou, Alain; Wang, Qiuping A


    Ranking is a ubiquitous phenomenon in the human society. By clicking the web pages of Forbes, you may find all kinds of rankings, such as world's most powerful people, world's richest people, top-paid tennis stars, and so on and so forth. Herewith, we study a specific kind, sports ranking systems in which players' scores and prize money are calculated based on their performances in attending various tournaments. A typical example is tennis. It is found that the distributions of both scores and prize money follow universal power laws, with exponents nearly identical for most sports fields. In order to understand the origin of this universal scaling we focus on the tennis ranking systems. By checking the data we find that, for any pair of players, the probability that the higher-ranked player will top the lower-ranked opponent is proportional to the rank difference between the pair. Such a dependence can be well fitted to a sigmoidal function. By using this feature, we propose a simple toy model which can simul...

  15. Building an advanced climate model: Program plan for the CHAMMP (Computer Hardware, Advanced Mathematics, and Model Physics) Climate Modeling Program

    Energy Technology Data Exchange (ETDEWEB)


    The issue of global warming and related climatic changes from increasing concentrations of greenhouse gases in the atmosphere has received prominent attention during the past few years. The Computer Hardware, Advanced Mathematics, and Model Physics (CHAMMP) Climate Modeling Program is designed to contribute directly to this rapid improvement. The goal of the CHAMMP Climate Modeling Program is to develop, verify, and apply a new generation of climate models within a coordinated framework that incorporates the best available scientific and numerical approaches to represent physical, biogeochemical, and ecological processes, that fully utilizes the hardware and software capabilities of new computer architectures, that probes the limits of climate predictability, and finally that can be used to address the challenging problem of understanding the greenhouse climate issue through the ability of the models to simulate time-dependent climatic changes over extended times and with regional resolution.

  16. DNAGEL: a computer program for determining DNA fragment sizes using a small computer equipped with a graphics tablet. (United States)

    Kieser, T


    The program DNAGEL is used to determine the size of DNA fragments run on agarose or polyacrylamide gels. The positions of the bands are read from gel photographs by means of a digitizer. Standard curves are calculated by the method of Southern (1979). The bands, as they are measured, are reproduced on the screen so that erroneous input can be recognized and corrected immediately. Similarly the estimated fragment sizes are printed in a table in the same relative positions as the bands on the gel. This makes it especially easy to relate fragment sizes with the bands on the gel picture. As an additional function the calculated positions of bands can be displayed on the screen. The program DNAGEL is written in APPLESOFT BASIC, suitable for APPLE II computers with 48K memory connected to a monitor, printer and a HOUSTON graphics tablet. PMID:6320102

  17. A computer program for two-particle generalized coefficients of fractional parentage (United States)

    Deveikis, A.; Juodagalvis, A.


    We present a FORTRAN90 program GCFP for the calculation of the generalized coefficients of fractional parentage (generalized CFPs or GCFP). The approach is based on the observation that the multi-shell CFPs can be expressed in terms of single-shell CFPs, while the latter can be readily calculated employing a simple enumeration scheme of antisymmetric A-particle states and an efficient method of construction of the idempotent matrix eigenvectors. The program provides fast calculation of GCFPs for a given particle number and produces results possessing numerical uncertainties below the desired tolerance. A single j-shell is defined by four quantum numbers, (e,l,j,t). A supplemental C++ program parGCFP allows calculation to be done in batches and/or in parallel. Program summaryProgram title:GCFP, parGCFP Catalogue identifier: AEBI_v1_0 Program summary URL: Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, No. of lines in distributed program, including test data, etc.: 17 199 No. of bytes in distributed program, including test data, etc.: 88 658 Distribution format: tar.gz Programming language: FORTRAN 77/90 ( GCFP), C++ ( parGCFP) Computer: Any computer with suitable compilers. The program GCFP requires a FORTRAN 77/90 compiler. The auxiliary program parGCFP requires GNU-C++ compatible compiler, while its parallel version additionally requires MPI-1 standard libraries Operating system: Linux (Ubuntu, Scientific) (all programs), also checked on Windows XP ( GCFP, serial version of parGCFP) RAM: The memory demand depends on the computation and output mode. If this mode is not 4, the program GCFP demands the following amounts of memory on a computer with Linux operating system. It requires around 2 MB of RAM for the A=12 system at E⩽2. Computation of the A=50 particle system requires around 60 MB of

  18. Applications of Out-of-Domain Knowledge in Students' Reasoning about Computer Program State (United States)

    Lewis, Colleen Marie

    To meet a growing demand and a projected deficit in the supply of computer professionals (NCWIT, 2009), it is of vital importance to expand students' access to computer science. However, many researchers in the computer science education community unproductively assume that some students lack an innate ability for computer science and therefore cannot be successful learning to program. In contrast, I hypothesize that the degree to which computer science students make productive use of their out-of-domain knowledge can better explain the range of success of novices learning to program. To investigate what non-programming knowledge supports students' success, I conducted and videotaped approximately 40 hours of clinical interviews with 30 undergraduate students enrolled in introductory programming courses. During each interview, a participant talked as they solved programming problems, many of which were multiple-choice problems that were highly correlated with success on an Advanced Placement Computer Science exam. In the analysis of the interviews I focused on students' strengths rather than the typical decision to focus on students' weaknesses. I documented specific competencies of the participants and applied analytic tools from the Knowledge in Pieces theoretical framework (diSessa, 1993) to attempt to understand the source and nature of these competencies. I found that participants appeared to build upon several kinds of out-of-domain knowledge. For example, many students used algebraic substitution techniques when tracing the state of recursive functions. Students appeared to use metaphors and their intuitive knowledge of both iteration and physics to understand infinite loops and base cases. On the level of an individual students' reasoning, a case study analysis illustrated the ways in which a participant integrated her linguistic knowledge of "and" into her reasoning about the computer science command "and." In addition to identifying these specific

  19. Rapid freeze-drying cycle optimization using computer programs developed based on heat and mass transfer models and facilitated by tunable diode laser absorption spectroscopy (TDLAS). (United States)

    Kuu, Wei Y; Nail, Steven L


    Computer programs in FORTRAN were developed to rapidly determine the optimal shelf temperature, T(f), and chamber pressure, P(c), to achieve the shortest primary drying time. The constraint for the optimization is to ensure that the product temperature profile, T(b), is below the target temperature, T(target). Five percent mannitol was chosen as the model formulation. After obtaining the optimal sets of T(f) and P(c), each cycle was assigned with a cycle rank number in terms of the length of drying time. Further optimization was achieved by dividing the drying time into a series of ramping steps for T(f), in a cascading manner (termed the cascading T(f) cycle), to further shorten the cycle time. For the purpose of demonstrating the validity of the optimized T(f) and P(c), four cycles with different predicted lengths of drying time, along with the cascading T(f) cycle, were chosen for experimental cycle runs. Tunable diode laser absorption spectroscopy (TDLAS) was used to continuously measure the sublimation rate. As predicted, maximum product temperatures were controlled slightly below the target temperature of -25 degrees C, and the cascading T(f)-ramping cycle is the most efficient cycle design. In addition, the experimental cycle rank order closely matches with that determined by modeling.

  20. Executing a gather operation on a parallel computer (United States)

    Archer, Charles J [Rochester, MN; Ratterman, Joseph D [Rochester, MN


    Methods, apparatus, and computer program products are disclosed for executing a gather operation on a parallel computer according to embodiments of the present invention. Embodiments include configuring, by the logical root, a result buffer or the logical root, the result buffer having positions, each position corresponding to a ranked node in the operational group and for storing contribution data gathered from that ranked node. Embodiments also include repeatedly for each position in the result buffer: determining, by each compute node of an operational group, whether the current position in the result buffer corresponds with the rank of the compute node, if the current position in the result buffer corresponds with the rank of the compute node, contributing, by that compute node, the compute node's contribution data, if the current position in the result buffer does not correspond with the rank of the compute node, contributing, by that compute node, a value of zero for the contribution data, and storing, by the logical root in the current position in the result buffer, results of a bitwise OR operation of all the contribution data by all compute nodes of the operational group for the current position, the results received through the global combining network.

  1. Node seniority ranking

    CERN Document Server

    Fioriti, Vincenzo


    Recent advances in graph theory suggest that is possible to identify the oldest nodes of a network using only the graph topology. Here we report on applications to heterogeneous real world networks. To this end, and in order to gain new insights, we propose the theoretical framework of the Estrada communicability. We apply it to two technological networks (an underground, the diffusion of a software worm in a LAN) and to a third network representing a cholera outbreak. In spite of errors introduced in the adjacency matrix of their graphs, the identification of the oldest nodes is feasible, within a small margin of error, and extremely simple. Utilizations include the search of the initial disease-spreader (patient zero problem), rumors in social networks, malware in computer networks, triggering events in blackouts, oldest urban sites recognition.

  2. An ensemble rank learning approach for gene prioritization. (United States)

    Lee, Po-Feng; Soo, Von-Wun


    Several different computational approaches have been developed to solve the gene prioritization problem. We intend to use the ensemble boosting learning techniques to combine variant computational approaches for gene prioritization in order to improve the overall performance. In particular we add a heuristic weighting function to the Rankboost algorithm according to: 1) the absolute ranks generated by the adopted methods for a certain gene, and 2) the ranking relationship between all gene-pairs from each prioritization result. We select 13 known prostate cancer genes in OMIM database as training set and protein coding gene data in HGNC database as test set. We adopt the leave-one-out strategy for the ensemble rank boosting learning. The experimental results show that our ensemble learning approach outperforms the four gene-prioritization methods in ToppGene suite in the ranking results of the 13 known genes in terms of mean average precision, ROC and AUC measures.

  3. Factors that Influence the Success of Male and Female Computer Programming Students in College (United States)

    Clinkenbeard, Drew A.

    As the demand for a technologically skilled work force grows, experience and skill in computer science have become increasingly valuable for college students. However, the number of students graduating with computer science degrees is not growing proportional to this need. Traditionally several groups are underrepresented in this field, notably women and students of color. This study investigated elements of computer science education that influence academic achievement in beginning computer programming courses. The goal of the study was to identify elements that increase success in computer programming courses. A 38-item questionnaire was developed and administered during the Spring 2016 semester at California State University Fullerton (CSUF). CSUF is an urban public university comprised of about 40,000 students. Data were collected from three beginning programming classes offered at CSUF. In total 411 questionnaires were collected resulting in a response rate of 58.63%. Data for the study were grouped into three broad categories of variables. These included academic and background variables; affective variables; and peer, mentor, and role-model variables. A conceptual model was developed to investigate how these variables might predict final course grade. Data were analyzed using statistical techniques such as linear regression, factor analysis, and path analysis. Ultimately this study found that peer interactions, comfort with computers, computer self-efficacy, self-concept, and perception of achievement were the best predictors of final course grade. In addition, the analyses showed that male students exhibited higher levels of computer self-efficacy and self-concept compared to female students, even when they achieved comparable course grades. Implications and explanations of these findings are explored, and potential policy changes are offered.

  4. Test rank of an abelian product of a free Lie algebra and a free ...

    Indian Academy of Sciences (India)

    Let be a free Lie algebra of rank ≥ 2 and be a free abelian Lie algebra of rank ≥ 2. We prove that the test rank of the abelian product F × A is . Morever we compute the test rank of the algebra F / k ( F ) ′ . Author Affiliations. Naime Ekici1 Nazar Şahin Öğüşlü1. Department of Mathematics, Çukurova University, ...

  5. An Efficient PageRank Approach for Urban Traffic Optimization

    Directory of Open Access Journals (Sweden)

    Florin Pop


    to determine optimal decisions for each traffic light, based on the solution given by Larry Page for page ranking in Web environment (Page et al. (1999. Our approach is similar with work presented by Sheng-Chung et al. (2009 and Yousef et al. (2010. We consider that the traffic lights are controlled by servers and a score for each road is computed based on efficient PageRank approach and is used in cost function to determine optimal decisions. We demonstrate that the cumulative contribution of each car in the traffic respects the main constrain of PageRank approach, preserving all the properties of matrix consider in our model.

  6. Google's pagerank and beyond the science of search engine rankings

    CERN Document Server

    Langville, Amy N


    Why doesn't your home page appear on the first page of search results, even when you query your own name? How do other Web pages always appear at the top? What creates these powerful rankings? And how? The first book ever about the science of Web page rankings, Google's PageRank and Beyond supplies the answers to these and other questions and more. The book serves two very different audiences: the curious science reader and the technical computational reader. The chapters build in mathematical sophistication, so that the first five are accessible to the general academic reader. While other cha

  7. VORSTAB: A computer program for calculating lateral-directional stability derivatives with vortex flow effect (United States)

    Lan, C. Edward


    A computer program based on the Quasi-Vortex-Lattice Method of Lan is presented for calculating longitudinal and lateral-directional aerodynamic characteristics of nonplanar wing-body combination. The method is based on the assumption of inviscid subsonic flow. Both attached and vortex-separated flows are treated. For the vortex-separated flow, the calculation is based on the method of suction analogy. The effect of vortex breakdown is accounted for by an empirical method. A summary of the theoretical method, program capabilities, input format, output variables and program job control set-up are described. Three test cases are presented as guides for potential users of the code.

  8. A computer program TRACK_P for studying proton tracks in PADC detectors

    Directory of Open Access Journals (Sweden)

    D. Nikezic


    Full Text Available A computer program for studying proton tracks in solid state nuclear track detectors was developed and described in this paper. The program was written in Fortran 90, with an additional tool for visualizing the track appearance as seen under the optical microscope in the transmission mode, which was written in the Python programming language. Measurable track parameters were determined and displayed in the application window and written in a data file. Three-dimensional representation of tracks was enabled. Examples of calculated tracks were also given in the present paper.

  9. UDATE1: A computer program for the calculation of uranium-series isotopic ages (United States)

    Rosenbauer, Robert J.

    UDATE1 is a FORTRAN-77 program with an interface for an Apple Macintosh computer that calculates isotope activities from measured count rates to date geologic materials by uranium-series disequilibria. Dates on pure samples can be determined directly by the accumulation of 230Th from 234U and of 231Pa from 235U. Dates for samples contaminated by clays containing abundant natural thorium can be corrected by the program using various mixing models. Input to the program and file management are made simple and user friendly by a series of Macintosh modal dialog boxes.

  10. DITTY - a computer program for calculating population dose integrated over ten thousand years

    Energy Technology Data Exchange (ETDEWEB)

    Napier, B.A.; Peloquin, R.A.; Strenge, D.L.


    The computer program DITTY (Dose Integrated Over Ten Thousand Years) was developed to determine the collective dose from long term nuclear waste disposal sites resulting from the ground-water pathways. DITTY estimates the time integral of collective dose over a ten-thousand-year period for time-variant radionuclide releases to surface waters, wells, or the atmosphere. This document includes the following information on DITTY: a description of the mathematical models, program designs, data file requirements, input preparation, output interpretations, sample problems, and program-generated diagnostic messages.

  11. Evaluation des Programmes d'Informatique (Evaluation of Computer Science Programs). (United States)

    Moisan, Claude

    In March 1994, Quebec's Commission on the Evaluation of Collegiate Teaching initiated an evaluation of computer science programs in province colleges. This report describes the evaluation process and presents results. The first section describes the following four stages of the evaluation: the formation of a consulting committee and evaluation…

  12. Development of a Computer Program for Teaching Periodontal Diagnosis Based on Clinical Epidemiological Principles. (United States)

    Fung, Kelvin; And Others


    Development of an inexpensive but powerful computer program to teach clinical periodontal diagnosis using epidemiological principles is described. Using probabilistic thinking, the student is guided from application of raw research data to derivation of likelihood ratios and how they affect clinical decision making. Student response was found to…

  13. Programmed Instruction versus Meaningful Learning Theory in Teaching Basic Structured Query Language (SQL) in Computer Lesson (United States)

    Efendioglu, Akin; Yelken, Tugba Yanpar


    The purpose of this study was to investigate the effects of two different methods on primary school teacher candidates' academic achievements and attitudes toward computer-based education, and to define their views on these methods. Both the first experimental group, programmed instruction (PI), and the second experimental group, meaningful…

  14. TimeSet: A computer program that accesses five atomic time services on two continents (United States)

    Petrakis, P. L.


    TimeSet is a shareware program for accessing digital time services by telephone. At its initial release, it was capable of capturing time signals only from the U.S. Naval Observatory to set a computer's clock. Later the ability to synchronize with the National Institute of Standards and Technology was added. Now, in Version 7.10, TimeSet is able to access three additional telephone time services in Europe - in Sweden, Austria, and Italy - making a total of five official services addressable by the program. A companion program, TimeGen, allows yet another source of telephone time data strings for callers equipped with TimeSet version 7.10. TimeGen synthesizes UTC time data strings in the Naval Observatory's format from an accurately set and maintained DOS computer clock, and transmits them to callers. This allows an unlimited number of 'freelance' time generating stations to be created. Timesetting from TimeGen is made feasible by the advent of Becker's RighTime, a shareware program that learns the drift characteristics of a computer's clock and continuously applies a correction to keep it accurate, and also brings .01 second resolution to the DOS clock. With clock regulation by RighTime and periodic update calls by the TimeGen station to an official time source via TimeSet, TimeGen offers the same degree of accuracy within the resolution of the computer clock as any official atomic time source.

  15. Study of a detuned accelerating section with the computer program PROGON

    Energy Technology Data Exchange (ETDEWEB)

    Heifets, S.A.; Kheifets, S.A.


    The longitudinal coupling impedance for a number of lower passbands, bunch to bunch energy variation due to longitudinal wake fields, the beam loading compensation, some effects of production errors, and the rf pulse transmission through a detuned disk loaded accelerating section with finite wall conductivity have been studied using the computer program PROGON.

  16. Dynamics of Population and Economic Growth: A Computer-Based Instruction Program. (United States)

    Roh, Chaisung; Handler, Paul

    A computer-assisted instructional (CAI) program at the University of Illinois is used to teach the dynamics of population growth. Socio-economic models are also developed to show the consequences of population growth upon variables such as income, productivity, and the demand for food. A one-sex population projection model allows students to…

  17. Assessment of the Accounting and Joint Accounting/Computer Information Systems Programs. (United States)

    Appiah, John; Cernigliaro, James; Davis, Jeffrey; Gordon, Millicent; Richards, Yves; Santamaria, Fernando; Siegel, Annette; Lytle, Namy; Wharton, Patrick

    This document presents City University of New York LaGuardia Community College's Department of Accounting and Managerial Studies assessment of its accounting and joint accounting/computer information systems programs report, and includes the following items: (1) description of the mission and goals of the Department of Accounting and Managerial…

  18. A Computer-Based Program to Teach Braille Reading to Sighted Individuals (United States)

    Scheithauer, Mindy C.; Tiger, Jeffrey H.


    Instructors of the visually impaired need efficient braille-training methods. This study conducted a preliminary evaluation of a computer-based program intended to teach the relation between braille characters and English letters using a matching-to-sample format with 4 sighted college students. Each participant mastered matching visual depictions…

  19. Phosphoric acid fuel cell power plant system performance model and computer program (United States)

    Alkasab, K. A.; Lu, C. Y.


    A FORTRAN computer program was developed for analyzing the performance of phosphoric acid fuel cell power plant systems. Energy mass and electrochemical analysis in the reformer, the shaft converters, the heat exchangers, and the fuel cell stack were combined to develop a mathematical model for the power plant for both atmospheric and pressurized conditions, and for several commercial fuels.

  20. Computer Programming and Biomolecular Structure Studies: A Step beyond Internet Bioinformatics (United States)

    Likic, Vladimir A.


    This article describes the experience of teaching structural bioinformatics to third year undergraduate students in a subject titled "Biomolecular Structure and Bioinformatics." Students were introduced to computer programming and used this knowledge in a practical application as an alternative to the well established Internet bioinformatics…