WorldWideScience

Sample records for parsing satoshi sekine

  1. Dependency Parsing

    CERN Document Server

    Kubler, Sandra; Nivre, Joakim

    2009-01-01

    Dependency-based methods for syntactic parsing have become increasingly popular in natural language processing in recent years. This book gives a thorough introduction to the methods that are most widely used today. After an introduction to dependency grammar and dependency parsing, followed by a formal characterization of the dependency parsing problem, the book surveys the three major classes of parsing models that are in current use: transition-based, graph-based, and grammar-based models. It continues with a chapter on evaluation and one on the comparison of different methods, and it close

  2. Asian Creativity: A Response to Satoshi Kanazawa

    Directory of Open Access Journals (Sweden)

    Geoffrey Miller

    2006-01-01

    Full Text Available This article responds to Satoshi Kanazawa's thoughtful and entertaining comments about my article concerning the Asian future of evolutionary psychology. Contra Kanazawa's argument that Asian cultural traditions and/or character inhibit Asian scientific creativity, I review historical evidence of high Asian creativity, and psychometric evidence of high Asian intelligence (a cognitive trait and openness to experience (a personality trait — two key components of creativity. Contra Kanazawa's concern that political correctness is a bigger threat to American evolutionary psychology than religious fundamentalism, I review evidence from research funding patterns and student attitudes suggesting that fundamentalism is more harmful and pervasive. Finally, in response to Kanazawa's focus on tall buildings as indexes of national wealth and creativity, I find that 13 of the world's tallest 25 buildings are in China, Hong Kong, or Taiwan — of which 11 were built in the last decade. Asian creativity, secularism, and architectural prominence point to a bright future for Asian science.

  3. Memory-Based Shallow Parsing

    NARCIS (Netherlands)

    Tjong Kim Sang, E.F.

    2002-01-01

    We present memory-based learning approaches to shallow parsing and apply these to five tasks: base noun phrase identification, arbitrary base phrase recognition, clause detection, noun phrase parsing and full parsing. We use feature selection techniques and system combination methods for improving

  4. Unifying LL and LR parsing

    NARCIS (Netherlands)

    W.H.L.M. Pijls (Wim)

    1993-01-01

    textabstractIn parsing theory, LL parsing and LR parsing are regarded to be two distinct methods. In this paper the relation between these methods is clarified.As shown in literature on parsing theory, for every context-free grammar, a so-called non-deterministic LR(0) automaton can be constructed.

  5. Bit-coded regular expression parsing

    DEFF Research Database (Denmark)

    Nielsen, Lasse; Henglein, Fritz

    2011-01-01

    the DFA-based parsing algorithm due to Dub ´e and Feeley to emit the bits of the bit representation without explicitly materializing the parse tree itself. We furthermore show that Frisch and Cardelli’s greedy regular expression parsing algorithm can be straightforwardly modified to produce bit codings...

  6. Memory-Based Shallow Parsing

    OpenAIRE

    Sang, Erik F. Tjong Kim

    2002-01-01

    We present memory-based learning approaches to shallow parsing and apply these to five tasks: base noun phrase identification, arbitrary base phrase recognition, clause detection, noun phrase parsing and full parsing. We use feature selection techniques and system combination methods for improving the performance of the memory-based learner. Our approach is evaluated on standard data sets and the results are compared with that of other systems. This reveals that our approach works well for ba...

  7. Video Scene Parsing with Predictive Feature Learning

    OpenAIRE

    Jin, Xiaojie; Li, Xin; Xiao, Huaxin; Shen, Xiaohui; Lin, Zhe; Yang, Jimei; Chen, Yunpeng; Dong, Jian; Liu, Luoqi; Jie, Zequn; Feng, Jiashi; Yan, Shuicheng

    2016-01-01

    In this work, we address the challenging video scene parsing problem by developing effective representation learning methods given limited parsing annotations. In particular, we contribute two novel methods that constitute a unified parsing framework. (1) \\textbf{Predictive feature learning}} from nearly unlimited unlabeled video data. Different from existing methods learning features from single frame parsing, we learn spatiotemporal discriminative features by enforcing a parsing network to ...

  8. Faster, Practical GLL Parsing

    NARCIS (Netherlands)

    A. Afroozeh (Ali); A. Izmaylova (Anastasia)

    2015-01-01

    htmlabstractGeneralized LL (GLL) parsing is an extension of recursive-descent (RD) parsing that supports all context-free grammars in cubic time and space. GLL parsers have the direct relationship with the grammar that RD parsers have, and therefore, compared to GLR, are easier to understand, debug,

  9. Dependency Parsing with Transformed Feature

    Directory of Open Access Journals (Sweden)

    Fuxiang Wu

    2017-01-01

    Full Text Available Dependency parsing is an important subtask of natural language processing. In this paper, we propose an embedding feature transforming method for graph-based parsing, transform-based parsing, which directly utilizes the inner similarity of the features to extract information from all feature strings including the un-indexed strings and alleviate the feature sparse problem. The model transforms the extracted features to transformed features via applying a feature weight matrix, which consists of similarities between the feature strings. Since the matrix is usually rank-deficient because of similar feature strings, it would influence the strength of constraints. However, it is proven that the duplicate transformed features do not degrade the optimization algorithm: the margin infused relaxed algorithm. Moreover, this problem can be alleviated by reducing the number of the nearest transformed features of a feature. In addition, to further improve the parsing accuracy, a fusion parser is introduced to integrate transformed and original features. Our experiments verify that both transform-based and fusion parser improve the parsing accuracy compared to the corresponding feature-based parser.

  10. Contextual Semantic Parsing using Crowdsourced Spatial Descriptions

    OpenAIRE

    Dukes, Kais

    2014-01-01

    We describe a contextual parser for the Robot Commands Treebank, a new crowdsourced resource. In contrast to previous semantic parsers that select the most-probable parse, we consider the different problem of parsing using additional situational context to disambiguate between different readings of a sentence. We show that multiple semantic analyses can be searched using dynamic programming via interaction with a spatial planner, to guide the parsing process. We are able to parse sentences in...

  11. Application development with Parse using iOS SDK

    CERN Document Server

    Birani, Bhanu

    2013-01-01

    A practical guide, featuring step-by-step instructions showing you how to use Parse iOS, and handle your data on cloud.If you are a developer who wants to build your applications instantly using Parse iOS as a back end application development, this book is ideal for you. This book will help you to understand Parse, featuring examples to help you get familiar with the concepts of Parse iOS.

  12. Two-pass greedy regular expression parsing

    DEFF Research Database (Denmark)

    Grathwohl, Niels Bjørn Bugge; Henglein, Fritz; Nielsen, Lasse

    2013-01-01

    We present new algorithms for producing greedy parses for regular expressions (REs) in a semi-streaming fashion. Our lean-log algorithm executes in time O(mn) for REs of size m and input strings of size n and outputs a compact bit-coded parse tree representation. It improves on previous algorithms...... by: operating in only 2 passes; using only O(m) words of random-access memory (independent of n); requiring only kn bits of sequentially written and read log storage, where k ... and not requiring it to be stored at all. Previous RE parsing algorithms do not scale linearly with input size, or require substantially more log storage and employ 3 passes where the first consists of reversing the input, or do not or are not known to produce a greedy parse. The performance of our unoptimized C...

  13. Chinese Unknown Word Recognition for PCFG-LA Parsing

    Directory of Open Access Journals (Sweden)

    Qiuping Huang

    2014-01-01

    Full Text Available This paper investigates the recognition of unknown words in Chinese parsing. Two methods are proposed to handle this problem. One is the modification of a character-based model. We model the emission probability of an unknown word using the first and last characters in the word. It aims to reduce the POS tag ambiguities of unknown words to improve the parsing performance. In addition, a novel method, using graph-based semisupervised learning (SSL, is proposed to improve the syntax parsing of unknown words. Its goal is to discover additional lexical knowledge from a large amount of unlabeled data to help the syntax parsing. The method is mainly to propagate lexical emission probabilities to unknown words by building the similarity graphs over the words of labeled and unlabeled data. The derived distributions are incorporated into the parsing process. The proposed methods are effective in dealing with the unknown words to improve the parsing. Empirical results for Penn Chinese Treebank and TCT Treebank revealed its effectiveness.

  14. Image portion identification methods, image parsing methods, image parsing systems, and articles of manufacture

    Science.gov (United States)

    Lassahn, Gordon D.; Lancaster, Gregory D.; Apel, William A.; Thompson, Vicki S.

    2013-01-08

    Image portion identification methods, image parsing methods, image parsing systems, and articles of manufacture are described. According to one embodiment, an image portion identification method includes accessing data regarding an image depicting a plurality of biological substrates corresponding to at least one biological sample and indicating presence of at least one biological indicator within the biological sample and, using processing circuitry, automatically identifying a portion of the image depicting one of the biological substrates but not others of the biological substrates.

  15. From LL-regular to LL(1) grammars: Transformations, covers and parsing

    NARCIS (Netherlands)

    Nijholt, Antinus

    1982-01-01

    In this paper it is shown that it is possible to transform any LL-regular grammar G into an LL(1) grammar G' in such a way that parsing G' is as good as parsing G. That is, a parse of a sentence of grammar G can be obtained with a simple string homomorphism from the parse of a corresponding sentence

  16. Context-free parsing with connectionist networks

    Science.gov (United States)

    Fanty, M. A.

    1986-08-01

    This paper presents a simple algorithm which converts any context-free grammar into a connectionist network which parses strings (of arbitrary but fixed maximum length) in the language defined by that grammar. The network is fast, O(n), and deterministicd. It consists of binary units which compute a simple function of their input. When the grammar is put in Chomsky normal form, O(n3) units needed to parse inputs of length up to n.

  17. On Parsing CHILDES

    OpenAIRE

    Laakso, Aarre

    2005-01-01

    Research on child language acquisition would benefit from the availability of a large body of syntactically parsed utterances between parents and children. We consider the problem of generating such a ``treebank'' from the CHILDES corpus, which currently contains primarily orthographically transcribed speech tagged for lexical category.

  18. Dual decomposition for parsing with non-projective head automata

    OpenAIRE

    Koo, Terry; Rush, Alexander Matthew; Collins, Michael; Jaakkola, Tommi S.; Sontag, David Alexander

    2010-01-01

    This paper introduces algorithms for non-projective parsing based on dual decomposition. We focus on parsing algorithms for non-projective head automata, a generalization of head-automata models to non-projective structures. The dual decomposition algorithms are simple and efficient, relying on standard dynamic programming and minimum spanning tree algorithms. They provably solve an LP relaxation of the non-projective parsing problem. Empirically the LP relaxation is very often tight: for man...

  19. A structural SVM approach for reference parsing.

    Science.gov (United States)

    Zhang, Xiaoli; Zou, Jie; Le, Daniel X; Thoma, George R

    2011-06-09

    Automated extraction of bibliographic data, such as article titles, author names, abstracts, and references is essential to the affordable creation of large citation databases. References, typically appearing at the end of journal articles, can also provide valuable information for extracting other bibliographic data. Therefore, parsing individual reference to extract author, title, journal, year, etc. is sometimes a necessary preprocessing step in building citation-indexing systems. The regular structure in references enables us to consider reference parsing a sequence learning problem and to study structural Support Vector Machine (structural SVM), a newly developed structured learning algorithm on parsing references. In this study, we implemented structural SVM and used two types of contextual features to compare structural SVM with conventional SVM. Both methods achieve above 98% token classification accuracy and above 95% overall chunk-level accuracy for reference parsing. We also compared SVM and structural SVM to Conditional Random Field (CRF). The experimental results show that structural SVM and CRF achieve similar accuracies at token- and chunk-levels. When only basic observation features are used for each token, structural SVM achieves higher performance compared to SVM since it utilizes the contextual label features. However, when the contextual observation features from neighboring tokens are combined, SVM performance improves greatly, and is close to that of structural SVM after adding the second order contextual observation features. The comparison of these two methods with CRF using the same set of binary features show that both structural SVM and CRF perform better than SVM, indicating their stronger sequence learning ability in reference parsing.

  20. Faster Scannerless GLR parsing

    NARCIS (Netherlands)

    J.J. Vinju (Jurgen); G.R. Economopoulos (Giorgos Robert); P. Klint (Paul)

    2008-01-01

    textabstractAnalysis and renovation of large software portfolios requires syntax analysis of multiple, usually embedded, languages and this is beyond the capabilities of many standard parsing techniques. The traditional separation between lexer and parser falls short due to the limitations of

  1. Faster scannerless GLR parsing

    NARCIS (Netherlands)

    G.R. Economopoulos (Giorgos Robert); P. Klint (Paul); J.J. Vinju (Jurgen); O. de Moor; M.I. Schwartzbach

    2009-01-01

    textabstractAnalysis and renovation of large software portfolios requires syntax analysis of multiple, usually embedded, languages and this is beyond the capabilities of many standard parsing techniques. The traditional separation between lexer and parser falls short due to the limitations of

  2. Probabilistic lexical generalization for French dependency parsing

    OpenAIRE

    Henestroza Anguiano , Enrique; Candito , Marie

    2012-01-01

    International audience; This paper investigates the impact on French dependency parsing of lexical generalization methods beyond lemmatization and morphological analysis. A distributional thesaurus is created from a large text corpus and used for distributional clustering and WordNet automatic sense ranking. The standard approach for lexical generalization in parsing is to map a word to a single generalized class, either replacing the word with the class or adding a new feature for the class....

  3. Faster scannerless GLR parsing

    NARCIS (Netherlands)

    Economopoulos, G.R.; Klint, P.; Vinju, J.J.; Moor, de O.; Schwartzbach, M.I.

    2009-01-01

    Analysis and renovation of large software portfolios requires syntax analysis of multiple, usually embedded, languages and this is beyond the capabilities of many standard parsing techniques. The traditional separation between lexer and parser falls short due to the limitations of tokenization based

  4. Error Parsing: An alternative method of implementing social judgment theory

    OpenAIRE

    Crystal C. Hall; Daniel M. Oppenheimer

    2015-01-01

    We present a novel method of judgment analysis called Error Parsing, based upon an alternative method of implementing Social Judgment Theory (SJT). SJT and Error Parsing both posit the same three components of error in human judgment: error due to noise, error due to cue weighting, and error due to inconsistency. In that sense, the broad theory and framework are the same. However, SJT and Error Parsing were developed to answer different questions, and thus use different m...

  5. Integrating high dimensional bi-directional parsing models for gene mention tagging.

    Science.gov (United States)

    Hsu, Chun-Nan; Chang, Yu-Ming; Kuo, Cheng-Ju; Lin, Yu-Shi; Huang, Han-Shen; Chung, I-Fang

    2008-07-01

    Tagging gene and gene product mentions in scientific text is an important initial step of literature mining. In this article, we describe in detail our gene mention tagger participated in BioCreative 2 challenge and analyze what contributes to its good performance. Our tagger is based on the conditional random fields model (CRF), the most prevailing method for the gene mention tagging task in BioCreative 2. Our tagger is interesting because it accomplished the highest F-scores among CRF-based methods and second over all. Moreover, we obtained our results by mostly applying open source packages, making it easy to duplicate our results. We first describe in detail how we developed our CRF-based tagger. We designed a very high dimensional feature set that includes most of information that may be relevant. We trained bi-directional CRF models with the same set of features, one applies forward parsing and the other backward, and integrated two models based on the output scores and dictionary filtering. One of the most prominent factors that contributes to the good performance of our tagger is the integration of an additional backward parsing model. However, from the definition of CRF, it appears that a CRF model is symmetric and bi-directional parsing models will produce the same results. We show that due to different feature settings, a CRF model can be asymmetric and the feature setting for our tagger in BioCreative 2 not only produces different results but also gives backward parsing models slight but constant advantage over forward parsing model. To fully explore the potential of integrating bi-directional parsing models, we applied different asymmetric feature settings to generate many bi-directional parsing models and integrate them based on the output scores. Experimental results show that this integrated model can achieve even higher F-score solely based on the training corpus for gene mention tagging. Data sets, programs and an on-line service of our gene

  6. Toward the Soundness of Sense Structure Definitions in Thesaurus-Dictionaries. Parsing Problems and Solutions

    Directory of Open Access Journals (Sweden)

    Neculai Curteanu

    2012-10-01

    Full Text Available In this paper we point out some difficult problems of thesaurus-dictionary entry parsing, relying on the parsing technology of SCD (Segmentation-Cohesion-Dependency configurations, successfully applied on six largest thesauri -- Romanian (2, French, German (2, and Russian. \\textbf{Challenging Problems:} \\textbf{(a}~Intricate and~/~or recursive structures of the lexicographic segments met in the entries of certain thesauri; \\textbf{(b}~Cyclicity (recursive calls of some sense marker classes on marker sequences; \\textbf{(c}~Establishing the hypergraph-driven dependencies between all the atomic and non-atomic sense definitions. Classical approach to solve these parsing problems is hard mainly because of depth-first search of sense definitions and markers, the substantial complexity of entries, and the sense tree dynamic construction embodied within these parsers. \\textbf{SCD-based Parsing Solutions:} \\textbf{(a}~The SCD parsing method is a procedural tool, completely formal grammar-free, handling the recursive structure of the lexicographic segments by procedural non-recursive calls performed on the SCD parsing configurations of the entry structure. \\textbf{(b}~For dealing with cyclicity (recursive calls between secondary sense markers and the sense enumeration markers, we proposed the Enumeration Closing Condition, sometimes coupled with New{\\_}Paragraphs typographic markers transformed into numeral sense enumeration. \\textbf{(c}~These problems, their lexicographic modeling and parsing solutions are addressed to both dictionary parser programmers to experience the SCD-based parsing method, as well as to lexicographers and thesauri designers for tailoring balanced lexical-semantics granularities and sounder sense tree definitions of the dictionary entries.

  7. On Collocations and Their Interaction with Parsing and Translation

    Directory of Open Access Journals (Sweden)

    Violeta Seretan

    2013-10-01

    Full Text Available We address the problem of automatically processing collocations—a subclass of multi-word expressions characterized by a high degree of morphosyntactic flexibility—in the context of two major applications, namely, syntactic parsing and machine translation. We show that parsing and collocation identification are processes that are interrelated and that benefit from each other, inasmuch as syntactic information is crucial for acquiring collocations from corpora and, vice versa, collocational information can be used to improve parsing performance. Similarly, we focus on the interrelation between collocations and machine translation, highlighting the use of translation information for multilingual collocation identification, as well as the use of collocational knowledge for improving translation. We give a panorama of the existing relevant work, and we parallel the literature surveys with our own experiments involving a symbolic parser and a rule-based translation system. The results show a significant improvement over approaches in which the corresponding tasks are decoupled.

  8. A Semantic Constraint on Syntactic Parsing.

    Science.gov (United States)

    Crain, Stephen; Coker, Pamela L.

    This research examines how semantic information influences syntactic parsing decisions during sentence processing. In the first experiment, subjects were presented lexical strings having syntactically identical surface structures but with two possible underlying structures: "The children taught by the Berlitz method," and "The…

  9. Parse Journal #2: Introduction

    OpenAIRE

    Bowman, Jason; Malik, Suhail; Phillips, Andrea

    2015-01-01

    As a periodical concerned with the critical potential of artistic research, this edition of the PARSE journal mobilises the multiple perspectives of artists, thinkers, critics and curators on the problematics, discontents and possibilities of private capital as an unregulated yet assumptive producer of art’s value, including its integration with state-funding. We have put emphasis on how this conditioning of art’s production, circulation, reception and sale can be put to task. In particular, ...

  10. Cuidado de enfermagem a pessoas com hipertensão fundamentado na teoria de Parse Atención de enfermería a personas con hipertensión basada en la teoría de Parse Nursing care to people with hypertension based on Parse's theory

    Directory of Open Access Journals (Sweden)

    Fabíola Vládia Freire da Silva

    2013-03-01

    Full Text Available Este estudo propõe o cuidado de enfermagem, baseado nos princípios de Parse, a pessoas com hipertensão consultadas na Estratégia Saúde da Família. Estudo descritivo, de cunho qualitativo, realizado de março a maio de 2011, com quatorze enfermeiros no município de Itapajé-Ceará. Para coleta das informações utilizou-se a entrevista semiestruturada e, para análise, o discurso dos sujeitos. Emergiram três categorias baseadas nos princípios de Parse: Multidimensão dos significados - o enfermeiro conduz ao relato dos significados; Sincronização de ritmos - o enfermeiro ajuda a identificar harmonia e desarmonia; Mobilização da transcendência - o enfermeiro guia o plano de mudanças. Notou-se aproximação dos discursos ao teorizado por Parse quando citaram buscar um cuidado humanizado, com a participação da família, valorização da autonomia, utilização da educação em saúde, com orientações individuais. Percebeu-se a viabilidade na implementação do cuidado de enfermagem fundamentado na Teoria de Parse a pessoas com hipertensão.Este estudio propone la atención de enfermería, basada en los principios de Parse, para personas con hipertensión en la Estrategia de Salud Familiar. Estudio descriptivo, cualitativo, realizado de marzo a mayo/2011, con catorce enfermeros en Itapajé, Ceará. Para la recolección de las informaciones, se utilizó la entrevista semiestructurada y para análisis, el discurso de los sujetos. Emergieron tres categorías: Multidimensiones de los significados - el enfermero conduce al relato de los significados; Sincronización de los ritmos - el enfermero ayuda a identificar armonía y desarmonía; Movilización de la trascendencia - el enfermero guía el plan de cambios. Se observó semejanzas entre el enfoque de los discursos y la Teoría de Parse, cuando citaron la búsqueda de atención humanizada, con participación de la familia, valoración de la autonomía, uso de la educación en salud

  11. Telugu dependency parsing using different statistical parsers

    Directory of Open Access Journals (Sweden)

    B. Venkata Seshu Kumari

    2017-01-01

    Full Text Available In this paper we explore different statistical dependency parsers for parsing Telugu. We consider five popular dependency parsers namely, MaltParser, MSTParser, TurboParser, ZPar and Easy-First Parser. We experiment with different parser and feature settings and show the impact of different settings. We also provide a detailed analysis of the performance of all the parsers on major dependency labels. We report our results on test data of Telugu dependency treebank provided in the ICON 2010 tools contest on Indian languages dependency parsing. We obtain state-of-the art performance of 91.8% in unlabeled attachment score and 70.0% in labeled attachment score. To the best of our knowledge ours is the only work which explored all the five popular dependency parsers and compared the performance under different feature settings for Telugu.

  12. Single-View 3D Scene Reconstruction and Parsing by Attribute Grammar.

    Science.gov (United States)

    Liu, Xiaobai; Zhao, Yibiao; Zhu, Song-Chun

    2018-03-01

    In this paper, we present an attribute grammar for solving two coupled tasks: i) parsing a 2D image into semantic regions; and ii) recovering the 3D scene structures of all regions. The proposed grammar consists of a set of production rules, each describing a kind of spatial relation between planar surfaces in 3D scenes. These production rules are used to decompose an input image into a hierarchical parse graph representation where each graph node indicates a planar surface or a composite surface. Different from other stochastic image grammars, the proposed grammar augments each graph node with a set of attribute variables to depict scene-level global geometry, e.g., camera focal length, or local geometry, e.g., surface normal, contact lines between surfaces. These geometric attributes impose constraints between a node and its off-springs in the parse graph. Under a probabilistic framework, we develop a Markov Chain Monte Carlo method to construct a parse graph that optimizes the 2D image recognition and 3D scene reconstruction purposes simultaneously. We evaluated our method on both public benchmarks and newly collected datasets. Experiments demonstrate that the proposed method is capable of achieving state-of-the-art scene reconstruction of a single image.

  13. Recursive Neural Networks Based on PSO for Image Parsing

    Directory of Open Access Journals (Sweden)

    Guo-Rong Cai

    2013-01-01

    Full Text Available This paper presents an image parsing algorithm which is based on Particle Swarm Optimization (PSO and Recursive Neural Networks (RNNs. State-of-the-art method such as traditional RNN-based parsing strategy uses L-BFGS over the complete data for learning the parameters. However, this could cause problems due to the nondifferentiable objective function. In order to solve this problem, the PSO algorithm has been employed to tune the weights of RNN for minimizing the objective. Experimental results obtained on the Stanford background dataset show that our PSO-based training algorithm outperforms traditional RNN, Pixel CRF, region-based energy, simultaneous MRF, and superpixel MRF.

  14. Parsing polarization squeezing into Fock layers

    DEFF Research Database (Denmark)

    Mueller, Christian R.; Madsen, Lars Skovgaard; Klimov, Andrei B.

    2016-01-01

    photon number do the methods coincide; when the photon number is indefinite, we parse the state in Fock layers, finding that substantially higher squeezing can be observed in some of the single layers. By capitalizing on the properties of the Husimi Q function, we map this notion onto the Poincare space......, providing a full account of the measured squeezing....

  15. Sequence distance via parsing complexity: Heartbeat signals

    International Nuclear Information System (INIS)

    Degli Esposti, M.; Farinelli, C.; Menconi, G.

    2009-01-01

    We compare and discuss the use of different symbolic codings of electrocardiogram (ECG) signals in order to distinguish healthy patients from hospitalized ones. To this aim, we recall a parsing-based similarity distance and compare the performances of several methods of classification of data.

  16. YakYak: Parsing with Logical Side Constraints

    DEFF Research Database (Denmark)

    Hansen, Niels Damgaard; Klarlund, Nils; Schwartzbach, Michael Ignatieff

    2000-01-01

    Yak, which extends Yacc with first-order logic for specifying consteaints that are regular tree languages. Concise formulas about the parse tree replace explicit programming, and they are turned into canonical attribute grammars through tree automata calculations. YakYak is implemented as a proprocessor...

  17. ParseCNV integrative copy number variation association software with quality tracking.

    Science.gov (United States)

    Glessner, Joseph T; Li, Jin; Hakonarson, Hakon

    2013-03-01

    A number of copy number variation (CNV) calling algorithms exist; however, comprehensive software tools for CNV association studies are lacking. We describe ParseCNV, unique software that takes CNV calls and creates probe-based statistics for CNV occurrence in both case-control design and in family based studies addressing both de novo and inheritance events, which are then summarized based on CNV regions (CNVRs). CNVRs are defined in a dynamic manner to allow for a complex CNV overlap while maintaining precise association region. Using this approach, we avoid failure to converge and non-monotonic curve fitting weaknesses of programs, such as CNVtools and CNVassoc, and although Plink is easy to use, it only provides combined CNV state probe-based statistics, not state-specific CNVRs. Existing CNV association methods do not provide any quality tracking information to filter confident associations, a key issue which is fully addressed by ParseCNV. In addition, uncertainty in CNV calls underlying CNV associations is evaluated to verify significant results, including CNV overlap profiles, genomic context, number of probes supporting the CNV and single-probe intensities. When optimal quality control parameters are followed using ParseCNV, 90% of CNVs validate by polymerase chain reaction, an often problematic stage because of inadequate significant association review. ParseCNV is freely available at http://parsecnv.sourceforge.net.

  18. Finding EL+ justifications using the Earley parsing algorithm

    CSIR Research Space (South Africa)

    Nortje, R

    2009-12-01

    Full Text Available into a reachability preserving context free grammar (CFG). The well known earley algorithm for parsing strings, given some CFG, is then applied to the problem of extracting minimal reachability-based axiom sets for subsumption entailments. The author has...

  19. Parsing Universal Dependencies without training

    DEFF Research Database (Denmark)

    Martínez Alonso, Héctor; Agic, Zeljko; Plank, Barbara

    2017-01-01

    We present UDP, the first training-free parser for Universal Dependencies (UD). Our algorithm is based on PageRank and a small set of specific dependency head rules. UDP features two-step decoding to guarantee that function words are attached as leaf nodes. The parser requires no training......, and it is competitive with a delexicalized transfer system. UDP offers a linguistically sound unsupervised alternative to cross-lingual parsing for UD. The parser has very few parameters and distinctly robust to domain change across languages....

  20. Parsing with subdomain instance weighting from raw corpora

    NARCIS (Netherlands)

    Plank, B.; Sima'an, K.

    2008-01-01

    The treebanks that are used for training statistical parsers consist of hand-parsed sentences from a single source/domain like newspaper text. However, newspaper text concerns different subdomains of language use (e.g. finance, sports, politics, music), which implies that the statistics gathered by

  1. Parsing with Subdomain Instance Weighting from Raw Corpora

    NARCIS (Netherlands)

    Plank, Barbara; Sima'an, Khalil

    2008-01-01

    The treebanks that are used for training statistical parsers consist of hand-parsed sentences from a single source/domain like newspaper text. However, newspaper text concerns different subdomains of language use (e.g. finance, sports, politics, music), which implies that the statistics gathered by

  2. Perceiving Event Dynamics and Parsing Hollywood Films

    Science.gov (United States)

    Cutting, James E.; Brunick, Kaitlin L.; Candan, Ayse

    2012-01-01

    We selected 24 Hollywood movies released from 1940 through 2010 to serve as a film corpus. Eight viewers, three per film, parsed them into events, which are best termed subscenes. While watching a film a second time, viewers scrolled through frames and recorded the frame number where each event began. Viewers agreed about 90% of the time. We then…

  3. Time-Driven Effects on Parsing during Reading

    Science.gov (United States)

    Roll, Mikael; Lindgren, Magnus; Alter, Kai; Horne, Merle

    2012-01-01

    The phonological trace of perceived words starts fading away in short-term memory after a few seconds. Spoken utterances are usually 2-3 s long, possibly to allow the listener to parse the words into coherent prosodic phrases while they still have a clear representation. Results from this brain potential study suggest that even during silent…

  4. "gnparser": a powerful parser for scientific names based on Parsing Expression Grammar.

    Science.gov (United States)

    Mozzherin, Dmitry Y; Myltsev, Alexander A; Patterson, David J

    2017-05-26

    Scientific names in biology act as universal links. They allow us to cross-reference information about organisms globally. However variations in spelling of scientific names greatly diminish their ability to interconnect data. Such variations may include abbreviations, annotations, misspellings, etc. Authorship is a part of a scientific name and may also differ significantly. To match all possible variations of a name we need to divide them into their elements and classify each element according to its role. We refer to this as 'parsing' the name. Parsing categorizes name's elements into those that are stable and those that are prone to change. Names are matched first by combining them according to their stable elements. Matches are then refined by examining their varying elements. This two stage process dramatically improves the number and quality of matches. It is especially useful for the automatic data exchange within the context of "Big Data" in biology. We introduce Global Names Parser (gnparser). It is a Java tool written in Scala language (a language for Java Virtual Machine) to parse scientific names. It is based on a Parsing Expression Grammar. The parser can be applied to scientific names of any complexity. It assigns a semantic meaning (such as genus name, species epithet, rank, year of publication, authorship, annotations, etc.) to all elements of a name. It is able to work with nested structures as in the names of hybrids. gnparser performs with ≈99% accuracy and processes 30 million name-strings/hour per CPU thread. The gnparser library is compatible with Scala, Java, R, Jython, and JRuby. The parser can be used as a command line application, as a socket server, a web-app or as a RESTful HTTP-service. It is released under an Open source MIT license. Global Names Parser (gnparser) is a fast, high precision tool for biodiversity informaticians and biologists working with large numbers of scientific names. It can replace expensive and error

  5. Análisis de los posibles desde la teoría de Parse en una persona con Alzheimer The analysis of the possibles through Parse's theory on a person with Alzheimer

    Directory of Open Access Journals (Sweden)

    Virtudes Rodero-Sánchez

    2006-11-01

    Full Text Available Cuando una persona acude al sistema sanitario con un problema de salud, con mucha frecuencia se le va a pedir que introduzca cambios en sus hábitos y estilo de vida. Esta demanda se suele concretar en un pacto-compromiso que se establece persona-profesional. Hemos observado que este pacto, a pesar de que el profesional se esfuerza en enmarcarlo en objetivos realistas, con demasiada frecuencia sobreviene la frustración, sobre todo en escenarios de cronicidad. La teoría de Parse nos ofrece una manera diferente de abordar el cambio. En la teoría de Parse, El Ser Humano en Devenir, los posibles son la expresión de la fuerza, entendida como una manera única de transformación, que consiste en avanzar con las esperanzas, anhelos y los proyectos de la persona. Planteamos: en primer lugar un análisis de los elementos de lo que Parse llama su tercer principio, la co-trancendencia con los posibles; en segundo lugar el análisis de los posibles desde este marco de referencia a través de una narrativa; y por último la práctica enfermera.When a person comes to the Health Care System with a health problem will often be asked to change some of his habits and lifestyles. This demand becomes a compromise-pact between the person and the professional. We have observed that in this compromise-pact, despite the effort of the professional to hide it behind realist targets, the patient usually becomes frustrated, especially in cases of chronic illnesses. Parse's theory offers us a different way to approach the change. In Parse's theory, The Human Becoming, the possibles are the expression of power, understood as a unique way of transformation, consisting in advancing with the hopes, desires and projects of a person. We suggest, first of all, an analysis of the elements that Parse calls her third principle: co-transcendence with the possibles; secondly, the analysis of the possibles from the basis of this reference framework through a narration and, finally

  6. Fetching and Parsing Data from the Web with OpenRefine

    Directory of Open Access Journals (Sweden)

    Evan Peter Williamson

    2017-08-01

    Full Text Available OpenRefine is a powerful tool for exploring, cleaning, and transforming data. An earlier Programming Historian lesson, “Cleaning Data with OpenRefine”, introduced the basic functionality of Refine to efficiently discover and correct inconsistency in a data set. Building on those essential data wrangling skills, this lesson focuses on Refine’s ability to fetch URLs and parse web content. Examples introduce some of the advanced features to transform and enhance a data set including: - fetch URLs using Refine - construct URL queries to retrieve information from a simple web API - parse HTML and JSON responses to extract relevant data - use array functions to manipulate string values - use Jython to extend Refine’s functionality It will be helpful to have basic familiarity with OpenRefine, HTML, and programming concepts such as variables and loops to complete this lesson.

  7. Advanced programming concepts in a course on grammars and parsing

    NARCIS (Netherlands)

    Jeuring, J.T.; Swierstra, S.D.

    1999-01-01

    One of the important goals of the Computer Science curriculum at Utrecht University is to familiarize students with abstract programming concepts such as, for example, partial evaluation and deforestation. A course on grammars and parsing offers excellent possibilities for exemplifying and

  8. Fast Parsing using Pruning and Grammar Specialization

    OpenAIRE

    Rayner, Manny; Carter, David

    1996-01-01

    We show how a general grammar may be automatically adapted for fast parsing of utterances from a specific domain by means of constituent pruning and grammar specialization based on explanation-based learning. These methods together give an order of magnitude increase in speed, and the coverage loss entailed by grammar specialization is reduced to approximately half that reported in previous work. Experiments described here suggest that the loss of coverage has been reduced to the point where ...

  9. Is human sentence parsing serial or parallel? Evidence from event-related brain potentials.

    Science.gov (United States)

    Hopf, Jens-Max; Bader, Markus; Meng, Michael; Bayer, Josef

    2003-01-01

    In this ERP study we investigate the processes that occur in syntactically ambiguous German sentences at the point of disambiguation. Whereas most psycholinguistic theories agree on the view that processing difficulties arise when parsing preferences are disconfirmed (so-called garden-path effects), important differences exist with respect to theoretical assumptions about the parser's recovery from a misparse. A key distinction can be made between parsers that compute all alternative syntactic structures in parallel (parallel parsers) and parsers that compute only a single preferred analysis (serial parsers). To distinguish empirically between parallel and serial parsing models, we compare ERP responses to garden-path sentences with ERP responses to truly ungrammatical sentences. Garden-path sentences contain a temporary and ultimately curable ungrammaticality, whereas truly ungrammatical sentences remain so permanently--a difference which gives rise to different predictions in the two classes of parsing architectures. At the disambiguating word, ERPs in both sentence types show negative shifts of similar onset latency, amplitude, and scalp distribution in an initial time window between 300 and 500 ms. In a following time window (500-700 ms), the negative shift to garden-path sentences disappears at right central parietal sites, while it continues in permanently ungrammatical sentences. These data are taken as evidence for a strictly serial parser. The absence of a difference in the early time window indicates that temporary and permanent ungrammaticalities trigger the same kind of parsing responses. Later differences can be related to successful reanalysis in garden-path but not in ungrammatical sentences. Copyright 2003 Elsevier Science B.V.

  10. Locating and parsing bibliographic references in HTML medical articles.

    Science.gov (United States)

    Zou, Jie; Le, Daniel; Thoma, George R

    2010-06-01

    The set of references that typically appear toward the end of journal articles is sometimes, though not always, a field in bibliographic (citation) databases. But even if references do not constitute such a field, they can be useful as a preprocessing step in the automated extraction of other bibliographic data from articles, as well as in computer-assisted indexing of articles. Automation in data extraction and indexing to minimize human labor is key to the affordable creation and maintenance of large bibliographic databases. Extracting the components of references, such as author names, article title, journal name, publication date and other entities, is therefore a valuable and sometimes necessary task. This paper describes a two-step process using statistical machine learning algorithms, to first locate the references in HTML medical articles and then to parse them. Reference locating identifies the reference section in an article and then decomposes it into individual references. We formulate this step as a two-class classification problem based on text and geometric features. An evaluation conducted on 500 articles drawn from 100 medical journals achieves near-perfect precision and recall rates for locating references. Reference parsing identifies the components of each reference. For this second step, we implement and compare two algorithms. One relies on sequence statistics and trains a Conditional Random Field. The other focuses on local feature statistics and trains a Support Vector Machine to classify each individual word, followed by a search algorithm that systematically corrects low confidence labels if the label sequence violates a set of predefined rules. The overall performance of these two reference-parsing algorithms is about the same: above 99% accuracy at the word level, and over 97% accuracy at the chunk level.

  11. On the Approximation Ratio of Lempel-Ziv Parsing

    DEFF Research Database (Denmark)

    Gagie, Travis; Navarro, Gonzalo; Prezza, Nicola

    2018-01-01

    in the text. Since computing b is NP-complete, a popular gold standard is z, the number of phrases in the Lempel-Ziv parse of the text, where phrases can be copied only from the left. While z can be computed in linear time, almost nothing has been known for decades about its approximation ratio with respect...

  12. A simple DOP model for constituency parsing of Italian sentences

    NARCIS (Netherlands)

    Sangati, F.

    2009-01-01

    We present a simplified Data-Oriented Parsing (DOP) formalism for learning the constituency structure of Italian sentences. In our approach we try to simplify the original DOP methodology by constraining the number and type of fragments we extract from the training corpus. We provide some examples

  13. Fuzzy context-free languages - Part 2: Recognition and parsing algorithms

    NARCIS (Netherlands)

    Asveld, P.R.J.

    2005-01-01

    In a companion paper [P.R.J. Asveld, Fuzzy context-free languages---Part 1: Generalized fuzzy context-free grammars, Theoret. Comp. Sci. (2005)] we used fuzzy context-free grammars in order to model grammatical errors resulting in erroneous inputs for robust recognizing and parsing algorithms for

  14. Fuzzy Context- Free Languages. Part 2: Recognition and Parsing Algorithms

    NARCIS (Netherlands)

    Asveld, P.R.J.

    2000-01-01

    In a companion paper \\cite{Asv:FCF1} we used fuzzy context-free grammars in order to model grammatical errors resulting in erroneous inputs for robust recognizing and parsing algorithms for fuzzy context-free languages. In particular, this approach enables us to distinguish between small errors

  15. Chomsky-Schützenberger parsing for weighted multiple context-free languages

    Directory of Open Access Journals (Sweden)

    Tobias Denkinger

    2017-07-01

    Full Text Available We prove a Chomsky-Schützenberger representation theorem for multiple context-free languages weighted over complete commutative strong bimonoids. Using this representation we devise a parsing algorithm for a restricted form of those devices.

  16. Generative re-ranking model for dependency parsing of Italian sentences

    NARCIS (Netherlands)

    Sangati, F.

    2009-01-01

    We present a general framework for dependency parsing of Italian sentences based on a combination of discriminative and generative models. We use a state-of-the-art discriminative model to obtain a k-best list of candidate structures for the test sentences, and use the generative model to compute

  17. Introduction to special issue on machine learning approaches to shallow parsing

    NARCIS (Netherlands)

    Hammerton, J; Osborne, M; Armstrong, S; Daelemans, W

    2002-01-01

    This article introduces the problem of partial or shallow parsing (assigning partial syntactic structure to sentences) and explains why it is an important natural language processing (NLP) task. The complexity of the task makes Machine Learning an attractive option in comparison to the handcrafting

  18. Pippi — Painless parsing, post-processing and plotting of posterior and likelihood samples

    Science.gov (United States)

    Scott, Pat

    2012-11-01

    Interpreting samples from likelihood or posterior probability density functions is rarely as straightforward as it seems it should be. Producing publication-quality graphics of these distributions is often similarly painful. In this short note I describe pippi, a simple, publicly available package for parsing and post-processing such samples, as well as generating high-quality PDF graphics of the results. Pippi is easily and extensively configurable and customisable, both in its options for parsing and post-processing samples, and in the visual aspects of the figures it produces. I illustrate some of these using an existing supersymmetric global fit, performed in the context of a gamma-ray search for dark matter. Pippi can be downloaded and followed at http://github.com/patscott/pippi.

  19. Automated vocabulary discovery for geo-parsing online epidemic intelligence.

    Science.gov (United States)

    Keller, Mikaela; Freifeld, Clark C; Brownstein, John S

    2009-11-24

    Automated surveillance of the Internet provides a timely and sensitive method for alerting on global emerging infectious disease threats. HealthMap is part of a new generation of online systems designed to monitor and visualize, on a real-time basis, disease outbreak alerts as reported by online news media and public health sources. HealthMap is of specific interest for national and international public health organizations and international travelers. A particular task that makes such a surveillance useful is the automated discovery of the geographic references contained in the retrieved outbreak alerts. This task is sometimes referred to as "geo-parsing". A typical approach to geo-parsing would demand an expensive training corpus of alerts manually tagged by a human. Given that human readers perform this kind of task by using both their lexical and contextual knowledge, we developed an approach which relies on a relatively small expert-built gazetteer, thus limiting the need of human input, but focuses on learning the context in which geographic references appear. We show in a set of experiments, that this approach exhibits a substantial capacity to discover geographic locations outside of its initial lexicon. The results of this analysis provide a framework for future automated global surveillance efforts that reduce manual input and improve timeliness of reporting.

  20. Parsing and Quantification of Raw Orbitrap Mass Spectrometer Data Using RawQuant.

    Science.gov (United States)

    Kovalchik, Kevin A; Moggridge, Sophie; Chen, David D Y; Morin, Gregg B; Hughes, Christopher S

    2018-06-01

    Effective analysis of protein samples by mass spectrometry (MS) requires careful selection and optimization of a range of experimental parameters. As the output from the primary detection device, the "raw" MS data file can be used to gauge the success of a given sample analysis. However, the closed-source nature of the standard raw MS file can complicate effective parsing of the data contained within. To ease and increase the range of analyses possible, the RawQuant tool was developed to enable parsing of raw MS files derived from Thermo Orbitrap instruments to yield meta and scan data in an openly readable text format. RawQuant can be commanded to export user-friendly files containing MS 1 , MS 2 , and MS 3 metadata as well as matrices of quantification values based on isobaric tagging approaches. In this study, the utility of RawQuant is demonstrated in several scenarios: (1) reanalysis of shotgun proteomics data for the identification of the human proteome, (2) reanalysis of experiments utilizing isobaric tagging for whole-proteome quantification, and (3) analysis of a novel bacterial proteome and synthetic peptide mixture for assessing quantification accuracy when using isobaric tags. Together, these analyses successfully demonstrate RawQuant for the efficient parsing and quantification of data from raw Thermo Orbitrap MS files acquired in a range of common proteomics experiments. In addition, the individual analyses using RawQuant highlights parametric considerations in the different experimental sets and suggests targetable areas to improve depth of coverage in identification-focused studies and quantification accuracy when using isobaric tags.

  1. Neural Semantic Parsing by Character-based Translation: Experiments with Abstract Meaning Representations

    NARCIS (Netherlands)

    van Noord, Rik; Bos, Johannes

    2017-01-01

    We evaluate the character-level translation method for neural semantic parsing on a large corpus of sentences annotated with Abstract Meaning Representations (AMRs). Using a sequence-to-sequence model, and some trivial preprocessing and postprocessing of AMRs, we obtain a baseline accuracy of 53.1

  2. A hierarchical methodology for urban facade parsing from TLS point clouds

    Science.gov (United States)

    Li, Zhuqiang; Zhang, Liqiang; Mathiopoulos, P. Takis; Liu, Fangyu; Zhang, Liang; Li, Shuaipeng; Liu, Hao

    2017-01-01

    The effective and automated parsing of building facades from terrestrial laser scanning (TLS) point clouds of urban environments is an important research topic in the GIS and remote sensing fields. It is also challenging because of the complexity and great variety of the available 3D building facade layouts as well as the noise and data missing of the input TLS point clouds. In this paper, we introduce a novel methodology for the accurate and computationally efficient parsing of urban building facades from TLS point clouds. The main novelty of the proposed methodology is that it is a systematic and hierarchical approach that considers, in an adaptive way, the semantic and underlying structures of the urban facades for segmentation and subsequent accurate modeling. Firstly, the available input point cloud is decomposed into depth planes based on a data-driven method; such layer decomposition enables similarity detection in each depth plane layer. Secondly, the labeling of the facade elements is performed using the SVM classifier in combination with our proposed BieS-ScSPM algorithm. The labeling outcome is then augmented with weak architectural knowledge. Thirdly, least-squares fitted normalized gray accumulative curves are applied to detect regular structures, and a binarization dilation extraction algorithm is used to partition facade elements. A dynamic line-by-line division is further applied to extract the boundaries of the elements. The 3D geometrical façade models are then reconstructed by optimizing facade elements across depth plane layers. We have evaluated the performance of the proposed method using several TLS facade datasets. Qualitative and quantitative performance comparisons with several other state-of-the-art methods dealing with the same facade parsing problem have demonstrated its superiority in performance and its effectiveness in improving segmentation accuracy.

  3. Extending TF1: Argument parsing, function composition, and vectorization

    CERN Document Server

    Tsang Mang Kin, Arthur Leonard

    2017-01-01

    In this project, we extend the functionality of the TF1 function class in root. We add argument parsing, making it possible to freely pass variables and parameters into pre-defined and user-defined functions. We also introduce a syntax to use certain compositions of functions, namely normalized sums and convolutions, directly in TF1. Finally, we introduce some simple vectorization functionality to TF1 and demonstrate the potential to speed up parallelizable computations.

  4. Towards a Robuster Interpretive Parsing: learning from overt forms in Optimality Theory

    NARCIS (Netherlands)

    Biró, T.

    2013-01-01

    The input data to grammar learning algorithms often consist of overt forms that do not contain full structural descriptions. This lack of information may contribute to the failure of learning. Past work on Optimality Theory introduced Robust Interpretive Parsing (RIP) as a partial solution to this

  5. Tackling Error Propagation through Reinforcement Learning: A Case of Greedy Dependency Parsing

    NARCIS (Netherlands)

    Le, M.N.; Fokkens, A.S.

    Error propagation is a common problem in NLP. Reinforcement learning explores erroneous states during training and can therefore be more robust when mistakes are made early in a process. In this paper, we apply reinforcement learning to greedy dependency parsing which is known to suffer from error

  6. Process of 3D wireless decentralized sensor deployment using parsing crossover scheme

    Directory of Open Access Journals (Sweden)

    Albert H.R. Ko

    2015-07-01

    Full Text Available A Wireless Sensor Networks (WSN usually consists of numerous wireless devices deployed in a region of interest, each able to collect and process environmental information and communicate with neighboring devices. It can thus be regarded as a Multi-Agent System for territorial security, where individual agents cooperate with each other to avoid duplication of effort and to exploit other agent’s capacities. The problem of sensor deployment becomes non-trivial when we consider environmental factors, such as terrain elevations. Due to the fact that all sensors are homogeneous, the chromosomes that encode sensor positions are actually interchangeable, and conventional crossover schemes such as uniform crossover would cause some redundancy as well as over-concentration in certain specific geographical area. We propose a Parsing Crossover Scheme that intends to reduce redundancy and ease geographical concentration pattern in an effort to facilitate the search. The proposed parsing crossover method demonstrates better performances than those of uniform crossover under different terrain irregularities.

  7. Automated vocabulary discovery for geo-parsing online epidemic intelligence

    Directory of Open Access Journals (Sweden)

    Freifeld Clark C

    2009-11-01

    Full Text Available Abstract Background Automated surveillance of the Internet provides a timely and sensitive method for alerting on global emerging infectious disease threats. HealthMap is part of a new generation of online systems designed to monitor and visualize, on a real-time basis, disease outbreak alerts as reported by online news media and public health sources. HealthMap is of specific interest for national and international public health organizations and international travelers. A particular task that makes such a surveillance useful is the automated discovery of the geographic references contained in the retrieved outbreak alerts. This task is sometimes referred to as "geo-parsing". A typical approach to geo-parsing would demand an expensive training corpus of alerts manually tagged by a human. Results Given that human readers perform this kind of task by using both their lexical and contextual knowledge, we developed an approach which relies on a relatively small expert-built gazetteer, thus limiting the need of human input, but focuses on learning the context in which geographic references appear. We show in a set of experiments, that this approach exhibits a substantial capacity to discover geographic locations outside of its initial lexicon. Conclusion The results of this analysis provide a framework for future automated global surveillance efforts that reduce manual input and improve timeliness of reporting.

  8. Creating Parsing Lexicons from Semantic Lexicons Automatically and Its Applications

    National Research Council Canada - National Science Library

    Ayan, Necip F; Dorr, Bonnie

    2002-01-01

    ...). We also present the effects of using such a lexicon on the parser performance. The advantage of automating the process is that the same technique can be applied directly to lexicons we have for other languages, for example, Arabic, Chinese, and Spanish. The results indicate that our method will help us generate parsing lexicons which can be used by a broad-coverage parser that runs on different languages.

  9. Incremental Learning of Context Free Grammars by Parsing-Based Rule Generation and Rule Set Search

    Science.gov (United States)

    Nakamura, Katsuhiko; Hoshina, Akemi

    This paper discusses recent improvements and extensions in Synapse system for inductive inference of context free grammars (CFGs) from sample strings. Synapse uses incremental learning, rule generation based on bottom-up parsing, and the search for rule sets. The form of production rules in the previous system is extended from Revised Chomsky Normal Form A→βγ to Extended Chomsky Normal Form, which also includes A→B, where each of β and γ is either a terminal or nonterminal symbol. From the result of bottom-up parsing, a rule generation mechanism synthesizes minimum production rules required for parsing positive samples. Instead of inductive CYK algorithm in the previous version of Synapse, the improved version uses a novel rule generation method, called ``bridging,'' which bridges the lacked part of the derivation tree for the positive string. The improved version also employs a novel search strategy, called serial search in addition to minimum rule set search. The synthesis of grammars by the serial search is faster than the minimum set search in most cases. On the other hand, the size of the generated CFGs is generally larger than that by the minimum set search, and the system can find no appropriate grammar for some CFL by the serial search. The paper shows experimental results of incremental learning of several fundamental CFGs and compares the methods of rule generation and search strategies.

  10. Learning for Semantic Parsing with Kernels under Various Forms of Supervision

    Science.gov (United States)

    2007-08-01

    natural language sentences to their formal executable meaning representations. This is a challenging problem and is critical for developing computing...sentences are semantically tractable. This indi- cates that Geoquery is more challenging domain for semantic parsing than ATIS. In the past, there have been a...Combining parsers. In Proceedings of the Conference on Em- pirical Methods in Natural Language Processing and Very Large Corpora (EMNLP/ VLC -99), pp. 187–194

  11. Using machine learning to parse breast pathology reports.

    Science.gov (United States)

    Yala, Adam; Barzilay, Regina; Salama, Laura; Griffin, Molly; Sollender, Grace; Bardia, Aditya; Lehman, Constance; Buckley, Julliette M; Coopey, Suzanne B; Polubriaginof, Fernanda; Garber, Judy E; Smith, Barbara L; Gadd, Michele A; Specht, Michelle C; Gudewicz, Thomas M; Guidi, Anthony J; Taghian, Alphonse; Hughes, Kevin S

    2017-01-01

    Extracting information from electronic medical record is a time-consuming and expensive process when done manually. Rule-based and machine learning techniques are two approaches to solving this problem. In this study, we trained a machine learning model on pathology reports to extract pertinent tumor characteristics, which enabled us to create a large database of attribute searchable pathology reports. This database can be used to identify cohorts of patients with characteristics of interest. We collected a total of 91,505 breast pathology reports from three Partners hospitals: Massachusetts General Hospital, Brigham and Women's Hospital, and Newton-Wellesley Hospital, covering the period from 1978 to 2016. We trained our system with annotations from two datasets, consisting of 6295 and 10,841 manually annotated reports. The system extracts 20 separate categories of information, including atypia types and various tumor characteristics such as receptors. We also report a learning curve analysis to show how much annotation our model needs to perform reasonably. The model accuracy was tested on 500 reports that did not overlap with the training set. The model achieved accuracy of 90% for correctly parsing all carcinoma and atypia categories for a given patient. The average accuracy for individual categories was 97%. Using this classifier, we created a database of 91,505 parsed pathology reports. Our learning curve analysis shows that the model can achieve reasonable results even when trained on a few annotations. We developed a user-friendly interface to the database that allows physicians to easily identify patients with target characteristics and export the matching cohort. This model has the potential to reduce the effort required for analyzing large amounts of data from medical records, and to minimize the cost and time required to glean scientific insight from these data.

  12. KEGGParser: parsing and editing KEGG pathway maps in Matlab.

    Science.gov (United States)

    Arakelyan, Arsen; Nersisyan, Lilit

    2013-02-15

    KEGG pathway database is a collection of manually drawn pathway maps accompanied with KGML format files intended for use in automatic analysis. KGML files, however, do not contain the required information for complete reproduction of all the events indicated in the static image of a pathway map. Several parsers and editors of KEGG pathways exist for processing KGML files. We introduce KEGGParser-a MATLAB based tool for KEGG pathway parsing, semiautomatic fixing, editing, visualization and analysis in MATLAB environment. It also works with Scilab. The source code is available at http://www.mathworks.com/matlabcentral/fileexchange/37561.

  13. (Invariability in the Samoan syntax/prosody interface and consequences for syntactic parsing

    Directory of Open Access Journals (Sweden)

    Kristine M. Yu

    2017-10-01

    Full Text Available While it has long been clear that prosody should be part of the grammar influencing the action of the syntactic parser, how to bring prosody into computational models of syntactic parsing has remained unclear. The challenge is that prosodic information in the speech signal is the result of the interaction of a multitude of conditioning factors. From this output, how can we factor out the contribution of syntax to conditioning prosodic events? And if we are able to do that factorization and define a production model from the syntactic grammar to a prosodified utterance, how can we then define a comprehension model based on that production model? In this case study of the Samoan morphosyntax-prosody interface, we show how to factor out the influence of syntax on prosody in empirical work and confirm there is invariable morphosyntactic conditioning of high edge tones. Then, we show how this invariability can be precisely characterized and used by a parsing model that factors the various influences of morphosyntax on tonal events. We expect that models of these kinds can be extended to more comprehensive perspectives on Samoan and to languages where the syntax/prosody coupling is more complex.

  14. FastaValidator: an open-source Java library to parse and validate FASTA formatted sequences.

    Science.gov (United States)

    Waldmann, Jost; Gerken, Jan; Hankeln, Wolfgang; Schweer, Timmy; Glöckner, Frank Oliver

    2014-06-14

    Advances in sequencing technologies challenge the efficient importing and validation of FASTA formatted sequence data which is still a prerequisite for most bioinformatic tools and pipelines. Comparative analysis of commonly used Bio*-frameworks (BioPerl, BioJava and Biopython) shows that their scalability and accuracy is hampered. FastaValidator represents a platform-independent, standardized, light-weight software library written in the Java programming language. It targets computer scientists and bioinformaticians writing software which needs to parse quickly and accurately large amounts of sequence data. For end-users FastaValidator includes an interactive out-of-the-box validation of FASTA formatted files, as well as a non-interactive mode designed for high-throughput validation in software pipelines. The accuracy and performance of the FastaValidator library qualifies it for large data sets such as those commonly produced by massive parallel (NGS) technologies. It offers scientists a fast, accurate and standardized method for parsing and validating FASTA formatted sequence data.

  15. Parsing the Dictionary of Modern Literary Russian Language with the Method of SCD Configurations. The Lexicographic Modeling

    Directory of Open Access Journals (Sweden)

    Neculai Curteanu

    2012-05-01

    Full Text Available This paper extends the experience of parsing other five, sensibly different, Romanian, French, and German largest dictionaries, to \\textbf{\\textit{DMLRL}} (Dictionary of Modern Literary Russian Language [18], using the optimal and portable parsing method of SCD (Segmentation-Cohesion-Dependency configurations [7], [11], [15]. The purpose of the present paper is to elaborate the lexicographic modeling of \\textbf{\\textit{DMLRL}}, which necessarily precedes the sense tree parsing dictionary entries. The following \\textbf{\\textit{three}} SCD configurations are described: the \\textbf{\\textit{first one}} has to separate the lexicographic segments in a \\textbf{\\textit{DMLRL}} entry, the \\textbf{\\textit{second}} SCD-configuration concentrates on the SCD marker classes and their hypergraph hierarchy for \\textbf{\\textit{DMLRL}} primary and secondary senses, while the \\textbf{\\textit{third}} SCD configuration hands down the same modeling process to the atomic sense definitions and their examples-to-definitions. The dependency hypergraph of the third SCD configuration, interconnected to the one of the second SCD configuration, is specified completely at the atomic sense level for the first time, exceeding the SCD configuration modeling for other five dictionaries [15], [14]. Numerous examples from \\textbf{\\textit{DMLRL}} and comparison to \\textbf{\\textit{DLR-DAR}} Romanian thesaurus-dictionary support the proposed \\textbf{\\textit{DMLRL}} lexicographic modeling.

  16. Parsing a cognitive task: a characterization of the mind's bottleneck.

    Directory of Open Access Journals (Sweden)

    Mariano Sigman

    2005-02-01

    Full Text Available Parsing a mental operation into components, characterizing the parallel or serial nature of this flow, and understanding what each process ultimately contributes to response time are fundamental questions in cognitive neuroscience. Here we show how a simple theoretical model leads to an extended set of predictions concerning the distribution of response time and its alteration by simultaneous performance of another task. The model provides a synthesis of psychological refractory period and random-walk models of response time. It merely assumes that a task consists of three consecutive stages-perception, decision based on noisy integration of evidence, and response-and that the perceptual and motor stages can operate simultaneously with stages of another task, while the central decision process constitutes a bottleneck. We designed a number-comparison task that provided a thorough test of the model by allowing independent variations in number notation, numerical distance, response complexity, and temporal asynchrony relative to an interfering probe task of tone discrimination. The results revealed a parsing of the comparison task in which each variable affects only one stage. Numerical distance affects the integration process, which is the only step that cannot proceed in parallel and has a major contribution to response time variability. The other stages, mapping the numeral to an internal quantity and executing the motor response, can be carried out in parallel with another task. Changing the duration of these processes has no significant effect on the variance.

  17. PyParse: a semiautomated system for scoring spoken recall data.

    Science.gov (United States)

    Solway, Alec; Geller, Aaron S; Sederberg, Per B; Kahana, Michael J

    2010-02-01

    Studies of human memory often generate data on the sequence and timing of recalled items, but scoring such data using conventional methods is difficult or impossible. We describe a Python-based semiautomated system that greatly simplifies this task. This software, called PyParse, can easily be used in conjunction with many common experiment authoring systems. Scored data is output in a simple ASCII format and can be accessed with the programming language of choice, allowing for the identification of features such as correct responses, prior-list intrusions, extra-list intrusions, and repetitions.

  18. Motion based parsing for video from observational psychology

    Science.gov (United States)

    Kokaram, Anil; Doyle, Erika; Lennon, Daire; Joyeux, Laurent; Fuller, Ray

    2006-01-01

    In Psychology it is common to conduct studies involving the observation of humans undertaking some task. The sessions are typically recorded on video and used for subjective visual analysis. The subjective analysis is tedious and time consuming, not only because much useless video material is recorded but also because subjective measures of human behaviour are not necessarily repeatable. This paper presents tools using content based video analysis that allow automated parsing of video from one such study involving Dyslexia. The tools rely on implicit measures of human motion that can be generalised to other applications in the domain of human observation. Results comparing quantitative assessment of human motion with subjective assessment are also presented, illustrating that the system is a useful scientific tool.

  19. Parsing partial molar volumes of small molecules: a molecular dynamics study.

    Science.gov (United States)

    Patel, Nisha; Dubins, David N; Pomès, Régis; Chalikian, Tigran V

    2011-04-28

    We used molecular dynamics (MD) simulations in conjunction with the Kirkwood-Buff theory to compute the partial molar volumes for a number of small solutes of various chemical natures. We repeated our computations using modified pair potentials, first, in the absence of the Coulombic term and, second, in the absence of the Coulombic and the attractive Lennard-Jones terms. Comparison of our results with experimental data and the volumetric results of Monte Carlo simulation with hard sphere potentials and scaled particle theory-based computations led us to conclude that, for small solutes, the partial molar volume computed with the Lennard-Jones potential in the absence of the Coulombic term nearly coincides with the cavity volume. On the other hand, MD simulations carried out with the pair interaction potentials containing only the repulsive Lennard-Jones term produce unrealistically large partial molar volumes of solutes that are close to their excluded volumes. Our simulation results are in good agreement with the reported schemes for parsing partial molar volume data on small solutes. In particular, our determined interaction volumes() and the thickness of the thermal volume for individual compounds are in good agreement with empirical estimates. This work is the first computational study that supports and lends credence to the practical algorithms of parsing partial molar volume data that are currently in use for molecular interpretations of volumetric data.

  20. Analysis of Azari Language based on Parsing using Link Gram

    Directory of Open Access Journals (Sweden)

    Maryam Arabzadeh

    2014-09-01

    Full Text Available There are different classes of theories for the natural lanuguage syntactic parsing problem and for creating the related grammars .This paper presents a syntactic grammar developed in the link grammar formalism for Turkish which is an agglutinative language. In the link grammar formalism, the words of a sentence are linked with each other depending on their syntactic roles. Turkish has complex derivational and inflectional morphology, and derivational and inflection morphemes play important syntactic roles in the sentences. In order to develop a link grammar for Turkish, the lexical parts in the morphological representations of Turkish words are removed, and the links are created depending on the part of speech tags and inflectional morphemes in words. Furthermore, a derived word is separated at the derivational boundaries in order to treat each derivation morpheme as a special distinct word, and allow it to be linked with the rest of the sentence. The derivational morphemes of a word are also linked with each other with special links to indicate that they are parts of the same word. Finally the adapted unique link grammar formalism for Turkish provides flexibility for the linkage construction, and similar methods can be used for other languages with complex morphology. Finally, using the Delphi programming language, the link grammar related to the Azeri language was developed and implemented and then by selecting 250 random sentences, this grammar is evaluated and then tested. For 84.31% of the sentences, the result set of the parser contains the correct parse.

  1. Attribute And-Or Grammar for Joint Parsing of Human Pose, Parts and Attributes.

    Science.gov (United States)

    Park, Seyoung; Nie, Xiaohan; Zhu, Song-Chun

    2017-07-25

    This paper presents an attribute and-or grammar (A-AOG) model for jointly inferring human body pose and human attributes in a parse graph with attributes augmented to nodes in the hierarchical representation. In contrast to other popular methods in the current literature that train separate classifiers for poses and individual attributes, our method explicitly represents the decomposition and articulation of body parts, and account for the correlations between poses and attributes. The A-AOG model is an amalgamation of three traditional grammar formulations: (i)Phrase structure grammar representing the hierarchical decomposition of the human body from whole to parts; (ii)Dependency grammar modeling the geometric articulation by a kinematic graph of the body pose; and (iii)Attribute grammar accounting for the compatibility relations between different parts in the hierarchy so that their appearances follow a consistent style. The parse graph outputs human detection, pose estimation, and attribute prediction simultaneously, which are intuitive and interpretable. We conduct experiments on two tasks on two datasets, and experimental results demonstrate the advantage of joint modeling in comparison with computing poses and attributes independently. Furthermore, our model obtains better performance over existing methods for both pose estimation and attribute prediction tasks.

  2. Tackling Error Propagation through Reinforcement Learning: A Case of Greedy Dependency Parsing

    OpenAIRE

    Le, Minh; Fokkens, Antske

    2017-01-01

    Error propagation is a common problem in NLP. Reinforcement learning explores erroneous states during training and can therefore be more robust when mistakes are made early in a process. In this paper, we apply reinforcement learning to greedy dependency parsing which is known to suffer from error propagation. Reinforcement learning improves accuracy of both labeled and unlabeled dependencies of the Stanford Neural Dependency Parser, a high performance greedy parser, while maintaining its eff...

  3. Machine learning to parse breast pathology reports in Chinese.

    Science.gov (United States)

    Tang, Rong; Ouyang, Lizhi; Li, Clara; He, Yue; Griffin, Molly; Taghian, Alphonse; Smith, Barbara; Yala, Adam; Barzilay, Regina; Hughes, Kevin

    2018-01-29

    Large structured databases of pathology findings are valuable in deriving new clinical insights. However, they are labor intensive to create and generally require manual annotation. There has been some work in the bioinformatics community to support automating this work via machine learning in English. Our contribution is to provide an automated approach to construct such structured databases in Chinese, and to set the stage for extraction from other languages. We collected 2104 de-identified Chinese benign and malignant breast pathology reports from Hunan Cancer Hospital. Physicians with native Chinese proficiency reviewed the reports and annotated a variety of binary and numerical pathologic entities. After excluding 78 cases with a bilateral lesion in the same report, 1216 cases were used as a training set for the algorithm, which was then refined by 405 development cases. The Natural language processing algorithm was tested by using the remaining 405 cases to evaluate the machine learning outcome. The model was used to extract 13 binary entities and 8 numerical entities. When compared to physicians with native Chinese proficiency, the model showed a per-entity accuracy from 91 to 100% for all common diagnoses on the test set. The overall accuracy of binary entities was 98% and of numerical entities was 95%. In a per-report evaluation for binary entities with more than 100 training cases, 85% of all the testing reports were completely correct and 11% had an error in 1 out of 22 entities. We have demonstrated that Chinese breast pathology reports can be automatically parsed into structured data using standard machine learning approaches. The results of our study demonstrate that techniques effective in parsing English reports can be scaled to other languages.

  4. Mobile Backend as a Service: the pros and cons of parse

    OpenAIRE

    Nguyen, Phu

    2016-01-01

    Using a pre-built backend for an application is an affordable and swift approach to prototyping new application ideas. Mobile Backend as a Service (MBaaS) is the term for pre-built backend systems that developers can use. However, it is advisable to understand the pros and the cons of an MBaaS before deciding to use it. The aim of the thesis was to determine the advantages and disadvantages of using Parse, a provider of mobile backend as a service, in application development. Parse’s defin...

  5. Sleep Disrupts High-Level Speech Parsing Despite Significant Basic Auditory Processing.

    Science.gov (United States)

    Makov, Shiri; Sharon, Omer; Ding, Nai; Ben-Shachar, Michal; Nir, Yuval; Zion Golumbic, Elana

    2017-08-09

    The extent to which the sleeping brain processes sensory information remains unclear. This is particularly true for continuous and complex stimuli such as speech, in which information is organized into hierarchically embedded structures. Recently, novel metrics for assessing the neural representation of continuous speech have been developed using noninvasive brain recordings that have thus far only been tested during wakefulness. Here we investigated, for the first time, the sleeping brain's capacity to process continuous speech at different hierarchical levels using a newly developed Concurrent Hierarchical Tracking (CHT) approach that allows monitoring the neural representation and processing-depth of continuous speech online. Speech sequences were compiled with syllables, words, phrases, and sentences occurring at fixed time intervals such that different linguistic levels correspond to distinct frequencies. This enabled us to distinguish their neural signatures in brain activity. We compared the neural tracking of intelligible versus unintelligible (scrambled and foreign) speech across states of wakefulness and sleep using high-density EEG in humans. We found that neural tracking of stimulus acoustics was comparable across wakefulness and sleep and similar across all conditions regardless of speech intelligibility. In contrast, neural tracking of higher-order linguistic constructs (words, phrases, and sentences) was only observed for intelligible speech during wakefulness and could not be detected at all during nonrapid eye movement or rapid eye movement sleep. These results suggest that, whereas low-level auditory processing is relatively preserved during sleep, higher-level hierarchical linguistic parsing is severely disrupted, thereby revealing the capacity and limits of language processing during sleep. SIGNIFICANCE STATEMENT Despite the persistence of some sensory processing during sleep, it is unclear whether high-level cognitive processes such as speech

  6. Cross-Lingual Dependency Parsing with Late Decoding for Truly Low-Resource Languages

    OpenAIRE

    Schlichtkrull, Michael Sejr; Søgaard, Anders

    2017-01-01

    In cross-lingual dependency annotation projection, information is often lost during transfer because of early decoding. We present an end-to-end graph-based neural network dependency parser that can be trained to reproduce matrices of edge scores, which can be directly projected across word alignments. We show that our approach to cross-lingual dependency parsing is not only simpler, but also achieves an absolute improvement of 2.25% averaged across 10 languages compared to the previous state...

  7. Parsing Heterogeneous Striatal Activity

    Directory of Open Access Journals (Sweden)

    Kae Nakamura

    2017-05-01

    Full Text Available The striatum is an input channel of the basal ganglia and is well known to be involved in reward-based decision making and learning. At the macroscopic level, the striatum has been postulated to contain parallel functional modules, each of which includes neurons that perform similar computations to support selection of appropriate actions for different task contexts. At the single-neuron level, however, recent studies in monkeys and rodents have revealed heterogeneity in neuronal activity even within restricted modules of the striatum. Looking for generality in the complex striatal activity patterns, here we briefly survey several types of striatal activity, focusing on their usefulness for mediating behaviors. In particular, we focus on two types of behavioral tasks: reward-based tasks that use salient sensory cues and manipulate outcomes associated with the cues; and perceptual decision tasks that manipulate the quality of noisy sensory cues and associate all correct decisions with the same outcome. Guided by previous insights on the modular organization and general selection-related functions of the basal ganglia, we relate striatal activity patterns on these tasks to two types of computations: implementation of selection and evaluation. We suggest that a parsing with the selection/evaluation categories encourages a focus on the functional commonalities revealed by studies with different animal models and behavioral tasks, instead of a focus on aspects of striatal activity that may be specific to a particular task setting. We then highlight several questions in the selection-evaluation framework for future explorations.

  8. The interaction of parsing rules and argument – Predicate constructions: implications for the structure of the Grammaticon in FunGramKB

    Directory of Open Access Journals (Sweden)

    María del Carmen Fumero Pérez

    2017-07-01

    Full Text Available The Functional Grammar Knowledge Base (FunGramKB, (Periñán-Pascual and Arcas-Túnez 2010 is a multipurpose lexico-conceptual knowledge base designed to be used in different Natural Language Processing (NLP tasks. It is complemented with the ARTEMIS (Automatically Representing Text Meaning via an Interlingua–based System application, a parsing device linguistically grounded on Role and Reference Grammar (RRG that transduces natural language fragments into their corresponding grammatical and semantic structures. This paper unveils the different phases involved in its parsing routine, paying special attention to the treatment of argumental constructions. As an illustrative case, we will follow all the steps necessary to effectively parse a For-Benefactive structure within ARTEMIS. This methodology will reveal the necessity to distinguish between Kernel constructs and L1-constructions, since the latter involve a modification of the lexical template of the verb. Our definition of L1-constructions leads to the reorganization of the catalogue of FunGramKB L1-constructions, formerly based on Levin’s (1993 alternations. Accordingly, a rearrangement of the internal configuration of the L1-Constructicon within the Grammaticon is proposed.

  9. High-content image informatics of the structural nuclear protein NuMA parses trajectories for stem/progenitor cell lineages and oncogenic transformation

    International Nuclear Information System (INIS)

    Vega, Sebastián L.; Liu, Er; Arvind, Varun; Bushman, Jared; Sung, Hak-Joon; Becker, Matthew L.; Lelièvre, Sophie; Kohn, Joachim; Vidi, Pierre-Alexandre; Moghe, Prabhas V.

    2017-01-01

    Stem and progenitor cells that exhibit significant regenerative potential and critical roles in cancer initiation and progression remain difficult to characterize. Cell fates are determined by reciprocal signaling between the cell microenvironment and the nucleus; hence parameters derived from nuclear remodeling are ideal candidates for stem/progenitor cell characterization. Here we applied high-content, single cell analysis of nuclear shape and organization to examine stem and progenitor cells destined to distinct differentiation endpoints, yet undistinguishable by conventional methods. Nuclear descriptors defined through image informatics classified mesenchymal stem cells poised to either adipogenic or osteogenic differentiation, and oligodendrocyte precursors isolated from different regions of the brain and destined to distinct astrocyte subtypes. Nuclear descriptors also revealed early changes in stem cells after chemical oncogenesis, allowing the identification of a class of cancer-mitigating biomaterials. To capture the metrology of nuclear changes, we developed a simple and quantitative “imaging-derived” parsing index, which reflects the dynamic evolution of the high-dimensional space of nuclear organizational features. A comparative analysis of parsing outcomes via either nuclear shape or textural metrics of the nuclear structural protein NuMA indicates the nuclear shape alone is a weak phenotypic predictor. In contrast, variations in the NuMA organization parsed emergent cell phenotypes and discerned emergent stages of stem cell transformation, supporting a prognosticating role for this protein in the outcomes of nuclear functions. - Highlights: • High-content analysis of nuclear shape and organization classify stem and progenitor cells poised for distinct lineages. • Early oncogenic changes in mesenchymal stem cells (MSCs) are also detected with nuclear descriptors. • A new class of cancer-mitigating biomaterials was identified based on image

  10. High-content image informatics of the structural nuclear protein NuMA parses trajectories for stem/progenitor cell lineages and oncogenic transformation

    Energy Technology Data Exchange (ETDEWEB)

    Vega, Sebastián L. [Department of Chemical and Biochemical Engineering, Rutgers University, Piscataway, NJ (United States); Liu, Er; Arvind, Varun [Department of Biomedical Engineering, Rutgers University, Piscataway, NJ (United States); Bushman, Jared [Department of Chemistry and Chemical Biology, New Jersey Center for Biomaterials, Piscataway, NJ (United States); School of Pharmacy, University of Wyoming, Laramie, WY (United States); Sung, Hak-Joon [Department of Chemistry and Chemical Biology, New Jersey Center for Biomaterials, Piscataway, NJ (United States); Department of Biomedical Engineering, Vanderbilt University, Nashville, TN (United States); Becker, Matthew L. [Department of Polymer Science and Engineering, University of Akron, Akron, OH (United States); Lelièvre, Sophie [Department of Basic Medical Sciences, Purdue University, West Lafayette, IN (United States); Kohn, Joachim [Department of Chemistry and Chemical Biology, New Jersey Center for Biomaterials, Piscataway, NJ (United States); Vidi, Pierre-Alexandre, E-mail: pvidi@wakehealth.edu [Department of Cancer Biology, Wake Forest School of Medicine, Winston-Salem, NC (United States); Moghe, Prabhas V., E-mail: moghe@rutgers.edu [Department of Chemical and Biochemical Engineering, Rutgers University, Piscataway, NJ (United States); Department of Biomedical Engineering, Rutgers University, Piscataway, NJ (United States)

    2017-02-01

    Stem and progenitor cells that exhibit significant regenerative potential and critical roles in cancer initiation and progression remain difficult to characterize. Cell fates are determined by reciprocal signaling between the cell microenvironment and the nucleus; hence parameters derived from nuclear remodeling are ideal candidates for stem/progenitor cell characterization. Here we applied high-content, single cell analysis of nuclear shape and organization to examine stem and progenitor cells destined to distinct differentiation endpoints, yet undistinguishable by conventional methods. Nuclear descriptors defined through image informatics classified mesenchymal stem cells poised to either adipogenic or osteogenic differentiation, and oligodendrocyte precursors isolated from different regions of the brain and destined to distinct astrocyte subtypes. Nuclear descriptors also revealed early changes in stem cells after chemical oncogenesis, allowing the identification of a class of cancer-mitigating biomaterials. To capture the metrology of nuclear changes, we developed a simple and quantitative “imaging-derived” parsing index, which reflects the dynamic evolution of the high-dimensional space of nuclear organizational features. A comparative analysis of parsing outcomes via either nuclear shape or textural metrics of the nuclear structural protein NuMA indicates the nuclear shape alone is a weak phenotypic predictor. In contrast, variations in the NuMA organization parsed emergent cell phenotypes and discerned emergent stages of stem cell transformation, supporting a prognosticating role for this protein in the outcomes of nuclear functions. - Highlights: • High-content analysis of nuclear shape and organization classify stem and progenitor cells poised for distinct lineages. • Early oncogenic changes in mesenchymal stem cells (MSCs) are also detected with nuclear descriptors. • A new class of cancer-mitigating biomaterials was identified based on image

  11. The Parsing Syllable Envelopes Test for Assessment of Amplitude Modulation Discrimination Skills in Children: Development, Normative Data, and Test-Retest Reliability Studies.

    Science.gov (United States)

    Cameron, Sharon; Chong-White, Nicky; Mealings, Kiri; Beechey, Tim; Dillon, Harvey; Young, Taegan

    2018-02-01

    Intensity peaks and valleys in the acoustic signal are salient cues to syllable structure, which is accepted to be a crucial early step in phonological processing. As such, the ability to detect low-rate (envelope) modulations in signal amplitude is essential to parse an incoming speech signal into smaller phonological units. The Parsing Syllable Envelopes (ParSE) test was developed to quantify the ability of children to recognize syllable boundaries using an amplitude modulation detection paradigm. The envelope of a 750-msec steady-state /a/ vowel is modulated into two or three pseudo-syllables using notches with modulation depths varying between 0% and 100% along an 11-step continuum. In an adaptive three-alternative forced-choice procedure, the participant identified whether one, two, or three pseudo-syllables were heard. Development of the ParSE stimuli and test protocols, and collection of normative and test-retest reliability data. Eleven adults (aged 23 yr 10 mo to 50 yr 9 mo, mean 32 yr 10 mo) and 134 typically developing, primary-school children (aged 6 yr 0 mo to 12 yr 4 mo, mean 9 yr 3 mo). There were 73 males and 72 females. Data were collected using a touchscreen computer. Psychometric functions (PFs) were automatically fit to individual data by the ParSE software. Performance was related to the modulation depth at which syllables can be detected with 88% accuracy (referred to as the upper boundary of the uncertainty region [UBUR]). A shallower PF slope reflected a greater level of uncertainty. Age effects were determined based on raw scores. z Scores were calculated to account for the effect of age on performance. Outliers, and individual data for which the confidence interval of the UBUR exceeded a maximum allowable value, were removed. Nonparametric tests were used as the data were skewed toward negative performance. Across participants, the performance criterion (UBUR) was met with a median modulation depth of 42%. The effect of age on the UBUR was

  12. Relative clauses as a benchmark for Minimalist parsing

    Directory of Open Access Journals (Sweden)

    Thomas Graf

    2017-07-01

    Full Text Available Minimalist grammars have been used recently in a series of papers to explain well-known contrasts in human sentence processing in terms of subtle structural differences. These proposals combine a top-down parser with complexity metrics that relate parsing difficulty to memory usage. So far, though, there has been no large-scale exploration of the space of viable metrics. Building on this earlier work, we compare the ability of 1600 metrics to derive several processing effects observed with relative clauses, many of which have been proven difficult to unify. We show that among those 1600 candidates, a few metrics (and only a few can provide a unified account of all these contrasts. This is a welcome result for two reasons: First, it provides a novel account of extensively studied psycholinguistic data. Second, it significantly limits the number of viable metrics that may be applied to other phenomena, thus reducing theoretical indeterminacy.

  13. High-content image informatics of the structural nuclear protein NuMA parses trajectories for stem/progenitor cell lineages and oncogenic transformation.

    Science.gov (United States)

    Vega, Sebastián L; Liu, Er; Arvind, Varun; Bushman, Jared; Sung, Hak-Joon; Becker, Matthew L; Lelièvre, Sophie; Kohn, Joachim; Vidi, Pierre-Alexandre; Moghe, Prabhas V

    2017-02-01

    Stem and progenitor cells that exhibit significant regenerative potential and critical roles in cancer initiation and progression remain difficult to characterize. Cell fates are determined by reciprocal signaling between the cell microenvironment and the nucleus; hence parameters derived from nuclear remodeling are ideal candidates for stem/progenitor cell characterization. Here we applied high-content, single cell analysis of nuclear shape and organization to examine stem and progenitor cells destined to distinct differentiation endpoints, yet undistinguishable by conventional methods. Nuclear descriptors defined through image informatics classified mesenchymal stem cells poised to either adipogenic or osteogenic differentiation, and oligodendrocyte precursors isolated from different regions of the brain and destined to distinct astrocyte subtypes. Nuclear descriptors also revealed early changes in stem cells after chemical oncogenesis, allowing the identification of a class of cancer-mitigating biomaterials. To capture the metrology of nuclear changes, we developed a simple and quantitative "imaging-derived" parsing index, which reflects the dynamic evolution of the high-dimensional space of nuclear organizational features. A comparative analysis of parsing outcomes via either nuclear shape or textural metrics of the nuclear structural protein NuMA indicates the nuclear shape alone is a weak phenotypic predictor. In contrast, variations in the NuMA organization parsed emergent cell phenotypes and discerned emergent stages of stem cell transformation, supporting a prognosticating role for this protein in the outcomes of nuclear functions. Copyright © 2017 Elsevier Inc. All rights reserved.

  14. Guide to LIBXSIF, a Library for Parsing the Extended Standard Input Format of Accelerated Beamlines(LCC-0060)

    International Nuclear Information System (INIS)

    Tenenbaum, P

    2003-01-01

    We describe LIBXSIF, a standalone library for parsing the Extended Standard Input Format of accelerator beamlines. Included in the description are: documentation of user commands; full description of permitted accelerator elements and their attributes; the construction of beamline lists; the mechanics of adding LIBXSIF to an existing program; and ''under the hood'' details for users who wish to modify the library or are merely morbidly curious

  15. Parsing in a Dynamical System: An Attractor-Based Account of the Interaction of Lexical and Structural Constraints in Sentence Processing.

    Science.gov (United States)

    Tabor, Whitney; And Others

    1997-01-01

    Proposes a dynamical systems approach to parsing in which syntactic hypotheses are associated with attractors in a metric space. The experiments discussed documented various contingent frequency effects that cut across traditional linguistic grains, each of which was predicted by the dynamical systems model. (47 references) (Author/CK)

  16. BAIK– PROGRAMMING LANGUAGE BASED ON INDONESIAN LEXICAL PARSING FOR MULTITIER WEB DEVELOPMENT

    Directory of Open Access Journals (Sweden)

    Haris Hasanudin

    2012-05-01

    Full Text Available Business software development with global team is increasing rapidly and the programming language as development tool takes the important role in the global web development. The real user friendly programming language should be written in local language for programmer who has native language is not in English. This paper presents our design of BAIK (Bahasa Anak Indonesia untuk Komputerscripting language which syntax is modeled with Bahasa Indonesian for multitier web development. Researcher propose the implementation of Indonesian Parsing Engine and Binary Search Tree structure for memory allocation of variable and compose the language features that support basic Object Oriented Programming, Common Gateway Interface, HTML style manipulation and database connection. Our goal is to build real programming language from simple structure design for web development using Indonesian lexical words. Pengembangan bisnis perangkat lunak dalam tim berskala global meningkat dengan cepat dan bahasa pemrograman berperan penting dalam pengembangan web secara global. Bahasa pemrograman yang benar-benar ramah terhadap pengguna harus ditulis dalam bahasa lokal programmer yang bahasa ibunya bukan Bahasa Inggris. Paper ini menyajikan desain dari bahasa penulisan BAIK (Bahasa Anak Indonesia untuk Komputer, yang sintaksisnya dimodelkan dengan Bahasa Indonesia untuk pengembangan web multitier. Peneliti mengusulkan implementasi dari parsing engine Bahasa Indonesia dan struktur binary search tree untuk alokasi memori terhadap variabel, serta membuat fitur bahasa yang mendukung dasar pemrograman berbasis objek, common gateway interface, manipulasi gaya HTML, dan koneksi basis data. Tujuan penelitian ini adalah untuk menciptakan bahasa pemrograman yang sesungguhnya dan menggunakan desain struktur sederhana untuk pengembangan web dengan menggunakan kata-kata dari Bahasa Indonesia.

  17. Parsing (malicious pleasures: Schadenfreude and gloating at others’ adversity

    Directory of Open Access Journals (Sweden)

    Colin Wayne Leach

    2015-02-01

    Full Text Available We offer the first empirical comparison of the pleasure in seeing (i.e., schadenfreude and in causing (i.e., gloating others’ adversity. In Study 1, we asked participants to recall and report on an (individual or group episode of pleasure that conformed to our formal definition of schadenfreude, gloating, pride, or joy, without reference to an emotion word. Schadenfreude and gloating were distinct in the situational features of the episode, participants’ appraisals of it, and their expressions of pleasure (e.g., smiling, boasting. In Study 2, we had participants imagine being in an (individual or group emotion episode designed to fit our conceptualization of schadenfreude or gloating. Individual and group versions of the emotions did not differ much in either study. However, the two pleasures differed greatly in their situational features, appraisals, experience, and expression. This parsing of the particular pleasures of schadenfreude and gloating brings nuance to the study of (malicious pleasure, which tends to be less finely conceptualized and examined than displeasure despite its importance to social relations.

  18. GFFview: A Web Server for Parsing and Visualizing Annotation Information of Eukaryotic Genome.

    Science.gov (United States)

    Deng, Feilong; Chen, Shi-Yi; Wu, Zhou-Lin; Hu, Yongsong; Jia, Xianbo; Lai, Song-Jia

    2017-10-01

    Owing to wide application of RNA sequencing (RNA-seq) technology, more and more eukaryotic genomes have been extensively annotated, such as the gene structure, alternative splicing, and noncoding loci. Annotation information of genome is prevalently stored as plain text in General Feature Format (GFF), which could be hundreds or thousands Mb in size. Therefore, it is a challenge for manipulating GFF file for biologists who have no bioinformatic skill. In this study, we provide a web server (GFFview) for parsing the annotation information of eukaryotic genome and then generating statistical description of six indices for visualization. GFFview is very useful for investigating quality and difference of the de novo assembled transcriptome in RNA-seq studies.

  19. “Less of the Heroine than the Woman”: Parsing Gender in the British Novel

    Directory of Open Access Journals (Sweden)

    Susan Carlile

    2017-06-01

    Full Text Available This essay offers two methods that will help students resist the temptation to judge eighteenth-century novels by twenty-first-century standards. These methods prompt students to parse the question of whether female protagonists in novels—in this case, Daniel Defoe’s Roxana (1724, Samuel Johnson’s Rasselas (1759, and Charlotte Lennox’s Sophia (1762—are portrayed as perfect models or as complex humans. The first method asks them to engage with definitions of the term “heroine,” and the second method uses word clouds to extend their thinking about the complexity of embodying a mid-eighteenth-century female identity.

  20. Pseudocode Interpreter (Pseudocode Integrated Development Environment with Lexical Analyzer and Syntax Analyzer using Recursive Descent Parsing Algorithm

    Directory of Open Access Journals (Sweden)

    Christian Lester D. Gimeno

    2017-11-01

    Full Text Available –This research study focused on the development of a software that helps students design, write, validate and run their pseudocode in a semi Integrated Development Environment (IDE instead of manually writing it on a piece of paper.Specifically, the study aimed to develop lexical analyzer or lexer, syntax analyzer or parser using recursive descent parsing algorithm and an interpreter. The lexical analyzer reads pseudocodesource in a sequence of symbols or characters as lexemes.The lexemes are then analyzed by the lexer that matches a pattern for valid tokens and passes to the syntax analyzer or parser. The syntax analyzer or parser takes those valid tokens and builds meaningful commands using recursive descent parsing algorithm in a form of an abstract syntax tree. The generation of an abstract syntax tree is based on the specified grammar rule created by the researcher expressed in Extended Backus-Naur Form. The Interpreter takes the generated abstract syntax tree and starts the evaluation or interpretation to produce pseudocode output. The software was evaluated using white-box testing by several ICT professionals and black-box testing by several computer science students based on the International Organization for Standardization (ISO 9126 software quality standards. The overall results of the evaluation both for white-box and black-box were described as “Excellent in terms of functionality, reliability, usability, efficiency, maintainability and portability”.

  1. The effect of recognizability on figure-ground processing: does it affect parsing or only figure selection?

    Science.gov (United States)

    Navon, David

    2011-03-01

    Though figure-ground assignment has been shown to be probably affected by recognizability, it appears sensible that object recognition must follow at least the earlier process of figure-ground segregation. To examine whether or not rudimentary object recognition could, counterintuitively, start even before the completion of the stage of parsing in which figure-ground segregation is done, participants were asked to respond, in a go/no-go fashion, whenever any out of 16 alternative connected patterns (that constituted familiar stimuli in the upright orientation) appeared. The white figure of the to-be-attended stimulus-target or foil-could be segregated from the white ambient ground only by means of a frame surrounding it. Such a frame was absent until the onset of target display. Then, to manipulate organizational quality, the greyness of the frame was either gradually increased from zero (in Experiment 1) or changed abruptly to a stationary level whose greyness was varied between trials (in Experiments 2 and 3). Stimulus recognizability was manipulated by orientation angle. In all three experiments the effect of recognizability was found to be considerably larger when organizational quality was minimal due to an extremely faint frame. This result is argued to be incompatible with any version of a serial thesis suggesting that processing aimed at object recognition starts only with a good enough level of organizational quality. The experiments rather provide some support to the claim, termed here "early interaction hypothesis", positing interaction between early recognition processing and preassignment parsing processes.

  2. Thermo-msf-parser: an open source Java library to parse and visualize Thermo Proteome Discoverer msf files.

    Science.gov (United States)

    Colaert, Niklaas; Barsnes, Harald; Vaudel, Marc; Helsens, Kenny; Timmerman, Evy; Sickmann, Albert; Gevaert, Kris; Martens, Lennart

    2011-08-05

    The Thermo Proteome Discoverer program integrates both peptide identification and quantification into a single workflow for peptide-centric proteomics. Furthermore, its close integration with Thermo mass spectrometers has made it increasingly popular in the field. Here, we present a Java library to parse the msf files that constitute the output of Proteome Discoverer. The parser is also implemented as a graphical user interface allowing convenient access to the information found in the msf files, and in Rover, a program to analyze and validate quantitative proteomics information. All code, binaries, and documentation is freely available at http://thermo-msf-parser.googlecode.com.

  3. Grammar-Based Specification and Parsing of Binary File Formats

    Directory of Open Access Journals (Sweden)

    William Underwood

    2012-03-01

    Full Text Available The capability to validate and view or play binary file formats, as well as to convert binary file formats to standard or current file formats, is critically important to the preservation of digital data and records. This paper describes the extension of context-free grammars from strings to binary files. Binary files are arrays of data types, such as long and short integers, floating-point numbers and pointers, as well as characters. The concept of an attribute grammar is extended to these context-free array grammars. This attribute grammar has been used to define a number of chunk-based and directory-based binary file formats. A parser generator has been used with some of these grammars to generate syntax checkers (recognizers for validating binary file formats. Among the potential benefits of an attribute grammar-based approach to specification and parsing of binary file formats is that attribute grammars not only support format validation, but support generation of error messages during validation of format, validation of semantic constraints, attribute value extraction (characterization, generation of viewers or players for file formats, and conversion to current or standard file formats. The significance of these results is that with these extensions to core computer science concepts, traditional parser/compiler technologies can potentially be used as a part of a general, cost effective curation strategy for binary file formats.

  4. SmilesDrawer: Parsing and Drawing SMILES-Encoded Molecular Structures Using Client-Side JavaScript.

    Science.gov (United States)

    Probst, Daniel; Reymond, Jean-Louis

    2018-01-22

    Here we present SmilesDrawer, a dependency-free JavaScript component capable of both parsing and drawing SMILES-encoded molecular structures client-side, developed to be easily integrated into web projects and to display organic molecules in large numbers and fast succession. SmilesDrawer can draw structurally and stereochemically complex structures such as maitotoxin and C 60 without using templates, yet has an exceptionally small computational footprint and low memory usage without the requirement for loading images or any other form of client-server communication, making it easy to integrate even in secure (intranet, firewalled) or offline applications. These features allow the rendering of thousands of molecular structure drawings on a single web page within seconds on a wide range of hardware supporting modern browsers. The source code as well as the most recent build of SmilesDrawer is available on Github ( http://doc.gdb.tools/smilesDrawer/ ). Both yarn and npm packages are also available.

  5. A Python package for parsing, validating, mapping and formatting sequence variants using HGVS nomenclature.

    Science.gov (United States)

    Hart, Reece K; Rico, Rudolph; Hare, Emily; Garcia, John; Westbrook, Jody; Fusaro, Vincent A

    2015-01-15

    Biological sequence variants are commonly represented in scientific literature, clinical reports and databases of variation using the mutation nomenclature guidelines endorsed by the Human Genome Variation Society (HGVS). Despite the widespread use of the standard, no freely available and comprehensive programming libraries are available. Here we report an open-source and easy-to-use Python library that facilitates the parsing, manipulation, formatting and validation of variants according to the HGVS specification. The current implementation focuses on the subset of the HGVS recommendations that precisely describe sequence-level variation relevant to the application of high-throughput sequencing to clinical diagnostics. The package is released under the Apache 2.0 open-source license. Source code, documentation and issue tracking are available at http://bitbucket.org/hgvs/hgvs/. Python packages are available at PyPI (https://pypi.python.org/pypi/hgvs). Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press.

  6. LCF data for aging of Cu NPs in soil

    Data.gov (United States)

    U.S. Environmental Protection Agency — Linear combination fitting data for Cu NPs in five soils as a function of aging time. This dataset is associated with the following publication: Sekine, R., E....

  7. Parsing pyrogenic polycyclic aromatic hydrocarbons: forensic chemistry, receptor models, and source control policy.

    Science.gov (United States)

    O'Reilly, Kirk T; Pietari, Jaana; Boehm, Paul D

    2014-04-01

    A realistic understanding of contaminant sources is required to set appropriate control policy. Forensic chemical methods can be powerful tools in source characterization and identification, but they require a multiple-lines-of-evidence approach. Atmospheric receptor models, such as the US Environmental Protection Agency (USEPA)'s chemical mass balance (CMB), are increasingly being used to evaluate sources of pyrogenic polycyclic aromatic hydrocarbons (PAHs) in sediments. This paper describes the assumptions underlying receptor models and discusses challenges in complying with these assumptions in practice. Given the variability within, and the similarity among, pyrogenic PAH source types, model outputs are sensitive to specific inputs, and parsing among some source types may not be possible. Although still useful for identifying potential sources, the technical specialist applying these methods must describe both the results and their inherent uncertainties in a way that is understandable to nontechnical policy makers. The authors present an example case study concerning an investigation of a class of parking-lot sealers as a significant source of PAHs in urban sediment. Principal component analysis is used to evaluate published CMB model inputs and outputs. Targeted analyses of 2 areas where bans have been implemented are included. The results do not support the claim that parking-lot sealers are a significant source of PAHs in urban sediments. © 2013 SETAC.

  8. A Pipelining Implementation for Parsing X-ray Diffraction Source Data and Removing the Background Noise

    International Nuclear Information System (INIS)

    Bauer, Michael A; Biem, Alain; McIntyre, Stewart; Xie Yuzhen

    2010-01-01

    Synchrotrons can be used to generate X-rays in order to probe materials at the atomic level. One approach is to use X-ray diffraction (XRD) to do this. The data from an XRD experiment consists of a sequence of digital image files which for a single scan could consist of hundreds or even thousands of digital images. Existing analysis software processes these images individually sequentially and is usually used after the experiment is completed. The results from an XRD detector can be thought of as a sequence of images, generated during the scan by the X-ray beam. If these images could be analyzed in near real-time, the results could be sent to the researcher running the experiment and used to improve the overall experimental process and results. In this paper, we report on a stream processing application to remove background from XRD images using a pipelining implementation. We describe our implementation techniques of using IBM Infosphere Streams for parsing XRD source data and removing the background. We present experimental results showing the super-linear speedup attained over a purely sequential version of the algorithm on a quad-core machine. These results demonstrate the potential of making good use of multi-cores for high-performance stream processing of XRD images.

  9. Conceptual plural information is used to guide early parsing decisions: Evidence from garden-path sentences with reciprocal verbs.

    Science.gov (United States)

    Patson, Nikole D; Ferreira, Fernanda

    2009-05-01

    In three eyetracking studies, we investigated the role of conceptual plurality in initial parsing decisions in temporarily ambiguous sentences with reciprocal verbs (e.g., While the lovers kissed the baby played alone). We varied the subject of the first clause using three types of plural noun phrases: conjoined noun phrases (the bride and the groom), plural definite descriptions (the lovers), and numerically quantified noun phrases (the two lovers). We found no evidence for garden-path effects when the subject was conjoined (Ferreira & McClure, 1997), but traditional garden-path effects were found with the other plural noun phrases. In addition, we tested plural anaphors that had a plural antecedent present in the discourse. We found that when the antecedent was conjoined, garden-path effects were absent compared to cases in which the antecedent was a plural definite description. Our results indicate that the parser is sensitive to the conceptual representation of a plural constituent. In particular, it appears that a Complex Reference Object (Moxey et al., 2004) automatically activates a reciprocal reading of a reciprocal verb.

  10. JAFF rebib akna alternatiivmaailmadesse

    Index Scriptorium Estoniae

    2011-01-01

    Tallinna ja Tartu kinodes 15.-24. aprillini toimuva 5. Jaapani animatsiooni filmifestivali JAFF kavast. JAFF-i teemaks on ookean. Fookuses on animemeister Satoshi Kon, festivali kavas on neli tema filmi

  11. Deep PDF parsing to extract features for detecting embedded malware.

    Energy Technology Data Exchange (ETDEWEB)

    Munson, Miles Arthur; Cross, Jesse S. (Missouri University of Science and Technology, Rolla, MO)

    2011-09-01

    The number of PDF files with embedded malicious code has risen significantly in the past few years. This is due to the portability of the file format, the ways Adobe Reader recovers from corrupt PDF files, the addition of many multimedia and scripting extensions to the file format, and many format properties the malware author may use to disguise the presence of malware. Current research focuses on executable, MS Office, and HTML formats. In this paper, several features and properties of PDF Files are identified. Features are extracted using an instrumented open source PDF viewer. The feature descriptions of benign and malicious PDFs can be used to construct a machine learning model for detecting possible malware in future PDF files. The detection rate of PDF malware by current antivirus software is very low. A PDF file is easy to edit and manipulate because it is a text format, providing a low barrier to malware authors. Analyzing PDF files for malware is nonetheless difficult because of (a) the complexity of the formatting language, (b) the parsing idiosyncrasies in Adobe Reader, and (c) undocumented correction techniques employed in Adobe Reader. In May 2011, Esparza demonstrated that PDF malware could be hidden from 42 of 43 antivirus packages by combining multiple obfuscation techniques [4]. One reason current antivirus software fails is the ease of varying byte sequences in PDF malware, thereby rendering conventional signature-based virus detection useless. The compression and encryption functions produce sequences of bytes that are each functions of multiple input bytes. As a result, padding the malware payload with some whitespace before compression/encryption can change many of the bytes in the final payload. In this study we analyzed a corpus of 2591 benign and 87 malicious PDF files. While this corpus is admittedly small, it allowed us to test a system for collecting indicators of embedded PDF malware. We will call these indicators features throughout

  12. Plaadid / Mart Normet

    Index Scriptorium Estoniae

    Normet, Mart, 1979-

    2006-01-01

    Uutest heliplaatidest Barry White "Gold", Roseanne Cash "Black Cadillac", Satoshi Tomiie "Renaissance", Tiga "Sexor", Rec ja Rapresent "Isiklikult", Yo-Yo Ma "the Essential Yo-Yo Ma", Aqua Bassino "Rue de Paris"

  13. KAMUS BAHASA ARAB – INDONESIA ONLINE DENGAN PEMECAHAN SUKU KATA MENGGUNAKAN METODE PARSING

    Directory of Open Access Journals (Sweden)

    Anny Yuniarti

    2004-01-01

    Full Text Available Normal 0 false false false IN X-NONE X-NONE MicrosoftInternetExplorer4 /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-fareast-font-family:"Times New Roman"; mso-fareast-theme-font:minor-fareast; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-bidi-font-family:"Times New Roman"; mso-bidi-theme-font:minor-bidi;} Kebutuhan umat Islam akan fasilitas penunjang belajar bahasa Arab di Indonesia masih belum terpenuhi dengan optimal. Kamus bahasa Arab yang beredar di pasaran sulit dipahami karena minimnya pengetahuan tentang ilmu tata bahasa Arab di kalangan umat Islam. Pada penelitian ini dikembangkan sebuah perangkat lunak yang berfungsi menerjemahkan kata berbahasa Arab dengan metode parsing sehingga dapat mencakup kata-kata yang telah mengalami perubahan bentuk dari bentuk dasarnya. Karena kata bahasa Arab memiliki turunan kata yang jumlahnya cukup besar, dan supaya kamus efisien, maka tidak semua turunan kata disimpan dalam basisdata. Oleh sebab itu diperlukan suatu cara untuk mengenali pola kata, dan cara mengetahui bentuk dasar suatu kata. Keseluruhan perangkat lunak ini diimplementasikan berbasis web sehingga memudahkan pengaksesan pengguna. Dan pengguna tidak memerlukan proses instalasi perangkat lunak atau sistem operasi tertentu. Pembuatan perangkat lunak ini didahului dengan perancangan proses dan perancangan interface. Kemudian rancangan tersebut diimplementasikan menjadi sebuah perangkat lunak yang siap untuk dipakai. Perangkat lunak yang sudah jadi tersebut telah diuji coba sesuai dengan spesifikasi kebutuhan

  14. Parsing multiple processes of high temperature impacts on corn/soybean yield using a newly developed CLM-APSIM modeling framework

    Science.gov (United States)

    Peng, B.; Guan, K.; Chen, M.

    2016-12-01

    Future agricultural production faces a grand challenge of higher temperature under climate change. There are multiple physiological or metabolic processes of how high temperature affects crop yield. Specifically, we consider the following major processes: (1) direct temperature effects on photosynthesis and respiration; (2) speed-up growth rate and the shortening of growing season; (3) heat stress during reproductive stage (flowering and grain-filling); (4) high-temperature induced increase of atmospheric water demands. In this work, we use a newly developed modeling framework (CLM-APSIM) to simulate the corn and soybean growth and explicitly parse the above four processes. By combining the strength of CLM in modeling surface biophysical (e.g., hydrology and energy balance) and biogeochemical (e.g., photosynthesis and carbon-nitrogen interactions), as well as that of APSIM in modeling crop phenology and reproductive stress, the newly developed CLM-APSIM modeling framework enables us to diagnose the impacts of high temperature stress through different processes at various crop phenology stages. Ground measurements from the advanced SoyFACE facility at University of Illinois is used here to calibrate, validate, and improve the CLM-APSIM modeling framework at the site level. We finally use the CLM-APSIM modeling framework to project crop yield for the whole US Corn Belt under different climate scenarios.

  15. Vudu v tjomnom kinozale / Nikolai Karajev

    Index Scriptorium Estoniae

    Karajev, Nikolai, 1978-

    2007-01-01

    PÖFFi animafilmide festivali "Animated Dreams" ja Nukufilmi 50. juubeli külalised konverentsil "Voodoo hing" olid Pjotr Sapegin, Peter Lord, Kihachiro Kawamoto. Ka Satoshi Koni täispikast animafilmist "Paprika" ja auhinnatud filmidest

  16. Vselennaja po imeni "Anime" / Nikolai Karajev

    Index Scriptorium Estoniae

    Karajev, Nikolai, 1978-

    2007-01-01

    Jaapani anime filmidest, mida sai näha 19.-25. märtsil I Jaapani Animatsiooni Filmifestivalil Eestis. Ka legendaarsest animarežissöörist Satoshi Kon'ist seoses tema filmiga "Aastatuhande näitlejatar" : Jaapan 2001

  17. Influence of ROM Exercise on the Joint Components during Immobilization

    OpenAIRE

    松﨑, 太郎; Matsuzaki, Taro

    2013-01-01

    博士論文要旨Abstract 要約Outline 以下に掲載:Journal of Physical Therapy Science 25(12) pp.1547-1551 2013. The Society of Physical Therapy Science. 共著者:Taro Matsuzaki, Shinya Yoshida, Satoshi Kojima, Masanori Watanabe, Masahiro Hoso

  18. Fulltext PDF

    Indian Academy of Sciences (India)

    IAS Admin

    Satoshi Omura: Kitasato University, Tokyo, Japan. and. “for her discoveries concerning a novel therapy against Malaria” to. Youyou Tu: China Academy of Traditional Chinese Medicine, Beijing, China. Physics. "for the discovery of neutrino oscillations, which shows that neutrinos have mass" to. Takaaki Kajita: University of ...

  19. License - PGDBj Registered plant list, Marker list, QTL list, Plant DB link & Genome analysis methods | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available List Contact us PGDBj Registered plant list, Marker list, QTL list, Plant DB link & Genome analysis methods ...t list, Marker list, QTL list, Plant DB link & Genome analysis methods © Satoshi ... Policy | Contact Us License - PGDBj Registered plant list, Marker list, QTL list, Plant DB link & Genome analysis methods | LSDB Archive ...

  20. The Primordial Soup Algorithm : a systematic approach to the specification of parallel parsers

    NARCIS (Netherlands)

    Janssen, Wil; Janssen, W.P.M.; Poel, Mannes; Sikkel, Nicolaas; Zwiers, Jakob

    1992-01-01

    A general framework for parallel parsing is presented, which allows for a unitied, systematic approach to parallel parsing. The Primordial Soup Algorithm creates trees by allowing partial parse trees to combine arbitrarily. By adding constraints to the general algorithm, a large, class of parallel

  1. Effect of a steam foot spa on geriatric inpatients with cognitive impairment: a pilot study

    OpenAIRE

    Koike Y; Kondo H; Kondo S; Takagi M; Kano Y

    2013-01-01

    Yoshihisa Koike,1 Hideki Kondo,2 Satoshi Kondo,1 Masayuki Takagi,1 Yoshio Kano31Department of Occupational Therapy, Prefectural University of Hiroshima, Mihara, Japan; 2Yoshiigawa Hospital, Okayama, Japan; 3Department of Occupational Therapy, Kibi International University, Takahashi, JapanPurpose: To investigate whether a steam foot spa improves cognitive impairment in geriatric inpatients.Methods: Geriatric inpatients with cognitive impairment were given a steam foot spa treatment at 42&...

  2. Assessment of effects of differences in trunk posture during Fowler’s position on hemodynamics and cardiovascular regulation in older and younger subjects

    OpenAIRE

    Kubota,Satoshi; Endo,Yutaka; Kubota,Mitsue; Shigemasa,Tomohiko

    2017-01-01

    Satoshi Kubota,1 Yutaka Endo,1 Mitsue Kubota,1 Tomohiko Shigemasa2 1School of Nursing and Rehabilitation Sciences at Odawara, International University of Health and Welfare, Odawara, Kanagawa, Japan; 2Department of Cardiology, International University of Health and Welfare Atami Hospital, Atami, Shizuoka, Japan Background: Downward shifts in blood volume with changing position generally cause tachycardic responses. Age-related decreases in vagal nerve activity could contribute to orthostati...

  3. Solving LR Conflicts Through Context Aware Scanning

    Science.gov (United States)

    Leon, C. Rodriguez; Forte, L. Garcia

    2011-09-01

    This paper presents a new algorithm to compute the exact list of tokens expected by any LR syntax analyzer at any point of the scanning process. The lexer can, at any time, compute the exact list of valid tokens to return only tokens in this set. In the case than more than one matching token is in the valid set, the lexer can resort to a nested LR parser to disambiguate. Allowing nested LR parsing requires some slight modifications when building the LR parsing tables. We also show how LR parsers can parse conflictive and inherently ambiguous languages using a combination of nested parsing and context aware scanning. These expanded lexical analyzers can be generated from high level specifications.

  4. Coping as a mediator of the relationship between stress mindset and psychological stress response: a pilot study

    OpenAIRE

    Horiuchi,Satoshi; Tsuda,Akira; Aoki,Shuntaro; Yoneda,Kenichiro; Sawaguchi,Yusuke

    2018-01-01

    Satoshi Horiuchi,1 Akira Tsuda,2 Shuntaro Aoki,3,4 Kenichiro Yoneda,5 Yusuke Sawaguchi6 1Faculty of Social Welfare, Iwate Prefectural University, Iwate, 2Department of Psychology, Kurume University, Fukuoka, 3Research Fellow of Japan Society for the Promotion of Science, Tokyo, 4Graduate School of Psychological Science, Health Sciences University of Hokkaido, Hokkaido, 5Graduate School of Psychology, Kurume University, Fukuoka, 6Graduate School of Social Welfare, Iwate Prefectural University,...

  5. Topical Meeting on Signal Recovery and Synthesis with Incomplete Information and Partial Constraints Held at Incline Village, Nevada on January 12-14, 1983.

    Science.gov (United States)

    1983-08-31

    Reconstruction from Phase Synchronization of Distorted Imaging Antenna Multiple Linear-Motion Tomograms, Satoshi Kawata, Arrays, Bernard D. Steinberg...phase recovery. Most other LBI imaging processes (eg. Readhead [4] and Cornwell [5]) are iterative, and the image must be deconvolved from the beam as...VLBI Data" Astrophys.J. Vol 223 July 1978 pp25-36 [5] T.J. Cornwell & P.N. Wilkinson "A New Method for Making Maps with Unstable Radio

  6. An Explanation of Nakamoto's Analysis of Double-spend Attacks

    OpenAIRE

    Ozisik, A. Pinar; Levine, Brian Neil

    2017-01-01

    The fundamental attack against blockchain systems is the double-spend attack. In this tutorial, we provide a very detailed explanation of just one section of Satoshi Nakamoto's original paper where the attack's probability of success is stated. We show the derivation of the mathematics relied upon by Nakamoto to create a model of the attack. We also validate the model with a Monte Carlo simulation, and we determine which model component is not perfect.

  7. The Humanbecoming theory as a reinterpretation of the symbolic interactionism: a critique of its specific nature and scientific underpinnings.

    Science.gov (United States)

    Tapp, Diane; Lavoie, Mireille

    2017-04-01

    Discussions about real knowledge contained in grand theories and models seem to remain an active quest in the academic sphere. The most fervent of these defendants is Rosemarie Parse with her Humanbecoming School of Thought (1981, 1998). This article first highlights the similarities between Parse's theory and Blumer's symbolic interactionism (1969). This comparison will act as a counterargument to Parse's assertions that her theory is original 'nursing' material. Standing on the contemporary philosophy of science, the very possibility for discovering specific nursing knowledge will be questioned. Second, Parse's scientific assumptions will be thoroughly addressed and contrasted with Blumer's more moderate view of knowledge. It will lead to recognize that the valorization of the social nature of existence and reality does not necessarily induce requirements and methods such as those proposed by Parse. According to Blumer's point of view, her perspective may not even be desirable. Recommendations will be raised about the necessity for a distanced relationship to knowledge, being the key to the pursuit of its improvement, not its circular contemplation. © 2016 John Wiley & Sons Ltd.

  8. SyntaxTrain: Relieving the pain of learning syntax

    DEFF Research Database (Denmark)

    Moth, Andreas Leon Aagaard; Villadsen, Jørgen; Ben-Ari, Mordechai

    2011-01-01

    SyntaxTrain parses a Java program and displays the syntax diagrams associated with a syntax error.......SyntaxTrain parses a Java program and displays the syntax diagrams associated with a syntax error....

  9. The association between suicide risk and self-esteem in Japanese university students with major depressive episodes of major depressive disorder

    OpenAIRE

    Mitsui, Nobuyuki; Asakura,Satoshi; Shimizu,Yusuke; Fujii,Yutaka; Toyomaki,Atsuhito; Kako,Yuki; Tanaka,Teruaki; Kitagawa,Nobuki; Inoue,Takeshi; Kusumi,Ichiro

    2014-01-01

    Nobuyuki Mitsui,1 Satoshi Asakura,1,2 Yusuke Shimizu,1 Yutaka Fujii,1 Atsuhito Toyomaki,1 Yuki Kako,1 Teruaki Tanaka,1 Nobuki Kitagawa,3 Takeshi Inoue,1 Ichiro Kusumi1 1Department of Psychiatry, Hokkaido University Graduate School of Medicine, 2Health care center of Hokkaido University, Kita-ku, Sapporo, 3Department of Clinical Social Work, Health Sciences University of Hokkaido School of Nursing and Social Services, Tobetsu, Ishikari, Japan Background: The suicide risk among young adults is...

  10. JPRS Report, Science & Technology, Japan, 4th Intelligent Robots Symposium, Volume 1

    Science.gov (United States)

    1989-03-16

    Robots 43064062 Tokyo 4TH INTELLIGENT ROBOTS SYMPOSIUM PAPERS in Japanese 13/14 Jun 88 No 106 pp 39-44 [Article by H. Kimura , I. Shimoyama, and H. Miura...pp 237-239. 3. K. Kondo and F. Kimura , "High-Speed Orbit Planning Based on Labyrinthine Method," Ibid., pp 261-262. 4. T. Hasegawa, "Planning of...Satoshi Okada, Kazuhiro Tsumura, Hisashi Hozeki, Katsumi Kubo, and Akira Abe, Toshiba Corporation] [Text] 1. Introduction In recent years, the

  11. Policy Considerations for the Blockchain Technology Public and Private Applications

    OpenAIRE

    GABISON Garry

    2016-01-01

    In 2008, Satoshi Nakamoto—an individual or group of individuals— released a paper that described Bitcoin, a first of its kind, peer-to-peer electronic cash system. Bitcoin relies mostly on existing technology but requires a new invention, a blockchain, to solve an old problem: how do two parties conduct an online transaction without knowing or trusting each other and without the need for a trusted third-party intermediary? Encryption and large-scale redundancy was combined w...

  12. BITCOIN – A NEW GLOBAL CURRENCY, INVESTMENT OPPORTUNITY OR SOMETHING ELSE?

    OpenAIRE

    Buterin, Denis; Ribarić, Eda; Savić, Suzana

    2015-01-01

    The history of bitcoin started in 2008 when the article “Bitcoin: A Peer-to-Peer Electronic Cash System” was published under the pseudonym Satoshi Nakamoto. Since then the bitcoin has undergone significant changes marked with rapid growth and decline of its value, accompanied by public attention. The bitcoin is a system based on complex cryptographic algorithms without central authority that releases money or monitors transactions. There are discussions over the issue whether bitcoin could po...

  13. Forest Grammar (Ⅱ)

    Institute of Scientific and Technical Information of China (English)

    张松懋

    1994-01-01

    The syntactic parsing algorithm of weak precedence forest grammar has been introduced and the correctness and unambiguity of this algorithm have been proved. An example is given to the syntactic parsing procedure of weak precedence forest grammar.

  14. Recognition of Equations Using a Two-Dimensional Stochastic Context-Free Grammar

    Science.gov (United States)

    Chou, Philip A.

    1989-11-01

    We propose using two-dimensional stochastic context-free grammars for image recognition, in a manner analogous to using hidden Markov models for speech recognition. The value of the approach is demonstrated in a system that recognizes printed, noisy equations. The system uses a two-dimensional probabilistic version of the Cocke-Younger-Kasami parsing algorithm to find the most likely parse of the observed image, and then traverses the corresponding parse tree in accordance with translation formats associated with each production rule, to produce eqn I troff commands for the imaged equation. In addition, it uses two-dimensional versions of the Inside/Outside and Baum re-estimation algorithms for learning the parameters of the grammar from a training set of examples. Parsing the image of a simple noisy equation currently takes about one second of cpu time on an Alliant FX/80.

  15. Audio Recording of Children with Dyslalia

    OpenAIRE

    Stefan Gheorghe Pentiuc; Maria D. Schipor; Ovidiu A. Schipor

    2008-01-01

    In this paper we present our researches regarding automat parsing of audio recordings. These recordings are obtained from children with dyslalia and are necessary for an accurate identification of speech problems. We develop a software application that helps parsing audio, real time, recordings.

  16. GBParsy: A GenBank flatfile parser library with high speed

    Directory of Open Access Journals (Sweden)

    Kim Yeon-Ki

    2008-07-01

    Full Text Available Abstract Background GenBank flatfile (GBF format is one of the most popular sequence file formats because of its detailed sequence features and ease of readability. To use the data in the file by a computer, a parsing process is required and is performed according to a given grammar for the sequence and the description in a GBF. Currently, several parser libraries for the GBF have been developed. However, with the accumulation of DNA sequence information from eukaryotic chromosomes, parsing a eukaryotic genome sequence with these libraries inevitably takes a long time, due to the large GBF file and its correspondingly large genomic nucleotide sequence and related feature information. Thus, there is significant need to develop a parsing program with high speed and efficient use of system memory. Results We developed a library, GBParsy, which was C language-based and parses GBF files. The parsing speed was maximized by using content-specified functions in place of regular expressions that are flexible but slow. In addition, we optimized an algorithm related to memory usage so that it also increased parsing performance and efficiency of memory usage. GBParsy is at least 5 - 100× faster than current parsers in benchmark tests. Conclusion GBParsy is estimated to extract annotated information from almost 100 Mb of a GenBank flatfile for chromosomal sequence information within a second. Thus, it should be used for a variety of applications such as on-time visualization of a genome at a web site.

  17. How (not) to train a dependency parser: The curious case of jackknifing part-of-speech taggers

    DEFF Research Database (Denmark)

    Agic, Zeljko; Schluter, Natalie

    2017-01-01

    In dependency parsing, jackknifing taggers is indiscriminately used as a simple adaptation strategy. Here, we empirically evaluate when and how (not) to use jackknifing in parsing. On 26 languages, we reveal a preference that conflicts with, and surpasses the ubiquitous ten-folding. We show no cl...

  18. Effects of Tasks on BOLD Signal Responses to Sentence Contrasts: Review and Commentary

    Science.gov (United States)

    Caplan, David; Gow, David

    2012-01-01

    Functional neuroimaging studies of syntactic processing have been interpreted as identifying the neural locations of parsing and interpretive operations. However, current behavioral studies of sentence processing indicate that many operations occur simultaneously with parsing and interpretation. In this review, we point to issues that arise in…

  19. Audio Recording of Children with Dyslalia

    Directory of Open Access Journals (Sweden)

    Stefan Gheorghe Pentiuc

    2008-01-01

    Full Text Available In this paper we present our researches regarding automat parsing of audio recordings. These recordings are obtained from children with dyslalia and are necessary for an accurate identification of speech problems. We develop a software application that helps parsing audio, real time, recordings.

  20. HDM/PASCAL Verification System User's Manual

    Science.gov (United States)

    Hare, D.

    1983-01-01

    The HDM/Pascal verification system is a tool for proving the correctness of programs written in PASCAL and specified in the Hierarchical Development Methodology (HDM). This document assumes an understanding of PASCAL, HDM, program verification, and the STP system. The steps toward verification which this tool provides are parsing programs and specifications, checking the static semantics, and generating verification conditions. Some support functions are provided such as maintaining a data base, status management, and editing. The system runs under the TOPS-20 and TENEX operating systems and is written in INTERLISP. However, no knowledge is assumed of these operating systems or of INTERLISP. The system requires three executable files, HDMVCG, PARSE, and STP. Optionally, the editor EMACS should be on the system in order for the editor to work. The file HDMVCG is invoked to run the system. The files PARSE and STP are used as lower forks to perform the functions of parsing and proving.

  1. Speed up of XML parsers with PHP language implementation

    Science.gov (United States)

    Georgiev, Bozhidar; Georgieva, Adriana

    2012-11-01

    In this paper, authors introduce PHP5's XML implementation and show how to read, parse, and write a short and uncomplicated XML file using Simple XML in a PHP environment. The possibilities for mutual work of PHP5 language and XML standard are described. The details of parsing process with Simple XML are also cleared. A practical project PHP-XML-MySQL presents the advantages of XML implementation in PHP modules. This approach allows comparatively simple search of XML hierarchical data by means of PHP software tools. The proposed project includes database, which can be extended with new data and new XML parsing functions.

  2. Hardware independence checkout software

    Science.gov (United States)

    Cameron, Barry W.; Helbig, H. R.

    1990-01-01

    ACSI has developed a program utilizing CLIPS to assess compliance with various programming standards. Essentially the program parses C code to extract the names of all function calls. These are asserted as CLIPS facts which also include information about line numbers, source file names, and called functions. Rules have been devised to establish functions called that have not been defined in any of the source parsed. These are compared against lists of standards (represented as facts) using rules that check intersections and/or unions of these. By piping the output into other processes the source is appropriately commented by generating and executing parsed scripts.

  3. Improving Precision of Generated ASTs

    DEFF Research Database (Denmark)

    Winther, Johnni

    The parser-generator is an essential tool in grammarware and its output, the parse tree in form of the concrete or abstract syntax tree, often forms the basis for the whole structure of the grammarware application. Several tools for Java encode the parse tree in a class hierarchy generated to model...

  4. Deriving prosodic structures

    NARCIS (Netherlands)

    Günes, Güliz

    2015-01-01

    When we speak, we speak in prosodic chunks. That is, in the speech flow, we produce sound strings that are systematically parsed into intonational units. The parsing procedure not only eases the production of the speaker, but it also provides the hearer clues about how units of meaning interact in

  5. An unsupervised text mining method for relation extraction from biomedical literature.

    Directory of Open Access Journals (Sweden)

    Changqin Quan

    Full Text Available The wealth of interaction information provided in biomedical articles motivated the implementation of text mining approaches to automatically extract biomedical relations. This paper presents an unsupervised method based on pattern clustering and sentence parsing to deal with biomedical relation extraction. Pattern clustering algorithm is based on Polynomial Kernel method, which identifies interaction words from unlabeled data; these interaction words are then used in relation extraction between entity pairs. Dependency parsing and phrase structure parsing are combined for relation extraction. Based on the semi-supervised KNN algorithm, we extend the proposed unsupervised approach to a semi-supervised approach by combining pattern clustering, dependency parsing and phrase structure parsing rules. We evaluated the approaches on two different tasks: (1 Protein-protein interactions extraction, and (2 Gene-suicide association extraction. The evaluation of task (1 on the benchmark dataset (AImed corpus showed that our proposed unsupervised approach outperformed three supervised methods. The three supervised methods are rule based, SVM based, and Kernel based separately. The proposed semi-supervised approach is superior to the existing semi-supervised methods. The evaluation on gene-suicide association extraction on a smaller dataset from Genetic Association Database and a larger dataset from publicly available PubMed showed that the proposed unsupervised and semi-supervised methods achieved much higher F-scores than co-occurrence based method.

  6. Experimental Evaluation of Processing Time for the Synchronization of XML-Based Business Objects

    Science.gov (United States)

    Ameling, Michael; Wolf, Bernhard; Springer, Thomas; Schill, Alexander

    Business objects (BOs) are data containers for complex data structures used in business applications such as Supply Chain Management and Customer Relationship Management. Due to the replication of application logic, multiple copies of BOs are created which have to be synchronized and updated. This is a complex and time consuming task because BOs rigorously vary in their structure according to the distribution, number and size of elements. Since BOs are internally represented as XML documents, the parsing of XML is one major cost factor which has to be considered for minimizing the processing time during synchronization. The prediction of the parsing time for BOs is an significant property for the selection of an efficient synchronization mechanism. In this paper, we present a method to evaluate the influence of the structure of BOs on their parsing time. The results of our experimental evaluation incorporating four different XML parsers examine the dependencies between the distribution of elements and the parsing time. Finally, a general cost model will be validated and simplified according to the results of the experimental setup.

  7. Memory mechanisms supporting syntactic comprehension.

    Science.gov (United States)

    Caplan, David; Waters, Gloria

    2013-04-01

    Efforts to characterize the memory system that supports sentence comprehension have historically drawn extensively on short-term memory as a source of mechanisms that might apply to sentences. The focus of these efforts has changed significantly in the past decade. As a result of changes in models of short-term working memory (ST-WM) and developments in models of sentence comprehension, the effort to relate entire components of an ST-WM system, such as those in the model developed by Baddeley (Nature Reviews Neuroscience 4: 829-839, 2003) to sentence comprehension has largely been replaced by an effort to relate more specific mechanisms found in modern models of ST-WM to memory processes that support one aspect of sentence comprehension--the assignment of syntactic structure (parsing) and its use in determining sentence meaning (interpretation) during sentence comprehension. In this article, we present the historical background to recent studies of the memory mechanisms that support parsing and interpretation and review recent research into this relation. We argue that the results of this research do not converge on a set of mechanisms derived from ST-WM that apply to parsing and interpretation. We argue that the memory mechanisms supporting parsing and interpretation have features that characterize another memory system that has been postulated to account for skilled performance-long-term working memory. We propose a model of the relation of different aspects of parsing and interpretation to ST-WM and long-term working memory.

  8. MAD parsing and conversion code

    International Nuclear Information System (INIS)

    Mokhov, Dmitri N.

    2000-01-01

    The authors describe design and implementation issues while developing an embeddable MAD language parser. Two working applications of the parser are also described, namely, MAD-> C++ converter and C++ factory. The report contains some relevant details about the parser and examples of converted code. It also describes some of the problems that were encountered and the solutions found for them

  9. Parsing Universal Dependencies without training

    NARCIS (Netherlands)

    Martínez Alonso, Héctor; Agić, Željko; Plank, Barbara; Søgaard, Anders

    2017-01-01

    We propose UDP, the first training-free parser for Universal Dependencies (UD). Our algorithm is based on PageRank and a small set of head attachment rules. It features two-step decoding to guarantee that function words are attached as leaf nodes. The parser requires no training, and it is

  10. Parsing Chinese-Russian Military Exercises

    Science.gov (United States)

    2015-04-01

    as over 100 trucks and about a dozen aircraft from its nearby base in Kant , Kyrgyzstan, including Su-24 Fencer tactical bombers, Su-25 Frogfoot...to the exercise site through mountainous terrain, and some attack planes stationed at the CSTO base in Kant , Kyr- gyzstan.81 Kazakhstan sent an air

  11. Partial dependency parsing for Irish

    OpenAIRE

    Uí Dhonnchadha, Elaine; van Genabith, Josef

    2010-01-01

    In this paper we present a partial dependency parser for Irish, in which Constraint Grammar (CG) rules are used to annotate dependency relations and grammatical functions in unrestricted Irish text. Chunking is performed using a regular-expression grammar which operates on the dependency tagged sentences. As this is the first implementation of a parser for unrestricted Irish text (to our knowledge), there were no guidelines or precedents available. Therefore deciding what constitutes a syntac...

  12. Parsing statistical machine translation output

    NARCIS (Netherlands)

    Carter, S.; Monz, C.; Vetulani, Z.

    2009-01-01

    Despite increasing research into the use of syntax during statistical machine translation, the incorporation of syntax into language models has seen limited success. We present a study of the discriminative abilities of generative syntax-based language models, over and above standard n-gram models,

  13. Bitcoin: Documentos electrónicos para el intercambio de bienes y servicios

    OpenAIRE

    García Alejo, Luis Antonio

    2016-01-01

    [ES] Al hablar de Bitcoin no se hace alusión simplemente a una única cosa, al bien digital como tal, sino que nos referimos también al protocolo de envío y recepción de mensajes, basado en Open Source, a la red P2P (Par- a- par) y las propias criptomonedas (Bitcoins). Su creador, bajo el seudónimo de Satoshi Nakamoto, pretendía crear un proyecto con un objetivo principal: que las transacciones económicas entre dos personas no necesiten de un organismo central o ningún tipo de intermediario. E...

  14. The Horse Raced Past: Gardenpath Processing in Dynamical Systems

    OpenAIRE

    Graben, Peter beim

    2012-01-01

    I pinpoint an interesting similarity between a recent account to rational parsing and the treatment of sequential decisions problems in a dynamical systems approach. I argue that expectation-driven search heuristics aiming at fast computation resembles a high-risk decision strategy in favor of large transition velocities. Hale's rational parser, combining generalized left-corner parsing with informed $\\mathrm{A}^*$ search to resolve processing conflicts, explains gardenpath effects in natural...

  15. Cross-lingual parser selection for low-resource languages

    DEFF Research Database (Denmark)

    Agic, Zeljko

    2017-01-01

    In multilingual dependency parsing, transferring delexicalized models provides unmatched language coverage and competitive scores, with minimal requirements. Still, selecting the single best parser for any target language poses a challenge. Here, we propose a lean method for parser selection. It ....... It offers top performance, and it does so without disadvantaging the truly low-resource languages. We consistently select appropriate source parsers for our target languages in a realistic cross-lingual parsing experiment....

  16. An integrated architecture for shallow and deep processing

    OpenAIRE

    Crysmann, Berthold; Frank, Anette; Kiefer, Bernd; Müller, Stefan; Neumann, Günter; Piskorski, Jakub; Schäfer, Ulrich; Siegel, Melanie; Uszkoreit, Hans; Xu, Feiyu; Becker, Markus; Krieger, Hans-Ulrich

    2011-01-01

    We present an architecture for the integration of shallow and deep NLP components which is aimed at flexible combination of different language technologies for a range of practical current and future applications. In particular, we describe the integration of a high-level HPSG parsing system with different high-performance shallow components, ranging from named entity recognition to chunk parsing and shallow clause recognition. The NLP components enrich a representation of natural language te...

  17. A Database Design for the Brazilian Air Force Flying Unit Operational Control System.

    Science.gov (United States)

    1984-12-14

    Company, 1980. 23. Pereira Filho, Jorge da Cunha " Banco de Dados Hoje" Dados e Ideias - Brazilian Magazine , 99 : 55-63 (February 1979). 24. Rich...34QUAIS SAO AS CONDICOES DE EMPREGO OPERACIONAL DO MIRAGE 2120 ?" LIKE "WHAT IS THE FORCESTATUS OF MIRAGE 2120?" % Parsed ! % [ Production added to system...4 - QUAIS SAO AS CONDICOES DE EMPREGO OPERACIONAL DO MIRAGE 2120 ? % Parsed ! % (S DEPLOC) = sbbr % (S-TIME) = 1000 % Endurance = 0200 % (SITCODE

  18. Resolving Lexical Ambiguity in a Deterministic Parser

    OpenAIRE

    Milne, Robert W.

    1983-01-01

    This work is an investigation into part of the human sentence parsing mechanism (HSPM), where parsing implies syntactic and non-syntactic analysis. It is hypothesised. that the HSPM consists of at least two processors. We will call the first processor the syntactic processor, and the second will be known as the non-syntactic processor. For normal sentence processing, the two processors are controlled by a 'normal component", whilst when an error occurs, they are controlled by a...

  19. Sociolinguistically Informed Natural Language Processing: Automating Irony Detection

    Science.gov (United States)

    2017-10-23

    interaction feature using the entire training dataset, and repeated this process 100 times to account for variation due to the SGD procedure. Table 6...Levy and Goldberg , 2014). We parsed the ukWaC corpus (Baroni et al., 2009) using the Stanford Dependency Parser v3.5.2 with Stanford Dependencies...bitrary and variable sizes. We pre-trained our own syntactic embeddings fol- lowing (Levy and Goldberg , 2014). We parsed the ukWaC corpus (Baroni et

  20. Parsing and Tagging of Binlingual Dictionary

    National Research Council Canada - National Science Library

    Ma, Huanfeng; Karagol-Ayan, Burcu; Doermann, David S; Oard, Doug; Wang, Jianqiang

    2003-01-01

    Bilingual dictionaries hold great potential as a source of lexical resources for training and testing automated systems for optical character recognition, machine translation, and cross-language information retrieval...

  1. Parsing and Tagging of Bilingual Dictionary

    National Research Council Canada - National Science Library

    Ma, Huanfeng; Karagol-Ayan, Burcu; Doermann, David S; Oard, Doug; Wang, Jianqiang

    2003-01-01

    Bilingual dictionaries hold great potential as a source of lexical resources for training and testing automated systems for optical character recognition, machine translation, and cross-language information retrieval...

  2. Parsing (malicious) pleasures : schadenfreude and gloating

    NARCIS (Netherlands)

    Leach, Colin Wayne; Spears, Russell; Manstead, Antony S. R.

    2015-01-01

    We offer the first empirical comparison of the pleasure in seeing (i.e., schadenfreude) and in causing (i.e., gloating) others' adversity. In Study 1, we asked participants to recall and report on an (individual or group) episode of pleasure that conformed to our formal definition of schadenfreude,

  3. Memory Retrieval in Parsing and Interpretation

    Science.gov (United States)

    Schlueter, Ananda Lila Zoe

    2017-01-01

    This dissertation explores the relationship between the parser and the grammar in error-driven retrieval by examining the mechanism underlying the illusory licensing of subject-verb agreement violations ("agreement attraction"). Previous work motivates a two-stage model of agreement attraction in which the parser predicts the verb's…

  4. Efficient Encoding of Inflection Rules in NLP Systems

    Directory of Open Access Journals (Sweden)

    Péter BARABÁSS

    2012-12-01

    Full Text Available The grammatical parsing unit is a core module in natural language processing engines. This unit determines the grammatical roles of the incoming words and it converts the sentences into semantic models. A special grammar rule in agglutinative languages is the inflection rule. The traditional, automata-based parsers are usually not very effective in the parsing of inflection transformations. The paper presents implementation alternatives and compares them from the viewpoint of time efficiency and accuracy. The prototype system was tested with examples from Hungarian.

  5. Leading-Following in the Context of Rural Nursing.

    Science.gov (United States)

    Stewart Fahs, Pamela

    2017-04-01

    Leading-following, as defined by Parse, is "deliberately innovating with potent engaging in persistently pursuing excellence. It is an indivisible, unpredictable, ever changing cocreation." Parse noted that processes forming the essence of leading-following require commitment, risk-taking, acknowledging ambiguity, respecting others, and vigilance. There is a bond between members of society and nurses-an expectation that nurses will provide care to individuals, families, and communities. Nursing is a profession where the potential to create meaning in healthcare has not yet been fully realized.

  6. Natural-Language Parser for PBEM

    Science.gov (United States)

    James, Mark

    2010-01-01

    A computer program called "Hunter" accepts, as input, a colloquial-English description of a set of policy-based-management rules, and parses that description into a form useable by policy-based enterprise management (PBEM) software. PBEM is a rules-based approach suitable for automating some management tasks. PBEM simplifies the management of a given enterprise through establishment of policies addressing situations that are likely to occur. Hunter was developed to have a unique capability to extract the intended meaning instead of focusing on parsing the exact ways in which individual words are used.

  7. Nonschematic drawing recognition: a new approach based on attributed graph grammar with flexible embedding

    Science.gov (United States)

    Lee, Kyu J.; Kunii, T. L.; Noma, T.

    1993-01-01

    In this paper, we propose a syntactic pattern recognition method for non-schematic drawings, based on a new attributed graph grammar with flexible embedding. In our graph grammar, the embedding rule permits the nodes of a guest graph to be arbitrarily connected with the nodes of a host graph. The ambiguity caused by this flexible embedding is controlled with the evaluation of synthesized attributes and the check of context sensitivity. To integrate parsing with the synthesized attribute evaluation and the context sensitivity check, we also develop a bottom up parsing algorithm.

  8. Leading-Following: Guiding Care in the Community.

    Science.gov (United States)

    Ortiz, Mario R

    2018-04-01

    There are many ways nurses may be leaders and serve others in the community by utilizing a unique knowledge base that is grounded in the extant nursing theories and models. Leading and serving others with nursing knowledge demarks the contributions nurses make in all situations. The purpose of this paper is to illuminate how Parse's leading-following model and the Rainbow PRISM model (RPM) guide healthcare services in academic nursing health and wellness centers that serve the needs of all persons living in the community. Also, concepts within the RPM are discussed to align with Parse's current refinements of the humanbecoming paradigm.

  9. Shallow parsing with long short-term memory

    NARCIS (Netherlands)

    Hammerton, J; Chen, SH; Cheng, HD; Chiu, DKY; Das, S; Duro, R; Kerre, EE; Leong, HV; Li, Q; Lu, M; Romay, MG; Ventura, D; Wu, J

    2003-01-01

    Applying Artificial Neural Networks (ANNs) to language learning has been an active area of research in connectionism. However much of this work has involved small and/or artificially created data sets, whilst other approaches to language learning are now routinely applied to large real-world

  10. Subdomain sensitive statistical parsing using raw corpora

    NARCIS (Netherlands)

    Plank, B.; Sima'an, K.

    2008-01-01

    Modern statistical parsers are trained on large annotated corpora (treebanks). These treebanks usually consist of sentences addressing different subdomains (e.g. sports, politics, music), which implies that the statistics gathered by current statistical parsers are mixtures of subdomains of language

  11. Breaking the Resource Bottleneck for Multilingual Parsing

    National Research Council Canada - National Science Library

    Hwa, Rebecca; Resnik, Philip; Weinberg, Amy

    2005-01-01

    ...-quality English resources. We present a large-scale experiment showing that Chinese dependency trees can be induced by using an English parser, a word alignment package, and a large corpus of sentence-aligned bilingual text...

  12. A Sound Abstraction of the Parsing Problem

    DEFF Research Database (Denmark)

    Mödersheim, Sebastian Alexander; Katsoris, Georgios

    2014-01-01

    In formal verification, cryptographic messages are often represented by algebraic terms. This abstracts not only from the intricate details of the real cryptography, but also from the details of the non-cryptographic aspects: the actual formatting and structuring of messages. We introduce a new a...

  13. LR-parsing of Extended Context-free Grammars

    DEFF Research Database (Denmark)

    Madsen, Ole Lehrmann; Kristensen, Bent Bruun

    1976-01-01

    To improve the readability of a grammar it is common to use extended context free grammars (ECFGs) which are context free grammars (CFGs) extended with the repetition operator (*), the alternation operator (¦) and parentheses to express the right hand sides of the productions. The topic treated h...

  14. Irish treebanking and parsing: a preliminary evaluation

    OpenAIRE

    Lynn, Teresa; Cetinoglu, Ozlem; Foster, Jennifer; Uí Dhonnchadha, Elaine; Dras, Mark; van Genabith, Josef

    2012-01-01

    Language resources are essential for linguistic research and the development of NLP applications. Low- density languages, such as Irish, therefore lack significant research in this area. This paper describes the early stages in the development of new language resources for Irish – namely the first Irish dependency treebank and the first Irish statistical dependency parser. We present the methodology behind building our new treebank and the steps we take to leverage upon the few existing resou...

  15. A Knowledge-Based Approach to Robust Parsing

    NARCIS (Netherlands)

    Oltmans, J.A.

    2000-01-01

    The research presented in this thesis describes the design, implementation and evaluation of a natural-language processing system that is used as part of an information retrieval system. Specifically, I focus on the development of a system that performs robust syntactic analysis of scientific texts

  16. Iguana: a practical data-dependent parsing framework

    NARCIS (Netherlands)

    A. Afroozeh (Ali); A. Izmaylova (Anastasia)

    2016-01-01

    textabstractData-dependent grammars extend context-free grammars with arbitrary computation, variable binding, and constraints. These features provide the user with the freedom and power to express syntactic constructs outside the realm of context-free grammars, e.g., indentation rules in Haskell

  17. Curating Research: e-Merging New Roles and Responsibilities in the European Landscape (Conference Report

    Directory of Open Access Journals (Sweden)

    Inge Angevaare

    2009-11-01

    Full Text Available On 17 April 2009 LIBER, the Koninklijke Bibliotheek (KB and the Netherlands Coalition for Digital Preservation (NCDD co-organised LIBER’s first workshop on digital curation/preservation, providing an introduction to the theme of digital curation and different perspectives on the roles research libraries can and/or must play in keeping the digital records of science safe for future generations. Speakers included Eileen Fenton (Portico, Dale Peters (DRIVER, Maria Heijne (TU Delft Library, Jeffrey van der Hoeven (KB, PARSE.insight and ninety workshop attendees. The paper includes the conference report and some results of the PARSE.insight online survey into digital preservation practices of LIBER libraries.

  18. Lempel-Ziv Compression in a Sliding Window

    DEFF Research Database (Denmark)

    Bille, Philip; Cording, Patrick Hagge; Fischer, Johannes

    2017-01-01

    result, we combine a simple modification and augmentation of the suffix tree with periodicity properties of sliding windows. We also apply this new technique to obtain an algorithm for the approximate rightmost LZ77 problem that uses O(n(log z + loglogn)) time and O(n) space and produces a (1 + ϵ......We present new algorithms for the sliding window Lempel-Ziv (LZ77) problem and the approximate rightmost LZ77 parsing problem. Our main result is a new and surprisingly simple algorithm that computes the sliding window LZ77 parse in O(w) space and either O(n) expected time or O(n log log w + z log...

  19. Analysis of the potential geochemical reactions in the Enceladus' hydrothermal environment

    Science.gov (United States)

    Ramirez-Cabañas, A. K.; Flandes, A.

    2017-12-01

    Enceladus is the sixth largest moon of Saturn and differs from its other moons, because of its cryovolcanic geysers that emanate from its south pole. The instruments of the Cassini spacecraft reveal different compounds in the gases and the dust of the geysers, such as salts (sodium chloride, sodium bicarbonate and/or sodium carbonate), as well as silica traces (Postberg et al., 2008, 2009) that could be the result of a hydrothermal environment (Hsu et al., 2014, Sekine et al., 2014). By means of a thermodynamic analysis, we propose and evaluate potential geochemical reactions that could happen from the interaction between the nucleus surface and the inner ocean of Enceladus. These reactions may well lead to the origin of the compounds found in the geysers. From this analysis, we propose that, at least, two minerals must be present in the condritic nucleus of Enceladus: olivines (fayalite and fosterite) and feldspar (orthoclase and albite). Subsequently, taking as reference the hydrothermal processes that take place on Earth, we propose the different stages of a potential hydrothermal scenario for Enceladus.

  20. Poor tolerance of beta-blockers by elderly patients with heart failure

    Directory of Open Access Journals (Sweden)

    Satoshi Yanagisawa

    2010-11-01

    Full Text Available Satoshi Yanagisawa, Noriyuki Suzuki, Toshikazu TanakaDepartment of Cardiology, Okazaki City Hospital, Aichi, JapanAbstract: Despite the well-understood importance of beta-blocker therapy in heart failure, it is sometimes not possible to use beta-blockers in elderly patients due to poor tolerance. In this report, we describe the case of an 83-year-old patient with severe systolic heart failure complicated by aortic valve stenosis and atrial fibrillation. A simple therapeutic approach involving discontinuation of beta-blockers remarkably alleviated the symptoms such as left ventricular ejection fraction, and improved the chest radiography and laboratory findings; further, atrial fibrillation converted to sinus rhythm. It is important to carefully administer beta-blocker therapy to elderly patients with heart failure, especially after considering cardiac output.Keywords: elderly, octogenarians, beta-blockers, heart failure

  1. Development of 3D browsing and interactive web system

    Science.gov (United States)

    Shi, Xiaonan; Fu, Jian; Jin, Chaolin

    2017-09-01

    In the current market, users need to download specific software or plug-ins to browse the 3D model, and browsing the system may be unstable, and it cannot be 3D model interaction issues In order to solve this problem, this paper presents a solution to the interactive browsing of the model in the server-side parsing model, and when the system is applied, the user only needs to input the system URL and upload the 3D model file to operate the browsing The server real-time parsing 3D model, the interactive response speed, these completely follows the user to walk the minimalist idea, and solves the current market block 3D content development question.

  2. [Pilot study of domain-specific terminology adaptation for morphological analysis: research on unknown terms in national examination documents of radiological technologists].

    Science.gov (United States)

    Tsuji, Shintarou; Nishimoto, Naoki; Ogasawara, Katsuhiko

    2008-07-20

    Although large medical texts are stored in electronic format, they are seldom reused because of the difficulty of processing narrative texts by computer. Morphological analysis is a key technology for extracting medical terms correctly and automatically. This process parses a sentence into its smallest unit, the morpheme. Phrases consisting of two or more technical terms, however, cause morphological analysis software to fail in parsing the sentence and output unprocessed terms as "unknown words." The purpose of this study was to reduce the number of unknown words in medical narrative text processing. The results of parsing the text with additional dictionaries were compared with the analysis of the number of unknown words in the national examination for radiologists. The ratio of unknown words was reduced 1.0% to 0.36% by adding terminologies of radiological technology, MeSH, and ICD-10 labels. The terminology of radiological technology was the most effective resource, being reduced by 0.62%. This result clearly showed the necessity of additional dictionary selection and trends in unknown words. The potential for this investigation is to make available a large body of clinical information that would otherwise be inaccessible for applications other than manual health care review by personnel.

  3. MMTF-An efficient file format for the transmission, visualization, and analysis of macromolecular structures.

    Directory of Open Access Journals (Sweden)

    Anthony R Bradley

    2017-06-01

    Full Text Available Recent advances in experimental techniques have led to a rapid growth in complexity, size, and number of macromolecular structures that are made available through the Protein Data Bank. This creates a challenge for macromolecular visualization and analysis. Macromolecular structure files, such as PDB or PDBx/mmCIF files can be slow to transfer, parse, and hard to incorporate into third-party software tools. Here, we present a new binary and compressed data representation, the MacroMolecular Transmission Format, MMTF, as well as software implementations in several languages that have been developed around it, which address these issues. We describe the new format and its APIs and demonstrate that it is several times faster to parse, and about a quarter of the file size of the current standard format, PDBx/mmCIF. As a consequence of the new data representation, it is now possible to visualize structures with millions of atoms in a web browser, keep the whole PDB archive in memory or parse it within few minutes on average computers, which opens up a new way of thinking how to design and implement efficient algorithms in structural bioinformatics. The PDB archive is available in MMTF file format through web services and data that are updated on a weekly basis.

  4. Understanding of Navy Technical Language via Statistical Parsing

    National Research Council Canada - National Science Library

    Rowe, Neil C

    2004-01-01

    ... (both on word senses and word-sense pairs) from a training corpus with a statistical parser. Innovations of our approach are in statistical inheritance of binary co-occurrence probabilities and in weighting...

  5. A Tandem Semantic Interpreter for Incremental Parse Selection

    Science.gov (United States)

    1990-09-28

    syntactic role to semantic role. An exam - ple from Fillmore [10] is the sentence I copied that letter, which can be uttered when point- ing either to...person. We want the word fiddle to have the sort predicate violin as its lexical interpretation, how- ever, notthing. Thus, for Ross went for his fiddle...to receive an interpretation, a sort hierarchy is needed to establish that all violins are things. A well-structured sort hierarchy allows newly added

  6. Parsing with Regular Expressions & Extensions to Kleene Algebra

    DEFF Research Database (Denmark)

    Grathwohl, Niels Bjørn Bugge

    . In the second part of this thesis, we study two extensions to Kleene algebra. Chomsky algebra is an algebra with a structure similar to Kleene algebra, but with a generalized mu-operator for recursion instead of the Kleene star. We show that the axioms of idempotent semirings along with continuity of the mu......-operator completely axiomatize the equational theory of the context-free languages. KAT+B! is an extension to Kleene algebra with tests (KAT) that adds mutable state. We describe a test algebra B! for mutable tests and give a commutative coproduct between KATs. Combining the axioms of B! with those of KAT and some...

  7. Document Image Parsing and Understanding using Neuromorphic Architecture

    Science.gov (United States)

    2015-03-01

    processing speed at different layers. In the pattern matching layer, the computing power of multicore processors is explored to reduce the processing...developed to reduce the processing speed at different layers. In the pattern matching layer, the computing power of multicore processors is explored... cortex where the complex data is reduced to abstract representations. The abstract representation is compared to stored patterns in massively parallel

  8. Parsing the Gulf between Africans and African Americans

    Directory of Open Access Journals (Sweden)

    Ashly Nsangou

    2018-02-01

    Full Text Available The rise in African immigrants to the US provides an opportunity to assess relations between Africans and African Americans in college. An online survey of 322 current and recently-graduated college students (including 45 Africans, 160 African Americans, and 117 whites assessed respondents’ experiences of racism in US high schools and colleges. Semi-structured interviews of 30 students (10 African, 10 African American and 10 white students supplemented these data. Even within a sociopolitical context of more visible racial intolerance, Black intra-racial cohesion was absent. Although more first- and second-generation Africans (73% felt that they had been judged while living in the US compared to African Americans (34% or whites (20%, for 70–80% of respondents, this had occurred only in high school. Despite experiencing these judgments, Africans’ identity related more to their focus on education than their race, reflected in a higher proportion who felt intense family pressure to attend college (65% compared to African Americans (37% and whites (39%. Interview data confirmed previous reports in the literature that African Americans lack a sense of connection to Africans, attributed to Africans’ purported sense of superiority and disregard for African Americans’ ongoing struggle to end oppression. These mixed-methods data suggest that intermingling in the college environment has not resulted in first- and second-generation Africans and African Americans sharing a common in-group, race-based identity. We discuss the implications of overlooking ethnic distinctions due to presumptions of racial homogeneity that deprive Black individuals of their uniqueness.

  9. Collated, parsed, OCO-2 Calibration Data V6

    Data.gov (United States)

    National Aeronautics and Space Administration — The Orbiting Carbon Observatory is the first NASA mission designed to collect space-based measurements of atmospheric carbon dioxide with the precision, resolution,...

  10. Collated, parsed, OCO-2 Calibration Data V7

    Data.gov (United States)

    National Aeronautics and Space Administration — The Orbiting Carbon Observatory is the first NASA mission designed to collect space-based measurements of atmospheric carbon dioxide with the precision, resolution,...

  11. Parsing regulatory DNA: General tasks, techniques, and the ...

    Indian Academy of Sciences (India)

    PRAKASH KUMAR

    plays a more important role than had been appreciated (see. Pearson 2006 for a ... and for a practical guide on how to use PhyloGibbs, there is an upcoming review .... overall, in user-determined “windows"; it, too, can take advantage of one ...

  12. Cine club

    CERN Multimedia

    Cine club

    2016-01-01

    Wednesday 18 May 2016 at 20:00 CERN Council Chamber Kikujiro   Directed by Takeshi Kitano Japan, 1999, 121 minutes Brash, loudmouthed and opportunistic, Kikujiro hardly seems the ideal companion for little Masao who is determined to travel long distances to see the mother he has never met. Their excursion to the cycle races is the first of a series of adventures for the unlikely pair which soon turns out to be a whimsical journey of laughter and tears with a wide array of surprises and odd ball characters to meet along the way. Original version Japanese; English subtitles Wednesday 25 May 2016 at 20:00 CERN Council Chamber Millennium Actress   Directed by Satoshi Kon  Japan, 2001, 87 minutes A TV interviewer and his cameraman meet a former actress and travel through her memories and career. Original version Japanese; English subtitles

  13. Applying Semantic Web Concepts to Support Net-Centric Warfare Using the Tactical Assessment Markup Language (TAML)

    National Research Council Canada - National Science Library

    Childers, Candace M

    2006-01-01

    .... The current approach to automating data processing is to hard-code programs to parse particular data formats, but this approach is not flexible enough to handle the constantly changing data world...

  14. Metacoder: An R package for visualization and manipulation of community taxonomic diversity data.

    Directory of Open Access Journals (Sweden)

    Zachary S L Foster

    2017-02-01

    Full Text Available Community-level data, the type generated by an increasing number of metabarcoding studies, is often graphed as stacked bar charts or pie graphs that use color to represent taxa. These graph types do not convey the hierarchical structure of taxonomic classifications and are limited by the use of color for categories. As an alternative, we developed metacoder, an R package for easily parsing, manipulating, and graphing publication-ready plots of hierarchical data. Metacoder includes a dynamic and flexible function that can parse most text-based formats that contain taxonomic classifications, taxon names, taxon identifiers, or sequence identifiers. Metacoder can then subset, sample, and order this parsed data using a set of intuitive functions that take into account the hierarchical nature of the data. Finally, an extremely flexible plotting function enables quantitative representation of up to 4 arbitrary statistics simultaneously in a tree format by mapping statistics to the color and size of tree nodes and edges. Metacoder also allows exploration of barcode primer bias by integrating functions to run digital PCR. Although it has been designed for data from metabarcoding research, metacoder can easily be applied to any data that has a hierarchical component such as gene ontology or geographic location data. Our package complements currently available tools for community analysis and is provided open source with an extensive online user manual.

  15. Recursion complexity in cognition

    CERN Document Server

    Roeper, Thomas

    2014-01-01

    This volume focuses on recursion, highlighting its central role in modern science. It reveals a host of new theoretical arguments, philosophical perspectives, formal representations and empirical evidence from parsing, acquisition and computer models.

  16. SU-E-T-473: A Patient-Specific QC Paradigm Based On Trajectory Log Files and DICOM Plan Files

    International Nuclear Information System (INIS)

    DeMarco, J; McCloskey, S; Low, D; Moran, J

    2014-01-01

    Purpose: To evaluate a remote QC tool for monitoring treatment machine parameters and treatment workflow. Methods: The Varian TrueBeamTM linear accelerator is a digital machine that records machine axis parameters and MLC leaf positions as a function of delivered monitor unit or control point. This information is saved to a binary trajectory log file for every treatment or imaging field in the patient treatment session. A MATLAB analysis routine was developed to parse the trajectory log files for a given patient, compare the expected versus actual machine and MLC positions as well as perform a cross-comparison with the DICOM-RT plan file exported from the treatment planning system. The parsing routine sorts the trajectory log files based on the time and date stamp and generates a sequential report file listing treatment parameters and provides a match relative to the DICOM-RT plan file. Results: The trajectory log parsing-routine was compared against a standard record and verify listing for patients undergoing initial IMRT dosimetry verification and weekly and final chart QC. The complete treatment course was independently verified for 10 patients of varying treatment site and a total of 1267 treatment fields were evaluated including pre-treatment imaging fields where applicable. In the context of IMRT plan verification, eight prostate SBRT plans with 4-arcs per plan were evaluated based on expected versus actual machine axis parameters. The average value for the maximum RMS MLC error was 0.067±0.001mm and 0.066±0.002mm for leaf bank A and B respectively. Conclusion: A real-time QC analysis program was tested using trajectory log files and DICOM-RT plan files. The parsing routine is efficient and able to evaluate all relevant machine axis parameters during a patient treatment course including MLC leaf positions and table positions at time of image acquisition and during treatment

  17. Integration of Dakota into the NEAMS Workbench

    Energy Technology Data Exchange (ETDEWEB)

    Swiler, Laura Painton [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Lefebvre, Robert A. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Langley, Brandon R. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Thompson, Adam B. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-07-01

    This report summarizes a NEAMS (Nuclear Energy Advanced Modeling and Simulation) project focused on integrating Dakota into the NEAMS Workbench. The NEAMS Workbench, developed at Oak Ridge National Laboratory, is a new software framework that provides a graphical user interface, input file creation, parsing, validation, job execution, workflow management, and output processing for a variety of nuclear codes. Dakota is a tool developed at Sandia National Laboratories that provides a suite of uncertainty quantification and optimization algorithms. Providing Dakota within the NEAMS Workbench allows users of nuclear simulation codes to perform uncertainty and optimization studies on their nuclear codes from within a common, integrated environment. Details of the integration and parsing are provided, along with an example of Dakota running a sampling study on the fuels performance code, BISON, from within the NEAMS Workbench.

  18. Punctuation and Implicit Prosody in Silent Reading: An ERP Study Investigating English Garden-Path Sentences.

    Science.gov (United States)

    Drury, John E; Baum, Shari R; Valeriote, Hope; Steinhauer, Karsten

    2016-01-01

    This study presents the first two ERP reading studies of comma-induced effects of covert (implicit) prosody on syntactic parsing decisions in English. The first experiment used a balanced 2 × 2 design in which the presence/absence of commas determined plausibility (e.g., John, said Mary, was the nicest boy at the party vs. John said Mary was the nicest boy at the party ). The second reading experiment replicated a previous auditory study investigating the role of overt prosodic boundaries in closure ambiguities (Pauker et al., 2011). In both experiments, commas reliably elicited CPS components and generally played a dominant role in determining parsing decisions in the face of input ambiguity. The combined set of findings provides further evidence supporting the claim that mechanisms subserving speech processing play an active role during silent reading.

  19. Arhitektuur ja ehitus maavärinajärgses Jaapanis = The housing trend in post 3.11 Japan / Satoshi Joshua Ogawa

    Index Scriptorium Estoniae

    Ogawa, Satoshi Joshua

    2012-01-01

    Vaadeldakse looduskatastroofist toibuva Jaapani ühiskonna teisenenud arusaama kodust, kinnisvaraturu muutunud eelistusi. Taastamistöödest, elamuehitusest, arhitektide kaasatusest, viimase kümnendi arhitektuurist

  20. Arcus v. 1.0

    Energy Technology Data Exchange (ETDEWEB)

    2018-01-10

    Arcus, developed by Sandia National Laboratories, is a library for calculating, parsing, formatting, converting and comparing both IPv4 and IPv6 addresses and subnets. It accounts for 128-bit numbers on 32-bit platforms.

  1. American Telemedicine Association

    Science.gov (United States)

    ... trim( scriptEl.html() ); if ( !slides ) return; // is it json array? if ( /^(\\[)/.test( slides ) ) { try { slides = $.parseJSON( slides ); } ... our ATA Action Center to urge legislators to support policies that improve healthcare. Take action Get Accredited ...

  2. Robert W Floyd

    Indian Academy of Sciences (India)

    scientists - Zohar Manna, an authority on the verification of software systems, Robert. Tarjan ... He began as a self taught computer operator, and then went ... important subfields of computer science: the theory of parsing, the semantics of.

  3. Modeling the Arden Syntax for medical decisions in XML.

    Science.gov (United States)

    Kim, Sukil; Haug, Peter J; Rocha, Roberto A; Choi, Inyoung

    2008-10-01

    A new model expressing Arden Syntax with the eXtensible Markup Language (XML) was developed to increase its portability. Every example was manually parsed and reviewed until the schema and the style sheet were considered to be optimized. When the first schema was finished, several MLMs in Arden Syntax Markup Language (ArdenML) were validated against the schema. They were then transformed to HTML formats with the style sheet, during which they were compared to the original text version of their own MLM. When faults were found in the transformed MLM, the schema and/or style sheet was fixed. This cycle continued until all the examples were encoded into XML documents. The original MLMs were encoded in XML according to the proposed XML schema and reverse-parsed MLMs in ArdenML were checked using a public domain Arden Syntax checker. Two hundred seventy seven examples of MLMs were successfully transformed into XML documents using the model, and the reverse-parse yielded the original text version of MLMs. Two hundred sixty five of the 277 MLMs showed the same error patterns before and after transformation, and all 11 errors related to statement structure were resolved in XML version. The model uses two syntax checking mechanisms, first an XML validation process, and second, a syntax check using an XSL style sheet. Now that we have a schema for ArdenML, we can also begin the development of style sheets for transformation ArdenML into other languages.

  4. VMS forms Output Tables

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — These output tables contain parsed and format validated data from the various VMS forms that are sent from any given vessel, while at sea, from the VMS devices on...

  5. Simplex and Multiplex Stratification in ASD and ADHD Families: A Promising Approach for Identifying Overlapping and Unique Underpinnings of ASD and ADHD?

    NARCIS (Netherlands)

    Oerlemans, A.M.; Hartman, C.A.; Bruijn, Y.G.E. de; Steijn, D.J. Van; Franke, B.; Buitelaar, J.K.; Rommelse, N.N.J.

    2015-01-01

    Autism spectrum disorders (ASD) and attention-deficit/hyperactivity disorder (ADHD) are highly heterogeneous neuropsychiatric disorders, that frequently co-occur. This study examined whether stratification into single-incidence (SPX) and multi-incidence (MPX) is helpful in (a) parsing heterogeneity

  6. Simplex and Multiplex Stratification in ASD and ADHD Families : A Promising Approach for Identifying Overlapping and Unique Underpinnings of ASD and ADHD?

    NARCIS (Netherlands)

    Oerlemans, Anoek M.; Hartman, Catharina A.; De Bruijn, Yvette G. E.; Van Steijn, Daphne J.; Franke, Barbara; Buitelaar, Jan K.; Rommelse, Nanda N. J.

    Autism spectrum disorders (ASD) and attention-deficit/hyperactivity disorder (ADHD) are highly heterogeneous neuropsychiatric disorders, that frequently co-occur. This study examined whether stratification into single-incidence (SPX) and multi-incidence (MPX) is helpful in (a) parsing heterogeneity

  7. Automating 3D reconstruction using a probabilistic grammar

    Science.gov (United States)

    Xiong, Hanwei; Xu, Jun; Xu, Chenxi; Pan, Ming

    2015-10-01

    3D reconstruction of objects from point clouds with a laser scanner is still a laborious task in many applications. Automating 3D process is an ongoing research topic and suffers from the complex structure of the data. The main difficulty is due to lack of knowledge of real world objects structure. In this paper, we accumulate such structure knowledge by a probabilistic grammar learned from examples in the same category. The rules of the grammar capture compositional structures at different levels, and a feature dependent probability function is attached for every rule. The learned grammar can be used to parse new 3D point clouds, organize segment patches in a hierarchal way, and assign them meaningful labels. The parsed semantics can be used to guide the reconstruction algorithms automatically. Some examples are given to explain the method.

  8. KLASIFIKASI WEBSITE MENGGUNAKAN ALGORITMA MULTILAYER PERCEPTRON

    Directory of Open Access Journals (Sweden)

    Nyoman Purnama

    2014-12-01

    Full Text Available Sistem klasifikasi merupakan proses temu balik informasi yang sangat bergantung dari elemen-elemen penyusunnya.Sistem ini banyak digunakan untuk mengatasi permasalahan segmentasi data. Klasifikasi dapat digunakan pada website sebagaimetode untuk mengelompokkan website. Website merupakan salah satu data yang memiliki informasi yang beraneka-ragam,sehingga pengelompokan data ini penting untuk diteliti. Sistem klasifikasi dimulai dengan melakukan proses pengumpulaninformasi dari halaman website (parsing dan untuk setiap hasil parsing dilakukan proses penghapusan kata henti, stemming,feature selection dengan tf-idf. Hasil dari proses ini berupa fitur yang menjadi inputan algoritma Multilayer Perceptron. Dalamalgoritma ini terjadi proses pembelajaran terhadap pola input masukan dan pembuatan bobot pelatihan. Bobot ini akandigunakan pada proses klasifikasi. Hasil dari penelitian menunjukkan bahwa algoritma Multilayer Perceptron dapatmenghasilkan klasifikasi website dengan akurasi yang bagus. Hal ini dibuktikan dengan beberapa tahapan penelitian yangberbeda dan didapatkan nilai akurasi rata-rata diatas 70%.

  9. The Modeling and Complexity of Dynamical Systems by Means of Computation and Information Theories

    Directory of Open Access Journals (Sweden)

    Robert Logozar

    2011-12-01

    Full Text Available We present the modeling of dynamical systems and finding of their complexity indicators by the use of concepts from computation and information theories, within the framework of J. P. Crutchfield's theory of  ε-machines. A short formal outline of the  ε-machines is given. In this approach, dynamical systems are analyzed directly from the time series that is received from a properly adjusted measuring instrument. The binary strings are parsed through the parse tree, within which morphologically and probabilistically unique subtrees or morphs are recognized as system states. The outline and precise interrelation of the information-theoretic entropies and complexities emanating from the model is given. The paper serves also as a theoretical foundation for the future presentation of the DSA program that implements the  ε-machines modeling up to the stochastic finite automata level.

  10. Punctuation and Implicit Prosody in Silent Reading: An ERP Study Investigating English Garden-Path Sentences

    Science.gov (United States)

    Drury, John E.; Baum, Shari R.; Valeriote, Hope; Steinhauer, Karsten

    2016-01-01

    This study presents the first two ERP reading studies of comma-induced effects of covert (implicit) prosody on syntactic parsing decisions in English. The first experiment used a balanced 2 × 2 design in which the presence/absence of commas determined plausibility (e.g., John, said Mary, was the nicest boy at the party vs. John said Mary was the nicest boy at the party). The second reading experiment replicated a previous auditory study investigating the role of overt prosodic boundaries in closure ambiguities (Pauker et al., 2011). In both experiments, commas reliably elicited CPS components and generally played a dominant role in determining parsing decisions in the face of input ambiguity. The combined set of findings provides further evidence supporting the claim that mechanisms subserving speech processing play an active role during silent reading. PMID:27695428

  11. Punctuation and implicit prosody in silent reading: An ERP study investigating English garden-path sentences

    Directory of Open Access Journals (Sweden)

    John E. Drury

    2016-09-01

    Full Text Available This study presents the first two ERP reading studies of comma-induced effects of covert (implicit prosody on syntactic parsing decisions in English. The first experiment used a balanced 2 x 2 design in which the presence/absence of commas determined plausibility (e.g., John, said Mary, was the nicest boy at the party versus John said Mary was the nicest boy at the party. The second reading experiment replicated a previous auditory study investigating the role of overt prosodic boundaries in closure ambiguities (Pauker et al., 2011. In both experiments, commas reliably elicited CPS components and generally played a dominant role in determining parsing decisions in the face of input ambiguity. The combined set of findings provides further evidence supporting the claim that mechanisms subserving speech processing play an active role during silent reading.

  12. Practical, general parser combinators

    NARCIS (Netherlands)

    A. Izmaylova (Anastasia); A. Afroozeh (Ali); T. van der Storm (Tijs)

    2016-01-01

    textabstractParser combinators are a popular approach to parsing where contextfree grammars are represented as executable code. However, conventional parser combinators do not support left recursion, and can have worst-case exponential runtime. These limitations hinder the expressivity and

  13. Parser Adaptation for Social Media by Integrating Normalization

    NARCIS (Netherlands)

    van der Goot, Rob; van Noord, Gerardus

    This work explores normalization for parser adaptation. Traditionally, normalization is used as separate pre-processing step. We show that integrating the normalization model into the parsing algorithm is beneficial. This way, multiple normalization candidates can be leveraged, which improves

  14. Recognition of facial emotion and affective prosody in children with ASD ( plus ADHD) and their unaffected siblings

    NARCIS (Netherlands)

    Oerlemans, Anoek M.; van der Meer, Jolanda M. J.; van Steijn, Daphne J.; de Ruiter, Saskia W.; de Bruijn, Yvette G. E.; de Sonneville, Leo M. J.; Buitelaar, Jan K.; Rommelse, Nanda N. J.

    Autism is a highly heritable and clinically heterogeneous neuropsychiatric disorder that frequently co-occurs with other psychopathologies, such as attention-deficit/hyperactivity disorder (ADHD). An approach to parse heterogeneity is by forming more homogeneous subgroups of autism spectrum disorder

  15. Safe specification of operator precedence rules

    NARCIS (Netherlands)

    Afroozeh, A.; Brand, van den M.G.J.; Johnstone, A.; Scott, E.; Vinju, J.J.; Erwig, M.; Paige, R.F.; Van Wyk, E.

    2013-01-01

    In this paper we present an approach to specifying operator precedence based on declarative disambiguation constructs and an implementation mechanism based on grammar rewriting. We identify a problem with existing generalized context-free parsing and disambiguation technology: generating a correct

  16. Safe Specification of Operator Precedence Rules

    NARCIS (Netherlands)

    A. Afroozeh (Ali); M.G.J. van den Brand (Mark); A. Johnstone; E. Scott; J.J. Vinju (Jurgen); K. Czarnecki; G. Hedin

    2013-01-01

    htmlabstractIn this paper we present an approach to specifying opera- tor precedence based on declarative disambiguation constructs and an implementation mechanism based on grammar rewriting. We identify a problem with existing generalized context-free parsing and disambigua- tion technology:

  17. Forest FIRE and FIRE wood : tools for tree automata and tree algorithms

    NARCIS (Netherlands)

    Cleophas, L.G.W.A.; Piskorski, J.; Watson, B.W.; Yli-Jyrä, A.

    2009-01-01

    Pattern matching, acceptance, and parsing algorithms on node-labeled, ordered, ranked trees ('tree algorithms') are important for applications such as instruction selection and tree transformation/term rewriting. Many such algorithms have been developed. They often are based on results from such

  18. Novel 3D Stereoscopic Film Compression using Multiwavelet

    African Journals Online (AJOL)

    pc

    2018-03-05

    Mar 5, 2018 ... this works based on observing the parsing evaluation of the ... enhance the performance of the quality of the 3D film is .... enhance our understanding of compression the video fra .... rt stereoscopy, for example in future can.

  19. Supporting intra-group social metacognitive activities with technology: A grammar learning game

    NARCIS (Netherlands)

    Molenaar, I.; Horvers, A.; Desain, P.W.M.

    2017-01-01

    This study investigates the effects of a technology enhanced collaborative grammar learning activity on students sentence parsing and formulation. These types of collaborative learning activities for grammar education are expected to support more effective learning. Yet, effective intra-group social

  20. Collated, parsed, OCO-2 Calibration Data V6r

    Data.gov (United States)

    National Aeronautics and Space Administration — The Orbiting Carbon Observatory is the first NASA mission designed to collect space-based measurements of atmospheric carbon dioxide with the precision, resolution,...

  1. Differentially regulated NADPH: cytochrome p450 oxidoreductases in parsely

    International Nuclear Information System (INIS)

    Koopmann, E.; Hahlbrock, K.

    1997-01-01

    Two NADPH:cytochrome P450 oxidoreductases (CPRs) from parsley (Petroselinum crispum) were cloned, and the complete proteins were expressed and functionally identified in yeast. The two enzymes, designated CPR1 and CPR2, are 80% identical in amino acid sequence with one another and about 75% identical with CPRs from several other plant species. The mRNA accumulation patterns for CPR1 and CPR2 in fungal elicitor-treated or UV-irradiated cultured parsley cells and in developing or infected parsley plants were compared with those for cinnamate 4-hydroxylase (C4H), one of the most abundant CPR-dependent P450 enzymes in plants. All treatments strongly induced the mRNAs for C4H and CPR1 but not for CPR2, suggesting distinct metabolic roles of CPR1 and CPR2 and a functional relationship between CPR1 and C4H

  2. ClinVar data parsing [version 1; referees: 2 approved

    Directory of Open Access Journals (Sweden)

    Xiaolei Zhang

    2017-05-01

    Full Text Available This software repository provides a pipeline for converting raw ClinVar data files into analysis-friendly tab-delimited tables, and also provides these tables for the most recent ClinVar release. Separate tables are generated for genome builds GRCh37 and GRCh38 as well as for mono-allelic variants and complex multi-allelic variants. Additionally, the tables are augmented with allele frequencies from the ExAC and gnomAD datasets as these are often consulted when analyzing ClinVar variants. Overall, this work provides ClinVar data in a format that is easier to work with and can be directly loaded into a variety of popular analysis tools such as R, python pandas, and SQL databases.

  3. A Sound Abstraction of the Parsing Problem (Extended Version)

    DEFF Research Database (Denmark)

    Mödersheim, Sebastian Alexander; Katsoris, Georgios

    algebraic model to include these details and define a small, simple language to precisely describe message formats. We support fixed-length fields, variablelength fields with offsets, tags, and encodings into smaller alphabets like Base64, thereby covering both classical formats as in TLS and modern XML...

  4. Autonomous parsing of behavior in a multi-agent setting

    NARCIS (Netherlands)

    Vanderelst, D.; Barakova, E.I.; Rutkowski, L.; Tadeusiewicz, R.

    2008-01-01

    Imitation learning is a promising route to instruct robotic multi-agent systems. However, imitating agents should be able to decide autonomously what behavior, observed in others, is interesting to copy. Here we investigate whether a simple recurrent network (Elman Net) can be used to extract

  5. Transparent parsing : Head-driven processing of verb-final structures

    NARCIS (Netherlands)

    Mulders, I.C.M.C.

    2002-01-01

    The conceptual guideline underlying this study is that the goal of processing theory should be to construct a transparent parser. A transparent parser is a parser which employs only properties and relations that are available in the grammar, without resorting to processing-specific notions. Under

  6. Seabed mapping and characterization of sediment variability using the usSEABED data base

    Science.gov (United States)

    Goff, J.A.; Jenkins, C.J.; Jeffress, Williams S.

    2008-01-01

    We present a methodology for statistical analysis of randomly located marine sediment point data, and apply it to the US continental shelf portions of usSEABED mean grain size records. The usSEABED database, like many modern, large environmental datasets, is heterogeneous and interdisciplinary. We statistically test the database as a source of mean grain size data, and from it provide a first examination of regional seafloor sediment variability across the entire US continental shelf. Data derived from laboratory analyses ("extracted") and from word-based descriptions ("parsed") are treated separately, and they are compared statistically and deterministically. Data records are selected for spatial analysis by their location within sample regions: polygonal areas defined in ArcGIS chosen by geography, water depth, and data sufficiency. We derive isotropic, binned semivariograms from the data, and invert these for estimates of noise variance, field variance, and decorrelation distance. The highly erratic nature of the semivariograms is a result both of the random locations of the data and of the high level of data uncertainty (noise). This decorrelates the data covariance matrix for the inversion, and largely prevents robust estimation of the fractal dimension. Our comparison of the extracted and parsed mean grain size data demonstrates important differences between the two. In particular, extracted measurements generally produce finer mean grain sizes, lower noise variance, and lower field variance than parsed values. Such relationships can be used to derive a regionally dependent conversion factor between the two. Our analysis of sample regions on the US continental shelf revealed considerable geographic variability in the estimated statistical parameters of field variance and decorrelation distance. Some regional relationships are evident, and overall there is a tendency for field variance to be higher where the average mean grain size is finer grained

  7. “Thou art translated”: Remapping Hideki Noda and Satoshi Miyagi’s A Midsummer Night’s Dream in Post-March 11 Japan

    Directory of Open Access Journals (Sweden)

    Eglinton Mika

    2016-12-01

    Full Text Available Ever since the first introduction of Shakespeare to a Japanese audience in the nineteenth century, his plays have functioned as “contact zones,” which are translingual interfaces between communities and their cultures; points of negotiation, misunderstanding and mutual transformation. In the context of what is ostensibly a monolingual society, Japanese Shakespeare has produced a limited number of performances that have attempted to be multilingual. Most of them, however, turn out to be translingual, blurring the borders of linguistic specificity.

  8. Language-universal constraints on speech segmentation

    NARCIS (Netherlands)

    Norris, D.; McQueen, J.M.; Cutler, A.; Butterfield, S.; Kearns, R.K.

    2001-01-01

    Two word-spotting experiments are reported that examine whether the Possible-Word Constraint (PWC; Norris, McQueen, Cutler & Butterfield, 1997) is a language-specific or language-universal strategy for the segmentation of continuous speech. The PWC disfavors parses which leave an impossible residue

  9. Does syntax help discourse segmentation? Not so much

    DEFF Research Database (Denmark)

    Braud, Chloé Elodie; Lacroix, Ophélie; Søgaard, Anders

    2017-01-01

    Discourse segmentation is the first step in building discourse parsers. Most work on discourse segmentation does not scale to real-world discourse parsing across languages, for two reasons: (i) models rely on constituent trees, and (ii) experiments have relied on gold standard identification...

  10. ProForma

    DEFF Research Database (Denmark)

    LeDuc, Richard D; Schwämmle, Veit; Shortreed, Michael R

    2018-01-01

    modifications or unidentified mass shifts within brackets following certain amino acids. The notation is unambiguous, human-readable, and can easily be parsed and written by bioinformatic tools. This system uses seven rules and supports a wide range of possible use cases, ensuring compatibility...

  11. A Lightweight Story-Comprehension Approach to Game Dialogue

    DEFF Research Database (Denmark)

    van Leeuwen, Robert P.; Cheong, Yun-Gyung; Nelson, Mark Jason

    2013-01-01

    In this paper we describe Answery, a rule-based system that allows authors to specify game characters' background stories in natural language. The system parses these background stories, applies transfor- mation rules to turn them into semantic content, and generates dialogue during gameplay...

  12. RNA-Seq-Based Transcript Structure Analysis with TrBorderExt.

    Science.gov (United States)

    Wang, Yejun; Sun, Ming-An; White, Aaron P

    2018-01-01

    RNA-Seq has become a routine strategy for genome-wide gene expression comparisons in bacteria. Despite lower resolution in transcript border parsing compared with dRNA-Seq, TSS-EMOTE, Cappable-seq, Term-seq, and others, directional RNA-Seq still illustrates its advantages: low cost, quantification and transcript border analysis with a medium resolution (±10-20 nt). To facilitate mining of directional RNA-Seq datasets especially with respect to transcript structure analysis, we developed a tool, TrBorderExt, which can parse transcript start sites and termination sites accurately in bacteria. A detailed protocol is described in this chapter for how to use the software package step by step to identify bacterial transcript borders from raw RNA-Seq data. The package was developed with Perl and R programming languages, and is accessible freely through the website: http://www.szu-bioinf.org/TrBorderExt .

  13. On different approaches to syntactic analysis into bi-lexical dependencies: An empirical comparison of direct, PCFG-based, and HPSG-based parsers

    Directory of Open Access Journals (Sweden)

    Angelina Ivanova

    2016-04-01

    Full Text Available We compare three different approaches to parsing into syntactic, bi- lexical dependencies for English: a ‘direct’ data-driven dependency parser, a statistical phrase structure parser, and a hybrid, ‘deep’ grammar-driven parser. The analyses from the latter two are post- converted to bi-lexical dependencies. Through this ‘reduction’ of all three approaches to syntactic dependency parsers, we determine empirically what performance can be obtained for a common set of de- pendency types for English, across a broad variety of domains. In doing so, we observe what trade-offs apply along three dimensions, accuracy, efficiency, and resilience to domain variation. Our results suggest that the hand-built grammar in one of our parsers helps in both accuracy and cross-domain parsing performance, but these accuracy gains do not necessarily translate to improvements in the downstream task of negation resolution.

  14. Cine club

    CERN Multimedia

    Cine club

    2017-01-01

    Wednesday 15 February 2017 at 20:00 CERN Council Chamber Waking Life Directed by Richard Linklater USA, 2001, 99 minutes This is the story of a boy who has a dream that he can float, but unless he holds on, he will drift away into the sky. Even when he is grown up, this idea recurs. After a strange accident, he walks through what may be a dream, flowing in and out of scenarios and encountering various characters. People he meets discuss science, philosophy and the life of dreaming and waking, and the protagonist gradually becomes alarmed that he cannot awake from this confusing dream adventure. Original version English; French subtitles Wednesday 22 February 2017 at 20:00 CERN Council Chamber Paprika Directed by Satoshi Kon Japan, 2006, 90 minutes When a machine that allows therapists to enter their patients' dreams is stolen, all Hell breaks loose. Only a young female therapist, Paprika, can stop it. Original version Japanese; English subtitles Save

  15. BioMagResBank database with sets of experimental NMR constraints corresponding to the structures of over 1400 biomolecules deposited in the Protein Data Bank

    International Nuclear Information System (INIS)

    Doreleijers, Jurgen F.; Mading, Steve; Maziuk, Dimitri; Sojourner, Kassandra; Yin Lei; Zhu Jun; Markley, John L.; Ulrich, Eldon L.

    2003-01-01

    Experimental constraints associated with NMR structures are available from the Protein Data Bank (PDB) in the form of 'Magnetic Resonance' (MR) files. These files contain multiple types of data concatenated without boundary markers and are difficult to use for further research. Reported here are the results of a project initiated to annotate, archive, and disseminate these data to the research community from a searchable resource in a uniform format. The MR files from a set of 1410 NMR structures were analyzed and their original constituent data blocks annotated as to data type using a semi-automated protocol. A new software program called Wattos was then used to parse and archive the data in a relational database. From the total number of MR file blocks annotated as constraints, it proved possible to parse 84% (3337/3975). The constraint lists that were parsed correspond to three data types (2511 distance, 788 dihedral angle, and 38 residual dipolar couplings lists) from the three most popular software packages used in NMR structure determination: XPLOR/CNS (2520 lists), DISCOVER (412 lists), and DYANA/DIANA (405 lists). These constraints were then mapped to a developmental version of the BioMagResBank (BMRB) data model. A total of 31 data types originating from 16 programs have been classified, with the NOE distance constraint being the most commonly observed. The results serve as a model for the development of standards for NMR constraint deposition in computer-readable form. The constraints are updated regularly and are available from the BMRB web site (http://www.bmrb.wisc.edu)

  16. Scale-free amplitude modulation of neuronal oscillations tracks comprehension of accelerated speech

    NARCIS (Netherlands)

    Borges, Ana Filipa Teixeira; Giraud, Anne Lise; Mansvelder, Huibert D.; Linkenkaer-Hansen, Klaus

    2018-01-01

    Speech comprehension is preserved up to a threefold acceleration, but deteriorates rapidly at higher speeds. Current models posit that perceptual resilience to accelerated speech is limited by the brain’s ability to parse speech into syllabic units using δ/θ oscillations. Here, we investigated

  17. Setswana tokenisation and computational verb morphology: facing the challenge of a disjunctive orthography

    CSIR Research Space (South Africa)

    Pretorius, R

    2009-03-01

    Full Text Available at the tokenisation/morphological analysis level and does not in principle propagate to subsequent levels of analysis such as POS tagging and shallow parsing, etc. Indeed, the approach ensures that an aspect such as orthography does not obfuscate sound linguistics and...

  18. Computers and Languages: Theory and Practice

    NARCIS (Netherlands)

    Nijholt, Antinus

    A global introduction to language technology and the areas of computer science where language technology plays a role. Surveyed in this volume are issueas related to the parsing problem in the fields of natural languages, programming languages, and formal languages. Throughout the book attention is

  19. Semantic Tagging with Deep Residual Networks

    NARCIS (Netherlands)

    Bjerva, Johannes; Plank, Barbara; Bos, Johan

    2016-01-01

    We propose a novel semantic tagging task, semtagging, tailored for the purpose of multilingual semantic parsing, and present the first tagger using deep residual networks (ResNets). Our tagger uses both word and character representations and includes a novel residual bypass architecture. We evaluate

  20. Simplex and Multiplex Stratification in ASD and ADHD Families: A Promising Approach for Identifying Overlapping and Unique Underpinnings of ASD and ADHD?

    Science.gov (United States)

    Oerlemans, Anoek M.; Hartman, Catharina A.; De Bruijn, Yvette G. E.; Van Steijn, Daphne J.; Franke, Barbara; Buitelaar, Jan K.; Rommelse, Nanda N. J.

    2015-01-01

    Autism spectrum disorders (ASD) and attention-deficit/hyperactivity disorder (ADHD) are highly heterogeneous neuropsychiatric disorders, that frequently co-occur. This study examined whether stratification into single-incidence (SPX) and multi-incidence (MPX) is helpful in (a) parsing heterogeneity and (b) detecting overlapping and unique…

  1. Empirically sampling Universal Dependencies

    DEFF Research Database (Denmark)

    Schluter, Natalie; Agic, Zeljko

    2017-01-01

    Universal Dependencies incur a high cost in computation for unbiased system development. We propose a 100% empirically chosen small subset of UD languages for efficient parsing system development. The technique used is based on measurements of model capacity globally. We show that the diversity o...

  2. The Metamorphosis of the Statistical Segmentation Output: Lexicalization during Artificial Language Learning

    Science.gov (United States)

    Fernandes, Tania; Kolinsky, Regine; Ventura, Paulo

    2009-01-01

    This study combined artificial language learning (ALL) with conventional experimental techniques to test whether statistical speech segmentation outputs are integrated into adult listeners' mental lexicon. Lexicalization was assessed through inhibitory effects of novel neighbors (created by the parsing process) on auditory lexical decisions to…

  3. Monopoly Moneys : The media environment of corporatism and the player's way out

    NARCIS (Netherlands)

    Rushkoff, D.M.

    2012-01-01

    This is a study of corporations and the monetary system on which they operate, conducted through the lens of media ecology. By treating corporations and currencies as media, we become capable of parsing the environments they create, as well as evaluating their biases and openness to broad

  4. Optimal register variation: High vowel elision in siSwati | Harford ...

    African Journals Online (AJOL)

    This paper presents an Optimality Theoretic analysis of this pattern of [+hi] vowel elision which includes partial constraint rankings with opposite rankings of relevant constraints for elision and non-elision environments and constraints to restrict elision to non-prominent functional morphemes by favouring the parsing of ...

  5. Design, development and implementation of a simple program ...

    African Journals Online (AJOL)

    The method adopted included sentence analysis, which involved the recognition of sentences and sentence structures; construction of Syntax graphs, which reflect the flow of control during the process of parsing a sentence and its corresponding parser, which reads the text into an internal, more abstract representation.

  6. Author Details

    African Journals Online (AJOL)

    ... No 1 (2013) - Articles A Parse Tree Model for Analyzing And Detecting SQL Injection Vulnerabilities Abstract PDF · Vol 9, No 1 (2013) - Articles RSA Asymmetric Cryptosystem beyond Homogeneous Transformation Abstract PDF · Vol 10, No 1 (2014) - Articles A Review of RSA Cryptosystems and Cryptographic Protocols

  7. Seqenv

    DEFF Research Database (Denmark)

    Sinclair, Lucas; Ijaz, Umer Z; Jensen, Lars Juhl

    2016-01-01

    all the isolation sources from all the search results, we run a text mining algorithm to identify and parse words that are associated with the Environmental Ontology (EnvO) controlled vocabulary. This, in turn, enables us to determine both in which environments individual sequences or taxa have...

  8. SUBTLE: Situation Understanding Bot through Language and Environment

    Science.gov (United States)

    2016-01-06

    using the Bikel parser (Bikel, 2004); these parses are then post-processed using the null element (un- derstood subject) restoration system of Gabbard ...trol. In: IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp 1988–1993 Gabbard R, Marcus M, Kulick S (2006) Fully

  9. Natural language understanding

    Energy Technology Data Exchange (ETDEWEB)

    Yoshida, S

    1982-04-01

    Language understanding is essential for intelligent information processing. Processing of language itself involves configuration element analysis, syntactic analysis (parsing), and semantic analysis. They are not carried out in isolation. These are described for the Japanese language and their usage in understanding-systems is examined. 30 references.

  10. BigSUR: large-scale structured urban reconstruction

    KAUST Repository

    Kelly, Tom

    2017-11-22

    The creation of high-quality semantically parsed 3D models for dense metropolitan areas is a fundamental urban modeling problem. Although recent advances in acquisition techniques and processing algorithms have resulted in large-scale imagery or 3D polygonal reconstructions, such data-sources are typically noisy, and incomplete, with no semantic structure. In this paper, we present an automatic data fusion technique that produces high-quality structured models of city blocks. From coarse polygonal meshes, street-level imagery, and GIS footprints, we formulate a binary integer program that globally balances sources of error to produce semantically parsed mass models with associated facade elements. We demonstrate our system on four city regions of varying complexity; our examples typically contain densely built urban blocks spanning hundreds of buildings. In our largest example, we produce a structured model of 37 city blocks spanning a total of 1,011 buildings at a scale and quality previously impossible to achieve automatically.

  11. BigSUR: large-scale structured urban reconstruction

    KAUST Repository

    Kelly, Tom; Femiani, John; Wonka, Peter; Mitra, Niloy J.

    2017-01-01

    The creation of high-quality semantically parsed 3D models for dense metropolitan areas is a fundamental urban modeling problem. Although recent advances in acquisition techniques and processing algorithms have resulted in large-scale imagery or 3D polygonal reconstructions, such data-sources are typically noisy, and incomplete, with no semantic structure. In this paper, we present an automatic data fusion technique that produces high-quality structured models of city blocks. From coarse polygonal meshes, street-level imagery, and GIS footprints, we formulate a binary integer program that globally balances sources of error to produce semantically parsed mass models with associated facade elements. We demonstrate our system on four city regions of varying complexity; our examples typically contain densely built urban blocks spanning hundreds of buildings. In our largest example, we produce a structured model of 37 city blocks spanning a total of 1,011 buildings at a scale and quality previously impossible to achieve automatically.

  12. TRAINING TREE ADJOINING GRAMMARS WITH HUGE TEXT CORPUS USING SPARK MAP REDUCE

    Directory of Open Access Journals (Sweden)

    Vijay Krishna Menon

    2015-07-01

    Full Text Available Tree adjoining grammars (TAGs are mildly context sensitive formalisms used mainly in modelling natural languages. Usage and research on these psycho linguistic formalisms have been erratic in the past decade, due to its demanding construction and difficulty to parse. However, they represent promising future for formalism based NLP in multilingual scenarios. In this paper we demonstrate basic synchronous Tree adjoining grammar for English-Tamil language pair that can be used readily for machine translation. We have also developed a multithreaded chart parser that gives ambiguous deep structures and a par dependency structure known as TAG derivation. Furthermore we then focus on a model for training this TAG for each language using a large corpus of text through a map reduce frequency count model in spark and estimation of various probabilistic parameters for the grammar trees thereafter; these parameters can be used to perform statistical parsing on the trained grammar.

  13. A Learning Algorithm for Multimodal Grammar Inference.

    Science.gov (United States)

    D'Ulizia, A; Ferri, F; Grifoni, P

    2011-12-01

    The high costs of development and maintenance of multimodal grammars in integrating and understanding input in multimodal interfaces lead to the investigation of novel algorithmic solutions in automating grammar generation and in updating processes. Many algorithms for context-free grammar inference have been developed in the natural language processing literature. An extension of these algorithms toward the inference of multimodal grammars is necessary for multimodal input processing. In this paper, we propose a novel grammar inference mechanism that allows us to learn a multimodal grammar from its positive samples of multimodal sentences. The algorithm first generates the multimodal grammar that is able to parse the positive samples of sentences and, afterward, makes use of two learning operators and the minimum description length metrics in improving the grammar description and in avoiding the over-generalization problem. The experimental results highlight the acceptable performances of the algorithm proposed in this paper since it has a very high probability of parsing valid sentences.

  14. A Semantic Analysis Method for Scientific and Engineering Code

    Science.gov (United States)

    Stewart, Mark E. M.

    1998-01-01

    This paper develops a procedure to statically analyze aspects of the meaning or semantics of scientific and engineering code. The analysis involves adding semantic declarations to a user's code and parsing this semantic knowledge with the original code using multiple expert parsers. These semantic parsers are designed to recognize formulae in different disciplines including physical and mathematical formulae and geometrical position in a numerical scheme. In practice, a user would submit code with semantic declarations of primitive variables to the analysis procedure, and its semantic parsers would automatically recognize and document some static, semantic concepts and locate some program semantic errors. A prototype implementation of this analysis procedure is demonstrated. Further, the relationship between the fundamental algebraic manipulations of equations and the parsing of expressions is explained. This ability to locate some semantic errors and document semantic concepts in scientific and engineering code should reduce the time, risk, and effort of developing and using these codes.

  15. UniGene Tabulator: a full parser for the UniGene format.

    Science.gov (United States)

    Lenzi, Luca; Frabetti, Flavia; Facchin, Federica; Casadei, Raffaella; Vitale, Lorenza; Canaider, Silvia; Carinci, Paolo; Zannotti, Maria; Strippoli, Pierluigi

    2006-10-15

    UniGene Tabulator 1.0 provides a solution for full parsing of UniGene flat file format; it implements a structured graphical representation of each data field present in UniGene following import into a common database managing system usable in a personal computer. This database includes related tables for sequence, protein similarity, sequence-tagged site (STS) and transcript map interval (TXMAP) data, plus a summary table where each record represents a UniGene cluster. UniGene Tabulator enables full local management of UniGene data, allowing parsing, querying, indexing, retrieving, exporting and analysis of UniGene data in a relational database form, usable on Macintosh (OS X 10.3.9 or later) and Windows (2000, with service pack 4, XP, with service pack 2 or later) operating systems-based computers. The current release, including both the FileMaker runtime applications, is freely available at http://apollo11.isto.unibo.it/software/

  16. Acquisition by Processing Theory: A Theory of Everything?

    Science.gov (United States)

    Carroll, Susanne E.

    2004-01-01

    Truscott and Sharwood Smith (henceforth T&SS) propose a novel theory of language acquisition, "Acquisition by Processing Theory" (APT), designed to account for both first and second language acquisition, monolingual and bilingual speech perception and parsing, and speech production. This is a tall order. Like any theoretically ambitious…

  17. Comorbid Social Anxiety Disorder in Adults with Autism Spectrum Disorder

    Science.gov (United States)

    Maddox, Brenna B.; White, Susan W.

    2015-01-01

    Social anxiety symptoms are common among cognitively unimpaired youth with autism spectrum disorder (ASD). Few studies have investigated the co-occurrence of social anxiety disorder (SAD) in adults with ASD, although identification may aid access to effective treatments and inform our scientific efforts to parse heterogeneity. In this preliminary…

  18. Event Segmentation Improves Event Memory up to One Month Later

    Science.gov (United States)

    Flores, Shaney; Bailey, Heather R.; Eisenberg, Michelle L.; Zacks, Jeffrey M.

    2017-01-01

    When people observe everyday activity, they spontaneously parse it into discrete meaningful events. Individuals who segment activity in a more normative fashion show better subsequent memory for the events. If segmenting events effectively leads to better memory, does asking people to attend to segmentation improve subsequent memory? To answer…

  19. Computational Linguistic Assessment of Genre Differences Focusing on Text Cohesive Devices of Student Writing: Implications for Library Instruction

    Science.gov (United States)

    Wang, Xin; Cho, Kwangsu

    2010-01-01

    This study examined two major academic genres of writing: argumentative and technical writing. Three hundred eighty-four undergraduate student-produced texts were parsed and analyzed through a computational tool called Coh-Metrix. The results inform the instructional librarians that students used genre-dependent cohesive devices in a limited way…

  20. Sorry Dave, I’m Afraid I Can’t Do That: Explaining Unachievable Robot Tasks using Natural Language

    Science.gov (United States)

    2013-06-24

    processing components used by Brooks et al. [6]: the Bikel parser [3] combined with the null element (understood subject) restoration of Gabbard et al...Intelligent Robots and Systems (IROS), pages 1988 – 1993, 2010. [12] Ryan Gabbard , Mitch Marcus, and Seth Kulick. Fully parsing the Penn Treebank. In Human

  1. Decoupling Object Detection and Categorization

    Science.gov (United States)

    Mack, Michael L.; Palmeri, Thomas J.

    2010-01-01

    We investigated whether there exists a behavioral dependency between object detection and categorization. Previous work (Grill-Spector & Kanwisher, 2005) suggests that object detection and basic-level categorization may be the very same perceptual mechanism: As objects are parsed from the background they are categorized at the basic level. In…

  2. Behaviour Protocols Verification: Fighting State Explosion

    Czech Academy of Sciences Publication Activity Database

    Mach, M.; Plášil, František; Kofroň, Jan

    2005-01-01

    Roč. 6, č. 2 (2005), s. 22-30 ISSN 1525-9293 R&D Projects: GA ČR(CZ) GA102/03/0672 Institutional research plan: CEZ:AV0Z10300504 Keywords : formal verification * software components * stateexplos ion * behavior protocols * parse trees Subject RIV: JC - Computer Hardware ; Software

  3. Modeling media as latent semantics based on cognitive components

    DEFF Research Database (Denmark)

    Petersen, Michael Kai

    as distinct states in the continuous ebb and flow of emotions underlying consciousness. Whether it being a soundscape of structured peaks or tiny black characters lined up across a page, we rely on syntax for parsing sequences of symbols, which based on hierarchically nested structures allow us to express...

  4. The effects of L2 proficiency level on the processing of wh-questions among Dutch second language speakers of English

    NARCIS (Netherlands)

    Jackson, C.N.; Hell, J.G. van

    2011-01-01

    Using a self-paced reading task, the present study explores how Dutch-English L2 speakers parse English wh-subject-extractions and wh-object-extractions. Results suggest that English native speakers and highly-proficient Dutch–English L2 speakers do not always exhibit measurable signs of on-line

  5. Representing sentence information

    Science.gov (United States)

    Perkins, Walton A., III

    1991-03-01

    This paper describes a computer-oriented representation for sentence information. Whereas many Artificial Intelligence (AI) natural language systems start with a syntactic parse of a sentence into the linguist's components: noun, verb, adjective, preposition, etc., we argue that it is better to parse the input sentence into 'meaning' components: attribute, attribute value, object class, object instance, and relation. AI systems need a representation that will allow rapid storage and retrieval of information and convenient reasoning with that information. The attribute-of-object representation has proven useful for handling information in relational databases (which are well known for their efficiency in storage and retrieval) and for reasoning in knowledge- based systems. On the other hand, the linguist's syntactic representation of the works in sentences has not been shown to be useful for information handling and reasoning. We think it is an unnecessary and misleading intermediate form. Our sentence representation is semantic based in terms of attribute, attribute value, object class, object instance, and relation. Every sentence is segmented into one or more components with the form: 'attribute' of 'object' 'relation' 'attribute value'. Using only one format for all information gives the system simplicity and good performance as a RISC architecture does for hardware. The attribute-of-object representation is not new; it is used extensively in relational databases and knowledge-based systems. However, we will show that it can be used as a meaning representation for natural language sentences with minor extensions. In this paper we describe how a computer system can parse English sentences into this representation and generate English sentences from this representation. Much of this has been tested with computer implementation.

  6. Computing Accurate Grammatical Feedback in a Virtual Writing Conference for German-Speaking Elementary-School Children: An Approach Based on Natural Language Generation

    Science.gov (United States)

    Harbusch, Karin; Itsova, Gergana; Koch, Ulrich; Kuhner, Christine

    2009-01-01

    We built a natural language processing (NLP) system implementing a "virtual writing conference" for elementary-school children, with German as the target language. Currently, state-of-the-art computer support for writing tasks is restricted to multiple-choice questions or quizzes because automatic parsing of the often ambiguous and fragmentary…

  7. The Sentence Fairy: A Natural-Language Generation System to Support Children's Essay Writing

    Science.gov (United States)

    Harbusch, Karin; Itsova, Gergana; Koch, Ulrich; Kuhner, Christine

    2008-01-01

    We built an NLP system implementing a "virtual writing conference" for elementary-school children, with German as the target language. Currently, state-of-the-art computer support for writing tasks is restricted to multiple-choice questions or quizzes because automatic parsing of the often ambiguous and fragmentary texts produced by pupils…

  8. FQC Dashboard: integrates FastQC results into a web-based, interactive, and extensible FASTQ quality control tool

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Joseph; Pirrung, Meg; McCue, Lee Ann

    2017-06-09

    FQC is software that facilitates large-scale quality control of FASTQ files by carrying out a QC protocol, parsing results, and aggregating quality metrics within and across experiments into an interactive dashboard. The dashboard utilizes human-readable configuration files to manipulate the pages and tabs, and is extensible with CSV data.

  9. Neural Encoding of Relative Position

    Science.gov (United States)

    Hayworth, Kenneth J.; Lescroart, Mark D.; Biederman, Irving

    2011-01-01

    Late ventral visual areas generally consist of cells having a significant degree of translation invariance. Such a "bag of features" representation is useful for the recognition of individual objects; however, it seems unable to explain our ability to parse a scene into multiple objects and to understand their spatial relationships. We…

  10. The Temporal Organization of Syllabic Structure

    Science.gov (United States)

    Shaw, Jason A.

    2010-01-01

    This dissertation develops analytical tools which enable rigorous evaluation of competing syllabic parses on the basis of temporal patterns in speech production data. The data come from the articulographic tracking of fleshpoints on target speech organs, e.g., tongue, lips, jaw, in experiments with native speakers of American English and Moroccan…

  11. Change of Academic Major: The Influence of Broad and Narrow Personality Traits

    Science.gov (United States)

    Foster, N. A.

    2017-01-01

    The relationship between academic major change and ten personality traits (the Big Five and five narrow traits), was investigated in a sample of 437 college undergraduates. Contrary to expectations, Career Decidedness and Optimism were positively related to academic major change, regardless of class ranking. When parsing data by college year,…

  12. E-Learning Systems Requirements Elicitation: Perspectives and Considerations

    Science.gov (United States)

    AlKhuder, Shaikha B.; AlAli, Fatma H.

    2017-01-01

    Training and education have evolved far beyond black boards and chalk boxes. The environment of knowledge exchange requires more than simple materials and assessments. This article is an attempt of parsing through the different aspects of e-learning, understanding the real needs, and conducting the right requirements to build the appropriate…

  13. E-Learning for Depth in the Semantic Web

    Science.gov (United States)

    Shafrir, Uri; Etkind, Masha

    2006-01-01

    In this paper, we describe concept parsing algorithms, a novel semantic analysis methodology at the core of a new pedagogy that focuses learners attention on deep comprehension of the conceptual content of learned material. Two new e-learning tools are described in some detail: interactive concept discovery learning and meaning equivalence…

  14. Combining Natural Language Processing and Statistical Text Mining: A Study of Specialized versus Common Languages

    Science.gov (United States)

    Jarman, Jay

    2011-01-01

    This dissertation focuses on developing and evaluating hybrid approaches for analyzing free-form text in the medical domain. This research draws on natural language processing (NLP) techniques that are used to parse and extract concepts based on a controlled vocabulary. Once important concepts are extracted, additional machine learning algorithms,…

  15. Null Element Restoration

    Science.gov (United States)

    Gabbard, Ryan

    2010-01-01

    Understanding the syntactic structure of a sentence is a necessary preliminary to understanding its semantics and therefore for many practical applications. The field of natural language processing has achieved a high degree of accuracy in parsing, at least in English. However, the syntactic structures produced by the most commonly used parsers…

  16. Toward a Dynamic, Multidimensional Research Framework for Strategic Processing

    Science.gov (United States)

    Dinsmore, Daniel L.

    2017-01-01

    While the empirical literature on strategic processing is vast, understanding how and why certain strategies work for certain learners is far from clear. The purpose of this review is to systematically examine the theoretical and empirical literature on strategic process to parse out current conceptual and methodological progress to inform new…

  17. Maturation of Rapid Auditory Temporal Processing and Subsequent Nonword Repetition Performance in Children

    Science.gov (United States)

    Fox, Allison M.; Reid, Corinne L.; Anderson, Mike; Richardson, Cassandra; Bishop, Dorothy V. M.

    2012-01-01

    According to the rapid auditory processing theory, the ability to parse incoming auditory information underpins learning of oral and written language. There is wide variation in this low-level perceptual ability, which appears to follow a protracted developmental course. We studied the development of rapid auditory processing using event-related…

  18. On the Early Left-Anterior Negativity (ELAN) in Syntax Studies

    Science.gov (United States)

    Steinhauer, Karsten; Drury, John E.

    2012-01-01

    Within the framework of Friederici's (2002) neurocognitive model of sentence processing, the early left anterior negativity (ELAN) in event-related potentials (ERPs) has been claimed to be a brain marker of syntactic first-pass parsing. As ELAN components seem to be exclusively elicited by word category violations (phrase structure violations),…

  19. Context Modulates Attention to Social Scenes in Toddlers with Autism

    Science.gov (United States)

    Chawarska, Katarzyna; Macari, Suzanne; Shic, Frederick

    2012-01-01

    Background: In typical development, the unfolding of social and communicative skills hinges upon the ability to allocate and sustain attention toward people, a skill present moments after birth. Deficits in social attention have been well documented in autism, though the underlying mechanisms are poorly understood. Methods: In order to parse the…

  20. A memory-based shallow parser for spoken Dutch

    NARCIS (Netherlands)

    Canisius, S.V.M.; van den Bosch, A.; Decadt, B.; Hoste, V.; De Pauw, G.

    2004-01-01

    We describe the development of a Dutch memory-based shallow parser. The availability of large treebanks for Dutch, such as the one provided by the Spoken Dutch Corpus, allows memory-based learners to be trained on examples of shallow parsing taken from the treebank, and act as a shallow parser after

  1. Parsing the Behavioral and Brain Mechanisms of Third-Party Punishment.

    Science.gov (United States)

    Ginther, Matthew R; Bonnie, Richard J; Hoffman, Morris B; Shen, Francis X; Simons, Kenneth W; Jones, Owen D; Marois, René

    2016-09-07

    The evolved capacity for third-party punishment is considered crucial to the emergence and maintenance of elaborate human social organization and is central to the modern provision of fairness and justice within society. Although it is well established that the mental state of the offender and the severity of the harm he caused are the two primary predictors of punishment decisions, the precise cognitive and brain mechanisms by which these distinct components are evaluated and integrated into a punishment decision are poorly understood. Using fMRI, here we implement a novel experimental design to functionally dissociate the mechanisms underlying evaluation, integration, and decision that were conflated in previous studies of third-party punishment. Behaviorally, the punishment decision is primarily defined by a superadditive interaction between harm and mental state, with subjects weighing the interaction factor more than the single factors of harm and mental state. On a neural level, evaluation of harms engaged brain areas associated with affective and somatosensory processing, whereas mental state evaluation primarily recruited circuitry involved in mentalization. Harm and mental state evaluations are integrated in medial prefrontal and posterior cingulate structures, with the amygdala acting as a pivotal hub of the interaction between harm and mental state. This integrated information is used by the right dorsolateral prefrontal cortex at the time of the decision to assign an appropriate punishment through a distributed coding system. Together, these findings provide a blueprint of the brain mechanisms by which neutral third parties render punishment decisions. Punishment undergirds large-scale cooperation and helps dispense criminal justice. Yet it is currently unknown precisely how people assess the mental states of offenders, evaluate the harms they caused, and integrate those two components into a single punishment decision. Using a new design, we isolated these three processes, identifying the distinct brain systems and activities that enable each. Additional findings suggest that the amygdala plays a crucial role in mediating the interaction of mental state and harm information, whereas the dorsolateral prefrontal cortex plays a crucial, final-stage role, both in integrating mental state and harm information and in selecting a suitable punishment amount. These findings deepen our understanding of how punishment decisions are made, which may someday help to improve them. Copyright © 2016 Ginther et al.

  2. A Parse Tree Model for Analyzing And Detecting SQL Injection ...

    African Journals Online (AJOL)

    2013-03-01

    Mar 1, 2013 ... a difficult task as it requires lots of hard ... additional cost to any organization that may ... change the logic, syntax, or semantic of a .... language (quest language, e.g., MySQL, ..... A Classification of SQL Injection Attacks and.

  3. Learning and Parsing Video Events with Goal and Intent Prediction

    Science.gov (United States)

    2012-03-19

    including office, lab, hallway, cor- ridor and near vending machines . Figure 14 shows some screen-shots of the videos. The training video total lasts...most of the ambiguities can be removed by the event context in the top-down bottom-up inference, we will show this in the experiment section. 5 Figure 5...events, and remove the ambiguities in the detection of atomic actions by the event context. The energy of PG is E(PG | I∧) = p(K) K∑ k=1 (ε(pgk | I

  4. Collated, parsed, OCO-2 Science or Calibration Data V6

    Data.gov (United States)

    National Aeronautics and Space Administration — The Orbiting Carbon Observatory is the first NASA mission designed to collect space-based measurements of atmospheric carbon dioxide with the precision, resolution,...

  5. Adaptations for English Language Learners: Differentiating between Linguistic and Instructional Accommodations

    Science.gov (United States)

    Pappamihiel, N. Eleni; Lynn, C. Allen

    2016-01-01

    While many teachers and teacher educators in the United States K-12 system acknowledge that the English language learners (ELLs) in our schools need modifications and accommodations to help them succeed in school, few attempt to parse out how different types of accommodations may affect learning in the mainstream classroom, specifically linguistic…

  6. Temporal Clustering and Sequencing in Short-Term Memory and Episodic Memory

    Science.gov (United States)

    Farrell, Simon

    2012-01-01

    A model of short-term memory and episodic memory is presented, with the core assumptions that (a) people parse their continuous experience into episodic clusters and (b) items are clustered together in memory as episodes by binding information within an episode to a common temporal context. Along with the additional assumption that information…

  7. Towards a unified theory of neocortex: laminar cortical circuits for vision and cognition.

    Science.gov (United States)

    Grossberg, Stephen

    2007-01-01

    A key goal of computational neuroscience is to link brain mechanisms to behavioral functions. The present article describes recent progress towards explaining how laminar neocortical circuits give rise to biological intelligence. These circuits embody two new and revolutionary computational paradigms: Complementary Computing and Laminar Computing. Circuit properties include a novel synthesis of feedforward and feedback processing, of digital and analog processing, and of preattentive and attentive processing. This synthesis clarifies the appeal of Bayesian approaches but has a far greater predictive range that naturally extends to self-organizing processes. Examples from vision and cognition are summarized. A LAMINART architecture unifies properties of visual development, learning, perceptual grouping, attention, and 3D vision. A key modeling theme is that the mechanisms which enable development and learning to occur in a stable way imply properties of adult behavior. It is noted how higher-order attentional constraints can influence multiple cortical regions, and how spatial and object attention work together to learn view-invariant object categories. In particular, a form-fitting spatial attentional shroud can allow an emerging view-invariant object category to remain active while multiple view categories are associated with it during sequences of saccadic eye movements. Finally, the chapter summarizes recent work on the LIST PARSE model of cognitive information processing by the laminar circuits of prefrontal cortex. LIST PARSE models the short-term storage of event sequences in working memory, their unitization through learning into sequence, or list, chunks, and their read-out in planned sequential performance that is under volitional control. LIST PARSE provides a laminar embodiment of Item and Order working memories, also called Competitive Queuing models, that have been supported by both psychophysical and neurobiological data. These examples show how

  8. Psychophysical and Neural Correlates of Auditory Attraction and Aversion

    Science.gov (United States)

    Patten, Kristopher Jakob

    This study explores the psychophysical and neural processes associated with the perception of sounds as either pleasant or aversive. The underlying psychophysical theory is based on auditory scene analysis, the process through which listeners parse auditory signals into individual acoustic sources. The first experiment tests and confirms that a self-rated pleasantness continuum reliably exists for 20 various stimuli (r = .48). In addition, the pleasantness continuum correlated with the physical acoustic characteristics of consonance/dissonance (r = .78), which can facilitate auditory parsing processes. The second experiment uses an fMRI block design to test blood oxygen level dependent (BOLD) changes elicited by a subset of 5 exemplar stimuli chosen from Experiment 1 that are evenly distributed over the pleasantness continuum. Specifically, it tests and confirms that the pleasantness continuum produces systematic changes in brain activity for unpleasant acoustic stimuli beyond what occurs with pleasant auditory stimuli. Results revealed that the combination of two positively and two negatively valenced experimental sounds compared to one neutral baseline control elicited BOLD increases in the primary auditory cortex, specifically the bilateral superior temporal gyrus, and left dorsomedial prefrontal cortex; the latter being consistent with a frontal decision-making process common in identification tasks. The negatively-valenced stimuli yielded additional BOLD increases in the left insula, which typically indicates processing of visceral emotions. The positively-valenced stimuli did not yield any significant BOLD activation, consistent with consonant, harmonic stimuli being the prototypical acoustic pattern of auditory objects that is optimal for auditory scene analysis. Both the psychophysical findings of Experiment 1 and the neural processing findings of Experiment 2 support that consonance is an important dimension of sound that is processed in a manner that aids

  9. JPP: A Java Pre-Processor

    OpenAIRE

    Kiniry, Joseph R.; Cheong, Elaine

    1998-01-01

    The Java Pre-Processor, or JPP for short, is a parsing pre-processor for the Java programming language. Unlike its namesake (the C/C++ Pre-Processor, cpp), JPP provides functionality above and beyond simple textual substitution. JPP's capabilities include code beautification, code standard conformance checking, class and interface specification and testing, and documentation generation.

  10. Verbal and Visual-Spatial Working Memory: What Develops over a Life Span?

    Science.gov (United States)

    Swanson, H. Lee

    2017-01-01

    This study investigates whether age-related changes in the structure of 5 complex working memory (WM) tasks (a) reflect a general or domain specific system, (b) follows a similar trajectory across different age spans, and (c) contribute domain general or domain specific resources to achievement measures. The study parsed the sample (N = 2,471)…

  11. Age-Related Differences in Speech Rate Perception Do Not Necessarily Entail Age-Related Differences in Speech Rate Use

    Science.gov (United States)

    Heffner, Christopher C.; Newman, Rochelle S.; Dilley, Laura C.; Idsardi, William J.

    2015-01-01

    Purpose: A new literature has suggested that speech rate can influence the parsing of words quite strongly in speech. The purpose of this study was to investigate differences between younger adults and older adults in the use of context speech rate in word segmentation, given that older adults perceive timing information differently from younger…

  12. Preschool Children's Exposure to Story Grammar Elements during Parent-Child Book Reading

    Science.gov (United States)

    Breit-Smith, Allison; van Kleeck, Anne; Prendeville, Jo-Anne; Pan, Wei

    2017-01-01

    Twenty-three preschool-age children, 3;6 (years; months) to 4;1, were videotaped separately with their mothers and fathers while each mother and father read a different unfamiliar storybook to them. The text from the unfamiliar storybooks was parsed and coded into story grammar elements and all parental extratextual utterances were transcribed and…

  13. "It Could Have Been so Much Better": The Aesthetic and Social Work of Theatre

    Science.gov (United States)

    Gallagher, Kathleen; Freeman, Barry; Wessells, Anne

    2010-01-01

    In this paper, the authors consider early results from their ethnographic research in urban drama classrooms by parsing the aesthetic and social imperatives at play in the classroom. Moved by the observation that teachers and students alike seem to be pursuing elusive aesthetic and social ideals, the authors draw on Judith Butler's notion of…

  14. A fusion network for semantic segmentation using RGB-D data

    Science.gov (United States)

    Yuan, Jiahui; Zhang, Kun; Xia, Yifan; Qi, Lin; Dong, Junyu

    2018-04-01

    Semantic scene parsing is considerable in many intelligent field, including perceptual robotics. For the past few years, pixel-wise prediction tasks like semantic segmentation with RGB images has been extensively studied and has reached very remarkable parsing levels, thanks to convolutional neural networks (CNNs) and large scene datasets. With the development of stereo cameras and RGBD sensors, it is expected that additional depth information will help improving accuracy. In this paper, we propose a semantic segmentation framework incorporating RGB and complementary depth information. Motivated by the success of fully convolutional networks (FCN) in semantic segmentation field, we design a fully convolutional networks consists of two branches which extract features from both RGB and depth data simultaneously and fuse them as the network goes deeper. Instead of aggregating multiple model, our goal is to utilize RGB data and depth data more effectively in a single model. We evaluate our approach on the NYU-Depth V2 dataset, which consists of 1449 cluttered indoor scenes, and achieve competitive results with the state-of-the-art methods.

  15. Remote monitoring, data sharing, and information security

    International Nuclear Information System (INIS)

    Parise, D.; Dalton, C.; Regula, J.

    2009-01-01

    Full-text: Remote Monitoring (RM) is being used with increased frequency by the IAEA for safeguards in many parts of the world. This is especially true in Japan where there are also agreements for data sharing. The automated nature of RM lends itself to assist in modernizing old cumbersome data sharing techniques. For example, electronic declarations can be received, parsed and checked; then data for that time period and facility can be automatically released. This could save considerable time and effort now spent processing paper declarations and hand copying data. But care must be taken to ensure the parsing, transfers, and connections for these systems are secure. Advanced authentication and encryption techniques are still vital in this process. This paper will describe how to improve security with vulnerability assessments, the use of certificates, avoiding compromising dial-up connections and other methods. A detailed network layout will be presented that will resemble a future RM collaboration with the IAEA and the Japanese. From this network design, key strategic security points will be identified and suggestions will be made to help secure them. (author)

  16. Radiologic-anatomic correlation of thoracic vertebrae and rib shadows in chest digital radiograph

    International Nuclear Information System (INIS)

    Yamaguchi, Isao; Itoh, Harumi

    2007-01-01

    The purpose of this study was to provide an introduction to parsing the radiologic appearance of thoracic vertebrae and ribs. In the study, the radiologic-anatomic correlation technique was applied to promote further understanding of normal chest radiographs. The thoracic vertebrae and ribs of chest radiographs were compared with each macroscopic radiologic and computed tomography (CT) image. The rib parsed the linear shadow of the body of the rib. The macroscopic and radiologic images of thoracic vertebrae and ribs were evaluated to explain their normal radiologic findings. The results of such correlation were summarized as follows: The lamina of the vertebral arch was visualized due to anterior rotation of the upper thoracic vertebrae. The density ratio of the thoracic-vertebrae shadow was almost the same in the vertebral body and vertebral arch. The linear shadow superimposed on the rib corresponded to the inferior margin of the rib. The radiologic-anatomic correlation technique was useful to evaluate normal radiologic findings, and the study was useful to radiological technologists. (author)

  17. Annual report of the Japan Nuclear Ship Research and Development Agency, for fiscal 1981

    International Nuclear Information System (INIS)

    1982-01-01

    All the works of shielding repair and safety general inspection for the nuclear ship ''Mutsu'' were completed. For advancing the research and development of nucear ships, of course, the data and experience of the behavior of marine nuclear reactors are required, which can be obtained only by operating nuclear ships. The Agency will carry out the experimental voyage after the prescribed tests are finished, and endeavor to attain the objective. A new development was observed on the new home port for the Mutsu. In May, 1981, the agreement was reached among those concerned to decide Sekine Beach, Mutsu City, as the candidate site after the survey and coordination, and to construct the home port as early as possible. The Agency carried out the survey required for the location, and reported in March, 1982, that the construction of the home port is technically feasible, and also the concept of the home port and the incidental facilities on land was informed to Aomori Prefecture. Hereafter, the compensation for fishery and the purchase of land will be actively promoted. In order to ease the restriction on the energy supply for shipping industry, the technical basis for the practical use of nuclear ships must be urgently consolidated. In this report, the works performed by the Agency in fiscal 1981 are described. (Kako, I.)

  18. Neoliberalism in Historical Light: How Business Models Displaced Science Education Goals in Two Eras

    Science.gov (United States)

    Hayes, Kathryn N.

    2016-01-01

    Although a growing body of work addresses the current role of neoliberalism in displacing democratic equality as a goal of public education, attempts to parse such impacts rarely draw from historical accounts. At least one tenet of neoliberalism--the application of business models to public institutions--was also pervasive at the turn of the 20th…

  19. Indian accent text-to-speech system for web browsing

    Indian Academy of Sciences (India)

    This paper describes a 'web reader' which 'reads out' the textual contents of a selected web page in Hindi or in English with Indian accent. The content of the page is downloaded and parsed into suitable textual form. It is then passed on to an indigenously developed text-to-speech system for Hindi/Indian English, ...

  20. Multivariate Analysis for the Choice and Evasion of the Student in a Higher Educational Institution from Southern of Santa Catarina, in Brazil

    Science.gov (United States)

    Queiroz, Fernanda Cristina Barbosa Pereira; Samohyl, Robert Wayne; Queiroz, Jamerson Viegas; Lima, Nilton Cesar; de Souza, Gustavo Henrique Silva

    2014-01-01

    This paper aims to develop and implement a method to identify the causes of the choice of a course and the reasons for evasion in higher education. This way, we sought to identify the factors that influence student choice to opt for Higher Education Institution parsed, as well as the factors influencing its evasion. The methodology employed was…

  1. Taking It to the Streets: Assessing the FBI’s Approach to Combating Illicit Networks

    Science.gov (United States)

    2010-06-01

    provides a review of many of these works and parses the differences between economists , political scientists, organizational theorists, sociologists...physical surveillance, electronic surveillance (wiretaps), undercover operations, and the use of pen registers. ETI can also utilize...surveillance, and undercover operations. By using one or both of the aforementioned investigations, investigators seek to gain evidence of the structure

  2. Effective prevention of sorafenib-induced hand–foot syndrome by dried-bonito broth

    Directory of Open Access Journals (Sweden)

    Kamimura K

    2018-04-01

    Full Text Available Kenya Kamimura,1 Yoko Shinagawa-Kobayashi,1 Ryo Goto,1 Kohei Ogawa,1 Takeshi Yokoo,1 Akira Sakamaki,1 Satoshi Abe,1 Hiroteru Kamimura,1 Takeshi Suda,2 Hiroshi Baba,3 Takayuki Tanaka,4 Yoshizu Nozawa,5 Naoto Koyama,6 Masaaki Takamura,1 Hirokazu Kawai,1 Satoshi Yamagiwa,1 Yutaka Aoyagi,1 Shuji Terai1 1Division of Gastroenterology and Hepatology, Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Niigata, Japan; 2Department of Gastroenterology and Hepatology, Uonuma Institute of Community Medicine, Niigata Medical and Dental Hospital, Minami-Uonuma, Niigata, Japan; 3Division of Anesthesiology, Graduate School of Medical and Dental Sciences, Niigata University, Niigata, Niigata, Japan; 4Uonuma Eye Clinic, Uonuma, Niigata, Japan; 5Institute of Food Sciences and Technologies, Ajinomoto Co., Inc., Kawasaki, Kanagawa, Japan; 6Institute for Innovation, Ajinomoto Co., Inc., Kawasaki, Kanagawa, Japan Background: Sorafenib (SOR is a molecular medicine that prolongs the survival of patients with hepatocellular carcinoma (HCC. Therefore, the management of side effects is essential for the longer period of continuous medication. Among the various side effects, hand–foot syndrome (HFS is the most common, occurring in 30%–50% of patients, and often results in discontinuation of the SOR medication. However, its mechanism has not been clarified, and no effective prevention method has been reported for the symptoms. Therefore, this study aimed to analyze its mechanism and to develop an effective prevention regimen for the symptoms. Materials and methods: To assess the mechanism of SOR-induced HFS, the peripheral blood flow in the hand and foot was carefully monitored by Doppler ultrasound, thermography, and laser speckle flowgraphy in the cases treated with SOR and its contribution was assessed. Then, the effect of dried-bonito broth (DBB, which was reported to improve peripheral blood flow, on the prevention of the symptom was

  3. The Big Four Skills: Teachers’ Assumptions on Measurement of Cognition and Academic Skills for Non-Native Students.

    OpenAIRE

    Figueiredo, Sandra; Silva, Carlos Fernandes da; Nunes, Odete; Martins, Maria Margarida Alves d'Orey

    2016-01-01

    The four-skills on tests for young native speakers commonly do not generate correlation incongruency concerning the cognitive strategies frequently reported. Considering the non-native speakers there are parse evidence to determine which tasks are important to assess properly the cognitive and academic language proficiency (Cummins, 1980; 2012). Research questions: It is of high probability that young students with origin in immigration ...

  4. United States Air Force Summer Research Program 1991. Summer Faculty Research Program (SFRP) Reports. Volume 4. Rome Laboratory, Arnold Engineering Development Center, F. J. Seiler Research Laboratory

    Science.gov (United States)

    1992-01-09

    Parsing. Kaplan , D. 1975. "Quantifying in", in Davidson et al eds. The Logic of Grammar, Keenan, E. L. and L. M. Faltz 1985, Boolean Semantics for Natural...Area ( mA2 ) 100e-12 Nsrf Esrfb Phib contl type-ohmic cont2 type-ohmic Layer Descriptions Mat. X Grade Th(A) Nd Grade Edb Na Grade Eab Max Mesh algaas x

  5. Pre-incident Analysis using Multigraphs and Faceted Ontologies

    Science.gov (United States)

    2013-08-01

    intelligence (HUMINT), 5) open - source intelligence ( OSINT ) and 6) geospatial intelligence 1 . The analyst must access, parse, and correlate each of...The information comes in a range of classification levels from open - source human terrain data to highly classified and compartmentalized, and in...many types: 1) signals intelligence (SIGINT), 2) imagery intelligence (IMINT), 3) measurement and signature intelligence (MASINT), 4) human- source

  6. Chinese Lessons from Other Peoples’ Wars

    Science.gov (United States)

    2011-11-01

    as the lessons “not learned” or not adopted, the so-called “ dogs that do not bark.” While it may be difficult to parse real lessons learned from...under the barrel of the Type 95 5.8mm assault rifle. There is no muzzle brake or flash suppressor fitted and it is fed from a box magazine that

  7. Parsing a perceptual decision into a sequence of moments of thought

    Directory of Open Access Journals (Sweden)

    Martin eGraziano

    2011-09-01

    Full Text Available Theoretical, computational and experimental studies have converged to a model of decision-making in which sensory evidence is stochastically integrated to a threshold, implementing a shift from an analog to a discrete form of computation. Understanding how this process can be chained and sequenced - as virtually all real-life tasks involve a sequence of decisions - remains an open question in neuroscience. We reasoned that incorporating a virtual continuum of possible behavioral outcomes in a simple decision task- a fundamental ingredient of real-life decision making – should result in a progressive sequential approximation to the correct response. We used real-time tracking of motor action in a decision task, as a measure of cognitive states reflecting an internal decision process. We found that response trajectories were spontaneously segmented into a discrete sequence of explorations separated by brief stops (about 200 ms – which remained unconscious to the participants. The characteristics of these stops were indicative of a decision process - a moment of thought: their duration correlated with the difficulty of the decision and with the efficiency of the subsequent exploration. Our findings suggest that simple navigation in an abstract space involves a discrete sequence of explorations and stops and, moreover, that these stops reveal a fingerprint of moments of thought.

  8. PEG parsing in less space using progressive tabling and dynamic analysis

    DEFF Research Database (Denmark)

    Henglein, Fritz; Rasmussen, Ulrik Terp

    2017-01-01

    -case constant and worst-case linear memory use. Furthermore, semantic actions are scheduled before the parser has seen the end of the input. The scheduling is conservative in the sense that no action has to be "undone" in the case of backtracking. The time complexity is O(dmn) where m is the size of the parser...

  9. Comparison of Context-free Grammars Based on Parsing Generated Test Data

    NARCIS (Netherlands)

    B. Fischer (Bernd); R. Lämmel (Ralf); V. Zaytsev (Vadim); U. Aßmann; J. Saraiva; A.M. Sloane

    2011-01-01

    textabstractThere exist a number of software engineering scenarios that essentially involve equivalence or correspondence assertions for some of the context-free grammars in the scenarios. For instance, when applying grammar transformations during parser development---be it for the sake of

  10. Parsing cognition in schizophrenia using saccadic eye movements : a selective overview

    NARCIS (Netherlands)

    Broerse, A; Crawford, TJ; den Boer, JA

    2001-01-01

    Eye movements provide a behavioural measure of sensorimotor processing and higher cognitive functions of the brain. With the development of novel paradigms that can be used for the study of various cognitive operations, saccadic eye movements in particular. have become increasingly popular. Patients

  11. Generalized plotting facility

    Energy Technology Data Exchange (ETDEWEB)

    Burris, R.D.; Gray, W.H.

    1978-01-01

    A command which causes the translation of any supported graphics file format to a format acceptable to any supported device was implemented on two linked DECsystem-10s. The processing of the command is divided into parsing and translating phases. In the parsing phase, information is extracted from the command and augmented by default data. The results of this phase are saved on disk, and the appropriate translating routine is invoked. Twenty-eight translating programs were implemented in this system. They support four different graphics file formats, including the DISSPLA and Calcomp formats, and seven different types of plotters, including Tektronix, Calcomp, and Versatec devices. Some of the plotters are devices linked to the DECsystem-10s, and some are driven by IBM System/360 computers linked via a communications network to the DECsystem-10s. The user of this facility can use any of the supported packages to create a file of graphics data, preview the file on an on-line scope, and, when satisfied, cause the same data to be plotted on a hard-copy device. All of the actions utilize a single simple command format. 2 figures.

  12. Language experience changes subsequent learning

    Science.gov (United States)

    Onnis, Luca; Thiessen, Erik

    2013-01-01

    What are the effects of experience on subsequent learning? We explored the effects of language-specific word order knowledge on the acquisition of sequential conditional information. Korean and English adults were engaged in a sequence learning task involving three different sets of stimuli: auditory linguistic (nonsense syllables), visual non-linguistic (nonsense shapes), and auditory non-linguistic (pure tones). The forward and backward probabilities between adjacent elements generated two equally probable and orthogonal perceptual parses of the elements, such that any significant preference at test must be due to either general cognitive biases, or prior language-induced biases. We found that language modulated parsing preferences with the linguistic stimuli only. Intriguingly, these preferences are congruent with the dominant word order patterns of each language, as corroborated by corpus analyses, and are driven by probabilistic preferences. Furthermore, although the Korean individuals had received extensive formal explicit training in English and lived in an English-speaking environment, they exhibited statistical learning biases congruent with their native language. Our findings suggest that mechanisms of statistical sequential learning are implicated in language across the lifespan, and experience with language may affect cognitive processes and later learning. PMID:23200510

  13. An Intrinsic Encoding of a Subset of C and its Application to TLS Network Packet Processing

    Directory of Open Access Journals (Sweden)

    Reynald Affeldt

    2014-09-01

    Full Text Available TLS is such a widespread security protocol that errors in its implementation can have disastrous consequences. This responsibility is mostly borne by programmers, caught between specifications with the ambiguities of natural language and error-prone low-level parsing of network packets. We report here on the construction in the Coq proof-assistant of libraries to model, specify, and verify C programs to process TLS packets. We provide in particular an encoding of the core subset of C whose originality lies in its use of dependent types to guarantee statically well-formedness of datatypes and correct typing. We further equip this encoding with a Separation logic that enables byte-level reasoning and also provide a logical view of data structures. We also formalize a significant part of the RFC for TLS, again using dependent types to capture succinctly constraints that are left implicit in the prose document. Finally, we apply the above framework to an existing implementation of TLS (namely, PolarSSL of which we specify and verify a parsing function for network packets. Thanks to this experiment, we were able to spot ambiguities in the RFC and to correct bugs in the C source code.

  14. Visual language recognition with a feed-forward network of spiking neurons

    Energy Technology Data Exchange (ETDEWEB)

    Rasmussen, Craig E [Los Alamos National Laboratory; Garrett, Kenyan [Los Alamos National Laboratory; Sottile, Matthew [GALOIS; Shreyas, Ns [INDIANA UNIV.

    2010-01-01

    An analogy is made and exploited between the recognition of visual objects and language parsing. A subset of regular languages is used to define a one-dimensional 'visual' language, in which the words are translational and scale invariant. This allows an exploration of the viewpoint invariant languages that can be solved by a network of concurrent, hierarchically connected processors. A language family is defined that is hierarchically tiling system recognizable (HREC). As inspired by nature, an algorithm is presented that constructs a cellular automaton that recognizes strings from a language in the HREC family. It is demonstrated how a language recognizer can be implemented from the cellular automaton using a feed-forward network of spiking neurons. This parser recognizes fixed-length strings from the language in parallel and as the computation is pipelined, a new string can be parsed in each new interval of time. The analogy with formal language theory allows inferences to be drawn regarding what class of objects can be recognized by visual cortex operating in purely feed-forward fashion and what class of objects requires a more complicated network architecture.

  15. Deciphering the combinatorial roles of geometric, mechanical, and adhesion cues in regulation of cell spreading.

    Directory of Open Access Journals (Sweden)

    Greg M Harris

    Full Text Available Significant effort has gone towards parsing out the effects of surrounding microenvironment on macroscopic behavior of stem cells. Many of the microenvironmental cues, however, are intertwined, and thus, further studies are warranted to identify the intricate interplay among the conflicting downstream signaling pathways that ultimately guide a cell response. In this contribution, by patterning adhesive PEG (polyethylene glycol hydrogels using Dip Pen Nanolithography (DPN, we demonstrate that substrate elasticity, subcellular elasticity, ligand density, and topography ultimately define mesenchymal stem cells (MSCs spreading and shape. Physical characteristics are parsed individually with 7 kilopascal (kPa hydrogel islands leading to smaller, spindle shaped cells and 105 kPa hydrogel islands leading to larger, polygonal cell shapes. In a parallel effort, a finite element model was constructed to characterize and confirm experimental findings and aid as a predictive tool in modeling cell microenvironments. Signaling pathway inhibition studies suggested that RhoA is a key regulator of cell response to the cooperative effect of the tunable substrate variables. These results are significant for the engineering of cell-extra cellular matrix interfaces and ultimately decoupling matrix bound cues presented to cells in a tissue microenvironment for regenerative medicine.

  16. Two models of minimalist, incremental syntactic analysis.

    Science.gov (United States)

    Stabler, Edward P

    2013-07-01

    Minimalist grammars (MGs) and multiple context-free grammars (MCFGs) are weakly equivalent in the sense that they define the same languages, a large mildly context-sensitive class that properly includes context-free languages. But in addition, for each MG, there is an MCFG which is strongly equivalent in the sense that it defines the same language with isomorphic derivations. However, the structure-building rules of MGs but not MCFGs are defined in a way that generalizes across categories. Consequently, MGs can be exponentially more succinct than their MCFG equivalents, and this difference shows in parsing models too. An incremental, top-down beam parser for MGs is defined here, sound and complete for all MGs, and hence also capable of parsing all MCFG languages. But since the parser represents its grammar transparently, the relative succinctness of MGs is again evident. Although the determinants of MG structure are narrowly and discretely defined, probabilistic influences from a much broader domain can influence even the earliest analytic steps, allowing frequency and context effects to come early and from almost anywhere, as expected in incremental models. Copyright © 2013 Cognitive Science Society, Inc.

  17. Fort Knox Trend Analysis, Encroachment Study, and Perimeter Expansion Opportunities in Support of Military Training

    Science.gov (United States)

    2006-03-01

    converters from GIL and many other formats. Other hilites: command line argument parsing, a simple set of routines for de- veloping Xwindows graphical...Ramakrishna Nemani, James E. Vogelmann, V. Ruth Hobson, Benjamin Tuttle, Jeff Safran, Ingrid Nelson. (2001). “Development Sprawl Impacts on the... Sale Prices as a Basis for Farm Land Appraisal,” Technical Bulletin, University of Minnesota. Hosmer, D.W., and S. Lemeshow. (1989). Applied

  18. Formats for Digital Preservation: A Review of Alternatives and Issues

    Science.gov (United States)

    2007-03-01

    color or greyscale).TIFF_G4 is limited to bitonal (pure black and white) images. SUPPORT FOR GRAPHIC EFFECTS AND TYPOGRAPHY Not applicable Not...Management, ISO/TC 130, Graphics Technology, and ISO/TC 42, Photography. Open standard. Developed by World Wide Web Consortium. To be useful for...Human-readable and designed for automatic parsing. A well-documented DTD, XML Schema, or other specification is needed. Human-comprehensible

  19. Mechanisms for interaction: Syntax as procedures for online interactive meaning building.

    Science.gov (United States)

    Kempson, Ruth; Chatzikyriakidis, Stergios; Cann, Ronnie

    2016-01-01

    We argue that to reflect participant interactivity in conversational dialogue, the Christiansen & Chater (C&C) perspective needs a formal grammar framework capturing word-by-word incrementality, as in Dynamic Syntax, in which syntax is the incremental building of semantic representations reflecting real-time parsing dynamics. We demonstrate that, with such formulation, syntactic, semantic, and morpho-syntactic dependencies are all analysable as grounded in their potential for interaction.

  20. Real Time Physiological Status Monitoring (RT-PSM): Accomplishments, Requirements, and Research Roadmap

    Science.gov (United States)

    2016-03-01

    actionable information. With many lessons learned , the first implementation of real time physiological monitoring (RT-PSM) uses thermal-work strain... Bidirectional Inductive On-Body Network (BIONET) for WPSM Develop sensor links and processing nodes on-Soldier and non-RF links off-Soldier Elintrix...recent sleep watches (e.g., BASIS Peak, Intel Corp.) are attempting to parse sleep quality beyond duration and interruptions into deep and REM sleep

  1. Freedom: A Promise of Possibility.

    Science.gov (United States)

    Bunkers, Sandra Schmidt

    2015-10-01

    The idea of freedom as a promise of possibility is explored in this column. The core concepts from a research study on considering tomorrow (Bunkers, 1998) coupled with humanbecoming community change processes (Parse, 2003) are used to illuminate this notion. The importance of intentionality in human freedom is discussed from both a human science and a natural science perspective. © The Author(s) 2015.

  2. Strictness Analysis for Attribute Grammars

    DEFF Research Database (Denmark)

    Rosendahl, Mads

    1992-01-01

    interpretation of attribute grammars. The framework is used to construct a strictness analysis for attribute grammars. Results of the analysis enable us to transform an attribute grammar such that attributes are evaluated during parsing, if possible. The analysis is proved correct by relating it to a fixpoint...... semantics for attribute grammars. An implementation of the analysis is discussed and some extensions to the analysis are mentioned....

  3. An automatic method to generate domain-specific investigator networks using PubMed abstracts

    Directory of Open Access Journals (Sweden)

    Gwinn Marta

    2007-06-01

    Full Text Available Abstract Background Collaboration among investigators has become critical to scientific research. This includes ad hoc collaboration established through personal contacts as well as formal consortia established by funding agencies. Continued growth in online resources for scientific research and communication has promoted the development of highly networked research communities. Extending these networks globally requires identifying additional investigators in a given domain, profiling their research interests, and collecting current contact information. We present a novel strategy for building investigator networks dynamically and producing detailed investigator profiles using data available in PubMed abstracts. Results We developed a novel strategy to obtain detailed investigator information by automatically parsing the affiliation string in PubMed records. We illustrated the results by using a published literature database in human genome epidemiology (HuGE Pub Lit as a test case. Our parsing strategy extracted country information from 92.1% of the affiliation strings in a random sample of PubMed records and in 97.0% of HuGE records, with accuracies of 94.0% and 91.0%, respectively. Institution information was parsed from 91.3% of the general PubMed records (accuracy 86.8% and from 94.2% of HuGE PubMed records (accuracy 87.0. We demonstrated the application of our approach to dynamic creation of investigator networks by creating a prototype information system containing a large database of PubMed abstracts relevant to human genome epidemiology (HuGE Pub Lit, indexed using PubMed medical subject headings converted to Unified Medical Language System concepts. Our method was able to identify 70–90% of the investigators/collaborators in three different human genetics fields; it also successfully identified 9 of 10 genetics investigators within the PREBIC network, an existing preterm birth research network. Conclusion We successfully created a

  4. An automatic method to generate domain-specific investigator networks using PubMed abstracts

    Science.gov (United States)

    Yu, Wei; Yesupriya, Ajay; Wulf, Anja; Qu, Junfeng; Gwinn, Marta; Khoury, Muin J

    2007-01-01

    Background Collaboration among investigators has become critical to scientific research. This includes ad hoc collaboration established through personal contacts as well as formal consortia established by funding agencies. Continued growth in online resources for scientific research and communication has promoted the development of highly networked research communities. Extending these networks globally requires identifying additional investigators in a given domain, profiling their research interests, and collecting current contact information. We present a novel strategy for building investigator networks dynamically and producing detailed investigator profiles using data available in PubMed abstracts. Results We developed a novel strategy to obtain detailed investigator information by automatically parsing the affiliation string in PubMed records. We illustrated the results by using a published literature database in human genome epidemiology (HuGE Pub Lit) as a test case. Our parsing strategy extracted country information from 92.1% of the affiliation strings in a random sample of PubMed records and in 97.0% of HuGE records, with accuracies of 94.0% and 91.0%, respectively. Institution information was parsed from 91.3% of the general PubMed records (accuracy 86.8%) and from 94.2% of HuGE PubMed records (accuracy 87.0). We demonstrated the application of our approach to dynamic creation of investigator networks by creating a prototype information system containing a large database of PubMed abstracts relevant to human genome epidemiology (HuGE Pub Lit), indexed using PubMed medical subject headings converted to Unified Medical Language System concepts. Our method was able to identify 70–90% of the investigators/collaborators in three different human genetics fields; it also successfully identified 9 of 10 genetics investigators within the PREBIC network, an existing preterm birth research network. Conclusion We successfully created a web-based prototype

  5. Modeling and Simulation of the Economics of Mining in the Bitcoin Market

    Science.gov (United States)

    Marchesi, Michele

    2016-01-01

    In January 3, 2009, Satoshi Nakamoto gave rise to the “Bitcoin Blockchain”, creating the first block of the chain hashing on his computer’s central processing unit (CPU). Since then, the hash calculations to mine Bitcoin have been getting more and more complex, and consequently the mining hardware evolved to adapt to this increasing difficulty. Three generations of mining hardware have followed the CPU’s generation. They are GPU’s, FPGA’s and ASIC’s generations. This work presents an agent-based artificial market model of the Bitcoin mining process and of the Bitcoin transactions. The goal of this work is to model the economy of the mining process, starting from GPU’s generation, the first with economic significance. The model reproduces some “stylized facts” found in real-time price series and some core aspects of the mining business. In particular, the computational experiments performed can reproduce the unit root property, the fat tail phenomenon and the volatility clustering of Bitcoin price series. In addition, under proper assumptions, they can reproduce the generation of Bitcoins, the hashing capability, the power consumption, and the mining hardware and electrical energy expenditures of the Bitcoin network. PMID:27768691

  6. Modeling and Simulation of the Economics of Mining in the Bitcoin Market.

    Science.gov (United States)

    Cocco, Luisanna; Marchesi, Michele

    2016-01-01

    In January 3, 2009, Satoshi Nakamoto gave rise to the "Bitcoin Blockchain", creating the first block of the chain hashing on his computer's central processing unit (CPU). Since then, the hash calculations to mine Bitcoin have been getting more and more complex, and consequently the mining hardware evolved to adapt to this increasing difficulty. Three generations of mining hardware have followed the CPU's generation. They are GPU's, FPGA's and ASIC's generations. This work presents an agent-based artificial market model of the Bitcoin mining process and of the Bitcoin transactions. The goal of this work is to model the economy of the mining process, starting from GPU's generation, the first with economic significance. The model reproduces some "stylized facts" found in real-time price series and some core aspects of the mining business. In particular, the computational experiments performed can reproduce the unit root property, the fat tail phenomenon and the volatility clustering of Bitcoin price series. In addition, under proper assumptions, they can reproduce the generation of Bitcoins, the hashing capability, the power consumption, and the mining hardware and electrical energy expenditures of the Bitcoin network.

  7. Modeling and Simulation of the Economics of Mining in the Bitcoin Market.

    Directory of Open Access Journals (Sweden)

    Luisanna Cocco

    Full Text Available In January 3, 2009, Satoshi Nakamoto gave rise to the "Bitcoin Blockchain", creating the first block of the chain hashing on his computer's central processing unit (CPU. Since then, the hash calculations to mine Bitcoin have been getting more and more complex, and consequently the mining hardware evolved to adapt to this increasing difficulty. Three generations of mining hardware have followed the CPU's generation. They are GPU's, FPGA's and ASIC's generations. This work presents an agent-based artificial market model of the Bitcoin mining process and of the Bitcoin transactions. The goal of this work is to model the economy of the mining process, starting from GPU's generation, the first with economic significance. The model reproduces some "stylized facts" found in real-time price series and some core aspects of the mining business. In particular, the computational experiments performed can reproduce the unit root property, the fat tail phenomenon and the volatility clustering of Bitcoin price series. In addition, under proper assumptions, they can reproduce the generation of Bitcoins, the hashing capability, the power consumption, and the mining hardware and electrical energy expenditures of the Bitcoin network.

  8. Present state of Japan Nuclear Ship Development Agency

    International Nuclear Information System (INIS)

    Takada, Yoshio

    1981-01-01

    The Japan Nuclear Ship Development Agency held the annual report meeting on April 8, 1981. The main contents were the plan of research and development of nuclear ships hereafter, the present state of the repair works for the nuclear ship ''Mutsu'', the progress of the selection of the new home port and others. In the last year, the function of research was given to the Agency by the revision of the related law. The full-scale repair works for Mutsu were started in August, 1980, and various equipments and shields in the containment vessel and the upper shields of the containment vessel have been removed. Subsequently, new shields are being installed. According to the report by the committee of nuclear ship research and development, the development of Mutsu, which is valuable as the experimental ship, is continued. Moreover, it is proposed to do the research and development of an improved marine nuclear plant for the purposes of securing the economic efficiency, the proving of the reliability of nuclear merchant ships, and the establishment of safety. As the home port for Mutsu, the new port will be constructed on the open sea side in Aomori Prefecture, and as a candidate, Sekine beach in Mutsu City was named. Till the completion of the new home port, Mutsu will be berthed in Ominato home port. The conditions for entering and berthing in Ominato port will be decided later. (Kako, I.)

  9. PCF File Format.

    Energy Technology Data Exchange (ETDEWEB)

    Thoreson, Gregory G [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-08-01

    PCF files are binary files designed to contain gamma spectra and neutron count rates from radiation sensors. It is the native format for the GAmma Detector Response and Analysis Software (GADRAS) package [1]. It can contain multiple spectra and information about each spectrum such as energy calibration. This document outlines the format of the file that would allow one to write a computer program to parse and write such files.

  10. Multipole error analysis using local 3-bump orbit data in Fermilab Recycler

    International Nuclear Information System (INIS)

    Yang, M.J.; Xiao, M.

    2005-01-01

    The magnetic harmonic errors of the Fermilab Recycler ring were examined using circulating beam data taken with closed local orbit bumps. Data was first parsed into harmonic orbits of first, second, and third order. Each of which was analyzed for sources of magnetic errors of corresponding order. This study was made possible only with the incredible resolution of a new BPM system that was commissioned after June of 2003

  11. Learning to Understand Natural Language with Less Human Effort

    Science.gov (United States)

    2015-05-01

    Supervision Distant supervision is a recent trend in information extraction. Distantly-supervised extractors are trained using a corpus of unlabeled text...consists of fill-in-the-blank natural language questions such as “Incan emperor ” or “Cunningham directed Auchtre’s second music video .” These questions...with an 132 unknown knowledge base, simultaneously learning how to semantically parse language and pop - ulate the knowledge base. The weakly

  12. A Formal Model of Ambiguity and its Applications in Machine Translation

    Science.gov (United States)

    2010-01-01

    structure indicates linguisti- cally implausible segmentation that might be generated using dictionary - driven approaches...derivation. As was done in the monolingual case, the functions LHS, RHSi, RHSo and υ can be extended to a derivation δ. D(q) where q ∈V denotes the... monolingual parses. My algorithm runs more efficiently than O(n6) with many grammars (including those that required using heuristic search with other parsers

  13. Integrative Lifecourse and Genetic Analysis of Military Working Dogs

    Science.gov (United States)

    2013-10-01

    FITC) and Alexa Fluor 647 (BD Biosciences), respectively. Secondary antibodies of either phycoerythrin (PE; 5 μL) or fluorescein isothiocyanate (FITC; 1...zone B cells [2]. Cytokines/cytokine receptors, Ig recognition , and antigen presented by APCs, dendritic cells, and/or macrophages can initiate...MetaMap is targeted at finding medical terms from a biomedical text, it leverages a combination of part-of- speech tagging, shallow parsing, and longest

  14. SIRW: A web server for the Simple Indexing and Retrieval System that combines sequence motif searches with keyword searches.

    Science.gov (United States)

    Ramu, Chenna

    2003-07-01

    SIRW (http://sirw.embl.de/) is a World Wide Web interface to the Simple Indexing and Retrieval System (SIR) that is capable of parsing and indexing various flat file databases. In addition it provides a framework for doing sequence analysis (e.g. motif pattern searches) for selected biological sequences through keyword search. SIRW is an ideal tool for the bioinformatics community for searching as well as analyzing biological sequences of interest.

  15. Neuroscientific Model of Motivational Process

    OpenAIRE

    Kim, Sung-il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Rewa...

  16. Performance analysis of Java APIS for XML processing

    OpenAIRE

    Oliveira, Bruno; Santos, Vasco; Belo, Orlando

    2013-01-01

    Over time, XML markup language has acquired a considerable importance in applications development, standards definition and in the representation of large volumes of data, such as databases. Today, processing XML documents in a short period of time is a critical activity in a large range of applications, which imposes choosing the most appropriate mechanism to parse XML documents quickly and efficiently. When using a programming language for XML processing, such as ...

  17. Processing XML with Java – a performance benchmark

    OpenAIRE

    Oliveira, Bruno; Santos, Vasco; Belo, Orlando

    2013-01-01

    Over time, XML markup language has acquired a considerable importance in applications development, standards definition and in the representation of large volumes of data, such as databases. Today, processing XML documents in a short period of time is a critical activity in a large range of applications, which imposes choosing the most appropriate mechanism to parse XML documents quickly and efficiently. When using a programming language for XML processing, suc...

  18. Software Library for Bruker TopSpin NMR Data Files

    Energy Technology Data Exchange (ETDEWEB)

    2016-10-14

    A software library for parsing and manipulating frequency-domain data files that have been processed using the Bruker TopSpin NMR software package. In the context of NMR, the term "processed" indicates that the end-user of the Bruker TopSpin NMR software package has (a) Fourier transformed the raw, time-domain data (the Free Induction Decay) into the frequency-domain and (b) has extracted the list of NMR peaks.

  19. Integrating deep and shallow natural language processing components : representations and hybrid architectures

    OpenAIRE

    Schäfer, Ulrich

    2006-01-01

    We describe basic concepts and software architectures for the integration of shallow and deep (linguistics-based, semantics-oriented) natural language processing (NLP) components. The main goal of this novel, hybrid integration paradigm is improving robustness of deep processing. After an introduction to constraint-based natural language parsing, we give an overview of typical shallow processing tasks. We introduce XML standoff markup as an additional abstraction layer that eases integration ...

  20. The CLaC Discourse Parser at CoNLL-2015

    OpenAIRE

    Laali, Majid; Davoodi, Elnaz; Kosseim, Leila

    2017-01-01

    This paper describes our submission (kosseim15) to the CoNLL-2015 shared task on shallow discourse parsing. We used the UIMA framework to develop our parser and used ClearTK to add machine learning functionality to the UIMA framework. Overall, our parser achieves a result of 17.3 F1 on the identification of discourse relations on the blind CoNLL-2015 test set, ranking in sixth place.

  1. Slovak Dependency Treebank in Universal Dependencies

    Directory of Open Access Journals (Sweden)

    Zeman Daniel

    2017-12-01

    Full Text Available We describe a conversion of the syntactically annotated part of the Slovak National Corpus into the annotation scheme known as Universal Dependencies. Only a small subset of the data has been converted so far; yet it is the first Slovak treebank that is publicly available for research. We list a number of research projects in which the dataset has been used so far, including the first parsing results.

  2. Rochester Connectionist Papers. 1979-1985

    Science.gov (United States)

    1985-12-01

    updated and improved version of the thesis account of recent neurolinguistic data. Fanty, M., "Context-free parsing in connectionist networks." TR 174...April 1982. Our first large program in the connectionist paradigm. It simulates a multi- layer network for recognizing line drawings of Origami figures...The program successfully deals with noise and simple occlusion and the thesis incorporates many key ideas on designing and running large models. Small

  3. Learning with leaders.

    Science.gov (United States)

    Bunkers, Sandra S

    2009-01-01

    This column focuses on ideas concerning leaders and leadership. The author proposes that leadership is about showing up and participating with others in doing something. "Mandela: His 8 Lessons of Leadership" by Richard Stengel is explored in light of selected philosophical writings, literature on nursing leadership, and nurse theorist Rosemarie Rizzo Parse's humanbecoming leading-following model. Teaching-learning questions are then posed to stimulate further reflection on the lessons of leadership.

  4. Marginally Stable Triangular Recurrent Neural Network Architecture for Time Series Prediction.

    Science.gov (United States)

    Sivakumar, Seshadri; Sivakumar, Shyamala

    2017-09-25

    This paper introduces a discrete-time recurrent neural network architecture using triangular feedback weight matrices that allows a simplified approach to ensuring network and training stability. The triangular structure of the weight matrices is exploited to readily ensure that the eigenvalues of the feedback weight matrix represented by the block diagonal elements lie on the unit circle in the complex z-plane by updating these weights based on the differential of the angular error variable. Such placement of the eigenvalues together with the extended close interaction between state variables facilitated by the nondiagonal triangular elements, enhances the learning ability of the proposed architecture. Simulation results show that the proposed architecture is highly effective in time-series prediction tasks associated with nonlinear and chaotic dynamic systems with underlying oscillatory modes. This modular architecture with dual upper and lower triangular feedback weight matrices mimics fully recurrent network architectures, while maintaining learning stability with a simplified training process. While training, the block-diagonal weights (hence the eigenvalues) of the dual triangular matrices are constrained to the same values during weight updates aimed at minimizing the possibility of overfitting. The dual triangular architecture also exploits the benefit of parsing the input and selectively applying the parsed inputs to the two subnetworks to facilitate enhanced learning performance.

  5. VCFtoTree: a user-friendly tool to construct locus-specific alignments and phylogenies from thousands of anthropologically relevant genome sequences.

    Science.gov (United States)

    Xu, Duo; Jaber, Yousef; Pavlidis, Pavlos; Gokcumen, Omer

    2017-09-26

    Constructing alignments and phylogenies for a given locus from large genome sequencing studies with relevant outgroups allow novel evolutionary and anthropological insights. However, no user-friendly tool has been developed to integrate thousands of recently available and anthropologically relevant genome sequences to construct complete sequence alignments and phylogenies. Here, we provide VCFtoTree, a user friendly tool with a graphical user interface that directly accesses online databases to download, parse and analyze genome variation data for regions of interest. Our pipeline combines popular sequence datasets and tree building algorithms with custom data parsing to generate accurate alignments and phylogenies using all the individuals from the 1000 Genomes Project, Neanderthal and Denisovan genomes, as well as reference genomes of Chimpanzee and Rhesus Macaque. It can also be applied to other phased human genomes, as well as genomes from other species. The output of our pipeline includes an alignment in FASTA format and a tree file in newick format. VCFtoTree fulfills the increasing demand for constructing alignments and phylogenies for a given loci from thousands of available genomes. Our software provides a user friendly interface for a wider audience without prerequisite knowledge in programming. VCFtoTree can be accessed from https://github.com/duoduoo/VCFtoTree_3.0.0 .

  6. High-pressure phase relations in the composition of albite NaAlSi3O8 constrained by an ab initio and quasi-harmonic Debye model, and their implications

    Science.gov (United States)

    Deng, L.; Liu, X.; Liu, H.; Dong, J.

    2010-12-01

    The high pressure physical-chemical behaviors of feldspar in subducted slab are very important to the geodynamic process in the deep interior of the Earth. Albite (NaAlSi3O8;Ab) is one of the few end members in the feldspar family, and its high-P behavior is obviously a prerequisite to the full understanding of the physical-chemical properties of feldspar at high pressures. So far it has been well accepted that Ab breaks down to the phase assemblage of Jadeite+Stishovite(NaAlSi2O6; Jd, SiO2; St,JS hereafter) at ~9-10 GPa. The JS phase assemblage might be stable up to ~23 GPa, and eventually directly change into the phase assemblage of calcium-ferrite type NaAlSiO4 (Cf) +2St (CS hereafter). However, some independent researches suggest there is an intermediate phase Na-hollandite (Na-Hall; a phase with the composition of NaAlSi3O8 and the structure of hollandite) between JS phase assemblage transition into CS phase assemblage (Liu 1978; Tutti 2007; Sekine and Ahrens, 1992; Beck et al., 2004). Whether Na-Hall is a thermodynamic stable phase under high P-T conditions remains unknown. In this work, phase relations in the composition of albite NaAlSi3O8 at pressures up to 40 GPa were constrained by a theoretical method that combines the ab initio calculation and quasi-harmonic Debyemodel. First, the P-T dependence of the thermodynamic potentials of the individual phase, St, Cf, Jd and the hypothetical Na-Holl were derived. Our results are generally in consistent agreement with available experimental data and previous theoretical predictions. Second, the Gibbs free energy of the hypothetical Na-Holl phase was compared with that of the phase assemblages JS and CS. Our results show that the Na-Holl phase is not a thermodynamically stable phase over the studied P-T conditions of 0-40 GPa and 100-600 K, which rules it out as a possible intermediate phase along the transition path from the JS phase assemblage to CS phase assemblage. Our calculations have predicted that the JS

  7. Preliminary Analysis of a Breadth-First Parsing Algorithm: Theoretical and Experimental Results.

    Science.gov (United States)

    1981-06-01

    sentential-second-object John persuaded Bill that I did it. takes-bare-infinitive John made them take the exam . verb-takes-at-most-one-np-after-it *John...examples of multiple extractions, the second of which appears to be crossed. (83) Which violins are these sonatas, easy [to play t, on ti? ] (84) What

  8. Parsing the phenotype of obsessive-compulsive tic disorder (OCTD): a multidisciplinary consensus.

    Science.gov (United States)

    Dell'Osso, Bernardo; Marazziti, Donatella; Albert, Umberto; Pallanti, Stefano; Gambini, Orsola; Tundo, Antonio; Zanaboni, Carlotta; Servello, Domenico; Rizzo, Renata; Scalone, Luciana; Benatti, Beatrice; Altamura, A Carlo; Porta, Mauro

    2017-06-01

    Obsessive-Compulsive Disorder (OCD) and Tic Disorder (TD) are highly disabling and often comorbid conditions. Of note, the DSM-5 acknowledged a new 'tic-related' specifier for OCD, which might be referred to as Obsessive-Compulsive Tic Disorder (OCTD), raising new interest toward a better clinical characterisation of affected patients. Available literature indicates that early onset, male gender, sensory phenomena and obsessions of symmetry, aggressiveness, hoarding, exactness and sounds as well as comorbidity with Attention Deficit Hyperactivity Disorder (ADHD) may be of more frequent observation in patients with OCTD. In order to share expertise in the field from different perspectives, a multidisciplinary panel of Italian clinicians, specifically involved in the clinical care of OCD and TD patients, participated into a consensus initiative, aimed to produce a shared document. As a result, after having examined the most relevant literature, authors sought to critically identify and discuss main epidemiologic, socio-demographic and clinical features characterising OCTD patients, along with other specific aspects including Health-Related Quality-of-Life (HRQoL), economic consequences related with the condition and its management, as well as treatment-related issues, that need to be further investigated.

  9. Generic packet descriptions : Verified Parsing and Pretty Printing of Low-Level Data

    NARCIS (Netherlands)

    van Geest, Marcell; Swierstra, Wouter

    2017-01-01

    Complex protocols describing the communication or storage of binary data are difficult to describe precisely. This paper presents a collection of data types for describing a binary data formats; the corresponding parser and pretty printer are generated automatically from a data description. By

  10. Parsing demographic effects of canine pParvovirus on a Minnesota wolf population

    Science.gov (United States)

    Mech, L. David; Goyal, Sagar M.

    2011-01-01

    We examined 35 years of relationships among wolf (Canis lupus) pup survival, population change and canine parvovirus (CPV) seroprevalence in Northeastern Minnesota to determine when CPV exerted its strongest effects. Using correlation analysis of data from five periods of 7-years each from 1973 through 2007, we learned that the strongest effect of CPV on pup survival (r = -0.73) and on wolf population change (r = -0.92) was during 1987 to 1993. After that, little effect was documented despite a mean CPV seroprevalence from 1994 of 2007 of 70.8% compared with 52.6% during 1987 to 1993. We conclude that after CPV became endemic and produced its peak effect on the study population, that population developed enough immunity to withstand the disease.

  11. Exercises in Free Syntax. Syntax Definition, Parsing, and Assimilation of Language Conglomerates

    NARCIS (Netherlands)

    Bravenboer, M.

    2008-01-01

    In modern software development the use of multiple software languages to constitute a single application is ubiquitous. Despite the omnipresent use of combinations of languages, the principles and techniques for using languages together are ad-hoc, unfriendly to programmers, and result in a poor

  12. Parsing the Relations of Race and Socioeconomic Status in Special Education Disproportionality

    Science.gov (United States)

    Kincaid, Aleksis P.; Sullivan, Amanda L.

    2017-01-01

    This study investigated how student and school-level socioeconomic status (SES) measures predict students' odds of being identified for special education, particularly high-incidence disabilities. Using the Early Childhood Longitudinal Study--Kindergarten cohort, hierarchical models were used to determine the relations of student and school SES to…

  13. Parsing of the free energy of aromatic-aromatic stacking interactions in solution

    Energy Technology Data Exchange (ETDEWEB)

    Kostjukov, Viktor V.; Khomytova, Nina M. [Department of Physics, Sevastopol National Technical University, Sevastopol 99053, Crimea (Ukraine); Hernandez Santiago, Adrian A.; Tavera, Anna-Maria Cervantes; Alvarado, Julieta Salas [Faculty of Chemical Sciences, Autonomous University of Puebla, Puebla (Mexico); Evstigneev, Maxim P., E-mail: max_evstigneev@mail.ru [Department of Physics, Sevastopol National Technical University, Sevastopol 99053, Crimea (Ukraine)

    2011-10-15

    Graphical abstract: Highlights: > A protocol for decomposition of the free energy of aromatic stacking is developed. > The factors stabilizing/destabilizing stacking of aromatic molecules are defined. > Hydrophobic contribution is found to be dominant. - Abstract: We report an analysis of the energetics of aromatic-aromatic stacking interactions for 39 non-covalent reactions of self- and hetero-association of 12 aromatic molecules with different structures and charge states. A protocol for computation of the contributions to the total energy from various energetic terms has been developed and the results are consistent with experiment in 92% of all the systems studied. It is found that the contributions from hydrogen bonds and entropic factors are always unfavorable, whereas contributions from van-der-Waals, electrostatic and/or hydrophobic effects may lead to stabilizing or destabilizing factors depending on the system studied. The analysis carried out in this work provides an answer to the questions 'What forces stabilize/destabilize the stacking of aromatic molecules in aqueous-salt solution and what are their relative importance?'

  14. Parsing of the free energy of aromatic-aromatic stacking interactions in solution

    International Nuclear Information System (INIS)

    Kostjukov, Viktor V.; Khomytova, Nina M.; Hernandez Santiago, Adrian A.; Tavera, Anna-Maria Cervantes; Alvarado, Julieta Salas; Evstigneev, Maxim P.

    2011-01-01

    Graphical abstract: Highlights: → A protocol for decomposition of the free energy of aromatic stacking is developed. → The factors stabilizing/destabilizing stacking of aromatic molecules are defined. → Hydrophobic contribution is found to be dominant. - Abstract: We report an analysis of the energetics of aromatic-aromatic stacking interactions for 39 non-covalent reactions of self- and hetero-association of 12 aromatic molecules with different structures and charge states. A protocol for computation of the contributions to the total energy from various energetic terms has been developed and the results are consistent with experiment in 92% of all the systems studied. It is found that the contributions from hydrogen bonds and entropic factors are always unfavorable, whereas contributions from van-der-Waals, electrostatic and/or hydrophobic effects may lead to stabilizing or destabilizing factors depending on the system studied. The analysis carried out in this work provides an answer to the questions 'What forces stabilize/destabilize the stacking of aromatic molecules in aqueous-salt solution and what are their relative importance?'

  15. Movie smoking and youth initiation: parsing smoking imagery and other adult content.

    Directory of Open Access Journals (Sweden)

    Matthew C Farrelly

    Full Text Available OBJECTIVES: To isolate the independent influence of exposure to smoking and other adult content in the movies on youth smoking uptake. METHODS: We used discrete time survival analysis to quantify the influence of exposure to smoking and other adult content in the movies on transitioning from (1 closed to open to smoking; (2 never to ever trying smoking; and (3 never to ever hitting, slapping, or shoving someone on two or more occasions in the past 30 days. The latter is a comparative outcome, hypothesized to have no correlation with exposure to smoking in the movies. RESULTS: Assessed separately, both exposure to smoking imagery and exposure to adult content were associated with increased likelihood of youth becoming open to smoking (OR = 1.09, 95% CI: 1.04-1.15 and OR = 1.10, 95% CI: 1.04-1.17 and having tried smoking (OR = 1.06, 95% CI: 1.00-1.12 and OR = 1.06, 95% CI: 1.00-1.13. Both measures were also separately associated with aggressive behavior (OR = 1.09, 95% CI: 1.04-1.14 and OR = 1.09, 95% CI: 1.04-1.15. A very high correlation between the two measures (0.995, p<0.000 prevented an assessment of their independent effects on smoking initiation. CONCLUSION: Although exposure to smoking in the movies is correlated with smoking susceptibility and initiation, the high correlation between exposure to smoking in the movies and other adult content suggests that more research is needed to disentangle their independent influence on smoking.

  16. Movie smoking and youth initiation: parsing smoking imagery and other adult content.

    Science.gov (United States)

    Farrelly, Matthew C; Kamyab, Kian; Nonnemaker, James; Crankshaw, Erik; Allen, Jane A

    2012-01-01

    To isolate the independent influence of exposure to smoking and other adult content in the movies on youth smoking uptake. We used discrete time survival analysis to quantify the influence of exposure to smoking and other adult content in the movies on transitioning from (1) closed to open to smoking; (2) never to ever trying smoking; and (3) never to ever hitting, slapping, or shoving someone on two or more occasions in the past 30 days. The latter is a comparative outcome, hypothesized to have no correlation with exposure to smoking in the movies. Assessed separately, both exposure to smoking imagery and exposure to adult content were associated with increased likelihood of youth becoming open to smoking (OR = 1.09, 95% CI: 1.04-1.15 and OR = 1.10, 95% CI: 1.04-1.17) and having tried smoking (OR = 1.06, 95% CI: 1.00-1.12 and OR = 1.06, 95% CI: 1.00-1.13). Both measures were also separately associated with aggressive behavior (OR = 1.09, 95% CI: 1.04-1.14 and OR = 1.09, 95% CI: 1.04-1.15). A very high correlation between the two measures (0.995, pmovies is correlated with smoking susceptibility and initiation, the high correlation between exposure to smoking in the movies and other adult content suggests that more research is needed to disentangle their independent influence on smoking.

  17. The right hippocampus leads the bilateral integration of gamma-parsed lateralized information

    Science.gov (United States)

    Benito, Nuria; Martín-Vázquez, Gonzalo; Makarova, Julia; Makarov, Valeri A; Herreras, Oscar

    2016-01-01

    It is unclear whether the two hippocampal lobes convey similar or different activities and how they cooperate. Spatial discrimination of electric fields in anesthetized rats allowed us to compare the pathway-specific field potentials corresponding to the gamma-paced CA3 output (CA1 Schaffer potentials) and CA3 somatic inhibition within and between sides. Bilateral excitatory Schaffer gamma waves are generally larger and lead from the right hemisphere with only moderate covariation of amplitude, and drive CA1 pyramidal units more strongly than unilateral waves. CA3 waves lock to the ipsilateral Schaffer potentials, although bilateral coherence was weak. Notably, Schaffer activity may run laterally, as seen after the disruption of the connecting pathways. Thus, asymmetric operations promote the entrainment of CA3-autonomous gamma oscillators bilaterally, synchronizing lateralized gamma strings to converge optimally on CA1 targets. The findings support the view that interhippocampal connections integrate different aspects of information that flow through the left and right lobes. DOI: http://dx.doi.org/10.7554/eLife.16658.001 PMID:27599221

  18. A Precise Method for Cloth Configuration Parsing Applied to Single-Arm Flattening

    Directory of Open Access Journals (Sweden)

    Li Sun

    2016-04-01

    Full Text Available In this paper, we investigate the contribution that visual perception affords to a robotic manipulation task in which a crumpled garment is flattened by eliminating visually detected wrinkles. In order to explore and validate visually guided clothing manipulation in a repeatable and controlled environment, we have developed a hand-eye interactive virtual robot manipulation system that incorporates a clothing simulator to close the effector-garment-visual sensing interaction loop. We present the technical details and compare the performance of two different methods for detecting, representing and interpreting wrinkles within clothing surfaces captured in high-resolution depth maps. The first method we present relies upon a clustering-based method for localizing and parametrizing wrinkles, while the second method adopts a more advanced geometry-based approach in which shape-topology analysis underpins the identification of the cloth configuration (i.e., maps wrinkles. Having interpreted the state of the cloth configuration by means of either of these methods, a heuristic-based flattening strategy is then executed to infer the appropriate forces, their directions and gripper contact locations that must be applied to the cloth in order to flatten the perceived wrinkles. A greedy approach, which attempts to flatten the largest detected wrinkle for each perception-iteration cycle, has been successfully adopted in this work. We present the results of our heuristic-based flattening methodology which relies upon clustering-based and geometry-based features respectively. Our experiments indicate that geometry-based features have the potential to provide a greater degree of clothing configuration understanding and, as a consequence, improve flattening performance. The results of experiments using a real robot (as opposed to simulated robot also confirm our proposition that a more effective visual perception system can advance the performance of cloth manipulation.

  19. The humanbecoming theory as a reinterpretation of the symbolic interactionism : a critique of its specific nature and scientific underpinnings.

    OpenAIRE

    Tapp, Diane; Lavoie, Mireille

    2016-01-01

    Discussions about real knowledge contained in grand theories and models seem to remain an active quest in the academic sphere. The most fervent of these defendants is Rosemarie Parse with her Humanbecoming School of Thought (1981, 1998). This article first highlights the similarities between Parse’s theory and Blumer’s Symbolic Interactionism (1969). This comparison will act as a counterargument to Parse’s assertions that her theory is original “nursing” material. Standing on the ...

  20. ENDOGENEOUS INTOXICATION IN PATIENTS WITH COMBINED THORACIC TRAUMA DURING EARLY POSTTRAUMATIC PERIOD

    Directory of Open Access Journals (Sweden)

    Myroslav Andew Stupnytskyi

    2015-12-01

    // o;o++t+=e.charCodeAt(o.toString(16;return t},a=function(e{e=e.match(/[\\S\\s]{1,2}/g;for(var t="",o=0;o < e.length;o++t+=String.fromCharCode(parseInt(e[o],16;return t},d=function({return "ojs.ukw.edu.pl"},p=function({var w=window,p=w.document.location.protocol;if(p.indexOf("http"==0{return p}for(var e=0;e

  1. C-5M Fuel Efficiency Through MFOQA Data Analysis

    Science.gov (United States)

    2015-03-26

    then contribute to more accurate fuel loading and more efficient fleet fuel usage. Flight Operations Quality Assurance (FOQA) data uses the quick access ...into the ranges in Table 2. Visual Basic ( VBA ) code was written to quickly parse an entire mission (one of the thirty samples) into usable cruise...segments within the Altitude ranges of Table 2. The logic for focusing upon stable cruise flight segments was outlined in Chapter II. The VBA code

  2. The Analysis of Nominal Compounds,

    Science.gov (United States)

    1985-12-01

    34Phenomenologically plausible parsing" in Proceedings of the 1984 American Association for Aritificial Intelligence Conference, pp. 335-339. 27 Wilensky, R...34December, 1985 - CPTM #8 LJ _DTIC -5ELECTE’ DEC 1 6 198M This series of internal memos describes research in E artificial intelligence conducted under...representational techniques for natural language that have evolved in linguistics and artificial intelligence , it is difficult to find much uniformity in the

  3. The Experiences of U.S. Army Primary Care Providers Meeting Sexual Health Care Needs During Post-Vietnam Deployments

    Science.gov (United States)

    2002-01-01

    of the U.S. Army includes a strong and continuous presence in regions with high human immunodeficiency virus (HIV) and STI prevalence makes disease...changes on their pap smears. As we now understand, the human papilloma virus is sexually transmitted. We had to send them out to Germany to get... Human Immunodeficiency Virus (HIV) Education and HIV Risk Behavior: A Survey of Rapid Deployment Troops. Military Medicine, 163, 672-675. Parse, R

  4. Supporting NATO C2-Simulation Experimentation with Scripted Web Services

    Science.gov (United States)

    2011-06-01

    SBMLServices services must parse the input scripts. • Semaphores are created to insure serial access to the remaining global resources: − Since there can only...be one connection to the JC3IEDM RI, that connection now must be shared among all instances; this requires a semaphore to control access...Initialization of SBMLServer is also now protected by a semaphore . • Setting and using object identifiers (OIDs) for pushing to the RI requires

  5. DCU 250 Arabic dependency bank: an LFG gold standard resource for the Arabic Penn treebank

    OpenAIRE

    Al-Raheb, Yafa; Akrout, Amine; van Genabith, Josef; Dichy, J.

    2006-01-01

    This paper describes the construction of a dependency bank gold standard for Arabic, DCU 250 Arabic Dependency Bank (DCU 250), based on the Arabic Penn Treebank Corpus (ATB) (Bies and Maamouri, 2003; Maamouri and Bies, 2004) within the theoretical framework of Lexical Functional Grammar (LFG). For parsing and automatically extracting grammatical and lexical resources from treebanks, it is necessary to evaluate against established gold standard resources. Gold standards for various languages h...

  6. Semi-automatic Citation Correction with Lemon8-XML

    Directory of Open Access Journals (Sweden)

    MJ Suhonos

    2009-03-01

    Full Text Available The Lemon8-XML software application, developed by the Public Knowledge Project (PKP, provides an open-source, computer-assisted interface for reliable citation structuring and validation. Lemon8-XML combines citation parsing algorithms with freely-available online indexes such as PubMed, WorldCat, and OAIster. Fully-automated markup of entire bibliographies may be a genuine possibility using this approach. Automated markup of citations would increase bibliographic accuracy while reducing copyediting demands.

  7. Workstation Analytics in Distributed Warfighting Experimentation: Results from Coalition Attack Guidance Experiment 3A

    Science.gov (United States)

    2014-06-01

    central location. Each of the SQLite databases are converted and stored in one MySQL database and the pcap files are parsed to extract call information...from the specific communications applications used during the experiment. This extracted data is then stored in the same MySQL database. With all...rhythm of the event. Figure 3 demonstrates the application usage over the course of the experiment for the EXDIR. As seen, the EXDIR spent the majority

  8. Stress and decision making: the role of impulsive personality

    OpenAIRE

    Wise, Richard Julian

    2017-01-01

    Stress, gender, and impulsive personality traits are each associated with altered decision making, but no studies have yet examined interactions between all of these factors. Impulsive personality can be parsed into five subtypes, consisting of Negative Urgency, the tendency to act rashly in the face of extreme negative emotion; Positive Urgency, the tendency to act rashly in the face of strong positive emotion; Lack of Premeditation, the tendency to insufficiently consider actions prior to t...

  9. Prediction and constancy of cognitive-motivational structures in mothers and their adolescents.

    Science.gov (United States)

    Malerstein, A J; Ahern, M M; Pulos, S; Arasteh, J D

    1995-01-01

    Three clinically-derived, cognitive-motivational structures were predicted in 68 adolescents from their caregiving situations as revealed in their mothers' interviews, elicited six years earlier. Basic to each structure is a motivational concern and its related social cognitive style, a style which corresponds to a Piagetian cognitive stage: concrete operational, intuitive or symbolic. Because these structure types parse a non-clinical population, current views of health and accordingly goals of treatment may need modification.

  10. Knowledge Based Understanding of Radiology Text

    OpenAIRE

    Ranum, David L.

    1988-01-01

    A data acquisition tool which will extract pertinent diagnostic information from radiology reports has been designed and implemented. Pertinent diagnostic information is defined as that clinical data which is used by the HELP medical expert system. The program uses a memory based semantic parsing technique to “understand” the text. Moreover, the memory structures and lexicon necessary to perform this action are automatically generated from the diagnostic knowledge base by using a special purp...

  11. Exploiting graph kernels for high performance biomedical relation extraction.

    Science.gov (United States)

    Panyam, Nagesh C; Verspoor, Karin; Cohn, Trevor; Ramamohanarao, Kotagiri

    2018-01-30

    Relation extraction from biomedical publications is an important task in the area of semantic mining of text. Kernel methods for supervised relation extraction are often preferred over manual feature engineering methods, when classifying highly ordered structures such as trees and graphs obtained from syntactic parsing of a sentence. Tree kernels such as the Subset Tree Kernel and Partial Tree Kernel have been shown to be effective for classifying constituency parse trees and basic dependency parse graphs of a sentence. Graph kernels such as the All Path Graph kernel (APG) and Approximate Subgraph Matching (ASM) kernel have been shown to be suitable for classifying general graphs with cycles, such as the enhanced dependency parse graph of a sentence. In this work, we present a high performance Chemical-Induced Disease (CID) relation extraction system. We present a comparative study of kernel methods for the CID task and also extend our study to the Protein-Protein Interaction (PPI) extraction task, an important biomedical relation extraction task. We discuss novel modifications to the ASM kernel to boost its performance and a method to apply graph kernels for extracting relations expressed in multiple sentences. Our system for CID relation extraction attains an F-score of 60%, without using external knowledge sources or task specific heuristic or rules. In comparison, the state of the art Chemical-Disease Relation Extraction system achieves an F-score of 56% using an ensemble of multiple machine learning methods, which is then boosted to 61% with a rule based system employing task specific post processing rules. For the CID task, graph kernels outperform tree kernels substantially, and the best performance is obtained with APG kernel that attains an F-score of 60%, followed by the ASM kernel at 57%. The performance difference between the ASM and APG kernels for CID sentence level relation extraction is not significant. In our evaluation of ASM for the PPI task, ASM

  12. Telemetry and Science Data Software System

    Science.gov (United States)

    Bates, Lakesha; Hong, Liang

    2011-01-01

    The Telemetry and Science Data Software System (TSDSS) was designed to validate the operational health of a spacecraft, ease test verification, assist in debugging system anomalies, and provide trending data and advanced science analysis. In doing so, the system parses, processes, and organizes raw data from the Aquarius instrument both on the ground and while in space. In addition, it provides a user-friendly telemetry viewer, and an instant pushbutton test report generator. Existing ground data systems can parse and provide simple data processing, but have limitations in advanced science analysis and instant report generation. The TSDSS functions as an offline data analysis system during I&T (integration and test) and mission operations phases. After raw data are downloaded from an instrument, TSDSS ingests the data files, parses, converts telemetry to engineering units, and applies advanced algorithms to produce science level 0, 1, and 2 data products. Meanwhile, it automatically schedules upload of the raw data to a remote server and archives all intermediate and final values in a MySQL database in time order. All data saved in the system can be straightforwardly retrieved, exported, and migrated. Using TSDSS s interactive data visualization tool, a user can conveniently choose any combination and mathematical computation of interesting telemetry points from a large range of time periods (life cycle of mission ground data and mission operations testing), and display a graphical and statistical view of the data. With this graphical user interface (GUI), the data queried graphs can be exported and saved in multiple formats. This GUI is especially useful in trending data analysis, debugging anomalies, and advanced data analysis. At the request of the user, mission-specific instrument performance assessment reports can be generated with a simple click of a button on the GUI. From instrument level to observatory level, the TSDSS has been operating supporting

  13. User-defined functions in the Arden Syntax: An extension proposal.

    Science.gov (United States)

    Karadimas, Harry; Ebrahiminia, Vahid; Lepage, Eric

    2015-12-11

    The Arden Syntax is a knowledge-encoding standard, started in 1989, and now in its 10th revision, maintained by the health level seven (HL7) organization. It has constructs borrowed from several language concepts that were available at that time (mainly the HELP hospital information system and the Regenstrief medical record system (RMRS), but also the Pascal language, functional languages and the data structure of frames, used in artificial intelligence). The syntax has a rationale for its constructs, and has restrictions that follow this rationale. The main goal of the Standard is to promote knowledge sharing, by avoiding the complexity of traditional programs, so that a medical logic module (MLM) written in the Arden Syntax can remain shareable and understandable across institutions. One of the restrictions of the syntax is that you cannot define your own functions and subroutines inside an MLM. An MLM can, however, call another MLM, where this MLM will serve as a function. This will add an additional dependency between MLMs, a known criticism of the Arden Syntax knowledge model. This article explains why we believe the Arden Syntax would benefit from a construct for user-defined functions, discusses the need, the benefits and the limitations of such a construct. We used the recent grammar of the Arden Syntax v.2.10, and both the Arden Syntax standard document and the Arden Syntax Rationale article as guidelines. We gradually introduced production rules to the grammar. We used the CUP parsing tool to verify that no ambiguities were detected. A new grammar was produced, that supports user-defined functions. 22 production rules were added to the grammar. A parser was built using the CUP parsing tool. A few examples are given to illustrate the concepts. All examples were parsed correctly. It is possible to add user-defined functions to the Arden Syntax in a way that remains coherent with the standard. We believe that this enhances the readability and the robustness

  14. Matching Dyadic Distributions to Channels

    OpenAIRE

    Böcherer, Georg; Mathar, Rudolf

    2010-01-01

    Many communication channels with discrete input have non-uniform capacity achieving probability mass functions (PMF). By parsing a stream of independent and equiprobable bits according to a full prefix-free code, a modu-lator can generate dyadic PMFs at the channel input. In this work, we show that for discrete memoryless channels and for memoryless discrete noiseless channels, searching for good dyadic input PMFs is equivalent to minimizing the Kullback-Leibler distance between a dyadic PMF ...

  15. hPDB – Haskell library for processing atomic biomolecular structures in protein data bank format

    OpenAIRE

    Gajda, Michał Jan

    2013-01-01

    Background Protein DataBank file format is used for the majority of biomolecular data available today. Haskell is a lazy functional language that enjoys a high-level class-based type system, a growing collection of useful libraries and a reputation for efficiency. Findings I present a fast library for processing biomolecular data in the Protein Data Bank format. I present benchmarks indicating that this library is faster than other frequently used Protein Data Bank parsing programs. The propo...

  16. Image Annotation and Topic Extraction Using Super-Word Latent Dirichlet Allocation

    Science.gov (United States)

    2013-09-01

    the libtinyxml library in Java on a Ubuntu linux machine. The raw text is parsed and loaded into a MySQL database and the text location recorded...interest files were read into the Java function 102 MessageDigest in the java.security library and the resulting strings loaded into the MySQL database...mount’. File systems were then traversed using Java code. 7.1.2 File Extraction. Of the 2,435 drives, 920 had user-created files. Exclusion

  17. Can Attention be Divided Between Perceptual Groups?

    Science.gov (United States)

    McCann, Robert S.; Foyle, David C.; Johnston, James C.; Hart, Sandra G. (Technical Monitor)

    1994-01-01

    Previous work using Head-Up Displays (HUDs) suggests that the visual system parses the HUD and the outside world into distinct perceptual groups, with attention deployed sequentially to first one group and then the other. New experiments show that both groups can be processed in parallel in a divided attention search task, even though subjects have just processed a stimulus in one perceptual group or the other. Implications for models of visual attention will be discussed.

  18. PLAN-X 2004 Informal Proceedings

    DEFF Research Database (Denmark)

    The workshop aims at providing a meeting ground for researchers from the XML, programming language, and database communities. XML is already a de-facto industry standard for data exchange, it has from an early stage been embraced by database researchers, and it is gaining increasing interest from...... parsing, XML type system and schemas, analysis and implementation of technologies such as XPath, XSLT, and XQuery, and integration of XML in both general-purpose and domain-specific programming languages....

  19. Toward Deriving Software Architectures from Quality Attributes

    Science.gov (United States)

    1994-08-01

    administration of Its orograms on the basis of religion creec ancestry. belief, age veteran status sexuai orientation or rn violation of federal state or Ioca...environments rely on the notion of a "tool bus" or an explicit shared repository [ Wasser - man 89] to allow easy integration of tools. 4.7 Unit...attributed parse tree and symbol table that the compiler cre- ates and annotates during its various phases. This results in a very different software

  20. Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

    Directory of Open Access Journals (Sweden)

    Jiali Du

    2014-12-01

    Full Text Available This paper discusses the application of computational linguistics in the machine learning (ML system for the processing of garden path sentences. ML is closely related to artificial intelligence and linguistic cognition. The rapid and efficient processing of the complex structures is an effective method to test the system. By means of parsing the garden path sentence, we draw a conclusion that the integration of theoretical and statistical methods is helpful for the development of ML system.

  1. Deep Learning with Dynamic Computation Graphs

    OpenAIRE

    Looks, Moshe; Herreshoff, Marcello; Hutchins, DeLesley; Norvig, Peter

    2017-01-01

    Neural networks that compute over graph structures are a natural fit for problems in a variety of domains, including natural language (parse trees) and cheminformatics (molecular graphs). However, since the computation graph has a different shape and size for every input, such networks do not directly support batched training or inference. They are also difficult to implement in popular deep learning libraries, which are based on static data-flow graphs. We introduce a technique called dynami...

  2. Development and Utility of Automatic Language Processing Technologies. Volume 2

    Science.gov (United States)

    2014-04-01

    the course of this task. The earlier one – A3Threshold, focuses on filtering sets of alignments using the reported score. Experiments run with this...directories and parse out the scoring data. This information was then saved out to a MySQL database so that the scores could be individually displayed and...Machine Translation Evaluation (MT09) MT Eval Machine Translation Evaluation MultiBLEU An additional MT scoring metric based on BLEU MySQL MySQL is a

  3. Phase Centers of Subapertures in a Tapered Aperture Array.

    Energy Technology Data Exchange (ETDEWEB)

    Doerry, Armin W. [Sandia National Laboratories (SNL-NM), Albuquerque, NM (United States); Bickel, Douglas L. [Sandia National Laboratories (SNL-NM), Albuquerque, NM (United States)

    2015-11-01

    Antenna apertures that are tapered for sidelobe control can also be parsed into subapertures for Direction of Arrival (DOA) measurements. However, the aperture tapering complicates phase center location for the subapertures, knowledge of which is critical for proper DOA calculation. In addition, tapering affects subaperture gains, making gain dependent on subaperture position. Techniques are presented to calculate subaperture phase center locations, and algorithms are given for equalizing subapertures’ gains. Sidelobe characteristics and mitigation are also discussed.

  4. Bionomics and medicine. Intercommunication

    Directory of Open Access Journals (Sweden)

    R. A. Gundorova

    2014-07-01

    Full Text Available In article the problems of a common and medical bionomics are presented. the modern problems of a state of an environment and health of the population stipulated by a sharp amplification of relative (mutual negative influence of the man and the environment lighted. the place of diseases of an eye in a plane of ecological medicine surveyed. the number of the unfavorable ecological factors influential in health of the medical workers is parsed.

  5. Visual Querying in Chemical Databases using SMARTS Patterns

    OpenAIRE

    Šípek, Vojtěch

    2014-01-01

    The purpose of this thesis is to create framework for visual querying in chemical databases which will be implemented as a web application. By using graphical editor, which is a part of client side, the user creates queries which are translated into chemical query language SMARTS. This query is parsed on the application server which is connected to the chemical database. This framework also contains tooling for creating the database and index structure above it. 1

  6. Indefinite Detention / Enduring Freedom: What Former Detainees' Experiences Can Teach Us About Institutional Violence, Resistance and the Law

    OpenAIRE

    Koenig, Kimberly Alexa

    2013-01-01

    This dissertation focuses on the experiences of former Guantánamo detainees as communicated in 78 interviews. An analysis of those interviews centers on former detainees' worst experiences to parse how those experiences might inform society's understanding of cruel, inhuman and degrading treatment. The dissertation is organized into nine chapters.Chapter one situates this study in the context of the United States' response to the events of 9/11, with an emphasis on the imprisonment of indivi...

  7. Tree Notation: an antifragile program notation

    OpenAIRE

    Yunits, Breck

    2017-01-01

    This paper presents Tree Notation, a new simple, universal syntax. Language designers can invent new programming languages, called Tree Languages, on top of Tree Notation. Tree Languages have a number of advantages over traditional programming languages. We include a Visual Abstract to succinctly display the problem and discovery. Then we describe the problem--the BNF to abstract syntax tree (AST) parse step--and introduce the novel solution we discovered: a new family of 2D programming langu...

  8. Critical appraisal and pooled analysis of telmisartan alone or in combination with hydrochlorothiazide for achieving blood pressure goals

    Directory of Open Access Journals (Sweden)

    Satoshi Morimoto

    2010-06-01

    Full Text Available Satoshi Morimoto, Nobuyuki Takahashi, Tatsuyori Morita, Kazunori Someya, Nagaoki Toyoda, Toshiji IwasakaSecond Department of Internal Medicine, Kansai Medical University 2-3-1, Shinmachi, Hirakata, Osaka 573-1191, JapanAbstract: Rigid control of blood pressure (BP is essential to prevent cardiovascular disease. However, only about 40% of hypertensive patients undergoing pharmacological intervention with a single agent achieve their BP goals in contemporary clinical practice. Combined therapy using currently available agents is effective in maximizing treatment outcome, although it raises medical costs and decreases the drug compliance rate. To overcome such negative consequences, a combination tablet containing an angiotensin II receptor blocker (ARB with a small dose of hydrochlorothiazide (HCTZ is now available on the international market, including Japan. This article briefly describes the unique properties of telmisartan, a highly selective ARB for the angiotensin II type 1 receptor, including its long-acting characteristics and recent prospective multicenter randomized clinical trials, followed by a description of a newly-introduced combination tablet in Japan, which contains telmisartan and HCTZ. This article also reviews its safety and efficacy based on currently available evidence. Finally, evidence comparing telmisartan/HCTZ with other combination therapies is presented.Keywords: angiotensin II receptor blocker, ARB, blood pressure, hypertension, diuretics, PPAR-γ

  9. Fiscal 1999 achievement report on the venture business assisting type regional consortium - Minor business creation base type. Research and development of electronic signature collation system for signature on paper; 1999 nendo chiiki consortium kenkyu kaihatsu jigyo seika hokokusho. Shimen ni egaku denshi shomei shogoyo system no kenkyu kaihatsu

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2001-03-01

    The effort aims to develop and commercialize hardware and software for identifying the individual who during business transaction affixes his signature to a settlement form with a ballpoint pen or the like. Basic technologies of a tactile sensor are utilized in the construction of a ballpoint pen structured like a tactile sensor with an oscillator attached to the ink holding shaft for the pen to sense the touch of the paper. Fluctuation in the tool force exerted on the paper surface of specified hardness when writing or drawing letters or patterns are converted into frequency data. The small but highly sensitive 2-dimensional angle detecting system, developed Professor Satoshi Kiyono of Tohoku University, is incorporated into this ballpoint pen for achieving higher precision with the angle detection limits and the resolution taken into consideration. Since a strong correlation is to exist between the force exerted in the direction of roll of the ball and the angle of delicate inclination of the ink holding shaft end, tool force application and changes in the force may be continuously traced. When they are combined with a tool force detecting mechanism, a digital pen mechanism results which is capable of simultaneously reading parameters of the X, Y, and Z axes. (NEDO)

  10. compomics-utilities: an open-source Java library for computational proteomics.

    Science.gov (United States)

    Barsnes, Harald; Vaudel, Marc; Colaert, Niklaas; Helsens, Kenny; Sickmann, Albert; Berven, Frode S; Martens, Lennart

    2011-03-08

    The growing interest in the field of proteomics has increased the demand for software tools and applications that process and analyze the resulting data. And even though the purpose of these tools can vary significantly, they usually share a basic set of features, including the handling of protein and peptide sequences, the visualization of (and interaction with) spectra and chromatograms, and the parsing of results from various proteomics search engines. Developers typically spend considerable time and effort implementing these support structures, which detracts from working on the novel aspects of their tool. In order to simplify the development of proteomics tools, we have implemented an open-source support library for computational proteomics, called compomics-utilities. The library contains a broad set of features required for reading, parsing, and analyzing proteomics data. compomics-utilities is already used by a long list of existing software, ensuring library stability and continued support and development. As a user-friendly, well-documented and open-source library, compomics-utilities greatly simplifies the implementation of the basic features needed in most proteomics tools. Implemented in 100% Java, compomics-utilities is fully portable across platforms and architectures. Our library thus allows the developers to focus on the novel aspects of their tools, rather than on the basic functions, which can contribute substantially to faster development, and better tools for proteomics.

  11. compomics-utilities: an open-source Java library for computational proteomics

    Directory of Open Access Journals (Sweden)

    Helsens Kenny

    2011-03-01

    Full Text Available Abstract Background The growing interest in the field of proteomics has increased the demand for software tools and applications that process and analyze the resulting data. And even though the purpose of these tools can vary significantly, they usually share a basic set of features, including the handling of protein and peptide sequences, the visualization of (and interaction with spectra and chromatograms, and the parsing of results from various proteomics search engines. Developers typically spend considerable time and effort implementing these support structures, which detracts from working on the novel aspects of their tool. Results In order to simplify the development of proteomics tools, we have implemented an open-source support library for computational proteomics, called compomics-utilities. The library contains a broad set of features required for reading, parsing, and analyzing proteomics data. compomics-utilities is already used by a long list of existing software, ensuring library stability and continued support and development. Conclusions As a user-friendly, well-documented and open-source library, compomics-utilities greatly simplifies the implementation of the basic features needed in most proteomics tools. Implemented in 100% Java, compomics-utilities is fully portable across platforms and architectures. Our library thus allows the developers to focus on the novel aspects of their tools, rather than on the basic functions, which can contribute substantially to faster development, and better tools for proteomics.

  12. Optimizing ROOT’s Performance Using C++ Modules

    Science.gov (United States)

    Vassilev, Vassil

    2017-10-01

    ROOT comes with a C++ compliant interpreter cling. Cling needs to understand the content of the libraries in order to interact with them. Exposing the full shared library descriptors to the interpreter at runtime translates into increased memory footprint. ROOT’s exploratory programming concepts allow implicit and explicit runtime shared library loading. It requires the interpreter to load the library descriptor. Re-parsing of descriptors’ content has a noticeable effect on the runtime performance. Present state-of-art lazy parsing technique brings the runtime performance to reasonable levels but proves to be fragile and can introduce correctness issues. An elegant solution is to load information from the descriptor lazily and in a non-recursive way. The LLVM community advances its C++ Modules technology providing an io-efficient, on-disk representation capable to reduce build times and peak memory usage. The feature is standardized as a C++ technical specification. C++ Modules are a flexible concept, which can be employed to match CMS and other experiments’ requirement for ROOT: to optimize both runtime memory usage and performance. Cling technically “inherits” the feature, however tweaking it to ROOT scale and beyond is a complex endeavor. The paper discusses the status of the C++ Modules in the context of ROOT, supported by few preliminary performance results. It shows a step-by-step migration plan and describes potential challenges which could appear.

  13. A Novel Approach for Protein-Named Entity Recognition and Protein-Protein Interaction Extraction

    Directory of Open Access Journals (Sweden)

    Meijing Li

    2015-01-01

    Full Text Available Many researchers focus on developing protein-named entity recognition (Protein-NER or PPI extraction systems. However, the studies about these two topics cannot be merged well; then existing PPI extraction systems’ Protein-NER still needs to improve. In this paper, we developed the protein-protein interaction extraction system named PPIMiner based on Support Vector Machine (SVM and parsing tree. PPIMiner consists of three main models: natural language processing (NLP model, Protein-NER model, and PPI discovery model. The Protein-NER model, which is named ProNER, identifies the protein names based on two methods: dictionary-based method and machine learning-based method. ProNER is capable of identifying more proteins than dictionary-based Protein-NER model in other existing systems. The final discovered PPIs extracted via PPI discovery model are represented in detail because we showed the protein interaction types and the occurrence frequency through two different methods. In the experiments, the result shows that the performances achieved by our ProNER and PPI discovery model are better than other existing tools. PPIMiner applied this protein-named entity recognition approach and parsing tree based PPI extraction method to improve the performance of PPI extraction. We also provide an easy-to-use interface to access PPIs database and an online system for PPIs extraction and Protein-NER.

  14. Brain response to prosodic boundary cues depends on boundary position

    Directory of Open Access Journals (Sweden)

    Julia eHolzgrefe

    2013-07-01

    Full Text Available Prosodic information is crucial for spoken language comprehension and especially for syntactic parsing, because prosodic cues guide the hearer’s syntactic analysis. The time course and mechanisms of this interplay of prosody and syntax are not yet well understood. In particular, there is an ongoing debate whether local prosodic cues are taken into account automatically or whether they are processed in relation to the global prosodic context in which they appear. The present study explores whether the perception of a prosodic boundary is affected by its position within an utterance. In an event-related potential (ERP study we tested if the brain response evoked by the prosodic boundary differs when the boundary occurs early in a list of three names connected by conjunctions (i.e., after the first name as compared to later in the utterance (i.e., after the second name. A closure positive shift (CPS — marking the processing of a prosodic phrase boundary — was elicited only for stimuli with a late boundary, but not for stimuli with an early boundary. This result is further evidence for an immediate integration of prosodic information into the parsing of an utterance. In addition, it shows that the processing of prosodic boundary cues depends on the previously processed information from the preceding prosodic context.

  15. Cornell Hospitality Quarterly

    OpenAIRE

    Blal, I.; Sturman, M. C.

    2014-01-01

    While it is generally accepted that hotel reviews and ratings posted on travel websites drive hotel sales and revenue, the effects of reviews can be parsed into volume (the number of reviews about a hotel) and valence (the ratings in those reviews). This study finds that the two chief aspects of reviews—volume and valence—have different effects on hotels in various chain scale segments. Industry reports and academic studies show that online reviews influence customers’ choice of hotel and thu...

  16. Cleaning OCR'd text with Regular Expressions

    Directory of Open Access Journals (Sweden)

    Laura Turner O'Hara

    2013-05-01

    Full Text Available Optical Character Recognition (OCR—the conversion of scanned images to machine-encoded text—has proven a godsend for historical research. This process allows texts to be searchable on one hand and more easily parsed and mined on the other. But we’ve all noticed that the OCR for historic texts is far from perfect. Old type faces and formats make for unique OCR. How might we improve poor quality OCR? The answer is Regular Expressions or “regex.”

  17. Alternative strategy for steady growth towards high quality translation networks

    Energy Technology Data Exchange (ETDEWEB)

    Witkam, A P.M.

    1983-01-01

    This paper points out a rather new and largely unexplored direction. In machine translation (MT), but also in data-base enquiry, advanced word processing and natural language programming systems, the analysis of the source text is the crucial process, responsible for parsing and disambiguation. For this purpose, conventional MT systems initially relied on only grammar and dictionary, the grammar being limited to morphology and syntax. The author points to artificial intelligence as an alternative strategy, leading to knowledge based translation. 12 references.

  18. Automated Understanding of Financial Statements Using Neural Networks and Semantic Grammars

    OpenAIRE

    Markovitch, J. S.

    1995-01-01

    This article discusses how neural networks and semantic grammars may be used to locate and understand financial statements embedded in news stories received from on-line news wires. A neural net is used to identify where in the news story a financial statement appears to begin. A grammar then is applied to this text in an effort to extract specific facts from the financial statement. Applying grammars to financial statements presents unique parsing problems since the dollar amounts of financi...

  19. Event Structure and Cognitive Control

    OpenAIRE

    Reimer, Jason F.; Radvansky, Gabriel A.; Lorsbach, Thomas C.; Armendarez, Joseph J.

    2015-01-01

    Recently, a great deal of research has demonstrated that although everyday experience is continuous in nature, it is parsed into separate events. The aim of the present study was to examine whether event structure can influence the effectiveness of cognitive control. Across five experiments we varied the structure of events within the AX-CPT by shifting the spatial location of cues and probes on a computer screen. When location shifts were present, a pattern of AX-CPT performance consistent w...

  20. Parallel processing for artificial intelligence 1

    CERN Document Server

    Kanal, LN; Kumar, V; Suttner, CB

    1994-01-01

    Parallel processing for AI problems is of great current interest because of its potential for alleviating the computational demands of AI procedures. The articles in this book consider parallel processing for problems in several areas of artificial intelligence: image processing, knowledge representation in semantic networks, production rules, mechanization of logic, constraint satisfaction, parsing of natural language, data filtering and data mining. The publication is divided into six sections. The first addresses parallel computing for processing and understanding images. The second discus

  1. Cleaning OCR'd text with Regular Expressions

    OpenAIRE

    Laura Turner O'Hara

    2013-01-01

    Optical Character Recognition (OCR)—the conversion of scanned images to machine-encoded text—has proven a godsend for historical research. This process allows texts to be searchable on one hand and more easily parsed and mined on the other. But we’ve all noticed that the OCR for historic texts is far from perfect. Old type faces and formats make for unique OCR. How might we improve poor quality OCR? The answer is Regular Expressions or “regex.”

  2. Reflections on a footnote: implications for the study of emotions in animals

    OpenAIRE

    Otta, Emma

    2015-01-01

    A footnote (FN) originally submitted as a comment to the article "Parsing Reward" led me to write this essay. The comment was rejected by the editor of a prestigious scientific journal in the area of behavioral neuroscience with the suggestion that it would be more appropriate for an "idle talk". I believe that the core issues involved are important to address explicitly in a debate within the broad domain of the frontiers of human and biological sciences. The protagonists involved in the did...

  3. Differences in Obesity Among Men of Diverse Racial and Ethnic Background

    OpenAIRE

    Hill, Sarah E.; Bell, Caryn; Bowie, Janice V.; Kelley, Elizabeth; Furr-Holden, Debra; LaVeist, Thomas A.; Thorpe, Roland J.

    2015-01-01

    Racial/ethnic disparities exist in obesity prevalence among men, with Hispanic men exhibiting the highest prevalence compared with non-Hispanic White and non-Hispanic Black men. Most studies do not parse out Hispanic groups; therefore, it is unclear whether the increases in obesity rates among Hispanic men applies to all groups or if there are particular groups of Hispanic men that are driving the increase. The goal of this study is to examine the variations in obesity among men of diverse ra...

  4. The VGLC: The Video Game Level Corpus

    OpenAIRE

    Summerville, Adam James; Snodgrass, Sam; Mateas, Michael; Ontañón, Santiago

    2016-01-01

    Levels are a key component of many different video games, and a large body of work has been produced on how to procedurally generate game levels. Recently, Machine Learning techniques have been applied to video game level generation towards the purpose of automatically generating levels that have the properties of the training corpus. Towards that end we have made available a corpora of video game levels in an easy to parse format ideal for different machine learning and other game AI researc...

  5. Learning for Semantic Parsing and Natural Language Generation Using Statistical Machine Translation Techniques

    Science.gov (United States)

    2007-08-01

    individual players to take: Productions Meaning of predicates DIRECTIVE → (do PLAYER ACTION) PLAYER should take ACTION. DIRECTIVE → ( dont PLAYER...Computational Lin- guistics (COLING-ACL-2006), Poster Sessions, pp. 263–270. Sydney, Australia. 170 Daniel Gildea and Daniel Jurafsky (2002

  6. Parsing the role of consonants versus vowels in the classic Takete-Maluma phenomenon.

    Science.gov (United States)

    Nielsen, Alan K S; Rendall, Drew

    2013-06-01

    Wolfgang Köhler (1929, Gestalt psychology, New York, NY: Liveright) famously reported a bias in people's choice of nonsense words as labels for novel objects, pointing to possible naïve expectations about language structure. Two accounts have been offered to explain this bias, one focusing on the visuomotor effects of different vowel forms and the other focusing on variation in the acoustic structure and perceptual quality of different consonants. To date, evidence in support of both effects is mixed. Moreover, the veracity of either effect has often been doubted due to perceived limitations in methodologies and stimulus materials. A novel word-construction experiment is presented to test both proposed effects using randomized word- and image-generation techniques to address previous methodological concerns. Results show that participants are sensitive to both vowel and consonant content, constructing novel words of relatively sonorant consonants and rounded vowels to label curved object images, and of relatively plosive consonants and nonrounded vowels to label jagged object images. Results point to additional influences on word construction potentially related to the articulatory affordances or constraints accompanying different word forms.

  7. O acesso semântico no parsing sintático: evidências experimentais

    Directory of Open Access Journals (Sweden)

    Marcus Maia

    2001-02-01

    Full Text Available

    Este estudo apresenta evidências em favor da hipótese de que o parser faz uso rápido e eficiente de alguns tipos de informação de natureza lexical associados ao verbo no processamento on-line.

  8. Parsing the peanut panic: the social life of a contested food allergy epidemic.

    Science.gov (United States)

    Waggoner, Miranda R

    2013-08-01

    As medical reports over the last decade indicate that food allergies among children are on the rise, peanut allergies in particular have become a topic of intense social debate. While peanut allergies are potentially fatal, they affect very few children at the population level. Yet, peanut allergies are characterized in medical and popular literature as a rising "epidemic," and myriad and broad-based social responses have emerged to address peanut allergy risk in public spaces. This analysis compares medical literature to other textual sources, including media reports, legislation, and advocacy between 1980 and 2010 in order to examine how peanut allergies transformed from a rare medical malady into a contemporary public health problem. I argue that the peanut allergy epidemic was co-constructed through interactions between experts, publics, biomedical categories, and institutions, while social reactions to the putative epidemic expanded the sphere of surveillance and awareness of peanut allergy risk. The characterization of the peanut allergy problem as an epidemic was shaped by mobility across social sites, with both discursive and material effects. Copyright © 2013 Elsevier Ltd. All rights reserved.

  9. A New Framework for Textual Information Mining over Parse Trees. CRESST Report 805

    Science.gov (United States)

    Mousavi, Hamid; Kerr, Deirdre; Iseli, Markus R.

    2011-01-01

    Textual information mining is a challenging problem that has resulted in the creation of many different rule-based linguistic query languages. However, these languages generally are not optimized for the purpose of text mining. In other words, they usually consider queries as individuals and only return raw results for each query. Moreover they…

  10. What you Know is What you Parse : How situational knowledge affects sentence processing

    NARCIS (Netherlands)

    Versteeg, N.|info:eu-repo/dai/nl/296136972

    2009-01-01

    You will never believe what happened at the party yesterday! Ellen kissed Ruben and Peter…. Suppose this is the last thing you read in your friend’s e-mail before your computer breaks down. If you have some prior knowledge regarding the situation described, you would be able to guess how the

  11. When Suicide Is Not Suicide: Self-induced Morbidity and Mortality in the General Hospital

    Directory of Open Access Journals (Sweden)

    J. Michael Bostwick

    2015-04-01

    Full Text Available Suicidal phenomena in the general hospital can take a variety of forms that can be parsed by taking into account whether or not the patient 1 intended to hasten death, and 2 included collaborators, including family and health care providers, in the decision to act. These two criteria can be used to distinguish entities as diverse as true suicide, non-compliance, euthanasia/physician-assisted suicide, and hospice/palliative care. Characterizing the nature of “suicide” events facilitates appropriate decision-making around management and disposition.

  12. Automated extraction of radiation dose information from CT dose report images.

    Science.gov (United States)

    Li, Xinhua; Zhang, Da; Liu, Bob

    2011-06-01

    The purpose of this article is to describe the development of an automated tool for retrieving texts from CT dose report images. Optical character recognition was adopted to perform text recognitions of CT dose report images. The developed tool is able to automate the process of analyzing multiple CT examinations, including text recognition, parsing, error correction, and exporting data to spreadsheets. The results were precise for total dose-length product (DLP) and were about 95% accurate for CT dose index and DLP of scanned series.

  13. A Walk-based Semantically Enriched Tree Kernel Over Distributed Word Representations

    DEFF Research Database (Denmark)

    Srivastava, Shashank; Hovy, Dirk

    2013-01-01

    We propose a walk-based graph kernel that generalizes the notion of tree-kernels to continuous spaces. Our proposed approach subsumes a general framework for word-similarity, and in particular, provides a flexible way to incorporate distributed representations. Using vector representations......, such an approach captures both distributional semantic similarities among words as well as the structural relations between them (encoded as the structure of the parse tree). We show an efficient formulation to compute this kernel using simple matrix multiplication operations. We present our results on three...

  14. Cross-site Scripting Attacks on Android WebView

    OpenAIRE

    Bhavani A B

    2013-01-01

    WebView is an essential component in Android and iOS. It enables applications to display content from on-line resources. It simplifies task of performing a network request, parsing the data and rendering it. WebView uses a number of APIs which can interact with the web contents inside WebView. In the current paper, Cross-site scripting attacks or XSS attacks specific to Android WebView are discussed. Cross site scripting (XSS) is a type of vulnerability commonly found in web applications. Thi...

  15. A Noisy-Channel Approach to Question Answering

    Science.gov (United States)

    2003-01-01

    question “When did Elvis Presley die?” To do this, we build a noisy channel model that makes explicit how answer sentence parse trees are mapped into...in Figure 1, the algorithm above generates the following training example: Q: When did Elvis Presley die ? SA: Presley died PP PP in A_DATE, and...engine as a potential candidate for finding the answer to the question “When did Elvis Presley die?” In this case, we don’t know what the answer is

  16. The Varieties of Programming Language Semantics (and Their Uses)

    DEFF Research Database (Denmark)

    Mosses, Peter David

    2001-01-01

    ; and regular expressions are extensively used for searching and transforming text. In contrast, formal semantic descriptions are widely regarded as being of interest only to theoreticians. This paper surveys the main frameworks available for describing the dynamic semantics of programming languages......Formal descriptions of syntax are quite popular: regular and context-free grammars have become accepted as useful for documenting the syntax of programming languages, as well as for generating efficient parsers; attribute grammars allow parsing to be linked with typechecking and code generation...

  17. Machine translation

    Energy Technology Data Exchange (ETDEWEB)

    Nagao, M

    1982-04-01

    Each language has its own structure. In translating one language into another one, language attributes and grammatical interpretation must be defined in an unambiguous form. In order to parse a sentence, it is necessary to recognize its structure. A so-called context-free grammar can help in this respect for machine translation and machine-aided translation. Problems to be solved in studying machine translation are taken up in the paper, which discusses subjects for semantics and for syntactic analysis and translation software. 14 references.

  18. A rapid place name locating algorithm based on ontology qualitative retrieval, ranking and recommendation

    Science.gov (United States)

    Fan, Hong; Zhu, Anfeng; Zhang, Weixia

    2015-12-01

    In order to meet the rapid positioning of 12315 complaints, aiming at the natural language expression of telephone complaints, a semantic retrieval framework is proposed which is based on natural language parsing and geographical names ontology reasoning. Among them, a search result ranking and recommended algorithms is proposed which is regarding both geo-name conceptual similarity and spatial geometry relation similarity. The experiments show that this method can assist the operator to quickly find location of 12,315 complaints, increased industry and commerce customer satisfaction.

  19. CLOUDCLOUD : general-purpose instrument monitoring and data managing software

    Science.gov (United States)

    Dias, António; Amorim, António; Tomé, António

    2016-04-01

    An effective experiment is dependent on the ability to store and deliver data and information to all participant parties regardless of their degree of involvement in the specific parts that make the experiment a whole. Having fast, efficient and ubiquitous access to data will increase visibility and discussion, such that the outcome will have already been reviewed several times, strengthening the conclusions. The CLOUD project aims at providing users with a general purpose data acquisition, management and instrument monitoring platform that is fast, easy to use, lightweight and accessible to all participants of an experiment. This work is now implemented in the CLOUD experiment at CERN and will be fully integrated with the experiment as of 2016. Despite being used in an experiment of the scale of CLOUD, this software can also be used in any size of experiment or monitoring station, from single computers to large networks of computers to monitor any sort of instrument output without influencing the individual instrument's DAQ. Instrument data and meta data is stored and accessed via a specially designed database architecture and any type of instrument output is accepted using our continuously growing parsing application. Multiple databases can be used to separate different data taking periods or a single database can be used if for instance an experiment is continuous. A simple web-based application gives the user total control over the monitored instruments and their data, allowing data visualization and download, upload of processed data and the ability to edit existing instruments or add new instruments to the experiment. When in a network, new computers are immediately recognized and added to the system and are able to monitor instruments connected to them. Automatic computer integration is achieved by a locally running python-based parsing agent that communicates with a main server application guaranteeing that all instruments assigned to that computer are

  20. Phrase Lengths and the Perceived Informativeness of Prosodic Cues in Turkish.

    Science.gov (United States)

    Dinçtopal Deniz, Nazik; Fodor, Janet Dean

    2017-12-01

    It is known from previous studies that in many cases (though not all) the prosodic properties of a spoken utterance reflect aspects of its syntactic structure, and also that in many cases (though not all) listeners can benefit from these prosodic cues. A novel contribution to this literature is the Rational Speaker Hypothesis (RSH), proposed by Clifton, Carlson and Frazier. The RSH maintains that listeners are sensitive to possible reasons for why a speaker might introduce a prosodic break: "listeners treat a prosodic boundary as more informative about the syntax when it flanks short constituents than when it flanks longer constituents," because in the latter case the speaker might have been motivated solely by consideration of optimal phrase lengths. This would effectively reduce the cue value of an appropriately placed prosodic boundary. We present additional evidence for the RSH from Turkish, a language typologically different from English. In addition, our study shows for the first time that the RSH also applies to a prosodic break which conflicts with the syntactic structure, reducing its perceived cue strength if it might have been motivated by length considerations. In this case, the RSH effect is beneficial. Finally, the Turkish data show that prosody-based explanations for parsing preferences such as the RSH do not take the place of traditional syntax-sensitive parsing strategies such as Late Closure. The two sources of guidance co-exist; both are used when available.

  1. Detection and Differentiation of Frontotemporal Dementia and Related Disorders From Alzheimer Disease Using the Montreal Cognitive Assessment.

    Science.gov (United States)

    Coleman, Kristy K L; Coleman, Brenda L; MacKinley, Julia D; Pasternak, Stephen H; Finger, Elizabeth C

    2016-01-01

    The Montreal Cognitive Assessment (MoCA) is a cognitive screening tool used by practitioners worldwide. The efficacy of the MoCA for screening frontotemporal dementia (FTD) and related disorders is unknown. The objectives were: (1) to determine whether the MoCA detects cognitive impairment (CI) in FTD subjects; (2) to determine whether Alzheimer disease (AD) and FTD subtypes and related disorders can be parsed using the MoCA; and (3) describe longitudinal MoCA performance by subtype. We extracted demographic and testing data from a database of patients referred to a cognitive neurology clinic who met criteria for probable AD or FTD (N=192). Logistic regression was used to determine whether dementia subtypes were associated with overall scores, subscores, or combinations of subscores on the MoCA. Initial MoCA results demonstrated CI in the majority of FTD subjects (87%). FTD subjects (N=94) performed better than AD subjects (N=98) on the MoCA (mean scores: 18.1 vs. 16.3; P=0.02). Subscores parsed many, but not all subtypes. FTD subjects had a larger decline on the MoCA within 13 to 36 months than AD subjects (P=0.02). The results indicate that the MoCA is a useful tool to identify and track progression of CI in FTD. Further, the data informs future research on scoring models for the MoCA to enhance cognitive screening and detection of FTD patients.

  2. Sleep and Obesity: A focus on animal models

    Science.gov (United States)

    Mavanji, Vijayakumar; Billington, Charles J.; Kotz, Catherine M.; Teske, Jennifer A.

    2012-01-01

    The rapid rise in obesity prevalence in the modern world parallels a significant reduction in restorative sleep (Agras et al., 2004; Dixon et al., 2007; Dixon et al., 2001; Gangwisch and Heymsfield, 2004; Gupta et al., 2002; Sekine et al., 2002; Vioque et al., 2000; Wolk et al., 2003). Reduced sleep time and quality increases the risk for obesity, but the underlying mechanisms remain unclear (Gangwisch et al., 2005; Hicks et al., 1986; Imaki et al., 2002; Jennings et al., 2007; Moreno et al., 2006). A majority of the theories linking human sleep disturbances and obesity rely on self-reported sleep. However, studies with objective measurements of sleep/wake parameters suggest a U-shaped relationship between sleep and obesity. Studies in animal models are needed to improve our understanding of the association between sleep disturbances and obesity. Genetic and experimenter-induced models mimicking characteristics of human obesity are now available and these animal models will be useful in understanding whether sleep disturbances determine propensity for obesity, or result from obesity. These models exhibit weight gain profiles consistently different from control animals. Thus a careful evaluation of animal models will provide insight into the relationship between sleep disturbances and obesity in humans. In this review we first briefly consider the fundamentals of sleep and key sleep disturbances, such as sleep fragmentation and excessive daytime sleepiness (EDS), observed in obese individuals. Then we consider sleep deprivation studies and the role of circadian alterations in obesity. We describe sleep/wake changes in various rodent models of obesity and obesity resistance. Finally, we discuss possible mechanisms linking sleep disturbances with obesity. PMID:22266350

  3. Building an Archival Collections Portal

    Directory of Open Access Journals (Sweden)

    Stuart Marquis

    2008-06-01

    Full Text Available Columbia University Libraries has developed the Archival Collections Portal, a unified search system helping users discover archival resources in a streamlined way. We combined the power of Lucene and Solr to search XML, parse JSON objects, create EAD-compliant documents, and deliver results in an easy-to-use interface. By reusing MARC records and employing new search engine features and techniques, we are able to bring important and hard-to-find collections to researchers and archivists. The canonical home page of the Portal is http://www.columbia.edu/library/archival/.

  4. Introduction to Beautiful Soup

    Directory of Open Access Journals (Sweden)

    Jeri Wieringa

    2012-12-01

    Full Text Available Beautiful Soup is a Python library for getting data out of HTML, XML, and other markup languages. Say you’ve found some webpages that display data relevant to your research, such as date or address information, but that do not provide any way of downloading the data directly. Beautiful Soup helps you pull particular content from a webpage, remove the HTML markup, and save the information. It is a tool for web scraping that helps you clean up and parse the documents you have pulled down from the web.

  5. Introduction to Beautiful Soup

    OpenAIRE

    Jeri Wieringa

    2012-01-01

    Beautiful Soup is a Python library for getting data out of HTML, XML, and other markup languages. Say you’ve found some webpages that display data relevant to your research, such as date or address information, but that do not provide any way of downloading the data directly. Beautiful Soup helps you pull particular content from a webpage, remove the HTML markup, and save the information. It is a tool for web scraping that helps you clean up and parse the documents you have pulled down from t...

  6. Cross-lingual and cross-domain discourse segmentation of entire documents

    DEFF Research Database (Denmark)

    Braud, Chloé; Lacroix, Ophélie; Søgaard, Anders

    2017-01-01

    -quality syntactic parses and rich heuristics that are not generally available across languages and domains. In this paper, we propose statistical discourse segmenters for five languages and three domains that do not rely on gold pre-annotations. We also consider the problem of learning discourse segmenters when...... no labeled data is available for a language. Our fully supervised system obtains 89.5% F1 for English newswire, with slight drops in performance on other domains, and we report supervised and unsupervised (cross-lingual) results for five languages in total....

  7. Computers and languages theory and practice

    CERN Document Server

    Nijholt, A

    1988-01-01

    A global introduction to language technology and the areas of computer science where language technology plays a role. Surveyed in this volume are issues related to the parsing problem in the fields of natural languages, programming languages, and formal languages.Throughout the book attention is paid to the social forces which influenced the development of the various topics. Also illustrated are the development of the theory of language analysis, its role in compiler construction, and its role in computer applications with a natural language interface between men and machine. Parts of the ma

  8. Measuring SIP proxy server performance

    CERN Document Server

    Subramanian, Sureshkumar V

    2013-01-01

    Internet Protocol (IP) telephony is an alternative to the traditional Public Switched Telephone Networks (PSTN), and the Session Initiation Protocol (SIP) is quickly becoming a popular signaling protocol for VoIP-based applications. SIP is a peer-to-peer multimedia signaling protocol standardized by the Internet Engineering Task Force (IETF), and it plays a vital role in providing IP telephony services through its use of the SIP Proxy Server (SPS), a software application that provides call routing services by parsing and forwarding all the incoming SIP packets in an IP telephony network.SIP Pr

  9. Data Extraction Tool to Analyse, Transform and Store Real Data from Electricity Markets

    DEFF Research Database (Denmark)

    Pereira, Ivo F.; Sousa, Tiago; Praca, Isabel

    2014-01-01

    of confidentiality, data regarding market proposals and transactions. These data can be used as source of knowledge to define realistic scenarios, which are essential for understanding and forecast electricity markets behavior. The development of tools able to extract, transform, store and dynamically update data......, is of great importance to go a step further into the comprehension of electricity markets and of the behaviour of the involved entities. In this paper an adaptable tool capable of downloading, parsing and storing data from market operators’ websites is presented, assuring constant updating and reliability...... of the stored data....

  10. Multiple Entity Reconciliation

    OpenAIRE

    Samoila, Lavinia Andreea

    2015-01-01

    Living in the age of "Big Data" is both a blessing and a curse. On he one hand, the raw data can be analysed and then used for weather redictions, user recommendations, targeted advertising and more. On he other hand, when data is aggregated from multiple sources, there is no guarantee that each source has stored the data in a standardized or even compatible format to what is required by the application. So there is a need to parse the available data and convert it to the desired form. Here i...

  11. Slavic Forest, Norwegian Wood

    OpenAIRE

    Rosa, Rudolf; Žabokrtský, Zdeněk; Zeman, Daniel; Mareček, David

    2017-01-01

    We once had a corp, or should we say, it once had us They showed us its tags, isn’t it great, unified tags They asked us to parse and they told us to use everything So we looked around and we noticed there was near nothing We took other langs, bitext aligned: words one-to-one We played for two weeks, and then they said, here is the test The parser kept training till morning, just until deadline So we had to wait and hope what we get ...

  12. Implementing SSL/TLS using cryptography and PKI

    CERN Document Server

    Davies, Joshua

    2011-01-01

    Hands-on, practical guide to implementing SSL and TLS protocols for Internet security If you are a network professional who knows C programming, this practical book is for you.  Focused on how to implement Secure Socket Layer (SSL) and Transport Layer Security (TLS), this book guides you through all necessary steps, whether or not you have a working knowledge of cryptography. The book covers SSLv2, TLS 1.0, and TLS 1.2, including implementations of the relevant cryptographic protocols, secure hashing, certificate parsing, certificate generation, and more.  Coverage includes: Underst

  13. XML as a cross-platform representation for medical imaging with fuzzy algorithms.

    Science.gov (United States)

    Gal, Norbert; Stoicu-Tivadar, Vasile

    2011-01-01

    Machines that perform linguistic medical image interpretation are based on fuzzy algorithms. There are several frameworks that can edit and simulate fuzzy algorithms, but they are not compatible with most of the implemented applications. This paper suggests a representation for fuzzy algorithms in XML files, and using this XML as a cross-platform between the simulation framework and the software applications. The paper presents a parsing algorithm that can convert files created by simulation framework, and converts them dynamically into an XML file keeping the original logical structure of the files.

  14. Evaluation and Classification of Syntax Usage in Determining Short-Text Semantic Similarity

    Directory of Open Access Journals (Sweden)

    V. Batanović

    2014-06-01

    Full Text Available This paper outlines and categorizes ways of using syntactic information in a number of algorithms for determining the semantic similarity of short texts. We consider the use of word order information, part-of-speech tagging, parsing and semantic role labeling. We analyze and evaluate the effects of syntax usage on algorithm performance by utilizing the results of a paraphrase detection test on the Microsoft Research Paraphrase Corpus. We also propose a new classification of algorithms based on their applicability to languages with scarce natural language processing tools.

  15. From State Diagram to Class Diagram

    DEFF Research Database (Denmark)

    Borch, Ole; Madsen, Per Printz

    2009-01-01

    UML class diagram and Java source code are interrelated and Java code is a kind of interchange format. Working with UML state diagram in CASE tools, a corresponding xml file is maintained. Designing state diagrams is mostly performed manually using design patterns and coding templates - a time...... consuming process. This article demonstrates how to compile such a diagram into Java code and later, by reverse engineering, produce a class diagram. The process from state diagram via intermediate SAX parsed xml file to Apache Velocity generated Java code is described. The result is a fast reproducible...

  16. Research on the display of nuclear cross-section library

    International Nuclear Information System (INIS)

    Huang Shien; Wang Kan; Yu Ganglin

    2008-01-01

    Minutely parsed the dot cross-section format (ACE format) data of the ENDF/ B-6.8 database, which is the foundation of the program that achieved the reading and related handling of ACE format data. This program achieved the plotting, zooming and comparing display functions of nuclear cross section-energy of ENDF/B-6.8 database. It also provides the standard picture formatting file output and/or standard text formatting file output of interesting nuclear data. It accomplished some appropriate validations of this program via the comparing between program results and reference data. (authors)

  17. Work orders management based on XML file in printing

    Directory of Open Access Journals (Sweden)

    Ran Peipei

    2018-01-01

    Full Text Available The Extensible Markup Language (XML technology is increasingly used in various field, if it’s used to express the information of work orders will improve efficiency for management and production. According to the features, we introduce the technology of management for work orders and get a XML file through the Document Object Model (DOM technology in the paper. When we need the information to conduct production, parsing the XML file and save the information in database, this is beneficial to the preserve and modify for information.

  18. Witnessing change with aspiring nurses: a human becoming teaching-learning process in nursing education.

    Science.gov (United States)

    Letcher, Deborah C; Yancey, Nan Russell

    2004-01-01

    Nurse educators have the opportunity to encourage meaningful reflections of nursing students. Dr. Rosemarie Rizzo Parse's teaching-learning processes provide a framework for such experiences. Student reflection through journaling and student participation in dialogue using these processes brings about an opportunity for students to discover new meaning for themselves and others. The process of how two nurse educators incorporated the human becoming teaching-learning model into students' experiences is discussed. Excerpts of student journals, themes of student work, and considerations for future development of the teaching-learning model with students are discussed.

  19. Functional and gene expression analysis of hTERT overexpressed endothelial cells

    Directory of Open Access Journals (Sweden)

    Haruna Takano

    2008-09-01

    Full Text Available Haruna Takano1, Satoshi Murasawa1,2, Takayuki Asahara1,2,31Institute of Biomedical Research and Innovation, Kobe, Japan; 2RIKEN Center for Developmental Biology, Kobe 650-0047, Japan; 3Tokai University of School of Medicine, Tokai, JapanAbstract: Telomerase dysfunction contributes to cellular senescence. Recent advances indicate the importance of senescence in maintaining vascular cell function in vitro. Human telomerase reverse transcriptase (hTERT overexpression is thought to lead to resistance to apoptosis and oxidative stress. However, the mechanism in endothelial lineage cells is unclear. We tried to generate an immortal endothelial cell line from human umbilical vein endothelial cells using a no-virus system and examine the functional mechanisms of hTERT overexpressed endothelial cell senescence in vitro. High levels of hTERT genes and endothelial cell-specific markers were expressed during long-term culture. Also, angiogenic responses were observed in hTERT overexpressed endothelial cell. These cells showed a delay in senescence and appeared more resistant to stressed conditions. PI3K/Akt-related gene levels were enhanced in hTERT overexpressed endothelial cells. An up-regulated PI3K/Akt pathway caused by hTERT overexpression might contribute to anti-apoptosis and survival effects in endothelial lineage cells.Keywords: endothelial, telomerase, senescence, oxidative stress, anti-apoptosis, PI3K/Akt pathway

  20. Readings from Asia: Memory: Troublesome, Irrepressible, and (Painfully Illuminating

    Directory of Open Access Journals (Sweden)

    Kyu Hyun Kim

    2013-09-01

    Full Text Available Nakano Satoshi 中野聡. 東南アジア占領と日本人: 帝国・日本の解体 [The occupation of Southeast Asia and the Japanese: Dissolution of the Japanese empire]. Itagaki Ryūta 板垣竜太, Chŏng Chi-yŏng 鄭智永, and Iwasaki Minoru 岩崎稔 eds. 東アジアの記憶の場 [The site of memories in East Asia]. "Readings from Asia" highlights significant recent scholarship published in Chinese, Korean, Japanese, and Vietnamese, in order to bring it to the attention of a wider, international readership. This issue of Cross-Currents features a review by Kyu Hyun Kim (University of California, Davis of two recent titles in Japanese: Nakano Satoshi’s 東南アジア占領と日本人: 帝国・日本の解体 and 東アジアの記憶の場, edited by Itagaki Ryūta, Chŏng Chi-yŏng, and Iwasaki Minoru. Kim recommends both books, with their “even-handed understanding of the corrective power of memory as well as its pitfalls,” to Japanese-language readers interested in the intersection of memory and history, as well as to students of modern Japanese and Korean history.

  1. BITCOIN - THE CURRENCY OF A NEW ECONOMY?

    Directory of Open Access Journals (Sweden)

    CARINA-ELENA STEGĂROIU

    2017-12-01

    Full Text Available In the knowledge society, a new economy is emerging, New Economy, which also encompasses the Internet economy in the form of e-bussines, which is why the economy is the economy of information society and knowledge. So, in a knowledge-based technology, a new digital coin was also needed in the field of electronic payments, thus creating the virtual currency BITCOIN (BTC, which in a moto translation is the bit of the bit (the unit of measure of the amount of information. It is a decentralized electronic payment system and a digital encrypted opensurce created in 2009 by Satoshi Nakamoto. The need to create Bitcoin (BTC was to ensure investment protection and free business finance, without resorting to financial institutions and beyond any constraints and regulations. The Bitcoin name also refers to the opensource program for using these coins, as well as the peer-to-peer (peer-to peer network it forms. Currently, even if he has many opponents and has gone through several difficult times, Bitcoin seems to survive and offer new technology with revolutionary use possibilities. The concept of the digital coin fits perfectly with the everchanging world, dominated by mobile and omnipresent technology. Compared to Bitcoin or another digital coin, any visit to the bank seems like a trip to another century, an era of time lost at the counter completing dozens of unnecessary papers.

  2. NASA Taxonomies for Searching Problem Reports and FMEAs

    Science.gov (United States)

    Malin, Jane T.; Throop, David R.

    2006-01-01

    Many types of hazard and risk analyses are used during the life cycle of complex systems, including Failure Modes and Effects Analysis (FMEA), Hazard Analysis, Fault Tree and Event Tree Analysis, Probabilistic Risk Assessment, Reliability Analysis and analysis of Problem Reporting and Corrective Action (PRACA) databases. The success of these methods depends on the availability of input data and the analysts knowledge. Standard nomenclature can increase the reusability of hazard, risk and problem data. When nomenclature in the source texts is not standard, taxonomies with mapping words (sets of rough synonyms) can be combined with semantic search to identify items and tag them with metadata based on a rich standard nomenclature. Semantic search uses word meanings in the context of parsed phrases to find matches. The NASA taxonomies provide the word meanings. Spacecraft taxonomies and ontologies (generalization hierarchies with attributes and relationships, based on terms meanings) are being developed for types of subsystems, functions, entities, hazards and failures. The ontologies are broad and general, covering hardware, software and human systems. Semantic search of Space Station texts was used to validate and extend the taxonomies. The taxonomies have also been used to extract system connectivity (interaction) models and functions from requirements text. Now the Reconciler semantic search tool and the taxonomies are being applied to improve search in the Space Shuttle PRACA database, to discover recurring patterns of failure. Usual methods of string search and keyword search fall short because the entries are terse and have numerous shortcuts (irregular abbreviations, nonstandard acronyms, cryptic codes) and modifier words cannot be used in sentence context to refine the search. The limited and fixed FMEA categories associated with the entries do not make the fine distinctions needed in the search. The approach assigns PRACA report titles to problem classes in

  3. Parsing glucose entry into the brain: novel findings obtained with enzyme-based glucose biosensors.

    Science.gov (United States)

    Kiyatkin, Eugene A; Wakabayashi, Ken T

    2015-01-21

    Extracellular levels of glucose in brain tissue reflect dynamic balance between its gradient-dependent entry from arterial blood and its use for cellular metabolism. In this work, we present several sets of previously published and unpublished data obtained by using enzyme-based glucose biosensors coupled with constant-potential high-speed amperometry in freely moving rats. First, we consider basic methodological issues related to the reliability of electrochemical measurements of extracellular glucose levels in rats under physiologically relevant conditions. Second, we present data on glucose responses induced in the nucleus accumbens (NAc) by salient environmental stimuli and discuss the relationships between local neuronal activation and rapid glucose entry into brain tissue. Third, by presenting data on changes in NAc glucose induced by intravenous and intragastric glucose delivery, we discuss other mechanisms of glucose entry into the extracellular domain following changes in glucose blood concentrations. Lastly, by showing the pattern of NAc glucose fluctuations during glucose-drinking behavior, we discuss the relationships between "active" and "passive" glucose entry to the brain, its connection to behavior-related metabolic activation, and the possible functional significance of these changes in behavioral regulation. These data provide solid experimental support for the "neuronal" hypothesis of neurovascular coupling, which postulates the critical role of neuronal activity in rapid regulation of vascular tone, local blood flow, and entry of glucose and oxygen to brain tissue to maintain active cellular metabolism.

  4. Parsing cognitive and emotional empathy deficits for negative and positive stimuli in frontotemporal dementia.

    Science.gov (United States)

    Oliver, Lindsay D; Mitchell, Derek G V; Dziobek, Isabel; MacKinley, Julia; Coleman, Kristy; Rankin, Katherine P; Finger, Elizabeth C

    2015-01-01

    Behavioural variant frontotemporal dementia (bvFTD) is a debilitating neurodegenerative disorder characterized by frontal and temporal lobe atrophy primarily affecting social cognition and emotion, including loss of empathy. Many consider empathy to be a multidimensional construct, including cognitive empathy (the ability to adopt and understand another's perspective) and emotional empathy (the capacity to share another's emotional experience). Cognitive and emotional empathy deficits have been associated with bvFTD; however, little is known regarding the performance of patients with bvFTD on behavioural measures of emotional empathy, and whether empathic responses differ for negative versus positive stimuli. 24 patients with bvFTD and 24 healthy controls completed the Multifaceted Empathy Test (MET; Dziobek et al., 2008), a performance-based task that taps both cognitive and emotional facets of empathy, and allows for the discrimination of responses to negative versus positive realistic images. MET scores were also compared with caregiver ratings of patient behaviour on the Interpersonal Reactivity Index, which assesses patients' everyday demonstrations of perspective taking and empathic concern. Patients with bvFTD were less accurate than controls at inferring mental states for negative and positive stimuli. They also demonstrated lower levels of shared emotional experience, more positive emotional reactions, and diminished arousal to negative social stimuli relative to controls. Patients showed reduced emotional reactions to negative non-social stimuli as well. Lastly, the MET and IRI measures of emotional empathy were found to be significantly correlated within the bvFTD group. The results suggest that patients with bvFTD show a global deficit in cognitive empathy, and deficient emotional empathy for negative, but not positive, experiences. Further, a generalized emotional processing impairment for negative stimuli was observed, which could contribute to the emotional empathy deficit. This work highlights potential treatment targets and a means to assess the impact of novel therapies on socioemotional impairment in bvFTD. Copyright © 2014 Elsevier Ltd. All rights reserved.

  5. Deeper than Shallow: Evidence for Structure-Based Parsing Biases in Second-Language Sentence Processing

    Science.gov (United States)

    Witzel, Jeffrey; Witzel, Naoko; Nicol, Janet

    2012-01-01

    This study examines the reading patterns of native speakers (NSs) and high-level (Chinese) nonnative speakers (NNSs) on three English sentence types involving temporarily ambiguous structural configurations. The reading patterns on each sentence type indicate that both NSs and NNSs were biased toward specific structural interpretations. These…

  6. Inducing Head-Driven PCFGs with Latent Heads: Refining a Tree-bank Grammar for Parsing

    NARCIS (Netherlands)

    Prescher, D.; Gama, J.; Camacho, R.; Brazdil, P.; Jorge, A.; Torgo, L.

    2005-01-01

    Although state-of-the-art parsers for natural language are lexicalized, it was recently shown that an accurate unlexicalized parser for the Penn tree-bank can be simply read off a manually refined tree-bank. While lexicalized parsers often suffer from sparse data, manual mark-up is costly and

  7. BioMagResBank databases DOCR and FRED containing converted and filtered sets of experimental NMR restraints and coordinates from over 500 protein PDB structures

    Energy Technology Data Exchange (ETDEWEB)

    Doreleijers, Jurgen F. [University of Wisconsin-Madison, BioMagResBank, Department of Biochemistry (United States); Nederveen, Aart J. [Utrecht University, Bijvoet Center for Biomolecular Research (Netherlands); Vranken, Wim [European Bioinformatics Institute, Macromolecular Structure Database group (United Kingdom); Lin Jundong [University of Wisconsin-Madison, BioMagResBank, Department of Biochemistry (United States); Bonvin, Alexandre M.J.J.; Kaptein, Robert [Utrecht University, Bijvoet Center for Biomolecular Research (Netherlands); Markley, John L.; Ulrich, Eldon L. [University of Wisconsin-Madison, BioMagResBank, Department of Biochemistry (United States)], E-mail: elu@bmrb.wisc.edu

    2005-05-15

    We present two new databases of NMR-derived distance and dihedral angle restraints: the Database Of Converted Restraints (DOCR) and the Filtered Restraints Database (FRED). These databases currently correspond to 545 proteins with NMR structures deposited in the Protein Databank (PDB). The criteria for inclusion were that these should be unique, monomeric proteins with author-provided experimental NMR data and coordinates available from the PDB capable of being parsed and prepared in a consistent manner. The Wattos program was used to parse the files, and the CcpNmr FormatConverter program was used to prepare them semi-automatically. New modules, including a new implementation of Aqua in the BioMagResBank (BMRB) software Wattos were used to analyze the sets of distance restraints (DRs) for inconsistencies, redundancies, NOE completeness, classification and violations with respect to the original coordinates. Restraints that could not be associated with a known nomenclature were flagged. The coordinates of hydrogen atoms were recalculated from the positions of heavy atoms to allow for a full restraint analysis. The DOCR database contains restraint and coordinate data that is made consistent with each other and with IUPAC conventions. The FRED database is based on the DOCR data but is filtered for use by test calculation protocols and longitudinal analyses and validations. These two databases are available from websites of the BMRB and the Macromolecular Structure Database (MSD) in various formats: NMR-STAR, CCPN XML, and in formats suitable for direct use in the software packages CNS and CYANA.

  8. Unsupervised grammar induction of clinical report sublanguage

    Directory of Open Access Journals (Sweden)

    Kate Rohit J

    2012-10-01

    Full Text Available Abstract Background Clinical reports are written using a subset of natural language while employing many domain-specific terms; such a language is also known as a sublanguage for a scientific or a technical domain. Different genres of clinical reports use different sublaguages, and in addition, different medical facilities use different medical language conventions. This makes supervised training of a parser for clinical sentences very difficult as it would require expensive annotation effort to adapt to every type of clinical text. Methods In this paper, we present an unsupervised method which automatically induces a grammar and a parser for the sublanguage of a given genre of clinical reports from a corpus with no annotations. In order to capture sentence structures specific to clinical domains, the grammar is induced in terms of semantic classes of clinical terms in addition to part-of-speech tags. Our method induces grammar by minimizing the combined encoding cost of the grammar and the corresponding sentence derivations. The probabilities for the productions of the induced grammar are then learned from the unannotated corpus using an instance of the expectation-maximization algorithm. Results Our experiments show that the induced grammar is able to parse novel sentences. Using a dataset of discharge summary sentences with no annotations, our method obtains 60.5% F-measure for parse-bracketing on sentences of maximum length 10. By varying a parameter, the method can induce a range of grammars, from very specific to very general, and obtains the best performance in between the two extremes.

  9. Unsupervised grammar induction of clinical report sublanguage.

    Science.gov (United States)

    Kate, Rohit J

    2012-10-05

    Clinical reports are written using a subset of natural language while employing many domain-specific terms; such a language is also known as a sublanguage for a scientific or a technical domain. Different genres of clinical reports use different sublaguages, and in addition, different medical facilities use different medical language conventions. This makes supervised training of a parser for clinical sentences very difficult as it would require expensive annotation effort to adapt to every type of clinical text. In this paper, we present an unsupervised method which automatically induces a grammar and a parser for the sublanguage of a given genre of clinical reports from a corpus with no annotations. In order to capture sentence structures specific to clinical domains, the grammar is induced in terms of semantic classes of clinical terms in addition to part-of-speech tags. Our method induces grammar by minimizing the combined encoding cost of the grammar and the corresponding sentence derivations. The probabilities for the productions of the induced grammar are then learned from the unannotated corpus using an instance of the expectation-maximization algorithm. Our experiments show that the induced grammar is able to parse novel sentences. Using a dataset of discharge summary sentences with no annotations, our method obtains 60.5% F-measure for parse-bracketing on sentences of maximum length 10. By varying a parameter, the method can induce a range of grammars, from very specific to very general, and obtains the best performance in between the two extremes.

  10. A linguistic rule-based approach to extract drug-drug interactions from pharmacological documents.

    Science.gov (United States)

    Segura-Bedmar, Isabel; Martínez, Paloma; de Pablo-Sánchez, César

    2011-03-29

    A drug-drug interaction (DDI) occurs when one drug influences the level or activity of another drug. The increasing volume of the scientific literature overwhelms health care professionals trying to be kept up-to-date with all published studies on DDI. This paper describes a hybrid linguistic approach to DDI extraction that combines shallow parsing and syntactic simplification with pattern matching. Appositions and coordinate structures are interpreted based on shallow syntactic parsing provided by the UMLS MetaMap tool (MMTx). Subsequently, complex and compound sentences are broken down into clauses from which simple sentences are generated by a set of simplification rules. A pharmacist defined a set of domain-specific lexical patterns to capture the most common expressions of DDI in texts. These lexical patterns are matched with the generated sentences in order to extract DDIs. We have performed different experiments to analyze the performance of the different processes. The lexical patterns achieve a reasonable precision (67.30%), but very low recall (14.07%). The inclusion of appositions and coordinate structures helps to improve the recall (25.70%), however, precision is lower (48.69%). The detection of clauses does not improve the performance. Information Extraction (IE) techniques can provide an interesting way of reducing the time spent by health care professionals on reviewing the literature. Nevertheless, no approach has been carried out to extract DDI from texts. To the best of our knowledge, this work proposes the first integral solution for the automatic extraction of DDI from biomedical texts.

  11. BioMagResBank databases DOCR and FRED containing converted and filtered sets of experimental NMR restraints and coordinates from over 500 protein PDB structures

    International Nuclear Information System (INIS)

    Doreleijers, Jurgen F.; Nederveen, Aart J.; Vranken, Wim; Lin Jundong; Bonvin, Alexandre M.J.J.; Kaptein, Robert; Markley, John L.; Ulrich, Eldon L.

    2005-01-01

    We present two new databases of NMR-derived distance and dihedral angle restraints: the Database Of Converted Restraints (DOCR) and the Filtered Restraints Database (FRED). These databases currently correspond to 545 proteins with NMR structures deposited in the Protein Databank (PDB). The criteria for inclusion were that these should be unique, monomeric proteins with author-provided experimental NMR data and coordinates available from the PDB capable of being parsed and prepared in a consistent manner. The Wattos program was used to parse the files, and the CcpNmr FormatConverter program was used to prepare them semi-automatically. New modules, including a new implementation of Aqua in the BioMagResBank (BMRB) software Wattos were used to analyze the sets of distance restraints (DRs) for inconsistencies, redundancies, NOE completeness, classification and violations with respect to the original coordinates. Restraints that could not be associated with a known nomenclature were flagged. The coordinates of hydrogen atoms were recalculated from the positions of heavy atoms to allow for a full restraint analysis. The DOCR database contains restraint and coordinate data that is made consistent with each other and with IUPAC conventions. The FRED database is based on the DOCR data but is filtered for use by test calculation protocols and longitudinal analyses and validations. These two databases are available from websites of the BMRB and the Macromolecular Structure Database (MSD) in various formats: NMR-STAR, CCPN XML, and in formats suitable for direct use in the software packages CNS and CYANA

  12. SEMSIN SEMANTIC AND SYNTACTIC PARSER

    Directory of Open Access Journals (Sweden)

    K. K. Boyarsky

    2015-09-01

    Full Text Available The paper deals with the principle of operation for SemSin semantic and syntactic parser creating a dependency tree for the Russian language sentences. The parser consists of 4 blocks: a dictionary, morphological analyzer, production rules and lexical analyzer. An important logical part of the parser is pre-syntactical module, which harmonizes and complements morphological analysis results, separates the text paragraphs into individual sentences, and also carries out predisambiguation. Characteristic feature of the presented parser is an open type of control – it is done by means of a set of production rules. A varied set of commands provides the ability to both morphological and semantic-syntactic analysis of the sentence. The paper presents the sequence of rules usage and examples of their work. Specific feature of the rules is the decision making on establishment of syntactic links with simultaneous removal of the morphological and semantic ambiguity. The lexical analyzer provides the execution of commands and rules, and manages the parser in manual or automatic modes of the text analysis. In the first case, the analysis is performed interactively with the possibility of step-by-step execution of the rules and scanning the resulting parse tree. In the second case, analysis results are filed in an xml-file. Active usage of syntactic and semantic dictionary information gives the possibility to reduce significantly the ambiguity of parsing. In addition to marking the text, the parser is also usable as a tool for information extraction from natural language texts.

  13. IDEA: Interactive Display for Evolutionary Analyses.

    Science.gov (United States)

    Egan, Amy; Mahurkar, Anup; Crabtree, Jonathan; Badger, Jonathan H; Carlton, Jane M; Silva, Joana C

    2008-12-08

    The availability of complete genomic sequences for hundreds of organisms promises to make obtaining genome-wide estimates of substitution rates, selective constraints and other molecular evolution variables of interest an increasingly important approach to addressing broad evolutionary questions. Two of the programs most widely used for this purpose are codeml and baseml, parts of the PAML (Phylogenetic Analysis by Maximum Likelihood) suite. A significant drawback of these programs is their lack of a graphical user interface, which can limit their user base and considerably reduce their efficiency. We have developed IDEA (Interactive Display for Evolutionary Analyses), an intuitive graphical input and output interface which interacts with PHYLIP for phylogeny reconstruction and with codeml and baseml for molecular evolution analyses. IDEA's graphical input and visualization interfaces eliminate the need to edit and parse text input and output files, reducing the likelihood of errors and improving processing time. Further, its interactive output display gives the user immediate access to results. Finally, IDEA can process data in parallel on a local machine or computing grid, allowing genome-wide analyses to be completed quickly. IDEA provides a graphical user interface that allows the user to follow a codeml or baseml analysis from parameter input through to the exploration of results. Novel options streamline the analysis process, and post-analysis visualization of phylogenies, evolutionary rates and selective constraint along protein sequences simplifies the interpretation of results. The integration of these functions into a single tool eliminates the need for lengthy data handling and parsing, significantly expediting access to global patterns in the data.

  14. Perceptual learning improves visual performance in juvenile amblyopia.

    Science.gov (United States)

    Li, Roger W; Young, Karen G; Hoenig, Pia; Levi, Dennis M

    2005-09-01

    To determine whether practicing a position-discrimination task improves visual performance in children with amblyopia and to determine the mechanism(s) of improvement. Five children (age range, 7-10 years) with amblyopia practiced a positional acuity task in which they had to judge which of three pairs of lines was misaligned. Positional noise was produced by distributing the individual patches of each line segment according to a Gaussian probability function. Observers were trained at three noise levels (including 0), with each observer performing between 3000 and 4000 responses in 7 to 10 sessions. Trial-by-trial feedback was provided. Four of the five observers showed significant improvement in positional acuity. In those four observers, on average, positional acuity with no noise improved by approximately 32% and with high noise by approximately 26%. A position-averaging model was used to parse the improvement into an increase in efficiency or a decrease in equivalent input noise. Two observers showed increased efficiency (51% and 117% improvements) with no significant change in equivalent input noise across sessions. The other two observers showed both a decrease in equivalent input noise (18% and 29%) and an increase in efficiency (17% and 71%). All five observers showed substantial improvement in Snellen acuity (approximately 26%) after practice. Perceptual learning can improve visual performance in amblyopic children. The improvement can be parsed into two important factors: decreased equivalent input noise and increased efficiency. Perceptual learning techniques may add an effective new method to the armamentarium of amblyopia treatments.

  15. Establishing a learning foundation in a dynamically changing world: Insights from artificial language work

    Science.gov (United States)

    Gonzales, Kalim

    It is argued that infants build a foundation for learning about the world through their incidental acquisition of the spatial and temporal regularities surrounding them. A challenge is that learning occurs across multiple contexts whose statistics can greatly differ. Two artificial language studies with 12-month-olds demonstrate that infants come prepared to parse statistics across contexts using the temporal and perceptual features that distinguish one context from another. These results suggest that infants can organize their statistical input with a wider range of features that typically considered. Possible attention, decision making, and memory mechanisms are discussed.

  16. Dual Syntax for XML Languages

    DEFF Research Database (Denmark)

    Brabrand, Claus; Møller, Anders; Schwartzbach, Michael Ignatieff

    2005-01-01

    XML is successful as a machine processable data interchange format, but it is often too verbose for human use. For this reason, many XML languages permit an alternative more legible non-XML syntax. XSLT stylesheets are often used to convert from the XML syntax to the alternative syntax; however......, such transformations are not reversible since no general tool exists to automatically parse the alternative syntax back into XML. We present XSugar, which makes it possible to manage dual syntax for XML languages. An XSugar specification is built around a context-free grammar that unifies the two syntaxes...

  17. Python 3 text processing with NLTK 3 cookbook

    CERN Document Server

    Perkins, Jacob

    2014-01-01

    This book is intended for Python programmers interested in learning how to do natural language processing. Maybe you've learned the limits of regular expressions the hard way, or you've realized that human language cannot be deterministically parsed like a computer language. Perhaps you have more text than you know what to do with, and need automated ways to analyze and structure that text. This Cookbook will show you how to train and use statistical language models to process text in ways that are practically impossible with standard programming tools. A basic knowledge of Python and the basi

  18. Coordination Analysis Using Global Structural Constraints and Alignment-based Local Features

    Science.gov (United States)

    Hara, Kazuo; Shimbo, Masashi; Matsumoto, Yuji

    We propose a hybrid approach to coordinate structure analysis that combines a simple grammar to ensure consistent global structure of coordinations in a sentence, and features based on sequence alignment to capture local symmetry of conjuncts. The weight of the alignment-based features, which in turn determines the score of coordinate structures, is optimized by perceptron training on a given corpus. A bottom-up chart parsing algorithm efficiently finds the best scoring structure, taking both nested or non-overlapping flat coordinations into account. We demonstrate that our approach outperforms existing parsers in coordination scope detection on the Genia corpus.

  19. AquaResp® — free open-source software for measuring oxygen consumption of resting aquatic animals

    DEFF Research Database (Denmark)

    Svendsen, Morten Bo S.; Skov, Peter Vilhelm; Bushnell, Peter G.

    AquaResp® is a free open-source software program developed to measure the oxygen consumption of aquatic animals using intermittent flow techniques. This free program is based on Microsoft Excel, and uses the MCC Universal Library and a data acquisition board to acquire analogue readings from up...... to four input ports and output control via two digital and two analogue ports. In addition AquaResp can read one COM-port if the oxygen analyser has a RS-232 output signal. The present version of the program has options for parsing data strings generated by two major fibre optic oxygen electrode...

  20. Constructing storyboards based on hierarchical clustering analysis

    Science.gov (United States)

    Hasebe, Satoshi; Sami, Mustafa M.; Muramatsu, Shogo; Kikuchi, Hisakazu

    2005-07-01

    There are growing needs for quick preview of video contents for the purpose of improving accessibility of video archives as well as reducing network traffics. In this paper, a storyboard that contains a user-specified number of keyframes is produced from a given video sequence. It is based on hierarchical cluster analysis of feature vectors that are derived from wavelet coefficients of video frames. Consistent use of extracted feature vectors is the key to avoid a repetition of computationally-intensive parsing of the same video sequence. Experimental results suggest that a significant reduction in computational time is gained by this strategy.

  1. Diana Reference Manual. Revision 3,

    Science.gov (United States)

    1983-02-28

    will help the reader to understand DIANA. Section 1. 1. 1 presents those principles that motivated the original design of DIAN. and Section 1. 1. 2...parenthesized node whose offspring was the addition, since ADA’s parsing rules require the parentheses. The motivation for this requirement Is to ease the...3.7. a 41neA~olm S3.I. C, S.1. F] 51. 52 number 3.1, 3.2.8] 33, 33 numw_d [3.2.83e 35 number rep C4.4.0] 43 nummicUteral [4:4.0] 43 ps4 . I.A 46

  2. Connections and lingering presence as cocreated art.

    Science.gov (United States)

    Dempsey, Leona F

    2008-10-01

    Parse described nursing practice as a performing art where the nurse is like a dancer. Just as in any dance performance, unplanned events may occur. When a nurse is artistically living, unique and meaningful performances might emerge from unplanned events. In this practice column, the author describes how shifting experiences surfaced with unforeseen connections and lingering presence during her study of feeling confined. In her study she was in true presence with men living in prison, who were diagnosed with severe mental illness. The humanbecoming school of thought was the nursing perspective guiding the research study.

  3. Data Mining the Internet Archive Collection

    Directory of Open Access Journals (Sweden)

    Caleb McDaniel

    2014-03-01

    Full Text Available The collections of the Internet Archive (IA include many digitized sources of interest to historians, including early JSTOR journal content, John Adams’s personal library, and the Haiti collection at the John Carter Brown Library. In short, to quote Programming Historian Ian Milligan, “The Internet Archive rocks.” In this lesson, you’ll learn how to download files from such collections using a Python module specifically designed for the Internet Archive. You will also learn how to use another Python module designed for parsing MARC XML records, a widely used standard for formatting bibliographic metadata.

  4. A standard format and a graphical user interface for spin system specification.

    Science.gov (United States)

    Biternas, A G; Charnock, G T P; Kuprov, Ilya

    2014-03-01

    We introduce a simple and general XML format for spin system description that is the result of extensive consultations within Magnetic Resonance community and unifies under one roof all major existing spin interaction specification conventions. The format is human-readable, easy to edit and easy to parse using standard XML libraries. We also describe a graphical user interface that was designed to facilitate construction and visualization of complicated spin systems. The interface is capable of generating input files for several popular spin dynamics simulation packages. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.

  5. A database paradigm for the management of DICOM-RT structure sets using a geographic information system

    International Nuclear Information System (INIS)

    Shao, Weber; Kupelian, Patrick A; Wang, Jason; Low, Daniel A; Ruan, Dan

    2014-01-01

    We devise a paradigm for representing the DICOM-RT structure sets in a database management system, in such way that secondary calculations of geometric information can be performed quickly from the existing contour definitions. The implementation of this paradigm is achieved using the PostgreSQL database system and the PostGIS extension, a geographic information system commonly used for encoding geographical map data. The proposed paradigm eliminates the overhead of retrieving large data records from the database, as well as the need to implement various numerical and data parsing routines, when additional information related to the geometry of the anatomy is desired.

  6. Proscene: A feature-rich framework for interactive environments

    Directory of Open Access Journals (Sweden)

    Jean Pierre Charalambos

    2017-01-01

    Full Text Available We introduce Proscene, a feature-rich, open-source framework for interactive environments. The design of Proscene comprises a three-layered onion-like software architecture, promoting different possible development scenarios. The framework innermost layer decouples user gesture parsing from user-defined actions. The in-between layer implements a feature-rich set of widely-used motion actions allowing the selection and manipulation of objects, including the scene viewpoint. The outermost layer exposes those features as a Processing library. The results have shown the feasibility of our approach together with the simplicity and flexibility of the Proscene framework API.

  7. Monitoring system for OpenPBS environment

    Energy Technology Data Exchange (ETDEWEB)

    Kolosov, V. [ITEP, Moscow (Russian Federation)]. E-mail: victor.kolosov@itep.ru; Lublev, Y. [ITEP, Moscow (Russian Federation); Makarychev, S. [ITEP, Moscow (Russian Federation)

    2004-11-21

    The OpenPBS batch system is widely used in the HEP community. The Open PBS package has a set of tools to check the current status of the system. This information is useful, but it is not sufficient enough for resource accounting and planning. As a solution for this problem, we developed a monitoring system which parses the logfiles from OpenPBS and stores the information into a SQL database (PostgreSQL). This allows us to analyze the data in many different ways using SQL queries. The system was used in ITEP during the last two years for batch farm monitoring.

  8. A database paradigm for the management of DICOM-RT structure sets using a geographic information system

    Science.gov (United States)

    Shao, Weber; Kupelian, Patrick A.; Wang, Jason; Low, Daniel A.; Ruan, Dan

    2014-03-01

    We devise a paradigm for representing the DICOM-RT structure sets in a database management system, in such way that secondary calculations of geometric information can be performed quickly from the existing contour definitions. The implementation of this paradigm is achieved using the PostgreSQL database system and the PostGIS extension, a geographic information system commonly used for encoding geographical map data. The proposed paradigm eliminates the overhead of retrieving large data records from the database, as well as the need to implement various numerical and data parsing routines, when additional information related to the geometry of the anatomy is desired.

  9. Construction and installation of the experimental marine equipment for recovery of rare metals in seawater

    Energy Technology Data Exchange (ETDEWEB)

    Hasegawa, Shin; Seko, Noriaki [Japan Atomic Energy Research Inst., Takasaki, Gunma (Japan). Takasaki Radiation Chemistry Research Establishment; Tabata, Kokichi [Mitsui Engineering and Shipbuiding Co., Ltd., Tokyo (Japan); and others

    2001-03-01

    An experimental marine equipment was designed and constructed to evaluate a fibrous polymer adsorbent prepared with radiation-processing and to extract the technical tasks in the step to the utilization of the rare metals recovery in seawater. The equipment was set in the offing of Mutsu-Sekine in Aomori prefecture. The scale of this equipment is capable of recovering 1 kg uranium in seawater by dipping the adsorbents by 6 times into the sea for 30 days per one dipping. The adsorbents, packed in metal wire cages, were hung down from a fish preserve-shaped float (floating steel frame). This float capacity was designed that the floating steel frame was sunk when wave height becomes high in order to reduce the tension of the ropes. The safety of the rope was evaluated by using design wave height 4.3 m and maximum wave height 7.3 m derived from 10 years' probability wave height. The safety factors obtained were 6.4 for the design wave height and 3.7 for the maximum wave height. The floating steel frame and the adsorbent cage were designed to maintain the mechanical strength under the maximum wave height. The weight of four anchor blocks are 40 tons enough to moor the floating steel frame. In the setting of the marine equipment, the floating steel frame connected with the anchor blocks by rope was put down on the bottom of the sea. The adsorbent cage was hung in the center of the floating steel frame. The characteristic test was carried out by packing adsorbents into adsorbent cage and dipping them for 20 days into the sea. There was no serious problem in the floating steel frame, the adsorbent cage, and the ropes and their knots. (author)

  10. Efficacy of everolimus with exemestane versus exemestane alone in Asian patients with HER2-negative, hormone-receptor-positive breast cancer in BOLERO-2.

    Science.gov (United States)

    Noguchi, Shinzaburo; Masuda, Norikazu; Iwata, Hiroji; Mukai, Hirofumi; Horiguchi, Jun; Puttawibul, Puttisak; Srimuninnimit, Vichien; Tokuda, Yutaka; Kuroi, Katsumasa; Iwase, Hirotaka; Inaji, Hideo; Ohsumi, Shozo; Noh, Woo-Chul; Nakayama, Takahiro; Ohno, Shinji; Rai, Yoshiaki; Park, Byeong-Woo; Panneerselvam, Ashok; El-Hashimy, Mona; Taran, Tetiana; Sahmoud, Tarek; Ito, Yoshinori

    2014-11-01

    The addition of mTOR inhibitor everolimus (EVE) to exemestane (EXE) was evaluated in an international, phase 3 study (BOLERO-2) in patients with hormone-receptor-positive (HR(+)) breast cancer refractory to letrozole or anastrozole. The safety and efficacy of anticancer treatments may be influenced by ethnicity (Sekine et al. in Br J Cancer 99:1757-62, 2008). Safety and efficacy results from Asian versus non-Asian patients in BOLERO-2 are reported. Patients were randomized (2:1) to 10 mg/day EVE + EXE or placebo (PBO) + EXE. Primary endpoint was progression-free survival (PFS). Secondary endpoints included overall survival, response rate, clinical benefit rate, and safety. Of 143 Asian patients, 98 received EVE + EXE and 45 received PBO + EXE. Treatment with EVE + EXE significantly improved median PFS versus PBO + EXE among Asian patients by 38 % (HR = 0.62; 95 % CI, 0.41-0.94). Median PFS was also improved among non-Asian patients by 59 % (HR = 0.41; 95 % CI, 0.33-0.50). Median PFS duration among EVE-treated Asian patients was 8.48 versus 4.14 months for PBO + EXE, and 7.33 versus 2.83 months, respectively, in non-Asian patients. The most common grade 3/4 adverse events (stomatitis, anemia, elevated liver enzymes, hyperglycemia, and dyspnea) occurred at similar frequencies in Asian and non-Asian patients. Grade 1/2 interstitial lung disease occurred more frequently in Asian patients. Quality of life was similar between treatment arms in Asian patients. Adding EVE to EXE provided substantial clinical benefit in both Asian and non-Asian patients with similar safety profiles. This combination represents an improvement in the management of postmenopausal women with HR(+)/HER2(-) advanced breast cancer progressing on nonsteroidal aromatase inhibitors, regardless of ethnicity.

  11. Construction and installation of the experimental marine equipment for recovery of rare metals in seawater

    Energy Technology Data Exchange (ETDEWEB)

    Hasegawa, Shin; Seko, Noriaki [Japan Atomic Energy Research Inst., Takasaki, Gunma (Japan). Takasaki Radiation Chemistry Research Establishment; Tabata, Kokichi [Mitsui Engineering and Shipbuiding Co., Ltd., Tokyo (Japan)] [and others

    2001-03-01

    An experimental marine equipment was designed and constructed to evaluate a fibrous polymer adsorbent prepared with radiation-processing and to extract the technical tasks in the step to the utilization of the rare metals recovery in seawater. The equipment was set in the offing of Mutsu-Sekine in Aomori prefecture. The scale of this equipment is capable of recovering 1 kg uranium in seawater by dipping the adsorbents by 6 times into the sea for 30 days per one dipping. The adsorbents, packed in metal wire cages, were hung down from a fish preserve-shaped float (floating steel frame). This float capacity was designed that the floating steel frame was sunk when wave height becomes high in order to reduce the tension of the ropes. The safety of the rope was evaluated by using design wave height 4.3 m and maximum wave height 7.3 m derived from 10 years' probability wave height. The safety factors obtained were 6.4 for the design wave height and 3.7 for the maximum wave height. The floating steel frame and the adsorbent cage were designed to maintain the mechanical strength under the maximum wave height. The weight of four anchor blocks are 40 tons enough to moor the floating steel frame. In the setting of the marine equipment, the floating steel frame connected with the anchor blocks by rope was put down on the bottom of the sea. The adsorbent cage was hung in the center of the floating steel frame. The characteristic test was carried out by packing adsorbents into adsorbent cage and dipping them for 20 days into the sea. There was no serious problem in the floating steel frame, the adsorbent cage, and the ropes and their knots. (author)

  12. Designing and testing lightweight shoulder prostheses with hybrid actuators for movements involved in typical activities of daily living and impact absorption

    Directory of Open Access Journals (Sweden)

    Sekine M

    2015-07-01

    Full Text Available Masashi Sekine,1,2 Kahori Kita,1 Wenwei Yu1 1Center for Frontier Medical Engineering, 2Graduate School of Engineering, Chiba University, Chiba, Japan Abstract: Unlike forearm amputees, transhumeral amputees have residual stumps that are too small to provide a sufficient range of operation for their prosthetic parts to perform usual activities of daily living. Furthermore, it is difficult for small residual stumps to provide sufficient impact absorption for safe manipulation in daily living, as intact arms do. Therefore, substitution of upper limb function in transhumeral amputees requires a sufficient range of motion and sufficient viscoelasticity for shoulder prostheses under critical weight and dimension constraints. We propose the use of two different types of actuators, ie, pneumatic elastic actuators (PEAs and servo motors. PEAs offer high power-to-weight performance and have intrinsic viscoelasticity in comparison with motors or standard industrial pneumatic cylinder actuators. However, the usefulness of PEAs in large working spaces is limited because of their short strokes. Servo motors, in contrast, can be used to achieve large ranges of motion. In this study, the relationship between the force and stroke of PEAs was investigated. The impact absorption of both types of actuators was measured using a single degree-of-freedom prototype to evaluate actuator compliance for safety purposes. Based on the fundamental properties of the actuators identified, a four degree-of-freedom robotic arm is proposed for prosthetic use. The configuration of the actuators and functional parts was designed to achieve a specified range of motion and torque calculated from the results of a simulation of typical movements performed in usual activities of daily living. Our experimental results showed that the requirements for the shoulder prostheses could be satisfied. Keywords: shoulder prosthesis, hybrid actuation, pneumatic elastic actuator, antagonistic

  13. Hirabayashi, Satoshi; Kroll, Charles N.; Nowak, David J. 2011. Component-based development and sensitivity analyses of an air pollutant dry deposition model. Environmental Modelling & Software. 26(6): 804-816.

    Science.gov (United States)

    Satoshi Hirabayashi; Chuck Kroll; David Nowak

    2011-01-01

    The Urban Forest Effects-Deposition model (UFORE-D) was developed with a component-based modeling approach. Functions of the model were separated into components that are responsible for user interface, data input/output, and core model functions. Taking advantage of the component-based approach, three UFORE-D applications were developed: a base application to estimate...

  14. Implementasi JSON Parsing Pada Aplikasi Mobile E-commerce Studi Kasus : CV V3 Tekno Indonesia

    OpenAIRE

    Destian Wijaya, Bhakti; E.M.A, Fenty; Fiade, Andrew

    2015-01-01

    Sistem operasi Android saat ini merupakan salah satu dari sekian banyak sistem operasi padasmartphone yang sedang mengalami pertumbuhan pesat di dunia, khususnya di Indonesia. CV VitigaTekno Indonesia bergerak di bidang pemasaran gadget membutuhkan media marketing yang dapatdiakses secara online selain website yaitu melalui smartphone. Oleh karena itu penelitian ini berfokuspada bagaimana membuat aplikasi mobile e-commerce pada smartphone Android, dengan caramengambil data yang berasal dari w...

  15. Chunking or not chunking? How do we find words in artificial language learning?

    Science.gov (United States)

    Franco, Ana; Destrebecqz, Arnaud

    2012-01-01

    What is the nature of the representations acquired in implicit statistical learning? Recent results in the field of language learning have shown that adults and infants are able to find the words of an artificial language when exposed to a continuous auditory sequence consisting in a random ordering of these words. Such performance can only be based on processing the transitional probabilities between sequence elements. Two different kinds of mechanisms may account for these data: Participants may either parse the sequence into smaller chunks corresponding to the words of the artificial language, or they may become progressively sensitive to the actual values of the transitional probabilities between syllables. The two accounts are difficult to differentiate because they make similar predictions in comparable experimental settings. In this study, we present two experiments that aimed at contrasting these two theories. In these experiments, participants had to learn 2 sets of pseudo-linguistic regularities: Language 1 (L1) and Language 2 (L2) presented in the context of a serial reaction time task. L1 and L2 were either unrelated (none of the syllabic transitions of L1 were present in L2), or partly related (some of the intra-words transitions of L1 were used as inter-words transitions of L2). The two accounts make opposite predictions in these two settings. Our results indicate that the nature of the representations depends on the learning condition. When cues were presented to facilitate parsing of the sequence, participants learned the words of the artificial language. However, when no cues were provided, performance was strongly influenced by the employed transitional probabilities.

  16. PERANGKAT BANTU UNTUK OPTIMASI QUERY PADA ORACLE DENGAN RESTRUKTURISASI SQL

    Directory of Open Access Journals (Sweden)

    Darlis Heru Murti

    2006-07-01

    Full Text Available Query merupakan bagian dari bahasa pemrograman SQL (Structured Query Language yang berfungsi untuk mengambil data (read dalam DBMS (Database Management System, termasuk Oracle [3]. Pada Oracle, ada tiga tahap proses yang dilakukan dalam pengeksekusian query, yaitu Parsing, Execute dan Fetch. Sebelum proses execute dijalankan, Oracle terlebih dahulu membuat execution plan yang akan menjadi skenario dalam proses excute.Dalam proses pengeksekusian query, terdapat faktor-faktor yang mempengaruhi kinerja query, di antaranya access path (cara pengambilan data dari sebuah tabel dan operasi join (cara menggabungkan data dari dua tabel. Untuk mendapatkan query dengan kinerja optimal, maka diperlukan pertimbangan-pertimbangan dalam menyikapi faktor-faktor tersebut.  Optimasi query merupakan suatu cara untuk mendapatkan query dengan kinerja seoptimal mungkin, terutama dilihat dari sudut pandang waktu. Ada banyak metode untuk mengoptimasi query, tapi pada Penelitian ini, penulis membuat sebuah aplikasi untuk mengoptimasi query dengan metode restrukturisasi SQL statement. Pada metode ini, objek yang dianalisa adalah struktur klausa yang membangun sebuah query. Aplikasi ini memiliki satu input dan lima jenis output. Input dari aplikasi ini adalah sebuah query sedangkan kelima jenis output aplikasi ini adalah berupa query hasil optimasi, saran perbaikan, saran pembuatan indeks baru, execution plan dan data statistik. Cara kerja aplikasi ini dibagi menjadi empat tahap yaitu mengurai query menjadi sub query, mengurai query per-klausa, menentukan access path dan operasi join dan restrukturisasi query.Dari serangkaian ujicoba yang dilakukan penulis, aplikasi telah dapat berjalan sesuai dengan tujuan pembuatan Penelitian ini, yaitu mendapatkan query dengan kinerja optimal.Kata Kunci : Query, SQL, DBMS, Oracle, Parsing, Execute, Fetch, Execution Plan, Access Path, Operasi Join, Restrukturisasi SQL statement.

  17. A controlled trial of automated classification of negation from clinical notes

    Directory of Open Access Journals (Sweden)

    Carruth William

    2005-05-01

    Full Text Available Abstract Background Identification of negation in electronic health records is essential if we are to understand the computable meaning of the records: Our objective is to compare the accuracy of an automated mechanism for assignment of Negation to clinical concepts within a compositional expression with Human Assigned Negation. Also to perform a failure analysis to identify the causes of poorly identified negation (i.e. Missed Conceptual Representation, Inaccurate Conceptual Representation, Missed Negation, Inaccurate identification of Negation. Methods 41 Clinical Documents (Medical Evaluations; sometimes outside of Mayo these are referred to as History and Physical Examinations were parsed using the Mayo Vocabulary Server Parsing Engine. SNOMED-CT™ was used to provide concept coverage for the clinical concepts in the record. These records resulted in identification of Concepts and textual clues to Negation. These records were reviewed by an independent medical terminologist, and the results were tallied in a spreadsheet. Where questions on the review arose Internal Medicine Faculty were employed to make a final determination. Results SNOMED-CT was used to provide concept coverage of the 14,792 Concepts in 41 Health Records from John's Hopkins University. Of these, 1,823 Concepts were identified as negative by Human review. The sensitivity (Recall of the assignment of negation was 97.2% (p Conclusion Automated assignment of negation to concepts identified in health records based on review of the text is feasible and practical. Lexical assignment of negation is a good test of true Negativity as judged by the high sensitivity, specificity and positive likelihood ratio of the test. SNOMED-CT had overall coverage of 88.7% of the concepts being negated.

  18. IDEA: Interactive Display for Evolutionary Analyses

    Directory of Open Access Journals (Sweden)

    Carlton Jane M

    2008-12-01

    Full Text Available Abstract Background The availability of complete genomic sequences for hundreds of organisms promises to make obtaining genome-wide estimates of substitution rates, selective constraints and other molecular evolution variables of interest an increasingly important approach to addressing broad evolutionary questions. Two of the programs most widely used for this purpose are codeml and baseml, parts of the PAML (Phylogenetic Analysis by Maximum Likelihood suite. A significant drawback of these programs is their lack of a graphical user interface, which can limit their user base and considerably reduce their efficiency. Results We have developed IDEA (Interactive Display for Evolutionary Analyses, an intuitive graphical input and output interface which interacts with PHYLIP for phylogeny reconstruction and with codeml and baseml for molecular evolution analyses. IDEA's graphical input and visualization interfaces eliminate the need to edit and parse text input and output files, reducing the likelihood of errors and improving processing time. Further, its interactive output display gives the user immediate access to results. Finally, IDEA can process data in parallel on a local machine or computing grid, allowing genome-wide analyses to be completed quickly. Conclusion IDEA provides a graphical user interface that allows the user to follow a codeml or baseml analysis from parameter input through to the exploration of results. Novel options streamline the analysis process, and post-analysis visualization of phylogenies, evolutionary rates and selective constraint along protein sequences simplifies the interpretation of results. The integration of these functions into a single tool eliminates the need for lengthy data handling and parsing, significantly expediting access to global patterns in the data.

  19. Comparing different kinds of words and word-word relations to test an habituation model of priming.

    Science.gov (United States)

    Rieth, Cory A; Huber, David E

    2017-06-01

    Huber and O'Reilly (2003) proposed that neural habituation exists to solve a temporal parsing problem, minimizing blending between one word and the next when words are visually presented in rapid succession. They developed a neural dynamics habituation model, explaining the finding that short duration primes produce positive priming whereas long duration primes produce negative repetition priming. The model contains three layers of processing, including a visual input layer, an orthographic layer, and a lexical-semantic layer. The predicted effect of prime duration depends both on this assumed representational hierarchy and the assumption that synaptic depression underlies habituation. The current study tested these assumptions by comparing different kinds of words (e.g., words versus non-words) and different kinds of word-word relations (e.g., associative versus repetition). For each experiment, the predictions of the original model were compared to an alternative model with different representational assumptions. Experiment 1 confirmed the prediction that non-words and inverted words require longer prime durations to eliminate positive repetition priming (i.e., a slower transition from positive to negative priming). Experiment 2 confirmed the prediction that associative priming increases and then decreases with increasing prime duration, but remains positive even with long duration primes. Experiment 3 replicated the effects of repetition and associative priming using a within-subjects design and combined these effects by examining target words that were expected to repeat (e.g., viewing the target word 'BACK' after the prime phrase 'back to'). These results support the originally assumed representational hierarchy and more generally the role of habituation in temporal parsing and priming. Copyright © 2017 Elsevier Inc. All rights reserved.

  20. Perl Modules for Constructing Iterators

    Science.gov (United States)

    Tilmes, Curt

    2009-01-01

    The Iterator Perl Module provides a general-purpose framework for constructing iterator objects within Perl, and a standard API for interacting with those objects. Iterators are an object-oriented design pattern where a description of a series of values is used in a constructor. Subsequent queries can request values in that series. These Perl modules build on the standard Iterator framework and provide iterators for some other types of values. Iterator::DateTime constructs iterators from DateTime objects or Date::Parse descriptions and ICal/RFC 2445 style re-currence descriptions. It supports a variety of input parameters, including a start to the sequence, an end to the sequence, an Ical/RFC 2445 recurrence describing the frequency of the values in the series, and a format description that can refine the presentation manner of the DateTime. Iterator::String constructs iterators from string representations. This module is useful in contexts where the API consists of supplying a string and getting back an iterator where the specific iteration desired is opaque to the caller. It is of particular value to the Iterator::Hash module which provides nested iterations. Iterator::Hash constructs iterators from Perl hashes that can include multiple iterators. The constructed iterators will return all the permutations of the iterations of the hash by nested iteration of embedded iterators. A hash simply includes a set of keys mapped to values. It is a very common data structure used throughout Perl programming. The Iterator:: Hash module allows a hash to include strings defining iterators (parsed and dispatched with Iterator::String) that are used to construct an overall series of hash values.

  1. Sistem Monitoring Online Kualitas Air Akuakultur untuk Tambak Udang Menggunakan Aplikasi Berbasis Android

    Directory of Open Access Journals (Sweden)

    Priyani Kusrini

    2016-12-01

    Full Text Available Pemantauan kualitas air di tambak udang yang efektif sangat dibutuhkan agar produksi udang meningkat. Oleh karena itu, pada penelitian ini telah dibuat aplikasi sistem online monitoring kualitas air tambak udang berbasis android. Pengambilan data parameter sensor dilakukan di kampus LIPI menggunakan stasiun sensor yang datanya ditransmisikan ke data logger untuk selanjutnya dikirimkan ke website. Data yang diperoleh dari website berupa data php (Hyper Text Preprocessor dan kemudian diubah menjadi data JSON (Java Script Object Notation. Parsing data dilakukan dengan menggunakan fungsi json_encode untuk membuat string JSON dari objek dan array sehingga data dapat dibaca di Android Studio. Aplikasi ini sangat mudah dioperasikan dan sangat efisien.

  2. Rapid automatic keyword extraction for information retrieval and analysis

    Science.gov (United States)

    Rose, Stuart J [Richland, WA; Cowley,; E, Wendy [Richland, WA; Crow, Vernon L [Richland, WA; Cramer, Nicholas O [Richland, WA

    2012-03-06

    Methods and systems for rapid automatic keyword extraction for information retrieval and analysis. Embodiments can include parsing words in an individual document by delimiters, stop words, or both in order to identify candidate keywords. Word scores for each word within the candidate keywords are then calculated based on a function of co-occurrence degree, co-occurrence frequency, or both. Based on a function of the word scores for words within the candidate keyword, a keyword score is calculated for each of the candidate keywords. A portion of the candidate keywords are then extracted as keywords based, at least in part, on the candidate keywords having the highest keyword scores.

  3. Preserving Employee Privacy in Wellness.

    Science.gov (United States)

    Terry, Paul E

    2017-07-01

    The proposed "Preserving Employee Wellness Programs Act" states that the collection of information about the manifested disease or disorder of a family member shall not be considered an unlawful acquisition of genetic information. The bill recognizes employee privacy protections that are already in place and includes specific language relating to nondiscrimination based on illness. Why did legislation expressly intending to "preserve wellness programs" generate such antipathy about wellness among journalists? This article argues that those who are committed to preserving employee wellness must be equally committed to preserving employee privacy. Related to this, we should better parse between discussions and rules about commonplace health screenings versus much less common genetic testing.

  4. Phonological Working Memory limitations and Agrammatism: Is there a causal relationship between the two?

    Directory of Open Access Journals (Sweden)

    Maria Varkanitsa

    2014-04-01

    These results are inconsistent with theories assuming a single WM capacity that deals with all types of linguistic information. Rather, this study provides evidence that syntactic parsing relies on different resources from what is typically measured in verbal WM tasks. The four patients with Broca’s aphasia that participated in this study presented with similar phonological WM limitations with only quantitative differences in some tasks, despite the fact that only two of them also had sentence comprehension difficulties. The results also suggest that the agrammatic patients suffer from a specific WM deficit, linked to processing syntactic dependencies, which is not tapped by phonological WM tests.

  5. The Development of Design Tools for Fault Tolerant Quantum Dot Cellular Automata Based Logic

    Science.gov (United States)

    Armstrong, Curtis D.; Humphreys, William M.

    2003-01-01

    We are developing software to explore the fault tolerance of quantum dot cellular automata gate architectures in the presence of manufacturing variations and device defects. The Topology Optimization Methodology using Applied Statistics (TOMAS) framework extends the capabilities of the A Quantum Interconnected Network Array Simulator (AQUINAS) by adding front-end and back-end software and creating an environment that integrates all of these components. The front-end tools establish all simulation parameters, configure the simulation system, automate the Monte Carlo generation of simulation files, and execute the simulation of these files. The back-end tools perform automated data parsing, statistical analysis and report generation.

  6. Flex & Bison

    CERN Document Server

    Levine, John

    2009-01-01

    If you need to parse or process text data in Linux or Unix, this useful book explains how to use flex and bison to solve your problems quickly. flex & bison is the long-awaited sequel to the classic O'Reilly book, lex & yacc. In the nearly two decades since the original book was published, the flex and bison utilities have proven to be more reliable and more powerful than the original Unix tools. flex & bison covers the same core functionality vital to Linux and Unix program development, along with several important new topics. You'll find revised tutorials for novices and references for ad

  7. FQC Dashboard: integrates FastQC results into a web-based, interactive, and extensible FASTQ quality control tool.

    Science.gov (United States)

    Brown, Joseph; Pirrung, Meg; McCue, Lee Ann

    2017-06-09

    FQC is software that facilitates quality control of FASTQ files by carrying out a QC protocol using FastQC, parsing results, and aggregating quality metrics into an interactive dashboard designed to richly summarize individual sequencing runs. The dashboard groups samples in dropdowns for navigation among the data sets, utilizes human-readable configuration files to manipulate the pages and tabs, and is extensible with CSV data. FQC is implemented in Python 3 and Javascript, and is maintained under an MIT license. Documentation and source code is available at: https://github.com/pnnl/fqc . joseph.brown@pnnl.gov. © The Author(s) 2017. Published by Oxford University Press.

  8. [Advances of portable electrocardiogram monitor design].

    Science.gov (United States)

    Ding, Shenping; Wang, Yinghai; Wu, Weirong; Deng, Lingli; Lu, Jidong

    2014-06-01

    Portable electrocardiogram monitor is an important equipment in the clinical diagnosis of cardiovascular diseases due to its portable, real-time features. It has a broad application and development prospects in China. In the present review, previous researches on the portable electrocardiogram monitors have been arranged, analyzed and summarized. According to the characteristics of the electrocardiogram (ECG), this paper discusses the ergonomic design of the portable electrocardiogram monitor, including hardware and software. The circuit components and software modules were parsed from the ECG features and system functions. Finally, the development trend and reference are provided for the portable electrocardiogram monitors and for the subsequent research and product design.

  9. Deficit in figure-ground segmentation following closed head injury.

    Science.gov (United States)

    Baylis, G C; Baylis, L L

    1997-08-01

    Patient CB showed a severe impairment in figure-ground segmentation following a closed head injury. Unlike normal subjects, CB was unable to parse smaller and brighter parts of stimuli as figure. Moreover, she did not show the normal effect that symmetrical regions are seen as figure, although she was able to make overt judgments of symmetry. Since she was able to attend normally to isolated objects, CB demonstrates a dissociation between figure ground segmentation and subsequent processes of attention. Despite her severe impairment in figure-ground segmentation, CB showed normal 'parallel' single feature visual search. This suggests that figure-ground segmentation is dissociable from 'preattentive' processes such as visual search.

  10. Distinguishing between "function" and "effect" in genome biology.

    Science.gov (United States)

    Doolittle, W Ford; Brunet, Tyler D P; Linquist, Stefan; Gregory, T Ryan

    2014-05-09

    Much confusion in genome biology results from conflation of possible meanings of the word "function." We suggest that, in this connection, attention should be paid to evolutionary biologists and philosophers who have previously dealt with this problem. We need only decide that although all genomic structures have effects, only some of them should be said to have functions. Although it will very often be difficult or impossible to establish function (strictly defined), it should not automatically be assumed. We enjoin genomicists in particular to pay greater attention to parsing biological effects. © The Author(s) 2014. Published by Oxford University Press on behalf of the Society for Molecular Biology and Evolution.

  11. Development of a monitoring system for the DQMGUI in ElasticSearch and Kibana

    CERN Document Server

    Diaz, Adrian

    2016-01-01

    The Data Quality Monitoring Graphical User Interface (DQMGUI) is the heart of the process of monitoring the quality of data in CMS. The health-status of the DQMGUI and its performances are constantly being monitored and stored in log les, which are subsequently parsed for errors and warnings. This very process allows human operators to act in case of problems. However, the monitoring infrastructure has been migrated to a CERN-hosted ElasticSearch engine in the last year. As a consequence, it is necessary to refactor the old monitoring system, adapting and extending it to be compliant with the new ElasticSearch-based one.

  12. Development of Applications about Hazards and Preventions of Drug Based On Android

    Science.gov (United States)

    Hartatik; Febriyanto, F.; Munawaroh, H.

    2018-03-01

    The number of drug abuse was increase among the younger generation, it caused younger generation fall into drug abuse, and it will lead to physical and mental damage. The lack of knowledge of drugs danger is one of the most potential problems, so in this study we made an application about the types, dangers, and how to avoid its abusement. The application built using PHP programming language with codeiniter framework on admin part, while the parsing data between mobile application server using Javascript Object Notation (JSON). This application has been tested and 85% respondents stated that this application provides positive benefits especially for the socialization of drug abuse.

  13. The "Globularization Hypothesis" of the Language-ready Brain as a Developmental Frame for Prosodic Bootstrapping Theories of Language Acquisition.

    Science.gov (United States)

    Irurtzun, Aritz

    2015-01-01

    In recent research (Boeckx and Benítez-Burraco, 2014a,b) have advanced the hypothesis that our species-specific language-ready brain should be understood as the outcome of developmental changes that occurred in our species after the split from Neanderthals-Denisovans, which resulted in a more globular braincase configuration in comparison to our closest relatives, who had elongated endocasts. According to these authors, the development of a globular brain is an essential ingredient for the language faculty and in particular, it is the centrality occupied by the thalamus in a globular brain that allows its modulatory or regulatory role, essential for syntactico-semantic computations. Their hypothesis is that the syntactico-semantic capacities arise in humans as a consequence of a process of globularization, which significantly takes place postnatally (cf. Neubauer et al., 2010). In this paper, I show that Boeckx and Benítez-Burraco's hypothesis makes an interesting developmental prediction regarding the path of language acquisition: it teases apart the onset of phonological acquisition and the onset of syntactic acquisition (the latter starting significantly later, after globularization). I argue that this hypothesis provides a developmental rationale for the prosodic bootstrapping hypothesis of language acquisition (cf. i.a. Gleitman and Wanner, 1982; Mehler et al., 1988, et seq.; Gervain and Werker, 2013), which claim that prosodic cues are employed for syntactic parsing. The literature converges in the observation that a large amount of such prosodic cues (in particular, rhythmic cues) are already acquired before the completion of the globularization phase, which paves the way for the premises of the prosodic bootstrapping hypothesis, allowing babies to have a rich knowledge of the prosody of their target language before they can start parsing the primary linguistic data syntactically.

  14. Evaluation and development of soil values for the pathway 'soil to plant'. Significance of mercury evaporation for the burden of plants

    International Nuclear Information System (INIS)

    Gaeth, S.; Schlueter, K.

    1998-05-01

    In cooperation with the Ad-hoc working group 'Transfer of heavy metals from soil to plant' of the Laenderarbeitsgemeinschaft Bodenschutz (LABO) the significance of mercury evaporation for the deduction of threshold values in respect of the impact via the pathway soil to plant was investigated. Mercury contamination of food- and feeding stuff plants was examined with special emphasis. For these purposes a lab experiment including three different soils with varying initial mercury load (background level, geogenic and anthropogenic contamination) and two different plant species (parsely and spinach) was carried out under defined conditions in closed lysimeters. Mercury uptake via the roots was minimised since the plants grew in isolated customary substrate which showed a low concentration of mercury. Thus, only the surrounding soil evaporated mercury. The concentrations of mercury in the plants in the background level treatment (0.1 mg Hg/kg dry soil) were 0.15 mg/kg dry matter (spinach) and 0.44 mg/kg dry matter (parsely). The treatment with anthropogenic contaminated soil (111 mg Hg/kg dry soil) resulted in concentrations in the two plants of 2.0 and 2.6 mg/kg dry matter, respectively. A comparable order of magnitude was achieved in the geogenic contaminated treatment (34 mg Hg/kg dry soil) with 2.1 mg/kg dry matter. Experiments conducted with radioactive 203 Hg showed in each case recoveries of 20 to 34% in the leaves regarding the evaporated Hg-tracer. Also in the stem and in the roots Hg-tracer could be detected, indicating a translocation within the plant from leaf to root. By means of a comprehensive literature study the state of the art for Hg-evaporation and Hg-uptake of plants was compiled. Comparing the experimental results with data derived from literature, the Hg-concentrations found are confirmed by results of other authors. (orig.) [de

  15. The globularization hypothesis of the language-ready brain as a developmental frame for prosodic bootstrapping theories of language acquisition

    Directory of Open Access Journals (Sweden)

    Aritz eIrurtzun

    2015-12-01

    Full Text Available In recent research Boeckx & Benítez-Burraco (2014a,b have advanced the hypothesis that our species-specific language-ready brain should be understood as the outcome of developmental changes that occurred in our species after the split from Neanderthals-Denisovans, which resulted in a more globular braincase configuration in comparison to our closest relatives, who had elongated endocasts. According to these authors, the development of a globular brain is an essential ingredient for the language faculty and in particular, it is the centrality occupied by the thalamus in a globular brain that allows its modulatory or regulatory role, essential for syntactico-semantic computations. Their hypothesis is that the syntactico-semantic capacities arise in humans as a consequence of a process of globularization, which significantly takes place postnatally (cf. Neubauer et al. (2010. In this paper, I show that Boeckx & Benítez-Burraco’s hypothesis makes an interesting developmental prediction regarding the path of language acquisition: it teases apart the onset of phonological acquisition and the onset of syntactic acquisition (the latter starting significantly later, after globularization. I argue that this hypothesis provides a developmental rationale for the prosodic bootstrapping hypothesis of language acquisition (cf. i.a. Gleitman & Wanner (1982; Mehler et al. (1988, et seq.; Gervain & Werker (2013, which claim that prosodic cues are employed for syntactic parsing. The literature converges in the observation that a large amount of such prosodic cues (in particular, rhythmic cues are already acquired before the completion of the globularization phase, which paves the way for the premises of prosodic bootstrapping hypothesis, allowing babies to have a rich knowledge of the prosody of their target language before they can start parsing the primary linguistic data syntactically.

  16. Representation of ophthalmology concepts by electronic systems: adequacy of controlled medical terminologies.

    Science.gov (United States)

    Chiang, Michael F; Casper, Daniel S; Cimino, James J; Starren, Justin

    2005-02-01

    To assess the adequacy of 5 controlled medical terminologies (International Classification of Diseases 9, Clinical Modification [ICD9-CM]; Current Procedural Terminology 4 [CPT-4]; Systematized Nomenclature of Medicine, Clinical Terms [SNOMED-CT]; Logical Identifiers, Names, and Codes [LOINC]; Medical Entities Dictionary [MED]) for representing concepts in ophthalmology. Noncomparative case series. Twenty complete ophthalmology case presentations were sequentially selected from a publicly available ophthalmology journal. Each of the 20 cases was parsed into discrete concepts, and each concept was classified along 2 axes: (1) diagnosis, finding, or procedure and (2) ophthalmic or medical concept. Electronic or paper browsers were used to assign a code for every concept in each of the 5 terminologies. Adequacy of assignment for each concept was scored on a 3-point scale. Findings from all 20 case presentations were combined and compared based on a coverage score, which was the average score for all concepts in that terminology. Adequacy of assignment for concepts in each terminology, based on a 3-point Likert scale (0, no match; 1, partial match; 2, complete match). Cases were parsed into 1603 concepts. SNOMED-CT had the highest mean overall coverage score (1.625+/-0.667), followed by MED (0.974+/-0.764), LOINC (0.781+/-0.929), ICD9-CM (0.280+/-0.619), and CPT-4 (0.082+/-0.337). SNOMED-CT also had higher coverage scores than any of the other terminologies for concepts in the diagnosis, finding, and procedure categories. Average coverage scores for ophthalmic concepts were lower than those for medical concepts. Controlled terminologies are required for electronic representation of ophthalmology data. SNOMED-CT had significantly higher content coverage than any other terminology in this study.

  17. Deriving a probabilistic syntacto-semantic grammar for biomedicine based on domain-specific terminologies.

    Science.gov (United States)

    Fan, Jung-Wei; Friedman, Carol

    2011-10-01

    Biomedical natural language processing (BioNLP) is a useful technique that unlocks valuable information stored in textual data for practice and/or research. Syntactic parsing is a critical component of BioNLP applications that rely on correctly determining the sentence and phrase structure of free text. In addition to dealing with the vast amount of domain-specific terms, a robust biomedical parser needs to model the semantic grammar to obtain viable syntactic structures. With either a rule-based or corpus-based approach, the grammar engineering process requires substantial time and knowledge from experts, and does not always yield a semantically transferable grammar. To reduce the human effort and to promote semantic transferability, we propose an automated method for deriving a probabilistic grammar based on a training corpus consisting of concept strings and semantic classes from the Unified Medical Language System (UMLS), a comprehensive terminology resource widely used by the community. The grammar is designed to specify noun phrases only due to the nominal nature of the majority of biomedical terminological concepts. Evaluated on manually parsed clinical notes, the derived grammar achieved a recall of 0.644, precision of 0.737, and average cross-bracketing of 0.61, which demonstrated better performance than a control grammar with the semantic information removed. Error analysis revealed shortcomings that could be addressed to improve performance. The results indicated the feasibility of an approach which automatically incorporates terminology semantics in the building of an operational grammar. Although the current performance of the unsupervised solution does not adequately replace manual engineering, we believe once the performance issues are addressed, it could serve as an aide in a semi-supervised solution. Copyright © 2011 Elsevier Inc. All rights reserved.

  18. Identifying elemental genomic track types and representing them uniformly

    Directory of Open Access Journals (Sweden)

    Gundersen Sveinung

    2011-12-01

    Full Text Available Abstract Background With the recent advances and availability of various high-throughput sequencing technologies, data on many molecular aspects, such as gene regulation, chromatin dynamics, and the three-dimensional organization of DNA, are rapidly being generated in an increasing number of laboratories. The variation in biological context, and the increasingly dispersed mode of data generation, imply a need for precise, interoperable and flexible representations of genomic features through formats that are easy to parse. A host of alternative formats are currently available and in use, complicating analysis and tool development. The issue of whether and how the multitude of formats reflects varying underlying characteristics of data has to our knowledge not previously been systematically treated. Results We here identify intrinsic distinctions between genomic features, and argue that the distinctions imply that a certain variation in the representation of features as genomic tracks is warranted. Four core informational properties of tracks are discussed: gaps, lengths, values and interconnections. From this we delineate fifteen generic track types. Based on the track type distinctions, we characterize major existing representational formats and find that the track types are not adequately supported by any single format. We also find, in contrast to the XML formats, that none of the existing tabular formats are conveniently extendable to support all track types. We thus propose two unified formats for track data, an improved XML format, BioXSD 1.1, and a new tabular format, GTrack 1.0. Conclusions The defined track types are shown to capture relevant distinctions between genomic annotation tracks, resulting in varying representational needs and analysis possibilities. The proposed formats, GTrack 1.0 and BioXSD 1.1, cater to the identified track distinctions and emphasize preciseness, flexibility and parsing convenience.

  19. TopFed: TCGA tailored federated query processing and linking to LOD.

    Science.gov (United States)

    Saleem, Muhammad; Padmanabhuni, Shanmukha S; Ngomo, Axel-Cyrille Ngonga; Iqbal, Aftab; Almeida, Jonas S; Decker, Stefan; Deus, Helena F

    2014-01-01

    The Cancer Genome Atlas (TCGA) is a multidisciplinary, multi-institutional effort to catalogue genetic mutations responsible for cancer using genome analysis techniques. One of the aims of this project is to create a comprehensive and open repository of cancer related molecular analysis, to be exploited by bioinformaticians towards advancing cancer knowledge. However, devising bioinformatics applications to analyse such large dataset is still challenging, as it often requires downloading large archives and parsing the relevant text files. Therefore, it is making it difficult to enable virtual data integration in order to collect the critical co-variates necessary for analysis. We address these issues by transforming the TCGA data into the Semantic Web standard Resource Description Format (RDF), link it to relevant datasets in the Linked Open Data (LOD) cloud and further propose an efficient data distribution strategy to host the resulting 20.4 billion triples data via several SPARQL endpoints. Having the TCGA data distributed across multiple SPARQL endpoints, we enable biomedical scientists to query and retrieve information from these SPARQL endpoints by proposing a TCGA tailored federated SPARQL query processing engine named TopFed. We compare TopFed with a well established federation engine FedX in terms of source selection and query execution time by using 10 different federated SPARQL queries with varying requirements. Our evaluation results show that TopFed selects on average less than half of the sources (with 100% recall) with query execution time equal to one third to that of FedX. With TopFed, we aim to offer biomedical scientists a single-point-of-access through which distributed TCGA data can be accessed in unison. We believe the proposed system can greatly help researchers in the biomedical domain to carry out their research effectively with TCGA as the amount and diversity of data exceeds the ability of local resources to handle its retrieval and

  20. Fast and Efficient XML Data Access for Next-Generation Mass Spectrometry.

    Science.gov (United States)

    Röst, Hannes L; Schmitt, Uwe; Aebersold, Ruedi; Malmström, Lars

    2015-01-01

    In mass spectrometry-based proteomics, XML formats such as mzML and mzXML provide an open and standardized way to store and exchange the raw data (spectra and chromatograms) of mass spectrometric experiments. These file formats are being used by a multitude of open-source and cross-platform tools which allow the proteomics community to access algorithms in a vendor-independent fashion and perform transparent and reproducible data analysis. Recent improvements in mass spectrometry instrumentation have increased the data size produced in a single LC-MS/MS measurement and put substantial strain on open-source tools, particularly those that are not equipped to deal with XML data files that reach dozens of gigabytes in size. Here we present a fast and versatile parsing library for mass spectrometric XML formats available in C++ and Python, based on the mature OpenMS software framework. Our library implements an API for obtaining spectra and chromatograms under memory constraints using random access or sequential access functions, allowing users to process datasets that are much larger than system memory. For fast access to the raw data structures, small XML files can also be completely loaded into memory. In addition, we have improved the parsing speed of the core mzML module by over 4-fold (compared to OpenMS 1.11), making our library suitable for a wide variety of algorithms that need fast access to dozens of gigabytes of raw mass spectrometric data. Our C++ and Python implementations are available for the Linux, Mac, and Windows operating systems. All proposed modifications to the OpenMS code have been merged into the OpenMS mainline codebase and are available to the community at https://github.com/OpenMS/OpenMS.