WorldWideScience

Sample records for linguistic information extraction

  1. Citizen-Centric Urban Planning through Extracting Emotion Information from Twitter in an Interdisciplinary Space-Time-Linguistics Algorithm

    Bernd Resch

    2016-07-01

    Full Text Available Traditional urban planning processes typically happen in offices and behind desks. Modern types of civic participation can enhance those processes by acquiring citizens’ ideas and feedback in participatory sensing approaches like “People as Sensors”. As such, citizen-centric planning can be achieved by analysing Volunteered Geographic Information (VGI data such as Twitter tweets and posts from other social media channels. These user-generated data comprise several information dimensions, such as spatial and temporal information, and textual content. However, in previous research, these dimensions were generally examined separately in single-disciplinary approaches, which does not allow for holistic conclusions in urban planning. This paper introduces TwEmLab, an interdisciplinary approach towards extracting citizens’ emotions in different locations within a city. More concretely, we analyse tweets in three dimensions (space, time, and linguistics, based on similarities between each pair of tweets as defined by a specific set of functional relationships in each dimension. We use a graph-based semi-supervised learning algorithm to classify the data into discrete emotions (happiness, sadness, fear, anger/disgust, none. Our proposed solution allows tweets to be classified into emotion classes in a multi-parametric approach. Additionally, we created a manually annotated gold standard that can be used to evaluate TwEmLab’s performance. Our experimental results show that we are able to identify tweets carrying emotions and that our approach bears extensive potential to reveal new insights into citizens’ perceptions of the city.

  2. Secure information management using linguistic threshold approach

    Ogiela, Marek R

    2013-01-01

    This book details linguistic threshold schemes for information sharing. It examines the opportunities of using these techniques to create new models of managing strategic information shared within a commercial organisation or a state institution.

  3. Information extraction system

    Lemmond, Tracy D; Hanley, William G; Guensche, Joseph Wendell; Perry, Nathan C; Nitao, John J; Kidwell, Paul Brandon; Boakye, Kofi Agyeman; Glaser, Ron E; Prenger, Ryan James

    2014-05-13

    An information extraction system and methods of operating the system are provided. In particular, an information extraction system for performing meta-extraction of named entities of people, organizations, and locations as well as relationships and events from text documents are described herein.

  4. Identification of threats using linguistics-based knowledge extraction.

    Chew, Peter A.

    2008-09-01

    One of the challenges increasingly facing intelligence analysts, along with professionals in many other fields, is the vast amount of data which needs to be reviewed and converted into meaningful information, and ultimately into rational, wise decisions by policy makers. The advent of the world wide web (WWW) has magnified this challenge. A key hypothesis which has guided us is that threats come from ideas (or ideology), and ideas are almost always put into writing before the threats materialize. While in the past the 'writing' might have taken the form of pamphlets or books, today's medium of choice is the WWW, precisely because it is a decentralized, flexible, and low-cost method of reaching a wide audience. However, a factor which complicates matters for the analyst is that material published on the WWW may be in any of a large number of languages. In 'Identification of Threats Using Linguistics-Based Knowledge Extraction', we have sought to use Latent Semantic Analysis (LSA) and other similar text analysis techniques to map documents from the WWW, in whatever language they were originally written, to a common language-independent vector-based representation. This then opens up a number of possibilities. First, similar documents can be found across language boundaries. Secondly, a set of documents in multiple languages can be visualized in a graphical representation. These alone offer potentially useful tools and capabilities to the intelligence analyst whose knowledge of foreign languages may be limited. Finally, we can test the over-arching hypothesis--that ideology, and more specifically ideology which represents a threat, can be detected solely from the words which express the ideology--by using the vector-based representation of documents to predict additional features (such as the ideology) within a framework based on supervised learning. In this report, we present the results of a three-year project of the same name. We believe

  5. Chemical-induced disease relation extraction with various linguistic features.

    Gu, Jinghang; Qian, Longhua; Zhou, Guodong

    2016-01-01

    Understanding the relations between chemicals and diseases is crucial in various biomedical tasks such as new drug discoveries and new therapy developments. While manually mining these relations from the biomedical literature is costly and time-consuming, such a procedure is often difficult to keep up-to-date. To address these issues, the BioCreative-V community proposed a challenging task of automatic extraction of chemical-induced disease (CID) relations in order to benefit biocuration. This article describes our work on the CID relation extraction task on the BioCreative-V tasks. We built a machine learning based system that utilized simple yet effective linguistic features to extract relations with maximum entropy models. In addition to leveraging various features, the hypernym relations between entity concepts derived from the Medical Subject Headings (MeSH)-controlled vocabulary were also employed during both training and testing stages to obtain more accurate classification models and better extraction performance, respectively. We demoted relation extraction between entities in documents to relation extraction between entity mentions. In our system, pairs of chemical and disease mentions at both intra- and inter-sentence levels were first constructed as relation instances for training and testing, then two classification models at both levels were trained from the training examples and applied to the testing examples. Finally, we merged the classification results from mention level to document level to acquire final relations between chemicals and diseases. Our system achieved promisingF-scores of 60.4% on the development dataset and 58.3% on the test dataset using gold-standard entity annotations, respectively. Database URL:https://github.com/JHnlp/BC5CIDTask. © The Author(s) 2016. Published by Oxford University Press.

  6. Describing linguistic information in a behavioural framework: Possible or not?

    De Cooman, G. [Universiteit Gent, Zwijnaarde (Belgium)

    1996-12-31

    The paper discusses important aspects of the representation of linguistic information, using imprecise probabilities with a behavioural interpretation. We define linguistic information as the information conveyed by statements in natural language, but restrict ourselves to simple affirmative statements of the type {open_quote}subject-is-predicate{close_quote}. Taking the behavioural stance, as it is described in detail, we investigate whether it is possible to give a mathematical model for this kind of information. In particular, we evaluate Zadeli`s suggestion that we should use possibility measures to this end. We come to tile conclusion that, generally speaking, possibility measures are possibility models for linguistic information, but that more work should be done in order to evaluate the suggestion that they may be the only ones.

  7. Full-fledged temporal processing: bridging the gap between deep linguistic processing and temporal extraction

    Francisco Costa

    2013-07-01

    Full Text Available The full-fledged processing of temporal information presents specific challenges. These difficulties largely stem from the fact that the temporal meaning conveyed by grammatical means interacts with many extra-linguistic factors (world knowledge, causality, calendar systems, reasoning. This article proposes a novel approach to this problem, based on a hybrid strategy that explores the complementarity of the symbolic and probabilistic methods. A specialized temporal extraction system is combined with a deep linguistic processing grammar. The temporal extraction system extracts eventualities, times and dates mentioned in text, and also temporal relations between them, in line with the tasks of the recent TempEval challenges; and uses machine learning techniques to draw from different sources of information (grammatical and extra-grammatical even if it is not explicitly known how these combine to produce the final temporal meaning being expressed. In turn, the deep computational grammar delivers richer truth-conditional meaning representations of input sentences, which include a principled representation of temporal information, on which higher level tasks, including reasoning, can be based. These deep semantic representations are extended and improved according to the output of the aforementioned temporal extraction module. The prototype implemented shows performance results that increase the quality of the temporal meaning representations and are better than the performance of each of the two components in isolation.

  8. The linguistic realization of information packaging

    Vallduví, Enric; Engdahl, Elisabet

    1996-01-01

    There is increasing awareness of the large degree of crosslinguistic diversity involved in the structural realization of information packaging (or information structure). Whereas English and many Germanic languages primarily exploit intonation for informational purposes, in other languages, like Catalan, syntax plays the primary role in the realization of information packaging and intonation is reduced to a secondary role. In yet another group of languages the primary structural correlate is ...

  9. Multimedia Information Extraction

    Maybury, Mark T

    2012-01-01

    The advent of increasingly large consumer collections of audio (e.g., iTunes), imagery (e.g., Flickr), and video (e.g., YouTube) is driving a need not only for multimedia retrieval but also information extraction from and across media. Furthermore, industrial and government collections fuel requirements for stock media access, media preservation, broadcast news retrieval, identity management, and video surveillance.  While significant advances have been made in language processing for information extraction from unstructured multilingual text and extraction of objects from imagery and vid

  10. A Knowledge Based Recommender System with Multigranular Linguistic Information

    Luis Martinez

    2008-08-01

    Full Text Available Recommender systems are applications that have emerged in the e-commerce area in order to assist users in their searches in electronic shops. These shops usually offer a wide range of items that cover the necessities of a great variety of users. Nevertheless, searching in such a wide range of items could be a very difficult and time-consuming task. Recommender systems assist users to find out suitable items by means of recommendations based on information provided by different sources such as: other users, experts, item features, etc. Most of the recommender systems force users to provide their preferences or necessities using an unique numerical scale of information fixed in advance. In spite of this information is usually related to opinions, tastes and perceptions, therefore, it seems that is usually better expressed in a qualitative way, with linguistic terms, than in a quantitative way, with precise numbers. We propose a Knowledge Based Recommender System that uses the fuzzy linguistic approach to define a flexible framework to capture the uncertainty of the user's preferences. Thus, this framework will allow users to express their necessities in scales closer to their own knowledge, and different from the scale utilized to describe the items.

  11. Challenges in Managing Information Extraction

    Shen, Warren H.

    2009-01-01

    This dissertation studies information extraction (IE), the problem of extracting structured information from unstructured data. Example IE tasks include extracting person names from news articles, product information from e-commerce Web pages, street addresses from emails, and names of emerging music bands from blogs. IE is all increasingly…

  12. A linguistic rule-based approach to extract drug-drug interactions from pharmacological documents.

    Segura-Bedmar, Isabel; Martínez, Paloma; de Pablo-Sánchez, César

    2011-03-29

    A drug-drug interaction (DDI) occurs when one drug influences the level or activity of another drug. The increasing volume of the scientific literature overwhelms health care professionals trying to be kept up-to-date with all published studies on DDI. This paper describes a hybrid linguistic approach to DDI extraction that combines shallow parsing and syntactic simplification with pattern matching. Appositions and coordinate structures are interpreted based on shallow syntactic parsing provided by the UMLS MetaMap tool (MMTx). Subsequently, complex and compound sentences are broken down into clauses from which simple sentences are generated by a set of simplification rules. A pharmacist defined a set of domain-specific lexical patterns to capture the most common expressions of DDI in texts. These lexical patterns are matched with the generated sentences in order to extract DDIs. We have performed different experiments to analyze the performance of the different processes. The lexical patterns achieve a reasonable precision (67.30%), but very low recall (14.07%). The inclusion of appositions and coordinate structures helps to improve the recall (25.70%), however, precision is lower (48.69%). The detection of clauses does not improve the performance. Information Extraction (IE) techniques can provide an interesting way of reducing the time spent by health care professionals on reviewing the literature. Nevertheless, no approach has been carried out to extract DDI from texts. To the best of our knowledge, this work proposes the first integral solution for the automatic extraction of DDI from biomedical texts.

  13. How age and linguistic competence affect memory for heard information

    Bruce A Schneider

    2016-05-01

    Full Text Available The short-term memory performance of a group of younger adults, for whom English was a second language (young EL2 listeners, was compared to that of younger and older adults for whom English was their first language (EL1 listeners. To-be-remembered words were presented in noise and in quiet. When presented in noise, the listening situation was adjusted to insure that the likelihood of recognizing the individual words was comparable for all groups. Previous studies which used the same paradigm found memory performance of older EL1 adults on this paired-associate task to be poorer than that of their younger EL1 counterparts both in quiet and in a background of babble. The purpose of the present study was to investigate whether the less well-established semantic and linguistic skills of EL2 listeners would also lead to memory deficits even after equating for word recognition as was done for the younger and older EL1 listeners. No significant differences in memory performance were found between young EL1 and EL2 listeners after equating for word recognition, indicating that the EL2 listeners’ poorer semantic and linguistic skills had little effect on their ability to memorize and recall paired associates. This result is consistent with the hypothesis that age-related declines in memory are primarily due to age-related declines in the perceptual and attentional processes required for extracting the word pairs from a background babble of voices. Such declines are likely to increase the load on higher-order (possibly limited cognitive processes supporting memory. The problems that these results pose for the comprehension of spoken language in these three groups are discussed.

  14. Exploring interdisciplinary relationships between linguistics and information retrieval from the 1960s to today

    Engerer, Volkmar Paul

    2017-01-01

    This article explores how linguistics has influenced information retrieval (IR) and attempts to explain the impact of linguistics through an analysis of internal developments in information science generally, and IR in particular. It notes that information science/IR has been evolving from a case...... science into a fully fledged, “disciplined”/disciplinary science. The article establishes correspondences between linguistics and information science/IR using the three established IR paradigms—physical, cognitive, and computational—as a frame of reference. The current relationship between information...... science/IR and linguistics is elucidated through discussion of some recent information science publications dealing with linguistic topics and a novel technique, “keyword collocation analysis,” is introduced. Insights from interdisciplinarity research and case theory are also discussed. It is demonstrated...

  15. Quantifying the information in the long-range order of words: semantic structures and universal linguistic constraints.

    Montemurro, Marcelo A

    2014-06-01

    We review some recent progress on the characterisation of long-range patterns of word use in language using methods from information theory. In particular, two levels of structure in language are considered. The first level corresponds to the patterns of words usage over different contextual domains. A direct application of information theory to quantify the specificity of words across different sections of a linguistic sequence leads to a measure of semantic information. Moreover, a natural scale emerges that characterises the typical size of semantic structures. Since the information measure is made up of additive contributions from individual words, it is possible to rank the words according to their overall weight in the total information. This allows the extraction of keywords most relevant to the semantic content of the sequence without any prior knowledge of the language. The second level considered is the complex structure of correlations among words in linguistic sequences. The degree of order in language can be quantified by means of the entropy. Reliable estimates of the entropy were obtained from corpora of texts from several linguistic families by means of lossless compression algorithms. The value of the entropy fluctuates across different languages since it depends on linguistic organisation at various levels. However, when a measure of relative entropy that specifically quantifies the degree of word ordering in language is estimated, it presents an almost constant value over all the linguistic families studied. This suggests that the entropy of word ordering is a novel quantitative linguistic universal. Copyright © 2013 Elsevier Ltd. All rights reserved.

  16. Scenario Customization for Information Extraction

    Yangarber, Roman

    2001-01-01

    Information Extraction (IE) is an emerging NLP technology, whose function is to process unstructured, natural language text, to locate specific pieces of information, or facts, in the text, and to use these facts to fill a database...

  17. Semi-automatic Term Extraction for an isiZulu Linguistic Terms ...

    user

    This paper advances the use of frequency analysis and the keyword analysis as strategies to extract terms for the compilation of the dictionary of isiZulu linguistic terms. The study uses the isiZulu. National Corpus (INC) of about 1,2 million tokens as a reference corpus as well as an LSP corpus of about 100,000 tokens as a ...

  18. Interface of Linguistic and Visual Information During Audience Design.

    Fukumura, Kumiko

    2015-08-01

    Evidence suggests that speakers can take account of the addressee's needs when referring. However, what representations drive the speaker's audience design has been less clear. This study aims to go beyond previous studies by investigating the interplay between the visual and linguistic context during audience design. Speakers repeated subordinate descriptions (e.g., firefighter) given in the prior linguistic context less and used basic-level descriptions (e.g., man) more when the addressee did not hear the linguistic context than when s/he did. But crucially, this effect happened only when the referent lacked the visual attributes associated with the expressions (e.g., the referent was in plain clothes rather than in a firefighter uniform), so there was no other contextual cue available for the identification of the referent. This suggests that speakers flexibly use different contextual cues to help their addressee map the referring expression onto the intended referent. In addition, speakers used fewer pronouns when the addressee did not hear the linguistic antecedent than when s/he did. This suggests that although speakers may be egocentric during anaphoric reference (Fukumura & Van Gompel, 2012), they can cooperatively avoid pronouns when the linguistic antecedents were not shared with their addressee during initial reference. © 2014 Cognitive Science Society, Inc.

  19. Extracting useful information from images

    Kucheryavskiy, Sergey

    2011-01-01

    The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic and heter......The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic...

  20. Information literacy and abstracting: interdisciplinary issues for linguists and information professionals

    Tibor Koltay

    2010-04-01

    Full Text Available Information literacy is a complex phenomenon that requires a multifaceted interdisciplinary approach as it is related to verbal communication, literacy, functional literacy and academic literacy, including issues of plagiarism. It also includes text authoring in a full range of genres, among others abstracts. Abstracting is a well-known act of verbal communication, and abstracts are a genre of written communication. The essence of abstracting is summarizing information making use of critical reading. Abstracting thus can be regarded as one of the instances of exercising information literacy on a higher level. Both information literacy and abstracting are of prime professional interest for linguists (among others in the field of ESP and information professionals.

  1. Incorporating linguistic, probabilistic, and possibilistic information in a risk-based approach for ranking contaminated sites.

    Zhang, Kejiang; Achari, Gopal; Pei, Yuansheng

    2010-10-01

    Different types of uncertain information-linguistic, probabilistic, and possibilistic-exist in site characterization. Their representation and propagation significantly influence the management of contaminated sites. In the absence of a framework with which to properly represent and integrate these quantitative and qualitative inputs together, decision makers cannot fully take advantage of the available and necessary information to identify all the plausible alternatives. A systematic methodology was developed in the present work to incorporate linguistic, probabilistic, and possibilistic information into the Preference Ranking Organization METHod for Enrichment Evaluation (PROMETHEE), a subgroup of Multi-Criteria Decision Analysis (MCDA) methods for ranking contaminated sites. The identification of criteria based on the paradigm of comparative risk assessment provides a rationale for risk-based prioritization. Uncertain linguistic, probabilistic, and possibilistic information identified in characterizing contaminated sites can be properly represented as numerical values, intervals, probability distributions, and fuzzy sets or possibility distributions, and linguistic variables according to their nature. These different kinds of representation are first transformed into a 2-tuple linguistic representation domain. The propagation of hybrid uncertainties is then carried out in the same domain. This methodology can use the original site information directly as much as possible. The case study shows that this systematic methodology provides more reasonable results. © 2010 SETAC.

  2. How Age and Linguistic Competence Affect Memory for Heard Information.

    Schneider, Bruce A; Avivi-Reich, Meital; Leung, Caterina; Heinrich, Antje

    2016-01-01

    The short-term memory performance of a group of younger adults, for whom English was a second language (young EL2 listeners), was compared to that of younger and older adults for whom English was their first language (EL1 listeners). To-be-remembered words were presented in noise and in quiet. When presented in noise, the listening situation was adjusted to ensure that the likelihood of recognizing the individual words was comparable for all groups. Previous studies which used the same paradigm found memory performance of older EL1 adults on this paired-associate task to be poorer than that of their younger EL1 counterparts both in quiet and in a background of babble. The purpose of the present study was to investigate whether the less well-established semantic and linguistic skills of EL2 listeners would also lead to memory deficits even after equating for word recognition as was done for the younger and older EL1 listeners. No significant differences in memory performance were found between young EL1 and EL2 listeners after equating for word recognition, indicating that the EL2 listeners' poorer semantic and linguistic skills had little effect on their ability to memorize and recall paired associates. This result is consistent with the hypothesis that age-related declines in memory are primarily due to age-related declines in higher-order processes supporting stream segregation and episodic memory. Such declines are likely to increase the load on higher-order (possibly limited) cognitive processes supporting memory. The problems that these results pose for the comprehension of spoken language in these three groups are discussed.

  3. Sensory Intelligence for Extraction of an Abstract Auditory Rule: A Cross-Linguistic Study.

    Guo, Xiao-Tao; Wang, Xiao-Dong; Liang, Xiu-Yuan; Wang, Ming; Chen, Lin

    2018-02-21

    In a complex linguistic environment, while speech sounds can greatly vary, some shared features are often invariant. These invariant features constitute so-called abstract auditory rules. Our previous study has shown that with auditory sensory intelligence, the human brain can automatically extract the abstract auditory rules in the speech sound stream, presumably serving as the neural basis for speech comprehension. However, whether the sensory intelligence for extraction of abstract auditory rules in speech is inherent or experience-dependent remains unclear. To address this issue, we constructed a complex speech sound stream using auditory materials in Mandarin Chinese, in which syllables had a flat lexical tone but differed in other acoustic features to form an abstract auditory rule. This rule was occasionally and randomly violated by the syllables with the rising, dipping or falling tone. We found that both Chinese and foreign speakers detected the violations of the abstract auditory rule in the speech sound stream at a pre-attentive stage, as revealed by the whole-head recordings of mismatch negativity (MMN) in a passive paradigm. However, MMNs peaked earlier in Chinese speakers than in foreign speakers. Furthermore, Chinese speakers showed different MMN peak latencies for the three deviant types, which paralleled recognition points. These findings indicate that the sensory intelligence for extraction of abstract auditory rules in speech sounds is innate but shaped by language experience. Copyright © 2018 IBRO. Published by Elsevier Ltd. All rights reserved.

  4. Linguistic Imperialism

    Phillipson, Robert

    2013-01-01

    The study of linguistic imperialism focuses on how and why certain languages dominate internationally, and attempts to account for such dominance in a theoretically informed way.......The study of linguistic imperialism focuses on how and why certain languages dominate internationally, and attempts to account for such dominance in a theoretically informed way....

  5. Models and Methods of Aggregating Linguistic Information in Multi-criteria Hierarchical Quality Assessment Systems

    Azarnova, T. V.; Titova, I. A.; Barkalov, S. A.

    2018-03-01

    The article presents an algorithm for obtaining an integral assessment of the quality of an organization from the perspective of customers, based on the method of aggregating linguistic information on a multilevel hierarchical system of quality assessment. The algorithm is of a constructive nature, it provides not only the possibility of obtaining an integral evaluation, but also the development of a quality improvement strategy based on the method of linguistic decomposition, which forms the minimum set of areas of work with clients whose quality change will allow obtaining the required level of integrated quality assessment.

  6. The effect of linguistic devices in information presentation messages on comprehension and recall

    Tietze, M.I.; Winterboer, A.; Moore, J.D.

    2009-01-01

    In this paper we examine the effect of linguistic devices on recall and comprehension in information presentation using both recall and eye-tracking data. In addition, the results were validated via an experiment using Amazon's Mechanical Turk micro-task environment.

  7. Extracting information from multiplex networks

    Iacovacci, Jacopo; Bianconi, Ginestra

    2016-06-01

    Multiplex networks are generalized network structures that are able to describe networks in which the same set of nodes are connected by links that have different connotations. Multiplex networks are ubiquitous since they describe social, financial, engineering, and biological networks as well. Extending our ability to analyze complex networks to multiplex network structures increases greatly the level of information that is possible to extract from big data. For these reasons, characterizing the centrality of nodes in multiplex networks and finding new ways to solve challenging inference problems defined on multiplex networks are fundamental questions of network science. In this paper, we discuss the relevance of the Multiplex PageRank algorithm for measuring the centrality of nodes in multilayer networks and we characterize the utility of the recently introduced indicator function Θ ˜ S for describing their mesoscale organization and community structure. As working examples for studying these measures, we consider three multiplex network datasets coming for social science.

  8. Prosody and informativity: A cross-linguistic investigation

    Ouyang, Iris Chuoying

    This dissertation aims to extend our knowledge of prosody -- in particular, what kinds of information may be conveyed through prosody, which prosodic dimensions may be used to convey them, and how individual speakers differ from one another in how they use prosody. Four production studies were conducted to examine how various factors interact with one another in shaping the prosody of an utterance and how prosody fulfills its multi-functional role. Experiments 1 explores the interaction between two types of informativity, namely information structure and information-theoretic properties. The results show that the prosodic consequences of new-information focus are modulated by the focused word's frequency, whereas the prosodic consequences of corrective focus are modulated by the focused word's probability in the context. Furthermore, f0 ranges appear to be more informative than f0 shapes in reflecting informativity across speakers. Specifically, speakers seem to have individual 'preferences' regarding f0 shapes, the f0 ranges they use for an utterance, and the magnitude of differences in f0 ranges by which they mark information-structural distinctions. In contrast, there is more cross-speaker validity in the actual directions of differences in f0 ranges between information-structural types. Experiments 2 and 3 further show that the interaction found between corrective focus and contextual probability depends on the interlocutor's knowledge state. When the interlocutor has no access to the crucial information concerning utterances' contextual probability, speakers prosodically emphasize contextually improbable corrections, but not contextually probable corrections. Furthermore, speakers prosodically emphasize the corrections in response to contextually probable misstatements, but not the corrections in response to contextually improbable misstatements. In contrast, completely opposite patterns are found when words' contextual probability is shared knowledge between

  9. Cloud decision model for selecting sustainable energy crop based on linguistic intuitionistic information

    Peng, Hong-Gang; Wang, Jian-Qiang

    2017-11-01

    In recent years, sustainable energy crop has become an important energy development strategy topic in many countries. Selecting the most sustainable energy crop is a significant problem that must be addressed during any biofuel production process. The focus of this study is the development of an innovative multi-criteria decision-making (MCDM) method to handle sustainable energy crop selection problems. Given that various uncertain data are encountered in the evaluation of sustainable energy crops, linguistic intuitionistic fuzzy numbers (LIFNs) are introduced to present the information necessary to the evaluation process. Processing qualitative concepts requires the effective support of reliable tools; then, a cloud model can be used to deal with linguistic intuitionistic information. First, LIFNs are converted and a novel concept of linguistic intuitionistic cloud (LIC) is proposed. The operations, score function and similarity measurement of the LICs are defined. Subsequently, the linguistic intuitionistic cloud density-prioritised weighted Heronian mean operator is developed, which served as the basis for the construction of an applicable MCDM model for sustainable energy crop selection. Finally, an illustrative example is provided to demonstrate the proposed method, and its feasibility and validity are further verified by comparing it with other existing methods.

  10. Transductive Pattern Learning for Information Extraction

    McLernon, Brian; Kushmerick, Nicholas

    2006-01-01

    .... We present TPLEX, a semi-supervised learning algorithm for information extraction that can acquire extraction patterns from a small amount of labelled text in conjunction with a large amount of unlabelled text...

  11. Linguistic Engineering and Linguistic of Engineering: Adaptation of Linguistic Paradigm for Circumstance of Engineering Epoch

    Natalya Halina

    2014-01-01

    The article is devoted to the problems of linguistic knowledge in the Engineering Epoch. Engineering Epoch is the time of adaptation to the information flows by knowledge management, The system of adaptation mechanisms is connected with linguistic and linguistic technologies, forming in new linguistic patterns Linguistic Engineering and Linguistic of Engineering.

  12. Automated Extraction of Substance Use Information from Clinical Texts.

    Wang, Yan; Chen, Elizabeth S; Pakhomov, Serguei; Arsoniadis, Elliot; Carter, Elizabeth W; Lindemann, Elizabeth; Sarkar, Indra Neil; Melton, Genevieve B

    2015-01-01

    Within clinical discourse, social history (SH) includes important information about substance use (alcohol, drug, and nicotine use) as key risk factors for disease, disability, and mortality. In this study, we developed and evaluated a natural language processing (NLP) system for automated detection of substance use statements and extraction of substance use attributes (e.g., temporal and status) based on Stanford Typed Dependencies. The developed NLP system leveraged linguistic resources and domain knowledge from a multi-site social history study, Propbank and the MiPACQ corpus. The system attained F-scores of 89.8, 84.6 and 89.4 respectively for alcohol, drug, and nicotine use statement detection, as well as average F-scores of 82.1, 90.3, 80.8, 88.7, 96.6, and 74.5 respectively for extraction of attributes. Our results suggest that NLP systems can achieve good performance when augmented with linguistic resources and domain knowledge when applied to a wide breadth of substance use free text clinical notes.

  13. Predicting panel scores by linguistic analysis

    Van den Besselaar, P.; Stout, L.; Gou, X

    2016-07-01

    In this paper we explore the use of text analysis for deriving quality indicators of project proposals. We do full text analysis of 3030 review reports. After term extraction, we aggregate the term occurrences to linguistic categories. Using thse linguistic categories as independent variables, we study how well these predict the grading by the review panels. Together, the different linguistic categories explain about 50% of the variance in the grading of the applications. The relative importance of the different linguistic categories inform us about the way the panels work. This can be used to develop altmetrics for the quality of the peer and panel review processes. (Author)

  14. The role of linguists in planning and making dictionaries in modern information society

    Bergenholtz, Henning

    2013-01-01

    , but not necessarily for all, e.g. not for meaning items, collocations or synonym items. This will be discussed outgoing from the description of a database and the concept for one polyfunctional and five monofunctional monolingual general dictionaries. The first monofunctional dictionary is a reception dictionary...... type of expert is best suited to make modern dictionaries in the information age? This question is quite complex. We have different kinds of dictionaries, e.g. general language and special language dictionaries. And we have different kinds of lexicographers, e.g. 1. metalexicographer, 2. practical...... lexicographer making the concept for a planned dictionary, 3. lexicographer making the concrete dictionary articles or parts of them. For (2) a linguist is of course not the natural choice. For (3) we need linguists for certain kind of dictionaries and certain data types in general languages...

  15. Information Extraction for Social Media

    Habib, M. B.; Keulen, M. van

    2014-01-01

    The rapid growth in IT in the last two decades has led to a growth in the amount of information available online. A new style for sharing information is social media. Social media is a continuously instantly updated source of information. In this position paper, we propose a framework for

  16. Informational and linguistic analysis of large genomic sequence collections via efficient Hadoop cluster algorithms.

    Ferraro Petrillo, Umberto; Roscigno, Gianluca; Cattaneo, Giuseppe; Giancarlo, Raffaele

    2018-06-01

    Information theoretic and compositional/linguistic analysis of genomes have a central role in bioinformatics, even more so since the associated methodologies are becoming very valuable also for epigenomic and meta-genomic studies. The kernel of those methods is based on the collection of k-mer statistics, i.e. how many times each k-mer in {A,C,G,T}k occurs in a DNA sequence. Although this problem is computationally very simple and efficiently solvable on a conventional computer, the sheer amount of data available now in applications demands to resort to parallel and distributed computing. Indeed, those type of algorithms have been developed to collect k-mer statistics in the realm of genome assembly. However, they are so specialized to this domain that they do not extend easily to the computation of informational and linguistic indices, concurrently on sets of genomes. Following the well-established approach in many disciplines, and with a growing success also in bioinformatics, to resort to MapReduce and Hadoop to deal with 'Big Data' problems, we present KCH, the first set of MapReduce algorithms able to perform concurrently informational and linguistic analysis of large collections of genomic sequences on a Hadoop cluster. The benchmarking of KCH that we provide indicates that it is quite effective and versatile. It is also competitive with respect to the parallel and distributed algorithms highly specialized to k-mer statistics collection for genome assembly problems. In conclusion, KCH is a much needed addition to the growing number of algorithms and tools that use MapReduce for bioinformatics core applications. The software, including instructions for running it over Amazon AWS, as well as the datasets are available at http://www.di-srv.unisa.it/KCH. umberto.ferraro@uniroma1.it. Supplementary data are available at Bioinformatics online.

  17. The role of linguistic experience in the processing of probabilistic information in production.

    Gustafson, Erin; Goldrick, Matthew

    2018-01-01

    Speakers track the probability that a word will occur in a particular context and utilize this information during phonetic processing. For example, content words that have high probability within a discourse tend to be realized with reduced acoustic/articulatory properties. Such probabilistic information may influence L1 and L2 speech processing in distinct ways (reflecting differences in linguistic experience across groups and the overall difficulty of L2 speech processing). To examine this issue, L1 and L2 speakers performed a referential communication task, describing sequences of simple actions. The two groups of speakers showed similar effects of discourse-dependent probabilistic information on production, suggesting that L2 speakers can successfully track discourse-dependent probabilities and use such information to modulate phonetic processing.

  18. A fuzzy MCDM approach for evaluating school performance based on linguistic information

    Musani, Suhaina; Jemain, Abdul Aziz

    2013-11-01

    Decision making is the process of finding the best option among the feasible alternatives. This process should consider a variety of criteria, but this study only focus on academic achievement. The data used is the percentage of candidates who obtained Malaysian Certificate of Education (SPM) in Melaka based on school academic achievement for each subject. 57 secondary schools in Melaka as listed by the Ministry of Education involved in this study. Therefore the school ranking can be done using MCDM (Multi Criteria Decision Making) methods. The objective of this study is to develop a rational method for evaluating school performance based on linguistic information. Since the information or level of academic achievement provided in linguistic manner, there is a possible chance of getting incomplete or uncertain problems. So in order to overcome the situation, the information could be provided as fuzzy numbers. Since fuzzy set represents the uncertainty in human perceptions. In this research, VIKOR (Multi Criteria Optimization and Compromise Solution) has been used as a MCDM tool for the school ranking process in fuzzy environment. Results showed that fuzzy set theory can solve the limitations of using MCDM when there is uncertainty problems exist in the data.

  19. Information Extraction From Chemical Patents

    Sandra Bergmann

    2012-01-01

    Full Text Available The development of new chemicals or pharmaceuticals is preceded by an indepth analysis of published patents in this field. This information retrieval is a costly and time inefficient step when done by a human reader, yet it is mandatory for potential success of an investment. The goal of the research project UIMA-HPC is to automate and hence speed-up the process of knowledge mining about patents. Multi-threaded analysis engines, developed according to UIMA (Unstructured Information Management Architecture standards, process texts and images in thousands of documents in parallel. UNICORE (UNiform Interface to COmputing Resources workflow control structures make it possible to dynamically allocate resources for every given task to gain best cpu-time/realtime ratios in an HPC environment.

  20. Forensic linguistics: Applications of forensic linguistics methods to anonymous letters

    NOVÁKOVÁ, Veronika

    2011-01-01

    The title of my bachelor work is ?Forensic linguistics: Applications of forensic linguistics methods to anonymous letters?. Forensic linguistics is young and not very known branch of applied linguistics. This bachelor work wants to introduce forensic linguistics and its method. The bachelor work has two parts ? theory and practice. The theoretical part informs about forensic linguistics in general. Its two basic aspects utilized in forensic science and respective methods. The practical part t...

  1. Preferences for (In)Formal Language: Correlations with Attitudes toward Linguistic Variation, Multilingualism, Tolerance of Ambiguity, and Residence Abroad

    van Compernolle, Rémi A.

    2017-01-01

    Drawing on data collected via a web-based survey, the study investigates the relationship between preferences for (in)formal language and attitudes toward linguistic variation among a large group of monolingual and multilingual adults (n = 379). Also explored are the links between preferences for (in)formal language and several secondary…

  2. Social adaptation in multi-agent model of linguistic categorization is affected by network information flow.

    Zubek, Julian; Denkiewicz, Michał; Barański, Juliusz; Wróblewski, Przemysław; Rączaszek-Leonardi, Joanna; Plewczynski, Dariusz

    2017-01-01

    This paper explores how information flow properties of a network affect the formation of categories shared between individuals, who are communicating through that network. Our work is based on the established multi-agent model of the emergence of linguistic categories grounded in external environment. We study how network information propagation efficiency and the direction of information flow affect categorization by performing simulations with idealized network topologies optimizing certain network centrality measures. We measure dynamic social adaptation when either network topology or environment is subject to change during the experiment, and the system has to adapt to new conditions. We find that both decentralized network topology efficient in information propagation and the presence of central authority (information flow from the center to peripheries) are beneficial for the formation of global agreement between agents. Systems with central authority cope well with network topology change, but are less robust in the case of environment change. These findings help to understand which network properties affect processes of social adaptation. They are important to inform the debate on the advantages and disadvantages of centralized systems.

  3. Semio-Linguistic Creative Actualization of the Concept “Information About the Future” in the Science Fiction Discourse

    Andrey Vladimirovich Olyanich

    2015-11-01

    Full Text Available The article deals with the cognitive category of "semio-linguistic creativity", that serves as a tool for implification of the concept "Information about the future" in the science fiction discourse. The correlation between the categories of future and information is studied in semio-linguistic aspect; the conceptual core, internal and external zones of the concept "Information about the future" are explored in connection with the concepts "Future", "Myths" and "Expectations" that are viewed as belonging to the science fiction discourse. The following issues are considered: coordination between axiological and imaginative spheres of the concept "Information about the future"; the mechanism of transforming information from present and past into the future by means of literary imagination, which is aimed at constructing the imaginary hyper-reality with the use of concepts that belong to contemporary reality; it is stated that such activity lays the basis for multiple forecasts. After the analysis of the novels by Vasily Golovachev, a famous Russian science fiction writer, the authors present their interpretation of the process of science-fiction discourse unfolding that involves groups of signs from the following semio-linguistic clusters (The Man as a species; Food; Space, Earth, their semantic content is directly related to the needs of the future. The proposed algorithm of analysis may be applied to studying other semio-linguistic clusters: "Habitat," "Communications", "Social Environment", "Transport", "Technology", that may explicate the concept "Information about the future".

  4. Fuzzy Linguistic Knowledge Based Behavior Extraction for Building Energy Management Systems

    Dumidu Wijayasekara; Milos Manic

    2013-08-01

    Significant portion of world energy production is consumed by building Heating, Ventilation and Air Conditioning (HVAC) units. Thus along with occupant comfort, energy efficiency is also an important factor in HVAC control. Modern buildings use advanced Multiple Input Multiple Output (MIMO) control schemes to realize these goals. However, since the performance of HVAC units is dependent on many criteria including uncertainties in weather, number of occupants, and thermal state, the performance of current state of the art systems are sub-optimal. Furthermore, because of the large number of sensors in buildings, and the high frequency of data collection, large amount of information is available. Therefore, important behavior of buildings that compromise energy efficiency or occupant comfort is difficult to identify. This paper presents an easy to use and understandable framework for identifying such behavior. The presented framework uses human understandable knowledge-base to extract important behavior of buildings and present it to users via a graphical user interface. The presented framework was tested on a building in the Pacific Northwest and was shown to be able to identify important behavior that relates to energy efficiency and occupant comfort.

  5. Extracting Information from Multimedia Meeting Collections

    Gatica-Perez, Daniel; Zhang, Dong; Bengio, Samy

    2005-01-01

    Multimedia meeting collections, composed of unedited audio and video streams, handwritten notes, slides, and electronic documents that jointly constitute a raw record of complex human interaction processes in the workplace, have attracted interest due to the increasing feasibility of recording them in large quantities, by the opportunities for information access and retrieval applications derived from the automatic extraction of relevant meeting information, and by the challenges that the ext...

  6. DKIE: Open Source Information Extraction for Danish

    Derczynski, Leon; Field, Camilla Vilhelmsen; Bøgh, Kenneth Sejdenfaden

    2014-01-01

    Danish is a major Scandinavian language spoken daily by around six million people. However, it lacks a unified, open set of NLP tools. This demonstration will introduce DKIE, an extensible open-source toolkit for processing Danish text. We implement an information extraction architecture for Danish...

  7. From Shared Contexts to Syntactic Categories: The Role of Distributional Information in Learning Linguistic Form-Classes

    Reeder, Patricia A.; Newport, Elissa L.; Aslin, Richard N.

    2013-01-01

    A fundamental component of language acquisition involves organizing words into grammatical categories. Previous literature has suggested a number of ways in which this categorization task might be accomplished. Here we ask whether the patterning of the words in a corpus of linguistic input ("distributional information") is sufficient, along with a…

  8. Developing written information for cancer survivors from culturally and linguistically diverse backgrounds: Lessons learnt

    Georgina Wiley

    2018-01-01

    Full Text Available Australia is a multicultural nation with a large migrant population. Migrants with cancer report inferior quality of life and the need for more information in their own language. This paper describes lessons learnt from developing culturally appropriate written information resources with and for Arabic, Italian, and Vietnamese cancer survivors and carers. The information needs of survivors from these language groups as well as guidelines for the development of written resources for culturally diverse populations were identified through literature review. Community consultation was undertaken with focus groups. The content was developed and tested with health professionals who spoke the appropriate language and focus group participants, ensuring relevance and appropriateness. Resource design and dissemination were informed through community consultation. A number of key tasks for developing resources were identified as follows: (1 community engagement and consultation; (2 culturally sensitive data collection; (3 focus group facilitators (recruitment and training; (4 content development; (5 translation and review process; (6 design; and (7 sustainability. This project reinforced literature review findings on the importance of cultural sensitivity in the development of resources. Engaging with community groups and incorporating culturally appropriate recruitment strategies optimises recruitment to focus groups and facilitates content development. Stakeholders and lay persons from the intended ethnic-minority communities should be involved in the development and formative evaluation of resources to ensure appropriateness and relevance and in the dissemination strategy to optimize penetration. We believe the lessons we have learnt will be relevant to any group intending to develop health information for culturally and linguistic diverse groups.

  9. Unsupervised information extraction by text segmentation

    Cortez, Eli

    2013-01-01

    A new unsupervised approach to the problem of Information Extraction by Text Segmentation (IETS) is proposed, implemented and evaluated herein. The authors' approach relies on information available on pre-existing data to learn how to associate segments in the input string with attributes of a given domain relying on a very effective set of content-based features. The effectiveness of the content-based features is also exploited to directly learn from test data structure-based features, with no previous human-driven training, a feature unique to the presented approach. Based on the approach, a

  10. Extracting the information backbone in online system.

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  11. Information extraction from muon radiography data

    Borozdin, K.N.; Asaki, T.J.; Chartrand, R.; Hengartner, N.W.; Hogan, G.E.; Morris, C.L.; Priedhorsky, W.C.; Schirato, R.C.; Schultz, L.J.; Sottile, M.J.; Vixie, K.R.; Wohlberg, B.E.; Blanpied, G.

    2004-01-01

    Scattering muon radiography was proposed recently as a technique of detection and 3-d imaging for dense high-Z objects. High-energy cosmic ray muons are deflected in matter in the process of multiple Coulomb scattering. By measuring the deflection angles we are able to reconstruct the configuration of high-Z material in the object. We discuss the methods for information extraction from muon radiography data. Tomographic methods widely used in medical images have been applied to a specific muon radiography information source. Alternative simple technique based on the counting of high-scattered muons in the voxels seems to be efficient in many simulated scenes. SVM-based classifiers and clustering algorithms may allow detection of compact high-Z object without full image reconstruction. The efficiency of muon radiography can be increased using additional informational sources, such as momentum estimation, stopping power measurement, and detection of muonic atom emission.

  12. Extracting the information backbone in online system.

    Qian-Ming Zhang

    Full Text Available Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  13. Extracting the Information Backbone in Online System

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such “less can be more” feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency. PMID:23690946

  14. Chaotic spectra: How to extract dynamic information

    Taylor, H.S.; Gomez Llorente, J.M.; Zakrzewski, J.; Kulander, K.C.

    1988-10-01

    Nonlinear dynamics is applied to chaotic unassignable atomic and molecular spectra with the aim of extracting detailed information about regular dynamic motions that exist over short intervals of time. It is shown how this motion can be extracted from high resolution spectra by doing low resolution studies or by Fourier transforming limited regions of the spectrum. These motions mimic those of periodic orbits (PO) and are inserts into the dominant chaotic motion. Considering these inserts and the PO as a dynamically decoupled region of space, resonant scattering theory and stabilization methods enable us to compute ladders of resonant states which interact with the chaotic quasi-continuum computed in principle from basis sets placed off the PO. The interaction of the resonances with the quasicontinuum explains the low resolution spectra seen in such experiments. It also allows one to associate low resolution features with a particular PO. The motion on the PO thereby supplies the molecular movements whose quantization causes the low resolution spectra. Characteristic properties of the periodic orbit based resonances are discussed. The method is illustrated on the photoabsorption spectrum of the hydrogen atom in a strong magnetic field and on the photodissociation spectrum of H 3 + . Other molecular systems which are currently under investigation using this formalism are also mentioned. 53 refs., 10 figs., 2 tabs

  15. Extraction of quantifiable information from complex systems

    Dahmen, Wolfgang; Griebel, Michael; Hackbusch, Wolfgang; Ritter, Klaus; Schneider, Reinhold; Schwab, Christoph; Yserentant, Harry

    2014-01-01

    In April 2007, the  Deutsche Forschungsgemeinschaft (DFG) approved the  Priority Program 1324 “Mathematical Methods for Extracting Quantifiable Information from Complex Systems.” This volume presents a comprehensive overview of the most important results obtained over the course of the program.   Mathematical models of complex systems provide the foundation for further technological developments in science, engineering and computational finance.  Motivated by the trend toward steadily increasing computer power, ever more realistic models have been developed in recent years. These models have also become increasingly complex, and their numerical treatment poses serious challenges.   Recent developments in mathematics suggest that, in the long run, much more powerful numerical solution strategies could be derived if the interconnections between the different fields of research were systematically exploited at a conceptual level. Accordingly, a deeper understanding of the mathematical foundations as w...

  16. Extraction of temporal information in functional MRI

    Singh, M.; Sungkarat, W.; Jeong, Jeong-Won; Zhou, Yongxia

    2002-10-01

    The temporal resolution of functional MRI (fMRI) is limited by the shape of the haemodynamic response function (hrf) and the vascular architecture underlying the activated regions. Typically, the temporal resolution of fMRI is on the order of 1 s. We have developed a new data processing approach to extract temporal information on a pixel-by-pixel basis at the level of 100 ms from fMRI data. Instead of correlating or fitting the time-course of each pixel to a single reference function, which is the common practice in fMRI, we correlate each pixel's time-course to a series of reference functions that are shifted with respect to each other by 100 ms. The reference function yielding the highest correlation coefficient for a pixel is then used as a time marker for that pixel. A Monte Carlo simulation and experimental study of this approach were performed to estimate the temporal resolution as a function of signal-to-noise ratio (SNR) in the time-course of a pixel. Assuming a known and stationary hrf, the simulation and experimental studies suggest a lower limit in the temporal resolution of approximately 100 ms at an SNR of 3. The multireference function approach was also applied to extract timing information from an event-related motor movement study where the subjects flexed a finger on cue. The event was repeated 19 times with the event's presentation staggered to yield an approximately 100-ms temporal sampling of the haemodynamic response over the entire presentation cycle. The timing differences among different regions of the brain activated by the motor task were clearly visualized and quantified by this method. The results suggest that it is possible to achieve a temporal resolution of /spl sim/200 ms in practice with this approach.

  17. Optical Aperture Synthesis Object's Information Extracting Based on Wavelet Denoising

    Fan, W J; Lu, Y

    2006-01-01

    Wavelet denoising is studied to improve OAS(optical aperture synthesis) object's Fourier information extracting. Translation invariance wavelet denoising based on Donoho wavelet soft threshold denoising is researched to remove Pseudo-Gibbs in wavelet soft threshold image. OAS object's information extracting based on translation invariance wavelet denoising is studied. The study shows that wavelet threshold denoising can improve the precision and the repetition of object's information extracting from interferogram, and the translation invariance wavelet denoising information extracting is better than soft threshold wavelet denoising information extracting

  18. Linguistic Polyphony

    Nølke, Henning

    on the Scandinavian variant of polyphony, ScaPoLine. ScaPoLine is a formal linguistic theory whose main purpose is to specify the instructions conveyed through linguistic form for the creation of polyphonic meaning. The theoretical introduction is followed by polyphonic analyses of linguistic phenomena...

  19. Respiratory Information Extraction from Electrocardiogram Signals

    Amin, Gamal El Din Fathy

    2010-12-01

    The Electrocardiogram (ECG) is a tool measuring the electrical activity of the heart, and it is extensively used for diagnosis and monitoring of heart diseases. The ECG signal reflects not only the heart activity but also many other physiological processes. The respiratory activity is a prominent process that affects the ECG signal due to the close proximity of the heart and the lungs. In this thesis, several methods for the extraction of respiratory process information from the ECG signal are presented. These methods allow an estimation of the lung volume and the lung pressure from the ECG signal. The potential benefit of this is to eliminate the corresponding sensors used to measure the respiration activity. A reduction of the number of sensors connected to patients will increase patients’ comfort and reduce the costs associated with healthcare. As a further result, the efficiency of diagnosing respirational disorders will increase since the respiration activity can be monitored with a common, widely available method. The developed methods can also improve the detection of respirational disorders that occur while patients are sleeping. Such disorders are commonly diagnosed in sleeping laboratories where the patients are connected to a number of different sensors. Any reduction of these sensors will result in a more natural sleeping environment for the patients and hence a higher sensitivity of the diagnosis.

  20. Computational Linguistics Applications

    Piasecki, Maciej; Jassem, Krzysztof; Fuglewicz, Piotr

    2013-01-01

    The ever-growing popularity of Google over the recent decade has required a specific method of man-machine communication: human query should be short, whereas the machine answer may take a form of a wide range of documents. This type of communication has triggered a rapid development in the domain of Information Extraction, aimed at providing the asker with a  more precise information. The recent success of intelligent personal assistants supporting users in searching or even extracting information and answers from large collections of electronic documents signals the onset of a new era in man-machine communication – we shall soon explain to our small devices what we need to know and expect valuable answers quickly and automatically delivered. The progress of man-machine communication is accompanied by growth in the significance of applied Computational Linguistics – we need machines to understand much more from the language we speak naturally than it is the case of up-to-date search systems. Moreover, w...

  1. Sample-based XPath Ranking for Web Information Extraction

    Jundt, Oliver; van Keulen, Maurice

    Web information extraction typically relies on a wrapper, i.e., program code or a configuration that specifies how to extract some information from web pages at a specific website. Manually creating and maintaining wrappers is a cumbersome and error-prone task. It may even be prohibitive as some

  2. Providing health information for culturally and linguistically diverse women: priorities and preferences of new migrants and refugees.

    Lee, Susan K; Sulaiman-Hill, Cheryl M R; Thompson, Sandra C

    2013-08-01

    Preferences for topics and means of access to health information among newly arrived, culturally and linguistically diverse women in Perth, Western Australia, were explored. A mixed-methods approach was adopted. Qualitative material obtained from focus groups and interviews with 22 service providers and 26 migrant women was used to develop a questionnaire, which was then administered to 268 newly arrived migrant and refugee women from 50 countries. Participants' information and support priorities were ascertained from a ranking exercise conducted in a non-threatening context. Responses of migrant and refugee women were compared quantitatively. Women's top priorities for information and support included employment advice, as well as information regarding mental health issues, women's health, exercise and nutrition, family violence and alcohol and other drug issues. Their preferred methods for receiving information were interactive talks or presentations, with written material support. Audiovisual and Web-based material were also considered useful. There were differences between refugee women's and other migrants' preferences for means of receiving information and topics of most concern. The use of a non-threatening ranking process encouraged women to prioritise sensitive topics, such as family violence, and revealed a need for such topics to be incorporated within general health information presentations. Internet-based technologies are becoming increasingly important methods for disseminating information to migrant women. SO WHAT? Differences between migrant and refugee women's priority health issues and their preferred methods for receiving information highlight the desirability of tailoring information to particular groups. Although advice on employment pathways and mental health concerns were top priorities, the study revealed a need for more discussion on other sensitive topics, such as family violence and alcohol-related issues, and that ideally these should

  3. The Agent of extracting Internet Information with Lead Order

    Mo, Zan; Huang, Chuliang; Liu, Aijun

    In order to carry out e-commerce better, advanced technologies to access business information are in need urgently. An agent is described to deal with the problems of extracting internet information that caused by the non-standard and skimble-scamble structure of Chinese websites. The agent designed includes three modules which respond to the process of extracting information separately. A method of HTTP tree and a kind of Lead algorithm is proposed to generate a lead order, with which the required web can be retrieved easily. How to transform the extracted information structuralized with natural language is also discussed.

  4. Probabilistic linguistics

    Bod, R.; Heine, B.; Narrog, H.

    2010-01-01

    Probabilistic linguistics takes all linguistic evidence as positive evidence and lets statistics decide. It allows for accurate modelling of gradient phenomena in production and perception, and suggests that rule-like behaviour is no more than a side effect of maximizing probability. This chapter

  5. Cause Information Extraction from Financial Articles Concerning Business Performance

    Sakai, Hiroyuki; Masuyama, Shigeru

    We propose a method of extracting cause information from Japanese financial articles concerning business performance. Our method acquires cause informtion, e. g. “_??__??__??__??__??__??__??__??__??__??_ (zidousya no uriage ga koutyou: Sales of cars were good)”. Cause information is useful for investors in selecting companies to invest. Our method extracts cause information as a form of causal expression by using statistical information and initial clue expressions automatically. Our method can extract causal expressions without predetermined patterns or complex rules given by hand, and is expected to be applied to other tasks for acquiring phrases that have a particular meaning not limited to cause information. We compared our method with our previous one originally proposed for extracting phrases concerning traffic accident causes and experimental results showed that our new method outperforms our previous one.

  6. On the concept of a linguistic variable

    Kerre, E.

    1996-01-01

    The concept of a linguistic variable plays a crucial role in the representation of imprecise knowledge in information sciences. A variable is called linguistic as soon as its values are linguistic terms rather than numerical ones. The power of daily communication and common sense reasoning lies in the use of such linguistic values. Even when exact numerical values are available, experts tend to transform these values into linguistic ones. A physician will usually translate a numerical measurement of a blood pressure into linguistic specifications such as normal, very high, too low... Zadeh has argued that the set of values for a linguistic variable assumes a more-or-less fixed structure. Starting from an atomic value and its antonym all remaining values are constructed using logical connectives on the one hand and linguistic hedges on the other hand. In this paper we will describe how to represent the value set of a linguistic variable in general and of linguistic hedges in particular

  7. Addressing the Issue of Cultural and Linguistic Diversity and Assessment: Informal Evaluation Measures for English Language Learners

    Spinelli, Cathleen G.

    2008-01-01

    Existing research indicates that there is a disproportionate number of students with cultural and linguistic differences, English Language Learners (ELL), who are misidentified as learning disabled when their problems are due to cultural and/or linguistic differences. As a consequence, these students do not receive appropriate services. With the…

  8. Can we replace curation with information extraction software?

    Karp, Peter D

    2016-01-01

    Can we use programs for automated or semi-automated information extraction from scientific texts as practical alternatives to professional curation? I show that error rates of current information extraction programs are too high to replace professional curation today. Furthermore, current IEP programs extract single narrow slivers of information, such as individual protein interactions; they cannot extract the large breadth of information extracted by professional curators for databases such as EcoCyc. They also cannot arbitrate among conflicting statements in the literature as curators can. Therefore, funding agencies should not hobble the curation efforts of existing databases on the assumption that a problem that has stymied Artificial Intelligence researchers for more than 60 years will be solved tomorrow. Semi-automated extraction techniques appear to have significantly more potential based on a review of recent tools that enhance curator productivity. But a full cost-benefit analysis for these tools is lacking. Without such analysis it is possible to expend significant effort developing information-extraction tools that automate small parts of the overall curation workflow without achieving a significant decrease in curation costs.Database URL. © The Author(s) 2016. Published by Oxford University Press.

  9. Mining knowledge from text repositories using information extraction ...

    Information extraction (IE); text mining; text repositories; knowledge discovery from .... general purpose English words. However ... of precision and recall, as extensive experimentation is required due to lack of public tagged corpora. 4. Mining ...

  10. Mars Target Encyclopedia: Information Extraction for Planetary Science

    Wagstaff, K. L.; Francis, R.; Gowda, T.; Lu, Y.; Riloff, E.; Singh, K.

    2017-06-01

    Mars surface targets / and published compositions / Seek and ye will find. We used text mining methods to extract information from LPSC abstracts about the composition of Mars surface targets. Users can search by element, mineral, or target.

  11. Extraction of Linguistic Information from Successive Words during Reading: Evidence for Spatially Distributed Lexical Processing

    Wang, Chin-An; Inhoff, Albrecht W.

    2013-01-01

    Two experiments examined whether word recognition progressed from one word to the next during reading, as maintained by sequential attention shift models such as the E-Z Reader model. The boundary technique was used to control the visibility of to-be-identified short target words, so that they were either previewed in the parafovea or masked. The…

  12. Integrating Information Extraction Agents into a Tourism Recommender System

    Esparcia, Sergio; Sánchez-Anguix, Víctor; Argente, Estefanía; García-Fornes, Ana; Julián, Vicente

    Recommender systems face some problems. On the one hand information needs to be maintained updated, which can result in a costly task if it is not performed automatically. On the other hand, it may be interesting to include third party services in the recommendation since they improve its quality. In this paper, we present an add-on for the Social-Net Tourism Recommender System that uses information extraction and natural language processing techniques in order to automatically extract and classify information from the Web. Its goal is to maintain the system updated and obtain information about third party services that are not offered by service providers inside the system.

  13. Addressing Information Proliferation: Applications of Information Extraction and Text Mining

    Li, Jingjing

    2013-01-01

    The advent of the Internet and the ever-increasing capacity of storage media have made it easy to store, deliver, and share enormous volumes of data, leading to a proliferation of information on the Web, in online libraries, on news wires, and almost everywhere in our daily lives. Since our ability to process and absorb this information remains…

  14. Information extraction from multi-institutional radiology reports.

    Hassanpour, Saeed; Langlotz, Curtis P

    2016-01-01

    The radiology report is the most important source of clinical imaging information. It documents critical information about the patient's health and the radiologist's interpretation of medical findings. It also communicates information to the referring physicians and records that information for future clinical and research use. Although efforts to structure some radiology report information through predefined templates are beginning to bear fruit, a large portion of radiology report information is entered in free text. The free text format is a major obstacle for rapid extraction and subsequent use of information by clinicians, researchers, and healthcare information systems. This difficulty is due to the ambiguity and subtlety of natural language, complexity of described images, and variations among different radiologists and healthcare organizations. As a result, radiology reports are used only once by the clinician who ordered the study and rarely are used again for research and data mining. In this work, machine learning techniques and a large multi-institutional radiology report repository are used to extract the semantics of the radiology report and overcome the barriers to the re-use of radiology report information in clinical research and other healthcare applications. We describe a machine learning system to annotate radiology reports and extract report contents according to an information model. This information model covers the majority of clinically significant contents in radiology reports and is applicable to a wide variety of radiology study types. Our automated approach uses discriminative sequence classifiers for named-entity recognition to extract and organize clinically significant terms and phrases consistent with the information model. We evaluated our information extraction system on 150 radiology reports from three major healthcare organizations and compared its results to a commonly used non-machine learning information extraction method. We

  15. Fine-grained information extraction from German transthoracic echocardiography reports.

    Toepfer, Martin; Corovic, Hamo; Fette, Georg; Klügl, Peter; Störk, Stefan; Puppe, Frank

    2015-11-12

    Information extraction techniques that get structured representations out of unstructured data make a large amount of clinically relevant information about patients accessible for semantic applications. These methods typically rely on standardized terminologies that guide this process. Many languages and clinical domains, however, lack appropriate resources and tools, as well as evaluations of their applications, especially if detailed conceptualizations of the domain are required. For instance, German transthoracic echocardiography reports have not been targeted sufficiently before, despite of their importance for clinical trials. This work therefore aimed at development and evaluation of an information extraction component with a fine-grained terminology that enables to recognize almost all relevant information stated in German transthoracic echocardiography reports at the University Hospital of Würzburg. A domain expert validated and iteratively refined an automatically inferred base terminology. The terminology was used by an ontology-driven information extraction system that outputs attribute value pairs. The final component has been mapped to the central elements of a standardized terminology, and it has been evaluated according to documents with different layouts. The final system achieved state-of-the-art precision (micro average.996) and recall (micro average.961) on 100 test documents that represent more than 90 % of all reports. In particular, principal aspects as defined in a standardized external terminology were recognized with f 1=.989 (micro average) and f 1=.963 (macro average). As a result of keyword matching and restraint concept extraction, the system obtained high precision also on unstructured or exceptionally short documents, and documents with uncommon layout. The developed terminology and the proposed information extraction system allow to extract fine-grained information from German semi-structured transthoracic echocardiography reports

  16. Extraction of Information of Audio-Visual Contents

    Carlos Aguilar

    2011-10-01

    Full Text Available In this article we show how it is possible to use Channel Theory (Barwise and Seligman, 1997 for modeling the process of information extraction realized by audiences of audio-visual contents. To do this, we rely on the concepts pro- posed by Channel Theory and, especially, its treatment of representational systems. We then show how the information that an agent is capable of extracting from the content depends on the number of channels he is able to establish between the content and the set of classifications he is able to discriminate. The agent can endeavor the extraction of information through these channels from the totality of content; however, we discuss the advantages of extracting from its constituents in order to obtain a greater number of informational items that represent it. After showing how the extraction process is endeavored for each channel, we propose a method of representation of all the informative values an agent can obtain from a content using a matrix constituted by the channels the agent is able to establish on the content (source classifications, and the ones he can understand as individual (destination classifications. We finally show how this representation allows reflecting the evolution of the informative items through the evolution of audio-visual content.

  17. Designing and Implementing a Cross-Language Information Retrieval System Using Linguistic Corpora

    Amin Nezarat

    2012-03-01

    Full Text Available Information retrieval (IR is a crucial area of natural language processing (NLP and can be defined as finding documents whose content is relevant to the query need of a user. Cross-language information retrieval (CLIR refers to a kind of information retrieval in which the language of the query and that of searched document are different. In fact, it is a retrieval process where the user presents queries in one language to retrieve documents in another language. This paper tried to construct a bilingual lexicon of parallel chunks of English and Persian from two very large monolingual corpora an English-Persian parallel corpus which could be directly applied to cross-language information retrieval tasks. For this purpose, a statistical measure known as Association Score (AS was used to compute the association value between every two corresponding chunks in the corpus using a couple of complicated algorithms. Once the CLIR system was developed using this bilingual lexicon, an experiment was performed on a set of one hundred English and Persian phrases and collocations to see to what extend this system was effective in assisting the users find the most relevant and suitable equivalents of their queries in either language.

  18. Semantic Information Extraction of Lanes Based on Onboard Camera Videos

    Tang, L.; Deng, T.; Ren, C.

    2018-04-01

    In the field of autonomous driving, semantic information of lanes is very important. This paper proposes a method of automatic detection of lanes and extraction of semantic information from onboard camera videos. The proposed method firstly detects the edges of lanes by the grayscale gradient direction, and improves the Probabilistic Hough transform to fit them; then, it uses the vanishing point principle to calculate the lane geometrical position, and uses lane characteristics to extract lane semantic information by the classification of decision trees. In the experiment, 216 road video images captured by a camera mounted onboard a moving vehicle were used to detect lanes and extract lane semantic information. The results show that the proposed method can accurately identify lane semantics from video images.

  19. Using newspaper collections in information institutions for research purposes: experiences of historians and linguists

    Maja Krtalić

    2015-04-01

    Full Text Available The aim of this paper is to present the opinions, experiences and needs of a sample of scientists in the field of the humanities – namely historians and philologists – and their use of newspaper collections in memory institutions during scientific research. A part of the results gathered in the research, which was conducted within the Newspapers as a source of scientific information in social sciences and humanities project, and is presented here. This project examined the extent and practices in the use of newspapers as a resource of information for scientific research in the social sciences and humanities field in Croatia. The quantitative and qualitative data about the methods and level of use of newspaper material as a resource in scientific research were gathered. The results of this project implied that the newspapers are recognised and used as an admissible resource in scientific work, especially in the fields of history and philology. The fact that scientists have specific information needs and search patterns should be considered in creating innovative and effective digital newspaper collections.

  20. Knowledge Dictionary for Information Extraction on the Arabic Text Data

    Wahyu Jauharis Saputra

    2013-04-01

    Full Text Available Information extraction is an early stage of a process of textual data analysis. Information extraction is required to get information from textual data that can be used for process analysis, such as classification and categorization. A textual data is strongly influenced by the language. Arabic is gaining a significant attention in many studies because Arabic language is very different from others, and in contrast to other languages, tools and research on the Arabic language is still lacking. The information extracted using the knowledge dictionary is a concept of expression. A knowledge dictionary is usually constructed manually by an expert and this would take a long time and is specific to a problem only. This paper proposed a method for automatically building a knowledge dictionary. Dictionary knowledge is formed by classifying sentences having the same concept, assuming that they will have a high similarity value. The concept that has been extracted can be used as features for subsequent computational process such as classification or categorization. Dataset used in this paper was the Arabic text dataset. Extraction result was tested by using a decision tree classification engine and the highest precision value obtained was 71.0% while the highest recall value was 75.0%. 

  1. Exploring culturally and linguistically diverse consumer needs in relation to medicines use and health information within the pharmacy setting.

    Mohammad, Annim; Saini, Bandana; Chaar, Betty Bouad

    2015-01-01

    Low health literacy may result in adverse health outcomes for patients and is a problem faced by countries with multi-ethnic demography. For those of culturally and linguistically diverse (CALD) backgrounds, this problem can be compounded by language barriers such as low English proficiency (LEP). The pharmacy is often the last point of health-care provider contact before patients begin taking their medicines and the first point of care for minor ailments. There is a paucity of data exploring or establishing the needs of this population with respect to general medicine use/health information and pharmacist assistance. This study aimed to investigate the needs of CALD Australians with low or negligible English proficiency, specifically in regards to their understanding of health and medicines and the role of pharmacy in achieving best medicine use outcomes for this population. A qualitative method was employed. Semi-structured interviews were conducted with individuals of CALD backgrounds with a self-reported low or negligible English proficiency. The interviews explored past experiences with medicines use and interaction with health care professionals. A grounded theory approach with the method of constant comparison was undertaken for analyzing the data. Interviews were conducted until there was a saturation of themes. Thirty-one interviews were conducted, and data analyses identified themes relating to medicine use of CALD community members which were broadly categorized into: (1) health information, (2) interactions with health care professionals, (3) social networks and (4) perceptions and beliefs influencing health-related behavior. In CALD communities there are significant barriers to patient understanding and optimal use of medicines. There is significant potential for pharmacy to facilitate in addressing these issues as currently pharmacy is largely playing the role of dispenser of medicines. Whilst timely access of medicines is being ensured, there seems

  2. Ontology-Based Information Extraction for Business Intelligence

    Saggion, Horacio; Funk, Adam; Maynard, Diana; Bontcheva, Kalina

    Business Intelligence (BI) requires the acquisition and aggregation of key pieces of knowledge from multiple sources in order to provide valuable information to customers or feed statistical BI models and tools. The massive amount of information available to business analysts makes information extraction and other natural language processing tools key enablers for the acquisition and use of that semantic information. We describe the application of ontology-based extraction and merging in the context of a practical e-business application for the EU MUSING Project where the goal is to gather international company intelligence and country/region information. The results of our experiments so far are very promising and we are now in the process of building a complete end-to-end solution.

  3. Physical Linguistics.

    Tice, Bradley S.

    Physical linguistics is defined as the use of treatments from the field of speech pathology to enhance first and second language production in healthy individuals, resulting in increased quality and strength of phonation and articulation. A series of exercises for treating dysarthria (weakness, paralysis, discoordination, primary and secondary…

  4. NAMED ENTITY RECOGNITION FROM BIOMEDICAL TEXT -AN INFORMATION EXTRACTION TASK

    N. Kanya

    2016-07-01

    Full Text Available Biomedical Text Mining targets the Extraction of significant information from biomedical archives. Bio TM encompasses Information Retrieval (IR and Information Extraction (IE. The Information Retrieval will retrieve the relevant Biomedical Literature documents from the various Repositories like PubMed, MedLine etc., based on a search query. The IR Process ends up with the generation of corpus with the relevant document retrieved from the Publication databases based on the query. The IE task includes the process of Preprocessing of the document, Named Entity Recognition (NER from the documents and Relationship Extraction. This process includes Natural Language Processing, Data Mining techniques and machine Language algorithm. The preprocessing task includes tokenization, stop word Removal, shallow parsing, and Parts-Of-Speech tagging. NER phase involves recognition of well-defined objects such as genes, proteins or cell-lines etc. This process leads to the next phase that is extraction of relationships (IE. The work was based on machine learning algorithm Conditional Random Field (CRF.

  5. A Two-Step Resume Information Extraction Algorithm

    Jie Chen

    2018-01-01

    Full Text Available With the rapid growth of Internet-based recruiting, there are a great number of personal resumes among recruiting systems. To gain more attention from the recruiters, most resumes are written in diverse formats, including varying font size, font colour, and table cells. However, the diversity of format is harmful to data mining, such as resume information extraction, automatic job matching, and candidates ranking. Supervised methods and rule-based methods have been proposed to extract facts from resumes, but they strongly rely on hierarchical structure information and large amounts of labelled data, which are hard to collect in reality. In this paper, we propose a two-step resume information extraction approach. In the first step, raw text of resume is identified as different resume blocks. To achieve the goal, we design a novel feature, Writing Style, to model sentence syntax information. Besides word index and punctuation index, word lexical attribute and prediction results of classifiers are included in Writing Style. In the second step, multiple classifiers are employed to identify different attributes of fact information in resumes. Experimental results on a real-world dataset show that the algorithm is feasible and effective.

  6. Optimum detection for extracting maximum information from symmetric qubit sets

    Mizuno, Jun; Fujiwara, Mikio; Sasaki, Masahide; Akiba, Makoto; Kawanishi, Tetsuya; Barnett, Stephen M.

    2002-01-01

    We demonstrate a class of optimum detection strategies for extracting the maximum information from sets of equiprobable real symmetric qubit states of a single photon. These optimum strategies have been predicted by Sasaki et al. [Phys. Rev. A 59, 3325 (1999)]. The peculiar aspect is that the detections with at least three outputs suffice for optimum extraction of information regardless of the number of signal elements. The cases of ternary (or trine), quinary, and septenary polarization signals are studied where a standard von Neumann detection (a projection onto a binary orthogonal basis) fails to access the maximum information. Our experiments demonstrate that it is possible with present technologies to attain about 96% of the theoretical limit

  7. Extracting Semantic Information from Visual Data: A Survey

    Qiang Liu

    2016-03-01

    Full Text Available The traditional environment maps built by mobile robots include both metric ones and topological ones. These maps are navigation-oriented and not adequate for service robots to interact with or serve human users who normally rely on the conceptual knowledge or semantic contents of the environment. Therefore, the construction of semantic maps becomes necessary for building an effective human-robot interface for service robots. This paper reviews recent research and development in the field of visual-based semantic mapping. The main focus is placed on how to extract semantic information from visual data in terms of feature extraction, object/place recognition and semantic representation methods.

  8. Rapid automatic keyword extraction for information retrieval and analysis

    Rose, Stuart J [Richland, WA; Cowley,; E, Wendy [Richland, WA; Crow, Vernon L [Richland, WA; Cramer, Nicholas O [Richland, WA

    2012-03-06

    Methods and systems for rapid automatic keyword extraction for information retrieval and analysis. Embodiments can include parsing words in an individual document by delimiters, stop words, or both in order to identify candidate keywords. Word scores for each word within the candidate keywords are then calculated based on a function of co-occurrence degree, co-occurrence frequency, or both. Based on a function of the word scores for words within the candidate keyword, a keyword score is calculated for each of the candidate keywords. A portion of the candidate keywords are then extracted as keywords based, at least in part, on the candidate keywords having the highest keyword scores.

  9. A Linguistic Foundation for Communicating Geo-Information in the context of BML and geoBML

    2010-03-23

    section, the C2LG is presented. C2LG is a formal grammar. As such, it follows the definition of formal grammars in general as proposed by Chomsky ...In: Mitkov, R. (Ed.), The Oxford Handbook of Computational Linguistics. Oxford, UK: Oxford University Press. Chomsky , N. (1957). Syntactic

  10. Evidence from Neurolinguistic Methodologies: Can It Actually Inform Linguistic/Language Acquisition Theories and Translate to Evidence-Based Applications?

    Roberts, Leah; González Alonso, Jorge; Pliatsikas, Christos; Rothman, Jason

    2018-01-01

    This special issue is a testament to the recent burgeoning interest by theoretical linguists, language acquisitionists and teaching practitioners in the neuroscience of language. It offers a highly valuable, state-of-the-art overview of the neurophysiological methods that are currently being applied to questions in the field of second language…

  11. A Study of a Collaborative Instructional Project Informed by Systemic Functional Linguistic Theory: Report Writing in Elementary Grades

    Brisk, Maria Estela; Hodgson-Drysdale, Tracy; O'Connor, Cheryl

    2011-01-01

    This study examined the teaching of report writing in PreK-5 through the lens of systemic functional linguistics theory. Teachers were part of a university-public school collaboration that included professional development on teaching genres, text organization, and language features. Grounded in this knowledge, teachers explicitly taught report…

  12. Cognitive linguistics.

    Evans, Vyvyan

    2012-03-01

    Cognitive linguistics is one of the fastest growing and influential perspectives on the nature of language, the mind, and their relationship with sociophysical (embodied) experience. It is a broad theoretical and methodological enterprise, rather than a single, closely articulated theory. Its primary commitments are outlined. These are the Cognitive Commitment-a commitment to providing a characterization of language that accords with what is known about the mind and brain from other disciplines-and the Generalization Commitment-which represents a dedication to characterizing general principles that apply to all aspects of human language. The article also outlines the assumptions and worldview which arises from these commitments, as represented in the work of leading cognitive linguists. WIREs Cogn Sci 2012, 3:129-141. doi: 10.1002/wcs.1163 For further resources related to this article, please visit the WIREs website. Copyright © 2012 John Wiley & Sons, Inc.

  13. Robust Vehicle and Traffic Information Extraction for Highway Surveillance

    Yeh Chia-Hung

    2005-01-01

    Full Text Available A robust vision-based traffic monitoring system for vehicle and traffic information extraction is developed in this research. It is challenging to maintain detection robustness at all time for a highway surveillance system. There are three major problems in detecting and tracking a vehicle: (1 the moving cast shadow effect, (2 the occlusion effect, and (3 nighttime detection. For moving cast shadow elimination, a 2D joint vehicle-shadow model is employed. For occlusion detection, a multiple-camera system is used to detect occlusion so as to extract the exact location of each vehicle. For vehicle nighttime detection, a rear-view monitoring technique is proposed to maintain tracking and detection accuracy. Furthermore, we propose a method to improve the accuracy of background extraction, which usually serves as the first step in any vehicle detection processing. Experimental results are given to demonstrate that the proposed techniques are effective and efficient for vision-based highway surveillance.

  14. Advanced applications of natural language processing for performing information extraction

    Rodrigues, Mário

    2015-01-01

    This book explains how can be created information extraction (IE) applications that are able to tap the vast amount of relevant information available in natural language sources: Internet pages, official documents such as laws and regulations, books and newspapers, and social web. Readers are introduced to the problem of IE and its current challenges and limitations, supported with examples. The book discusses the need to fill the gap between documents, data, and people, and provides a broad overview of the technology supporting IE. The authors present a generic architecture for developing systems that are able to learn how to extract relevant information from natural language documents, and illustrate how to implement working systems using state-of-the-art and freely available software tools. The book also discusses concrete applications illustrating IE uses.   ·         Provides an overview of state-of-the-art technology in information extraction (IE), discussing achievements and limitations for t...

  15. Improving information extraction using a probability-based approach

    Kim, S.; Ahmed, Saeema; Wallace, K.

    2007-01-01

    Information plays a crucial role during the entire life-cycle of a product. It has been shown that engineers frequently consult colleagues to obtain the information they require to solve problems. However, the industrial world is now more transient and key personnel move to other companies...... or retire. It is becoming essential to retrieve vital information from archived product documents, if it is available. There is, therefore, great interest in ways of extracting relevant and sharable information from documents. A keyword-based search is commonly used, but studies have shown...... the recall, while maintaining the high precision, a learning approach that makes identification decisions based on a probability model, rather than simply looking up the presence of the pre-defined variations, looks promising. This paper presents the results of developing such a probability-based entity...

  16. The linguistically aware teacher and the teacher-aware linguist.

    McCartney, Elspeth; Ellis, Sue

    2013-07-01

    This review evaluates issues of teacher linguistic knowledge relating to their work with children with speech, language and communication difficulties (SLCD). Information is from Ellis and McCartney [(2011a). Applied linguistics and primary school teaching. Cambridge: Cambridge University Press], a state-of-the-art text deriving from a British Association of Applied Linguistics/Cambridge University Press expert seminar series that details: linguistic research underpinning primary school curricula and pedagogy; the form of linguistic knowledge useful for teachers supporting children with SLCD in partnership with speech and language therapists; and how and when teachers acquire and learn to apply such knowledge. Critical analysis of the options presented for teacher learning indicate that policy enjoinders now include linguistic application as an expected part of teachers' professional knowledge, for all children including those with SLCD, but there is a large unmet learning need. It is concluded that there is a role for clinical linguists to disseminate useable knowledge to teachers in an accessible format. Ways of achieving this are considered.

  17. Transliteration normalization for Information Extraction and Machine Translation

    Yuval Marton

    2014-12-01

    Full Text Available Foreign name transliterations typically include multiple spelling variants. These variants cause data sparseness and inconsistency problems, increase the Out-of-Vocabulary (OOV rate, and present challenges for Machine Translation, Information Extraction and other natural language processing (NLP tasks. This work aims to identify and cluster name spelling variants using a Statistical Machine Translation method: word alignment. The variants are identified by being aligned to the same “pivot” name in another language (the source-language in Machine Translation settings. Based on word-to-word translation and transliteration probabilities, as well as the string edit distance metric, names with similar spellings in the target language are clustered and then normalized to a canonical form. With this approach, tens of thousands of high-precision name transliteration spelling variants are extracted from sentence-aligned bilingual corpora in Arabic and English (in both languages. When these normalized name spelling variants are applied to Information Extraction tasks, improvements over strong baseline systems are observed. When applied to Machine Translation tasks, a large improvement potential is shown.

  18. Linguistic relativity.

    Wolff, Phillip; Holmes, Kevin J

    2011-05-01

    The central question in research on linguistic relativity, or the Whorfian hypothesis, is whether people who speak different languages think differently. The recent resurgence of research on this question can be attributed, in part, to new insights about the ways in which language might impact thought. We identify seven categories of hypotheses about the possible effects of language on thought across a wide range of domains, including motion, color, spatial relations, number, and false belief understanding. While we do not find support for the idea that language determines the basic categories of thought or that it overwrites preexisting conceptual distinctions, we do find support for the proposal that language can make some distinctions difficult to avoid, as well as for the proposal that language can augment certain types of thinking. Further, we highlight recent evidence suggesting that language may induce a relatively schematic mode of thinking. Although the literature on linguistic relativity remains contentious, there is growing support for the view that language has a profound effect on thought. WIREs Cogni Sci 2011 2 253-265 DOI: 10.1002/wcs.104 For further resources related to this article, please visit the WIREs website. Copyright © 2010 John Wiley & Sons, Ltd.

  19. Knowledge discovery: Extracting usable information from large amounts of data

    Whiteson, R.

    1998-01-01

    The threat of nuclear weapons proliferation is a problem of world wide concern. Safeguards are the key to nuclear nonproliferation and data is the key to safeguards. The safeguards community has access to a huge and steadily growing volume of data. The advantages of this data rich environment are obvious, there is a great deal of information which can be utilized. The challenge is to effectively apply proven and developing technologies to find and extract usable information from that data. That information must then be assessed and evaluated to produce the knowledge needed for crucial decision making. Efficient and effective analysis of safeguards data will depend on utilizing technologies to interpret the large, heterogeneous data sets that are available from diverse sources. With an order-of-magnitude increase in the amount of data from a wide variety of technical, textual, and historical sources there is a vital need to apply advanced computer technologies to support all-source analysis. There are techniques of data warehousing, data mining, and data analysis that can provide analysts with tools that will expedite their extracting useable information from the huge amounts of data to which they have access. Computerized tools can aid analysts by integrating heterogeneous data, evaluating diverse data streams, automating retrieval of database information, prioritizing inputs, reconciling conflicting data, doing preliminary interpretations, discovering patterns or trends in data, and automating some of the simpler prescreening tasks that are time consuming and tedious. Thus knowledge discovery technologies can provide a foundation of support for the analyst. Rather than spending time sifting through often irrelevant information, analysts could use their specialized skills in a focused, productive fashion. This would allow them to make their analytical judgments with more confidence and spend more of their time doing what they do best

  20. Evolving spectral transformations for multitemporal information extraction using evolutionary computation

    Momm, Henrique; Easson, Greg

    2011-01-01

    Remote sensing plays an important role in assessing temporal changes in land features. The challenge often resides in the conversion of large quantities of raw data into actionable information in a timely and cost-effective fashion. To address this issue, research was undertaken to develop an innovative methodology integrating biologically-inspired algorithms with standard image classification algorithms to improve information extraction from multitemporal imagery. Genetic programming was used as the optimization engine to evolve feature-specific candidate solutions in the form of nonlinear mathematical expressions of the image spectral channels (spectral indices). The temporal generalization capability of the proposed system was evaluated by addressing the task of building rooftop identification from a set of images acquired at different dates in a cross-validation approach. The proposed system generates robust solutions (kappa values > 0.75 for stage 1 and > 0.4 for stage 2) despite the statistical differences between the scenes caused by land use and land cover changes coupled with variable environmental conditions, and the lack of radiometric calibration between images. Based on our results, the use of nonlinear spectral indices enhanced the spectral differences between features improving the clustering capability of standard classifiers and providing an alternative solution for multitemporal information extraction.

  1. Recognition techniques for extracting information from semistructured documents

    Della Ventura, Anna; Gagliardi, Isabella; Zonta, Bruna

    2000-12-01

    Archives of optical documents are more and more massively employed, the demand driven also by the new norms sanctioning the legal value of digital documents, provided they are stored on supports that are physically unalterable. On the supply side there is now a vast and technologically advanced market, where optical memories have solved the problem of the duration and permanence of data at costs comparable to those for magnetic memories. The remaining bottleneck in these systems is the indexing. The indexing of documents with a variable structure, while still not completely automated, can be machine supported to a large degree with evident advantages both in the organization of the work, and in extracting information, providing data that is much more detailed and potentially significant for the user. We present here a system for the automatic registration of correspondence to and from a public office. The system is based on a general methodology for the extraction, indexing, archiving, and retrieval of significant information from semi-structured documents. This information, in our prototype application, is distributed among the database fields of sender, addressee, subject, date, and body of the document.

  2. What Does Corpus Linguistics Have to Offer to Language Assessment?

    Xi, Xiaoming

    2017-01-01

    In recent years, continuing advances in technology have increased the capacity to automate the extraction of a range of linguistic features of texts and thus have provided the impetus for the substantial growth of corpus linguistics. While corpus linguistic tools and methods have been used extensively in second language learning research, they…

  3. Heritage language and linguistic theory

    Scontras, Gregory; Fuchs, Zuzanna; Polinsky, Maria

    2015-01-01

    This paper discusses a common reality in many cases of multilingualism: heritage speakers, or unbalanced bilinguals, simultaneous or sequential, who shifted early in childhood from one language (their heritage language) to their dominant language (the language of their speech community). To demonstrate the relevance of heritage linguistics to the study of linguistic competence more broadly defined, we present a series of case studies on heritage linguistics, documenting some of the deficits and abilities typical of heritage speakers, together with the broader theoretical questions they inform. We consider the reorganization of morphosyntactic feature systems, the reanalysis of atypical argument structure, the attrition of the syntax of relativization, and the simplification of scope interpretations; these phenomena implicate diverging trajectories and outcomes in the development of heritage speakers. The case studies also have practical and methodological implications for the study of multilingualism. We conclude by discussing more general concepts central to linguistic inquiry, in particular, complexity and native speaker competence. PMID:26500595

  4. Automated extraction of chemical structure information from digital raster images

    Shedden Kerby A

    2009-02-01

    Full Text Available Abstract Background To search for chemical structures in research articles, diagrams or text representing molecules need to be translated to a standard chemical file format compatible with cheminformatic search engines. Nevertheless, chemical information contained in research articles is often referenced as analog diagrams of chemical structures embedded in digital raster images. To automate analog-to-digital conversion of chemical structure diagrams in scientific research articles, several software systems have been developed. But their algorithmic performance and utility in cheminformatic research have not been investigated. Results This paper aims to provide critical reviews for these systems and also report our recent development of ChemReader – a fully automated tool for extracting chemical structure diagrams in research articles and converting them into standard, searchable chemical file formats. Basic algorithms for recognizing lines and letters representing bonds and atoms in chemical structure diagrams can be independently run in sequence from a graphical user interface-and the algorithm parameters can be readily changed-to facilitate additional development specifically tailored to a chemical database annotation scheme. Compared with existing software programs such as OSRA, Kekule, and CLiDE, our results indicate that ChemReader outperforms other software systems on several sets of sample images from diverse sources in terms of the rate of correct outputs and the accuracy on extracting molecular substructure patterns. Conclusion The availability of ChemReader as a cheminformatic tool for extracting chemical structure information from digital raster images allows research and development groups to enrich their chemical structure databases by annotating the entries with published research articles. Based on its stable performance and high accuracy, ChemReader may be sufficiently accurate for annotating the chemical database with links

  5. Information Extraction, Data Integration, and Uncertain Data Management: The State of The Art

    Habib, Mena Badieh; van Keulen, Maurice

    2011-01-01

    Information Extraction, data Integration, and uncertain data management are different areas of research that got vast focus in the last two decades. Many researches tackled those areas of research individually. However, information extraction systems should have integrated with data integration

  6. Integrated model of assisted parking system and performance evaluation with entropy weight extended analytic hierarchy process and two-tuple linguistic information

    Yiding Hua

    2016-06-01

    Full Text Available Evaluating comprehensive performance of assisted parking system has been a very important issue for car companies for years, because the overall performance of assisted parking system directly influences car intellectualization and customers’ degree of satisfaction. Therefore, this article proposes two-tuple linguistic analytic hierarchy process to evaluate assisted parking system so as to avoid information loss during the processes of evaluation integration. The performance evaluation attributes for assisted parking system are established initially. Subsequently, the information entropy theory is proposed to improve the evaluation attribute weight determined by analytic hierarchy process for the influencing factors of the randomness in parking test process. Furthermore, the evaluation attribute measure values of comprehensive performance are calculated and the assisted parking system evaluation results are obtained with ordered weighted averaging operator. Finally, numerical examples of vehicle types equipped with eight different assisted parking systems and computational results are presented.

  7. Comprehensive Evaluation on Employee Satisfaction of Mine Occupational Health and Safety Management System Based on Improved AHP and 2-Tuple Linguistic Information

    Jiangdong Bao

    2017-01-01

    Full Text Available In order to comprehensively evaluate the employee satisfaction of mine occupational health and safety management system, an analytic method based on fuzzy analytic hierarchy process and 2-tuple linguistic model was established. Based on the establishment of 5 first-grade indicators and 20 second-grade ones, method of improved AHP and the time-ordered Weighted Averaging Operator (T-OWA model is constructed. The results demonstrate that the employee satisfaction of the mine occupational health and safety management system is of the ‘general’ rank. The method including the evaluation of employee satisfaction and the quantitative analysis of language evaluation information ensures the authenticity of the language evaluation information.

  8. Information Extraction for Clinical Data Mining: A Mammography Case Study.

    Nassif, Houssam; Woods, Ryan; Burnside, Elizabeth; Ayvaci, Mehmet; Shavlik, Jude; Page, David

    2009-01-01

    Breast cancer is the leading cause of cancer mortality in women between the ages of 15 and 54. During mammography screening, radiologists use a strict lexicon (BI-RADS) to describe and report their findings. Mammography records are then stored in a well-defined database format (NMD). Lately, researchers have applied data mining and machine learning techniques to these databases. They successfully built breast cancer classifiers that can help in early detection of malignancy. However, the validity of these models depends on the quality of the underlying databases. Unfortunately, most databases suffer from inconsistencies, missing data, inter-observer variability and inappropriate term usage. In addition, many databases are not compliant with the NMD format and/or solely consist of text reports. BI-RADS feature extraction from free text and consistency checks between recorded predictive variables and text reports are crucial to addressing this problem. We describe a general scheme for concept information retrieval from free text given a lexicon, and present a BI-RADS features extraction algorithm for clinical data mining. It consists of a syntax analyzer, a concept finder and a negation detector. The syntax analyzer preprocesses the input into individual sentences. The concept finder uses a semantic grammar based on the BI-RADS lexicon and the experts' input. It parses sentences detecting BI-RADS concepts. Once a concept is located, a lexical scanner checks for negation. Our method can handle multiple latent concepts within the text, filtering out ultrasound concepts. On our dataset, our algorithm achieves 97.7% precision, 95.5% recall and an F 1 -score of 0.97. It outperforms manual feature extraction at the 5% statistical significance level.

  9. INFORMATION EXTRACTION IN TOMB PIT USING HYPERSPECTRAL DATA

    X. Yang

    2018-04-01

    Full Text Available Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  10. Information Extraction in Tomb Pit Using Hyperspectral Data

    Yang, X.; Hou, M.; Lyu, S.; Ma, S.; Gao, Z.; Bai, S.; Gu, M.; Liu, Y.

    2018-04-01

    Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  11. Domain-independent information extraction in unstructured text

    Irwin, N.H. [Sandia National Labs., Albuquerque, NM (United States). Software Surety Dept.

    1996-09-01

    Extracting information from unstructured text has become an important research area in recent years due to the large amount of text now electronically available. This status report describes the findings and work done during the second year of a two-year Laboratory Directed Research and Development Project. Building on the first-year`s work of identifying important entities, this report details techniques used to group words into semantic categories and to output templates containing selective document content. Using word profiles and category clustering derived during a training run, the time-consuming knowledge-building task can be avoided. Though the output still lacks in completeness when compared to systems with domain-specific knowledge bases, the results do look promising. The two approaches are compatible and could complement each other within the same system. Domain-independent approaches retain appeal as a system that adapts and learns will soon outpace a system with any amount of a priori knowledge.

  12. Extracting and Using Photon Polarization Information in Radiative B Decays

    Grossman, Yuval

    2000-05-09

    The authors discuss the uses of conversion electron pairs for extracting photon polarization information in weak radiative B decays. Both cases of leptons produced through a virtual and real photon are considered. Measurements of the angular correlation between the (K-pi) and (e{sup +}e{sup {minus}}) decay planes in B --> K*(--> K-pi)gamma (*)(--> e{sup +}e{sup {minus}}) decays can be used to determine the helicity amplitudes in the radiative B --> K*gamma decays. A large right-handed helicity amplitude in B-bar decays is a signal of new physics. The time-dependent CP asymmetry in the B{sup 0} decay angular correlation is shown to measure sin 2-beta and cos 2-beta with little hadronic uncertainty.

  13. Extraction of neutron spectral information from Bonner-Sphere data

    Haney, J H; Zaidins, C S

    1999-01-01

    We have extended a least-squares method of extracting neutron spectral information from Bonner-Sphere data which was previously developed by Zaidins et al. (Med. Phys. 5 (1978) 42). A pulse-height analysis with background stripping is employed which provided a more accurate count rate for each sphere. Newer response curves by Mares and Schraube (Nucl. Instr. and Meth. A 366 (1994) 461) were included for the moderating spheres and the bare detector which comprise the Bonner spectrometer system. Finally, the neutron energy spectrum of interest was divided using the philosophy of fuzzy logic into three trapezoidal regimes corresponding to slow, moderate, and fast neutrons. Spectral data was taken using a PuBe source in two different environments and the analyzed data is presented for these cases as slow, moderate, and fast neutron fluences. (author)

  14. ONTOGRABBING: Extracting Information from Texts Using Generative Ontologies

    Nilsson, Jørgen Fischer; Szymczak, Bartlomiej Antoni; Jensen, P.A.

    2009-01-01

    We describe principles for extracting information from texts using a so-called generative ontology in combination with syntactic analysis. Generative ontologies are introduced as semantic domains for natural language phrases. Generative ontologies extend ordinary finite ontologies with rules...... for producing recursively shaped terms representing the ontological content (ontological semantics) of NL noun phrases and other phrases. We focus here on achieving a robust, often only partial, ontology-driven parsing of and ascription of semantics to a sentence in the text corpus. The aim of the ontological...... analysis is primarily to identify paraphrases, thereby achieving a search functionality beyond mere keyword search with synsets. We further envisage use of the generative ontology as a phrase-based rather than word-based browser into text corpora....

  15. Information extraction and knowledge graph construction from geoscience literature

    Wang, Chengbin; Ma, Xiaogang; Chen, Jianguo; Chen, Jingwen

    2018-03-01

    Geoscience literature published online is an important part of open data, and brings both challenges and opportunities for data analysis. Compared with studies of numerical geoscience data, there are limited works on information extraction and knowledge discovery from textual geoscience data. This paper presents a workflow and a few empirical case studies for that topic, with a focus on documents written in Chinese. First, we set up a hybrid corpus combining the generic and geology terms from geology dictionaries to train Chinese word segmentation rules of the Conditional Random Fields model. Second, we used the word segmentation rules to parse documents into individual words, and removed the stop-words from the segmentation results to get a corpus constituted of content-words. Third, we used a statistical method to analyze the semantic links between content-words, and we selected the chord and bigram graphs to visualize the content-words and their links as nodes and edges in a knowledge graph, respectively. The resulting graph presents a clear overview of key information in an unstructured document. This study proves the usefulness of the designed workflow, and shows the potential of leveraging natural language processing and knowledge graph technologies for geoscience.

  16. Speech-language assessment in a linguistically diverse setting: Preliminary exploration of the possible impact of informal ‘solutions’ within the South African context

    Joanne Barratt

    2012-12-01

    Full Text Available Speech-language therapists (SLTs working in the context of cultural and linguistic diversity face considerable challenges in providing equitable services to all clients. This is complicated by the fact that the majority of SLTs in South Africa are English or Afrikaans speakers, while the majority of the population have a home language other than English/Afrikaans. Consequently, SLTs are often forced to call on untrained personnel to act as interpreters or translators, and to utilise informally translated materials in the assessment and management of clients with communication impairments. However, variations in translation have the potential to considerably alter intervention plans. This study explored whether the linguistic complexity conveyed in translation of the Western Aphasia Battery (WAB test changed when translated from English to isiZulu by five different first-language IsiZulu speakers. A qualitative comparative research design was adopted and results were analysed using comparative data analysis. Results revealed notable differences in the translations, with most differences relating to vocabulary and semantics. This finding holds clinical implications for the use of informal translators as well as for the utilisation of translated material in the provision of speech-language therapy services in multilingual contexts. This study highlights the need for cautious use of translators and/or translated materials that are not appropriately and systematically adapted for local usage. Further recommendations include a call for intensified efforts in the transformation of the profession within the country, specifically by attracting greater numbers of students who are fluent in African languages.

  17. Data Assimilation to Extract Soil Moisture Information from SMAP Observations

    Jana Kolassa

    2017-11-01

    Full Text Available This study compares different methods to extract soil moisture information through the assimilation of Soil Moisture Active Passive (SMAP observations. Neural network (NN and physically-based SMAP soil moisture retrievals were assimilated into the National Aeronautics and Space Administration (NASA Catchment model over the contiguous United States for April 2015 to March 2017. By construction, the NN retrievals are consistent with the global climatology of the Catchment model soil moisture. Assimilating the NN retrievals without further bias correction improved the surface and root zone correlations against in situ measurements from 14 SMAP core validation sites (CVS by 0.12 and 0.16, respectively, over the model-only skill, and reduced the surface and root zone unbiased root-mean-square error (ubRMSE by 0.005 m 3 m − 3 and 0.001 m 3 m − 3 , respectively. The assimilation reduced the average absolute surface bias against the CVS measurements by 0.009 m 3 m − 3 , but increased the root zone bias by 0.014 m 3 m − 3 . Assimilating the NN retrievals after a localized bias correction yielded slightly lower surface correlation and ubRMSE improvements, but generally the skill differences were small. The assimilation of the physically-based SMAP Level-2 passive soil moisture retrievals using a global bias correction yielded similar skill improvements, as did the direct assimilation of locally bias-corrected SMAP brightness temperatures within the SMAP Level-4 soil moisture algorithm. The results show that global bias correction methods may be able to extract more independent information from SMAP observations compared to local bias correction methods, but without accurate quality control and observation error characterization they are also more vulnerable to adverse effects from retrieval errors related to uncertainties in the retrieval inputs and algorithm. Furthermore, the results show that using global bias correction approaches without a

  18. Multi-Filter String Matching and Human-Centric Entity Matching for Information Extraction

    Sun, Chong

    2012-01-01

    More and more information is being generated in text documents, such as Web pages, emails and blogs. To effectively manage this unstructured information, one broadly used approach includes locating relevant content in documents, extracting structured information and integrating the extracted information for querying, mining or further analysis. In…

  19. Term Bases and Linguistic Linked Open Data

    for pursuing their work. The theme of this year’s TKE is ‘Term Bases and Linguistic Linked Open Data’. Mono- and multi-lingual term bases, which contain information about concepts (terms, definitions, examples of use, references, comments on equivalence etc.), have always made up valuable linguistic resources...

  20. Hypertextuality, Complexity, Creativity: Using Linguistic Software Tools to Uncover New Information about the Food and Drink of Historic Mayans

    Rose Lema

    2012-05-01

    Full Text Available In this essay on natural language I present a computer-supported study of words, sentences and hypertexts concerning bromatology (the study of food and drink in a XVI century Maya-Spanish Calepin—the most complete and extended dictionary ever written on the culture of the constructors of the wonderful and prestigious Mayan cities of Uxmal, Kalakmul, Chichén-Itzá (ARZÁPALO, 1995. For constructing a complex corpus, I apply concepts of the three-body and the fractal dimension theories (POINCARÉ, 1908; MANDELBROT, 1975. First, I register an initial body of text by simply searching via the find key for abbreviations of bromatology and botany already recorded by the citation word in the Calepin. Then, I arbitrarily shorten the Spanish form corresponding to tasty and gather it through the whole dictionary. This way I obtain three bodies of interpretative meaning, lexias (BARTHES, 2002. Second, I establish the second and the third dimensional hypertextual relations between the gleaned words or sentences of text as well as their co-occurrences by using the comprehensive linguistics software, Tropes, a lexical and content analysis mixed tool, which brings up the qualitative and quantitative data pertinent to the research. Third, to bring back the colonial Maya voices of the Calepin, I surf the Internet and add to both written bodies of text a third text composed of beautiful colored images presenting food, drinks and tasty dishes that are still enjoyed by the Maya today and have been appreciated for almost five centuries. Notwithstanding the above, neither one of the three bodies (corpora nested fractally one inside the other is exhaustive. Nonetheless, the study of their interrelations could lead to the deepening of our knowledge on the complex juxtaposition between Siglo de Oro and Maya languages and cultures in the Yucatán Peninsula. URN: http://nbn-resolving.de/urn:nbn:de:0114-fqs1202215

  1. Earth Science Data Analytics: Preparing for Extracting Knowledge from Information

    Kempler, Steven; Barbieri, Lindsay

    2016-01-01

    Data analytics is the process of examining large amounts of data of a variety of types to uncover hidden patterns, unknown correlations and other useful information. Data analytics is a broad term that includes data analysis, as well as an understanding of the cognitive processes an analyst uses to understand problems and explore data in meaningful ways. Analytics also include data extraction, transformation, and reduction, utilizing specific tools, techniques, and methods. Turning to data science, definitions of data science sound very similar to those of data analytics (which leads to a lot of the confusion between the two). But the skills needed for both, co-analyzing large amounts of heterogeneous data, understanding and utilizing relevant tools and techniques, and subject matter expertise, although similar, serve different purposes. Data Analytics takes on a practitioners approach to applying expertise and skills to solve issues and gain subject knowledge. Data Science, is more theoretical (research in itself) in nature, providing strategic actionable insights and new innovative methodologies. Earth Science Data Analytics (ESDA) is the process of examining, preparing, reducing, and analyzing large amounts of spatial (multi-dimensional), temporal, or spectral data using a variety of data types to uncover patterns, correlations and other information, to better understand our Earth. The large variety of datasets (temporal spatial differences, data types, formats, etc.) invite the need for data analytics skills that understand the science domain, and data preparation, reduction, and analysis techniques, from a practitioners point of view. The application of these skills to ESDA is the focus of this presentation. The Earth Science Information Partners (ESIP) Federation Earth Science Data Analytics (ESDA) Cluster was created in recognition of the practical need to facilitate the co-analysis of large amounts of data and information for Earth science. Thus, from a to

  2. Etymology and Modern Linguistics

    Malkiel, Yakov

    1975-01-01

    Discusses the estrangement between etymology and modern linguistics, and concludes that a reconciliation between spatio-temporal linguistics and etymology must occur, because without it, both disciplines are doomed to inanition. (Author/AM)

  3. Testing the reliability of information extracted from ancient zircon

    Kielman, Ross; Whitehouse, Martin; Nemchin, Alexander

    2015-04-01

    Studies combining zircon U-Pb chronology, trace element distribution as well as O and Hf isotope systematics are a powerful way to gain understanding of the processes shaping Earth's evolution, especially in detrital populations where constraints from the original host are missing. Such studies of the Hadean detrital zircon population abundant in sedimentary rocks in Western Australia have involved analysis of an unusually large number of individual grains, but also highlighted potential problems with the approach, only apparent when multiple analyses are obtained from individual grains. A common feature of the Hadean as well as many early Archaean zircon populations is their apparent inhomogeneity, which reduces confidence in conclusions based on studies combining chemistry and isotopic characteristics of zircon. In order to test the reliability of information extracted from early Earth zircon, we report results from one of the first in-depth multi-method study of zircon from a relatively simple early Archean magmatic rock, used as an analogue to ancient detrital zircon. The approach involves making multiple SIMS analyses in individual grains in order to be comparable to the most advanced studies of detrital zircon populations. The investigated sample is a relatively undeformed, non-migmatitic ca. 3.8 Ga tonalite collected a few kms south of the Isua Greenstone Belt, southwest Greenland. Extracted zircon grains can be combined into three different groups based on the behavior of their U-Pb systems: (i) grains that show internally consistent and concordant ages and define an average age of 3805±15 Ma, taken to be the age of the rock, (ii) grains that are distributed close to the concordia line, but with significant variability between multiple analyses, suggesting an ancient Pb loss and (iii) grains that have multiple analyses distributed along a discordia pointing towards a zero intercept, indicating geologically recent Pb-loss. This overall behavior has

  4. What Is Applied Linguistics?

    James, Carl

    1993-01-01

    Ostensive and expository definitions of applied linguistics are assessed. It is suggested that the key to a meaningful definition lies in the dual articulation of applied linguistics: it is an interface between linguistics and practicality. Its role as an "expert system" is suggested. (45 references) (Author/LB)

  5. Ninth international conference on computational linguistics Coling 82

    1983-01-01

    This paper presents the summary reports presented at the concluding session and evaluating the state of the art, trends and perspectives as reflected in the papers presented at Coling 82 in six domains: machine translation, grammatico-semantic analysis, linguistics in its relations to computational linguistics, question answering, artificial intelligence and knowledge representation, and information retrieval and linguistic data bases.

  6. Formal monkey linguistics : The debate

    Schlenker, Philippe; Chemla, Emmanuel; Schel, Anne M.|info:eu-repo/dai/nl/413333450; Fuller, James; Gautier, Jean Pierre; Kuhn, Jeremy; Veselinović, Dunja; Arnold, Kate; Cäsar, Cristiane; Keenan, Sumir; Lemasson, Alban; Ouattara, Karim; Ryder, Robin; Zuberbühler, Klaus

    2016-01-01

    We explain why general techniques from formal linguistics can and should be applied to the analysis of monkey communication - in the areas of syntax and especially semantics. An informed look at our recent proposals shows that such techniques needn't rely excessively on categories of human language:

  7. Stellenbosch Papers in Linguistics Plus

    ). The SPiL Plus series has two main aims. Firstly, it serves as a vehicle for the distribution of new and relatively inaccessible information in the field of modern linguistics. Secondly, it aims to stimulate critical discussion in Southern African ...

  8. Extraction of CT dose information from DICOM metadata: automated Matlab-based approach.

    Dave, Jaydev K; Gingold, Eric L

    2013-01-01

    The purpose of this study was to extract exposure parameters and dose-relevant indexes of CT examinations from information embedded in DICOM metadata. DICOM dose report files were identified and retrieved from a PACS. An automated software program was used to extract from these files information from the structured elements in the DICOM metadata relevant to exposure. Extracting information from DICOM metadata eliminated potential errors inherent in techniques based on optical character recognition, yielding 100% accuracy.

  9. Medicaid Analytic eXtract (MAX) General Information

    U.S. Department of Health & Human Services — The Medicaid Analytic eXtract (MAX) data is a set of person-level data files on Medicaid eligibility, service utilization, and payments. The MAX data are created to...

  10. Lancaster Summer School in Corpus Linguistics

    Jaka Čibej

    2016-11-01

    Full Text Available Med 12. in 15. julijem je na Univerzi v Lancastru potekala poletna šola korpusnega jezikoslovja Lancaster Summer Schools in Corpus Linguistics and Other Digital Methods. Poletno šolo so organizirali UCREL (University Centre for Computer Corpus Research on Language, ERC (Evropski svet za raziskave – European Research Council, CASS (ESRC Centre for Corpus Approaches to Social Science in ESRC (Economic and Social Research Council, razdeljena pa je bila na šest programov, prilagojenih različnim področjem: Korpusno jezikoslovje za proučevanje jezikov (Corpus Linguistics for Language Studies, Korpusno jezikoslovje za družbene vede (Corpus Linguistics for Social Science, Korpusno jezikoslovje za humanistiko (Corpus Linguistics for Humanities, Statistika za korpusno jezikoslovje (Statistics for Corpus Linguistics, Geografski informacijski sistemi za digitalno humanistiko (Geographical Information Systems for the Digital Humanities in Korpusno podprta obdelava naravnih jezikov (Corpus-based Natural Language Processing.

  11. On Linguistic Abilities, Multilingualism, and Linguistic Justice

    Iannàccaro Gabriele

    2016-10-01

    Full Text Available The notion of linguistic justice should be related to the concept of linguistic ease, by which we mean the full social and communicative freedom of concern of the speaker in a given social interaction involving the use of language(s present in the society, according to the social norms of use. To acquire an acceptable degree of linguistic ease, the knowledge of at least one L2 is considered important. But the acquisition of a L2 is interfered by the previous linguistic skills of the learner/speaker who, in many cases, does not have a suitable competence even of the languages of the society in which he/she lives.

  12. Information Extraction with Character-level Neural Networks and Free Noisy Supervision

    Meerkamp, Philipp; Zhou, Zhengyi

    2016-01-01

    We present an architecture for information extraction from text that augments an existing parser with a character-level neural network. The network is trained using a measure of consistency of extracted data with existing databases as a form of noisy supervision. Our architecture combines the ability of constraint-based information extraction systems to easily incorporate domain knowledge and constraints with the ability of deep neural networks to leverage large amounts of data to learn compl...

  13. Semantics-based information extraction for detecting economic events

    A.C. Hogenboom (Alexander); F. Frasincar (Flavius); K. Schouten (Kim); O. van der Meer

    2013-01-01

    textabstractAs today's financial markets are sensitive to breaking news on economic events, accurate and timely automatic identification of events in news items is crucial. Unstructured news items originating from many heterogeneous sources have to be mined in order to extract knowledge useful for

  14. Tagline: Information Extraction for Semi-Structured Text Elements in Medical Progress Notes

    Finch, Dezon Kile

    2012-01-01

    Text analysis has become an important research activity in the Department of Veterans Affairs (VA). Statistical text mining and natural language processing have been shown to be very effective for extracting useful information from medical documents. However, neither of these techniques is effective at extracting the information stored in…

  15. An Effective Approach to Biomedical Information Extraction with Limited Training Data

    Jonnalagadda, Siddhartha

    2011-01-01

    In the current millennium, extensive use of computers and the internet caused an exponential increase in information. Few research areas are as important as information extraction, which primarily involves extracting concepts and the relations between them from free text. Limitations in the size of training data, lack of lexicons and lack of…

  16. Health Information National Trends Survey in American Sign Language (HINTS-ASL): Protocol for the Cultural Adaptation and Linguistic Validation of a National Survey.

    Kushalnagar, Poorna; Harris, Raychelle; Paludneviciene, Raylene; Hoglind, TraciAnn

    2017-09-13

    The Health Information National Trends Survey (HINTS) collects nationally representative data about the American's public use of health-related information. This survey is available in English and Spanish, but not in American Sign Language (ASL). Thus, the exclusion of ASL users from these national health information survey studies has led to a significant gap in knowledge of Internet usage for health information access in this underserved and understudied population. The objectives of this study are (1) to culturally adapt and linguistically translate the HINTS items to ASL (HINTS-ASL); and (2) to gather information about deaf people's health information seeking behaviors across technology-mediated platforms. We modified the standard procedures developed at the US National Center for Health Statistics Cognitive Survey Laboratory to culturally adapt and translate HINTS items to ASL. Cognitive interviews were conducted to assess clarity and delivery of these HINTS-ASL items. Final ASL video items were uploaded to a protected online survey website. The HINTS-ASL online survey has been administered to over 1350 deaf adults (ages 18 to 90 and up) who use ASL. Data collection is ongoing and includes deaf adult signers across the United States. Some items from HINTS item bank required cultural adaptation for use with deaf people who use accessible services or technology. A separate item bank for deaf-related experiences was created, reflecting deaf-specific technology such as sharing health-related ASL videos through social network sites and using video remote interpreting services in health settings. After data collection is complete, we will conduct a series of analyses on deaf people's health information seeking behaviors across technology-mediated platforms. HINTS-ASL is an accessible health information national trends survey, which includes a culturally appropriate set of items that are relevant to the experiences of deaf people who use ASL. The final HINTS

  17. A rapid extraction of landslide disaster information research based on GF-1 image

    Wang, Sai; Xu, Suning; Peng, Ling; Wang, Zhiyi; Wang, Na

    2015-08-01

    In recent years, the landslide disasters occurred frequently because of the seismic activity. It brings great harm to people's life. It has caused high attention of the state and the extensive concern of society. In the field of geological disaster, landslide information extraction based on remote sensing has been controversial, but high resolution remote sensing image can improve the accuracy of information extraction effectively with its rich texture and geometry information. Therefore, it is feasible to extract the information of earthquake- triggered landslides with serious surface damage and large scale. Taking the Wenchuan county as the study area, this paper uses multi-scale segmentation method to extract the landslide image object through domestic GF-1 images and DEM data, which uses the estimation of scale parameter tool to determine the optimal segmentation scale; After analyzing the characteristics of landslide high-resolution image comprehensively and selecting spectrum feature, texture feature, geometric features and landform characteristics of the image, we can establish the extracting rules to extract landslide disaster information. The extraction results show that there are 20 landslide whose total area is 521279.31 .Compared with visual interpretation results, the extraction accuracy is 72.22%. This study indicates its efficient and feasible to extract earthquake landslide disaster information based on high resolution remote sensing and it provides important technical support for post-disaster emergency investigation and disaster assessment.

  18. Towards an information extraction and knowledge formation framework based on Shannon entropy

    Iliescu Dragoș

    2017-01-01

    Full Text Available Information quantity subject is approached in this paperwork, considering the specific domain of nonconforming product management as information source. This work represents a case study. Raw data were gathered from a heavy industrial works company, information extraction and knowledge formation being considered herein. Involved method for information quantity estimation is based on Shannon entropy formula. Information and entropy spectrum are decomposed and analysed for extraction of specific information and knowledge-that formation. The result of the entropy analysis point out the information needed to be acquired by the involved organisation, this being presented as a specific knowledge type.

  19. A New Hesitant Fuzzy Linguistic TOPSIS Method for Group Multi-Criteria Linguistic Decision Making

    Fangling Ren

    2017-11-01

    Full Text Available Hesitant fuzzy linguistic decision making is a focus point in linguistic decision making, in which the main method is based on preference ordering. This paper develops a new hesitant fuzzy linguistic TOPSIS method for group multi-criteria linguistic decision making; the method is inspired by the TOPSIS method and the preference degree between two hesitant fuzzy linguistic term sets (HFLTSs. To this end, we first use the preference degree to define a pseudo-distance between two HFLTSs and analyze its properties. Then we present the positive (optimistic and negative (pessimistic information of each criterion provided by each decision maker and aggregate these by using weights of decision makers to obtain the hesitant fuzzy linguistic positive and negative ideal solutions. On the basis of the proposed pseudo-distance, we finally obtain the positive (negative ideal separation matrix and a new relative closeness degree to rank alternatives. We also design an algorithm based on the provided method to carry out hesitant fuzzy linguistic decision making. An illustrative example shows the elaboration of the proposed method and comparison with the symbolic aggregation-based method, the hesitant fuzzy linguistic TOPSIS method and the hesitant fuzzy linguistic VIKOR method; it seems that the proposed method is a useful and alternative decision-making method.

  20. Linguistic Structure Prediction

    Smith, Noah A

    2011-01-01

    A major part of natural language processing now depends on the use of text data to build linguistic analyzers. We consider statistical, computational approaches to modeling linguistic structure. We seek to unify across many approaches and many kinds of linguistic structures. Assuming a basic understanding of natural language processing and/or machine learning, we seek to bridge the gap between the two fields. Approaches to decoding (i.e., carrying out linguistic structure prediction) and supervised and unsupervised learning of models that predict discrete structures as outputs are the focus. W

  1. Extracting local information from crowds through betting markets

    Weijs, Steven

    2015-04-01

    In this research, a set-up is considered in which users can bet against a forecasting agency to challenge their probabilistic forecasts. From an information theory standpoint, a reward structure is considered that either provides the forecasting agency with better information, paying the successful providers of information for their winning bets, or funds excellent forecasting agencies through users that think they know better. Especially for local forecasts, the approach may help to diagnose model biases and to identify local predictive information that can be incorporated in the models. The challenges and opportunities for implementing such a system in practice are also discussed.

  2. Spoken Language Understanding Systems for Extracting Semantic Information from Speech

    Tur, Gokhan

    2011-01-01

    Spoken language understanding (SLU) is an emerging field in between speech and language processing, investigating human/ machine and human/ human communication by leveraging technologies from signal processing, pattern recognition, machine learning and artificial intelligence. SLU systems are designed to extract the meaning from speech utterances and its applications are vast, from voice search in mobile devices to meeting summarization, attracting interest from both commercial and academic sectors. Both human/machine and human/human communications can benefit from the application of SLU, usin

  3. Sifting Through Chaos: Extracting Information from Unstructured Legal Opinions.

    Oliveira, Bruno Miguel; Guimarães, Rui Vasconcellos; Antunes, Luís; Rodrigues, Pedro Pereira

    2018-01-01

    Abiding to the law is, in some cases, a delicate balance between the rights of different players. Re-using health records is such a case. While the law grants reuse rights to public administration documents, in which health records produced in public health institutions are included, it also grants privacy to personal records. To safeguard a correct usage of data, public hospitals in Portugal employ jurists that are responsible for allowing or withholding access rights to health records. To help decision making, these jurists can consult the legal opinions issued by the national committee on public administration documents usage. While these legal opinions are of undeniable value, due to their doctrine contribution, they are only available in a format best suited from printing, forcing individual consultation of each document, with no option, whatsoever of clustered search, filtering or indexing, which are standard operations nowadays in a document management system. When having to decide on tens of data requests a day, it becomes unfeasible to consult the hundreds of legal opinions already available. With the objective to create a modern document management system, we devised an open, platform agnostic system that extracts and compiles the legal opinions, ex-tracts its contents and produces metadata, allowing for a fast searching and filtering of said legal opinions.

  4. Stellenbosch Papers in Linguistics

    Stellenbosch Papers in Linguistics (SPiL) is an annual/biannual open access, peer-reviewed international journal, published by the Department of General Linguistics, Stellenbosch University. The papers published in SPiL are ... Poetry in South African Sign Language: What is different? EMAIL FREE FULL TEXT EMAIL ...

  5. Logic Programming for Linguistics

    Christiansen, Henning

    2010-01-01

    This article gives a short introduction on how to get started with logic pro- gramming in Prolog that does not require any previous programming expe- rience. The presentation is aimed at students of linguistics, but it does not go deeper into linguistics than any student who has some ideas of what...

  6. Linguistic Communications 1.

    Monash Univ., Clayton, Victoria (Australia).

    The present compilation of papers on linguistics is the result of joint efforts by the Classical Studies, French, Japanese, Linguistics, and Russian Departments of Monash University. Selections in the Pre-Prints and Articles section include: "For/Arabic Bilingualism in the Zalingei Area," by B. Jernudd; "Prosodic Problems in a Generative Phonology…

  7. Linguistic Extensions of Topic Models

    2010-09-01

    Movie Legally Multiplex Heralded As Linchpin To Growth The Shape of Cinema , Transformed At the Click of a Mouse A Peaceful Crew Puts Muppets...Linguistic Representation of Multiple Languages The formalism of WordNet has been applied to many languages from different language families, e.g. Japanese ...could be also share information gleaned from 100 reviews on Amazon.com’s Japanese and German language sites. 6.2.3 Learning Deeper Structures and Testing

  8. Information extraction from FN plots of tungsten microemitters

    Mussa, Khalil O. [Department of Physics, Mu' tah University, Al-Karak (Jordan); Mousa, Marwan S., E-mail: mmousa@mutah.edu.jo [Department of Physics, Mu' tah University, Al-Karak (Jordan); Fischer, Andreas, E-mail: andreas.fischer@physik.tu-chemnitz.de [Institut für Physik, Technische Universität Chemnitz, Chemnitz (Germany)

    2013-09-15

    Tungsten based microemitter tips have been prepared both clean and coated with dielectric materials. For clean tungsten tips, apex radii have been varied ranging from 25 to 500 nm. These tips were manufactured by electrochemical etching a 0.1 mm diameter high purity (99.95%) tungsten wire at the meniscus of two molar NaOH solution. Composite micro-emitters considered here are consisting of a tungsten core coated with different dielectric materials—such as magnesium oxide (MgO), sodium hydroxide (NaOH), tetracyanoethylene (TCNE), and zinc oxide (ZnO). It is worthwhile noting here, that the rather unconventional NaOH coating has shown several interesting properties. Various properties of these emitters were measured including current–voltage (IV) characteristics and the physical shape of the tips. A conventional field emission microscope (FEM) with a tip (cathode)–screen (anode) separation standardized at 10 mm was used to electrically characterize the electron emitters. The system was evacuated down to a base pressure of ∼10{sup −8}mbar when baked at up to ∼180°C overnight. This allowed measurements of typical field electron emission (FE) characteristics, namely the IV characteristics and the emission images on a conductive phosphorus screen (the anode). Mechanical characterization has been performed through a FEI scanning electron microscope (SEM). Within this work, the mentioned experimental results are connected to the theory for analyzing Fowler–Nordheim (FN) plots. We compared and evaluated the data extracted from clean tungsten tips of different radii and determined deviations between the results of different extraction methods applied. In particular, we derived the apex radii of several clean and coated tungsten tips by both SEM imaging and analyzing FN plots. The aim of this analysis is to support the ongoing discussion on recently developed improvements of the theory for analyzing FN plots related to metal field electron emitters, which in

  9. Information extraction from FN plots of tungsten microemitters

    Mussa, Khalil O.; Mousa, Marwan S.; Fischer, Andreas

    2013-01-01

    Tungsten based microemitter tips have been prepared both clean and coated with dielectric materials. For clean tungsten tips, apex radii have been varied ranging from 25 to 500 nm. These tips were manufactured by electrochemical etching a 0.1 mm diameter high purity (99.95%) tungsten wire at the meniscus of two molar NaOH solution. Composite micro-emitters considered here are consisting of a tungsten core coated with different dielectric materials—such as magnesium oxide (MgO), sodium hydroxide (NaOH), tetracyanoethylene (TCNE), and zinc oxide (ZnO). It is worthwhile noting here, that the rather unconventional NaOH coating has shown several interesting properties. Various properties of these emitters were measured including current–voltage (IV) characteristics and the physical shape of the tips. A conventional field emission microscope (FEM) with a tip (cathode)–screen (anode) separation standardized at 10 mm was used to electrically characterize the electron emitters. The system was evacuated down to a base pressure of ∼10 −8 mbar when baked at up to ∼180°C overnight. This allowed measurements of typical field electron emission (FE) characteristics, namely the IV characteristics and the emission images on a conductive phosphorus screen (the anode). Mechanical characterization has been performed through a FEI scanning electron microscope (SEM). Within this work, the mentioned experimental results are connected to the theory for analyzing Fowler–Nordheim (FN) plots. We compared and evaluated the data extracted from clean tungsten tips of different radii and determined deviations between the results of different extraction methods applied. In particular, we derived the apex radii of several clean and coated tungsten tips by both SEM imaging and analyzing FN plots. The aim of this analysis is to support the ongoing discussion on recently developed improvements of the theory for analyzing FN plots related to metal field electron emitters, which in

  10. When do we communicate stereotypes? Influence of the social context on the linguistic expectancy bias

    Wigboldus, DHJ; Spears, R; Semin, GR

    The linguistic expectancy bias (LEB) refers to the tendency to describe expectancy consistent information at a higher level of linguistic abstraction than expectancy inconsistent information. Two experiments examined the influence of the social communicative context on the production of this

  11. Towards a theoretical framework for analyzing complex linguistic networks

    Lücking, Andy; Banisch, Sven; Blanchard, Philippe; Job, Barbara

    2016-01-01

    The aim of this book is to advocate and promote network models of linguistic systems that are both based on thorough mathematical models and substantiated in terms of linguistics. In this way, the book contributes first steps towards establishing a statistical network theory as a theoretical basis of linguistic network analysis the boarder of the natural sciences and the humanities.This book addresses researchers who want to get familiar with theoretical developments, computational models and their empirical evaluation in the field of complex linguistic networks. It is intended to all those who are interested in statisticalmodels of linguistic systems from the point of view of network research. This includes all relevant areas of linguistics ranging from phonological, morphological and lexical networks on the one hand and syntactic, semantic and pragmatic networks on the other. In this sense, the volume concerns readers from many disciplines such as physics, linguistics, computer science and information scien...

  12. Optimal Information Extraction of Laser Scanning Dataset by Scale-Adaptive Reduction

    Zang, Y.; Yang, B.

    2018-04-01

    3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  13. OPTIMAL INFORMATION EXTRACTION OF LASER SCANNING DATASET BY SCALE-ADAPTIVE REDUCTION

    Y. Zang

    2018-04-01

    Full Text Available 3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  14. Information extraction from FN plots of tungsten microemitters.

    Mussa, Khalil O; Mousa, Marwan S; Fischer, Andreas

    2013-09-01

    Tungsten based microemitter tips have been prepared both clean and coated with dielectric materials. For clean tungsten tips, apex radii have been varied ranging from 25 to 500 nm. These tips were manufactured by electrochemical etching a 0.1 mm diameter high purity (99.95%) tungsten wire at the meniscus of two molar NaOH solution. Composite micro-emitters considered here are consisting of a tungsten core coated with different dielectric materials-such as magnesium oxide (MgO), sodium hydroxide (NaOH), tetracyanoethylene (TCNE), and zinc oxide (ZnO). It is worthwhile noting here, that the rather unconventional NaOH coating has shown several interesting properties. Various properties of these emitters were measured including current-voltage (IV) characteristics and the physical shape of the tips. A conventional field emission microscope (FEM) with a tip (cathode)-screen (anode) separation standardized at 10 mm was used to electrically characterize the electron emitters. The system was evacuated down to a base pressure of ∼10(-8) mbar when baked at up to ∼180 °C overnight. This allowed measurements of typical field electron emission (FE) characteristics, namely the IV characteristics and the emission images on a conductive phosphorus screen (the anode). Mechanical characterization has been performed through a FEI scanning electron microscope (SEM). Within this work, the mentioned experimental results are connected to the theory for analyzing Fowler-Nordheim (FN) plots. We compared and evaluated the data extracted from clean tungsten tips of different radii and determined deviations between the results of different extraction methods applied. In particular, we derived the apex radii of several clean and coated tungsten tips by both SEM imaging and analyzing FN plots. The aim of this analysis is to support the ongoing discussion on recently developed improvements of the theory for analyzing FN plots related to metal field electron emitters, which in particular

  15. Study on methods and techniques of aeroradiometric weak information extraction for sandstone-hosted uranium deposits based on GIS

    Han Shaoyang; Ke Dan; Hou Huiqun

    2005-01-01

    The weak information extraction is one of the important research contents in the current sandstone-type uranium prospecting in China. This paper introduces the connotation of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information and establishes some effective mathematic models for weak information extraction. Models for weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are completed in known uranium mineralized areas. Research results prove that the prospective areas of sandstone-type uranium deposits can be rapidly delineated by extracting aeroradiometric weak information. (authors)

  16. Extraction of Graph Information Based on Image Contents and the Use of Ontology

    Kanjanawattana, Sarunya; Kimura, Masaomi

    2016-01-01

    A graph is an effective form of data representation used to summarize complex information. Explicit information such as the relationship between the X- and Y-axes can be easily extracted from a graph by applying human intelligence. However, implicit knowledge such as information obtained from other related concepts in an ontology also resides in…

  17. Extracting information of fixational eye movements through pupil tracking

    Xiao, JiangWei; Qiu, Jian; Luo, Kaiqin; Peng, Li; Han, Peng

    2018-01-01

    Human eyes are never completely static even when they are fixing a stationary point. These irregular, small movements, which consist of micro-tremors, micro-saccades and drifts, can prevent the fading of the images that enter our eyes. The importance of researching the fixational eye movements has been experimentally demonstrated recently. However, the characteristics of fixational eye movements and their roles in visual process have not been explained clearly, because these signals can hardly be completely extracted by now. In this paper, we developed a new eye movement detection device with a high-speed camera. This device includes a beam splitter mirror, an infrared light source and a high-speed digital video camera with a frame rate of 200Hz. To avoid the influence of head shaking, we made the device wearable by fixing the camera on a safety helmet. Using this device, the experiments of pupil tracking were conducted. By localizing the pupil center and spectrum analysis, the envelope frequency spectrum of micro-saccades, micro-tremors and drifts are shown obviously. The experimental results show that the device is feasible and effective, so that the device can be applied in further characteristic analysis.

  18. Extracting Social Networks and Contact Information From Email and the Web

    Culotta, Aron; Bekkerman, Ron; McCallum, Andrew

    2005-01-01

    ...-suited for such information extraction tasks. By recursively calling itself on new people discovered on the Web, the system builds a social network with multiple degrees of separation from the user...

  19. Lithium NLP: A System for Rich Information Extraction from Noisy User Generated Text on Social Media

    Bhargava, Preeti; Spasojevic, Nemanja; Hu, Guoning

    2017-01-01

    In this paper, we describe the Lithium Natural Language Processing (NLP) system - a resource-constrained, high- throughput and language-agnostic system for information extraction from noisy user generated text on social media. Lithium NLP extracts a rich set of information including entities, topics, hashtags and sentiment from text. We discuss several real world applications of the system currently incorporated in Lithium products. We also compare our system with existing commercial and acad...

  20. Applied Linguistics and the "Annual Review of Applied Linguistics."

    Kaplan, Robert B.; Grabe, William

    2000-01-01

    Examines the complexities and differences involved in granting disciplinary status to the role of applied linguistics, discusses the role of the "Annual Review of Applied Linguistics" as a contributor to the development of applied linguistics, and highlights a set of publications for the future of applied linguistics. (Author/VWL)

  1. Information Extraction of High Resolution Remote Sensing Images Based on the Calculation of Optimal Segmentation Parameters

    Zhu, Hongchun; Cai, Lijie; Liu, Haiying; Huang, Wei

    2016-01-01

    Multi-scale image segmentation and the selection of optimal segmentation parameters are the key processes in the object-oriented information extraction of high-resolution remote sensing images. The accuracy of remote sensing special subject information depends on this extraction. On the basis of WorldView-2 high-resolution data, the optimal segmentation parameters methodof object-oriented image segmentation and high-resolution image information extraction, the following processes were conducted in this study. Firstly, the best combination of the bands and weights was determined for the information extraction of high-resolution remote sensing image. An improved weighted mean-variance method was proposed andused to calculatethe optimal segmentation scale. Thereafter, the best shape factor parameter and compact factor parameters were computed with the use of the control variables and the combination of the heterogeneity and homogeneity indexes. Different types of image segmentation parameters were obtained according to the surface features. The high-resolution remote sensing images were multi-scale segmented with the optimal segmentation parameters. Ahierarchical network structure was established by setting the information extraction rules to achieve object-oriented information extraction. This study presents an effective and practical method that can explain expert input judgment by reproducible quantitative measurements. Furthermore the results of this procedure may be incorporated into a classification scheme. PMID:27362762

  2. Overview of ImageCLEF 2017: information extraction from images

    Ionescu, Bogdan; Müller, Henning; Villegas, Mauricio; Arenas, Helbert; Boato, Giulia; Dang Nguyen, Duc Tien; Dicente Cid, Yashin; Eickhoff, Carsten; Seco de Herrera, Alba G.; Gurrin, Cathal; Islam, Bayzidul; Kovalev, Vassili; Liauchuk, Vitali; Mothe, Josiane; Piras, Luca

    2017-01-01

    This paper presents an overview of the ImageCLEF 2017 evaluation campaign, an event that was organized as part of the CLEF (Conference and Labs of the Evaluation Forum) labs 2017. ImageCLEF is an ongoing initiative (started in 2003) that promotes the evaluation of technologies for annotation, indexing and retrieval for providing information access to collections of images in various usage scenarios and domains. In 2017, the 15th edition of ImageCLEF, three main tasks were proposed and one pil...

  3. Statistical techniques to extract information during SMAP soil moisture assimilation

    Kolassa, J.; Reichle, R. H.; Liu, Q.; Alemohammad, S. H.; Gentine, P.

    2017-12-01

    Statistical techniques permit the retrieval of soil moisture estimates in a model climatology while retaining the spatial and temporal signatures of the satellite observations. As a consequence, the need for bias correction prior to an assimilation of these estimates is reduced, which could result in a more effective use of the independent information provided by the satellite observations. In this study, a statistical neural network (NN) retrieval algorithm is calibrated using SMAP brightness temperature observations and modeled soil moisture estimates (similar to those used to calibrate the SMAP Level 4 DA system). Daily values of surface soil moisture are estimated using the NN and then assimilated into the NASA Catchment model. The skill of the assimilation estimates is assessed based on a comprehensive comparison to in situ measurements from the SMAP core and sparse network sites as well as the International Soil Moisture Network. The NN retrieval assimilation is found to significantly improve the model skill, particularly in areas where the model does not represent processes related to agricultural practices. Additionally, the NN method is compared to assimilation experiments using traditional bias correction techniques. The NN retrieval assimilation is found to more effectively use the independent information provided by SMAP resulting in larger model skill improvements than assimilation experiments using traditional bias correction techniques.

  4. Research on Crowdsourcing Emergency Information Extraction of Based on Events' Frame

    Yang, Bo; Wang, Jizhou; Ma, Weijun; Mao, Xi

    2018-01-01

    At present, the common information extraction method cannot extract the structured emergency event information accurately; the general information retrieval tool cannot completely identify the emergency geographic information; these ways also do not have an accurate assessment of these results of distilling. So, this paper proposes an emergency information collection technology based on event framework. This technique is to solve the problem of emergency information picking. It mainly includes emergency information extraction model (EIEM), complete address recognition method (CARM) and the accuracy evaluation model of emergency information (AEMEI). EIEM can be structured to extract emergency information and complements the lack of network data acquisition in emergency mapping. CARM uses a hierarchical model and the shortest path algorithm and allows the toponomy pieces to be joined as a full address. AEMEI analyzes the results of the emergency event and summarizes the advantages and disadvantages of the event framework. Experiments show that event frame technology can solve the problem of emergency information drawing and provides reference cases for other applications. When the emergency disaster is about to occur, the relevant departments query emergency's data that has occurred in the past. They can make arrangements ahead of schedule which defense and reducing disaster. The technology decreases the number of casualties and property damage in the country and world. This is of great significance to the state and society.

  5. [Extraction of management information from the national quality assurance program].

    Stausberg, Jürgen; Bartels, Claus; Bobrowski, Christoph

    2007-07-15

    Starting with clinically motivated projects, the national quality assurance program has established a legislative obligatory framework. Annual feedback of results is an important means of quality control. The annual reports cover quality-related information with high granularity. A synopsis for corporate management is missing, however. Therefore, the results of the University Clinics in Greifswald, Germany, have been analyzed and aggregated to support hospital management. Strengths were identified by the ranking of results within the state for each quality indicator, weaknesses by the comparison with national reference values. The assessment was aggregated per clinical discipline and per category (indication, process, and outcome). A composition of quality indicators was claimed multiple times. A coherent concept is still missing. The method presented establishes a plausible summary of strengths and weaknesses of a hospital from the point of view of the national quality assurance program. Nevertheless, further adaptation of the program is needed to better assist corporate management.

  6. Siblings as Mediators of Literacy in Linguistic Minority Communities.

    Gregory, Eve

    1998-01-01

    Argues for need to move beyond model of "parental" involvement in reading, which presently informs home/school reading programmes for linguistic-minority children in UK. First examines literature informing current model showing marked absence of studies on role played by siblings as mediators of literacy in new linguistic and cultural…

  7. The Perilous Life of a Linguistic Genre Convention

    Borchmann, Simon

    2014-01-01

    , the descriptions are more informative than the structures hitherto described by text linguistics. Secondly, as historical norms, they are a testimony to the development and change of language use. Thirdly, the descriptions contribute to language users’ awareness of the origin of standards, their understanding......The primary, theoretical aim of the article is to present a linguistic text analysis that differs from standard text linguistic approaches by being informative with regard to the linguistic choices and textual organisation that characterise a text as a social act. The analysis is exemplified...... by using texts of a relatively new Danish journalistic genre nyhedsanalyse (news analysis). The secondary, empirical aim of the article is to present a corpus-based, linguistic analysis of central elements of the genre nyhedsanalyse within the Danish system of newspaper genres. Text linguistics is based...

  8. Extracting of implicit information in English advertising texts with phonetic and lexical-morphological means

    Traikovskaya Natalya Petrovna

    2015-12-01

    Full Text Available The article deals with phonetic and lexical-morphological language means participating in the process of extracting implicit information in English-speaking advertising texts for men and women. The functioning of phonetic means of the English language is not the basis for implication of information in advertising texts. Lexical and morphological means play the role of markers of relevant information, playing the role of the activator ofimplicit information in the texts of advertising.

  9. Saussure and Linguistic Geography.

    Harris, Roy

    1993-01-01

    Discusses Saussures's "Cours de linguistique generale," which was published in 1916, and devotes specific attention to the significance of Part VI, which is devoted to linguistic geography. (16 references) (Author/VWL)

  10. Language Works. Linguistic Journal

    Hartling, Anna Sofie; Nørreby, Thomas Rørbeck; Skovse, Astrid Ravn

    2016-01-01

    Language works! – and with this initiative and this journal we want to give the opportunity to many more students to present their linguistic research to each other, to the scientific community and to all interested.......Language works! – and with this initiative and this journal we want to give the opportunity to many more students to present their linguistic research to each other, to the scientific community and to all interested....

  11. Mathematics and linguistics

    Landauer, C.; Bellman, K.L.

    1996-12-31

    In this paper, we study foundational issues that we believe will help us develop a theoretically sound approach to constructing complex systems. The two theoretical approaches that have helped us understand and develop computational systems in the past are mathematics and linguistics. We describe some differences and strengths of the approaches, and propose a research program to combine the richness of linguistic reasoning with the precision of mathematics.

  12. Post-processing of Deep Web Information Extraction Based on Domain Ontology

    PENG, T.

    2013-11-01

    Full Text Available Many methods are utilized to extract and process query results in deep Web, which rely on the different structures of Web pages and various designing modes of databases. However, some semantic meanings and relations are ignored. So, in this paper, we present an approach for post-processing deep Web query results based on domain ontology which can utilize the semantic meanings and relations. A block identification model (BIM based on node similarity is defined to extract data blocks that are relevant to specific domain after reducing noisy nodes. Feature vector of domain books is obtained by result set extraction model (RSEM based on vector space model (VSM. RSEM, in combination with BIM, builds the domain ontology on books which can not only remove the limit of Web page structures when extracting data information, but also make use of semantic meanings of domain ontology. After extracting basic information of Web pages, a ranking algorithm is adopted to offer an ordered list of data records to users. Experimental results show that BIM and RSEM extract data blocks and build domain ontology accurately. In addition, relevant data records and basic information are extracted and ranked. The performances precision and recall show that our proposed method is feasible and efficient.

  13. Can delusions be understood linguistically?

    Hinzen, Wolfram; Rosselló, Joana; McKenna, Peter

    2016-01-01

    ABSTRACT Delusions are widely believed to reflect disturbed cognitive function, but the nature of this remains elusive. The “un-Cartesian” cognitive-linguistic hypothesis maintains (a) that there is no thought separate from language, that is, there is no distinct mental space removed from language where “thinking” takes place; and (b) that a somewhat broadened concept of grammar is responsible for bestowing meaning on propositions, and this among other things gives them their quality of being true or false. It is argued that a loss of propositional meaning explains why delusions are false, impossible and sometimes fantastic. A closely related abnormality, failure of linguistic embedding, can additionally account for why delusions are held with fixed conviction and are not adequately justified by the patient. The un-Cartesian linguistic approach to delusions has points of contact with Frith’s theory that inability to form meta-representations underlies a range of schizophrenic symptoms. It may also be relevant to the nature of the “second factor” in monothematic delusions in neurological disease. Finally, it can inform the current debate about whether or not delusions really are beliefs. PMID:27322493

  14. Can delusions be understood linguistically?

    Hinzen, Wolfram; Rosselló, Joana; McKenna, Peter

    2016-07-01

    Delusions are widely believed to reflect disturbed cognitive function, but the nature of this remains elusive. The "un-Cartesian" cognitive-linguistic hypothesis maintains (a) that there is no thought separate from language, that is, there is no distinct mental space removed from language where "thinking" takes place; and (b) that a somewhat broadened concept of grammar is responsible for bestowing meaning on propositions, and this among other things gives them their quality of being true or false. It is argued that a loss of propositional meaning explains why delusions are false, impossible and sometimes fantastic. A closely related abnormality, failure of linguistic embedding, can additionally account for why delusions are held with fixed conviction and are not adequately justified by the patient. The un-Cartesian linguistic approach to delusions has points of contact with Frith's theory that inability to form meta-representations underlies a range of schizophrenic symptoms. It may also be relevant to the nature of the "second factor" in monothematic delusions in neurological disease. Finally, it can inform the current debate about whether or not delusions really are beliefs.

  15. Having Linguistic Rules and Knowing Linguistic Facts

    Peter Ludlow

    2010-11-01

    Full Text Available

    'Knowledge' doesn't correctly describe our relation to linguistic rules. It is too thick a notion (for example, we don't believe linguistic rules. On the other hand, 'cognize', without further elaboration, is too thin a notion, which is to say that it is too thin to play a role in a competence theory. One advantage of the term 'knowledge'-and presumably Chomsky's original motivation for using it-is that knowledge would play the right kind of role in a competence theory: Our competence would consist in a body of knowledge which we have and which we may or may not act upon-our performance need not conform to the linguistic rules that we know.

    Is there a way out of the dilemma? I'm going to make the case that the best way to talk about grammatical rules is simply to say that we have them. That doesn't sound very deep, I know, but saying that we have individual rules leaves room for individual norm guidance in a way that 'cognize' does not. Saying we have a rule like subjacency is also thicker than merely saying we cognize it. Saying I have such a rule invites the interpretation that it is a rule for me-that I am normatively guided by it. The competence theory thus becomes a theory of the rules that we have. Whether we follow those rules is another matter entirely.

  16. a Statistical Texture Feature for Building Collapse Information Extraction of SAR Image

    Li, L.; Yang, H.; Chen, Q.; Liu, X.

    2018-04-01

    Synthetic Aperture Radar (SAR) has become one of the most important ways to extract post-disaster collapsed building information, due to its extreme versatility and almost all-weather, day-and-night working capability, etc. In view of the fact that the inherent statistical distribution of speckle in SAR images is not used to extract collapsed building information, this paper proposed a novel texture feature of statistical models of SAR images to extract the collapsed buildings. In the proposed feature, the texture parameter of G0 distribution from SAR images is used to reflect the uniformity of the target to extract the collapsed building. This feature not only considers the statistical distribution of SAR images, providing more accurate description of the object texture, but also is applied to extract collapsed building information of single-, dual- or full-polarization SAR data. The RADARSAT-2 data of Yushu earthquake which acquired on April 21, 2010 is used to present and analyze the performance of the proposed method. In addition, the applicability of this feature to SAR data with different polarizations is also analysed, which provides decision support for the data selection of collapsed building information extraction.

  17. A method for automating the extraction of specialized information from the web

    Lin, L.; Liotta, A.; Hippisley, A.; Hao, Y.; Liu, J.; Wang, Y.; Cheung, Y-M.; Yin, H.; Jiao, L.; Ma, j.; Jiao, Y-C.

    2005-01-01

    The World Wide Web can be viewed as a gigantic distributed database including millions of interconnected hosts some of which publish information via web servers or peer-to-peer systems. We present here a novel method for the extraction of semantically rich information from the web in a fully

  18. Information analysis of iris biometrics for the needs of cryptology key extraction

    Adamović Saša

    2013-01-01

    Full Text Available The paper presents a rigorous analysis of iris biometric information for the synthesis of an optimized system for the extraction of a high quality cryptology key. Estimations of local entropy and mutual information were identified as segments of the iris most suitable for this purpose. In order to optimize parameters, corresponding wavelets were transformed, in order to obtain the highest possible entropy and mutual information lower in the transformation domain, which set frameworks for the synthesis of systems for the extraction of truly random sequences of iris biometrics, without compromising authentication properties. [Projekat Ministarstva nauke Republike Srbije, br. TR32054 i br. III44006

  19. Automated Linguistic Personality Description and Recognition Methods

    Danylyuk Illya

    2016-12-01

    Full Text Available Background: The relevance of our research, above all, is theoretically motivated by the development of extraordinary scientific and practical interest in the possibilities of language processing of huge amount of data generated by people in everyday professional and personal life in the electronic forms of communication (e-mail, sms, voice, audio and video blogs, social networks, etc.. Purpose: The purpose of the article is to describe the theoretical and practical framework of the project "Communicative-pragmatic and discourse-grammatical lingvopersonology: structuring linguistic identity and computer modeling". The description of key techniques is given, such as machine learning for language modeling, speech synthesis, handwriting simulation. Results: Lingvopersonology developed some great theoretical foundations, its methods, tools, and significant achievements let us predict that the newest promising trend is a linguistic identity modeling by means of information technology, including language. We see three aspects of the modeling: 1 modeling the semantic level of linguistic identity – by means of the use of corpus linguistics; 2 sound level formal modeling of linguistic identity – with the help of speech synthesis; 3 formal graphic level modeling of linguistic identity – with the help of image synthesis (handwriting. For the first case, we suppose to use machine learning technics and vector-space (word2vec algorithm for textual speech modeling. Hybrid CUTE method for personality speech modeling will be applied to the second case. Finally, trained with the person handwriting images neural network can be an instrument for the last case. Discussion: The project "Communicative-pragmatic, discourse, and grammatical lingvopersonology: structuring linguistic identity and computer modeling", which is implementing by the Department of General and Applied Linguistics and Slavonic philology, selected a task to model Yuriy Shevelyov (Sherekh

  20. Words Get in the Way: Linguistic Effects on Talker Discrimination.

    Narayan, Chandan R; Mak, Lorinda; Bialystok, Ellen

    2017-07-01

    A speech perception experiment provides evidence that the linguistic relationship between words affects the discrimination of their talkers. Listeners discriminated two talkers' voices with various linguistic relationships between their spoken words. Listeners were asked whether two words were spoken by the same person or not. Word pairs varied with respect to the linguistic relationship between the component words, forming either: phonological rhymes, lexical compounds, reversed compounds, or unrelated pairs. The degree of linguistic relationship between the words affected talker discrimination in a graded fashion, revealing biases listeners have regarding the nature of words and the talkers that speak them. These results indicate that listeners expect a talker's words to be linguistically related, and more generally, indexical processing is affected by linguistic information in a top-down fashion even when listeners are not told to attend to it. Copyright © 2016 Cognitive Science Society, Inc.

  1. MedTime: a temporal information extraction system for clinical narratives.

    Lin, Yu-Kai; Chen, Hsinchun; Brown, Randall A

    2013-12-01

    Temporal information extraction from clinical narratives is of critical importance to many clinical applications. We participated in the EVENT/TIMEX3 track of the 2012 i2b2 clinical temporal relations challenge, and presented our temporal information extraction system, MedTime. MedTime comprises a cascade of rule-based and machine-learning pattern recognition procedures. It achieved a micro-averaged f-measure of 0.88 in both the recognitions of clinical events and temporal expressions. We proposed and evaluated three time normalization strategies to normalize relative time expressions in clinical texts. The accuracy was 0.68 in normalizing temporal expressions of dates, times, durations, and frequencies. This study demonstrates and evaluates the integration of rule-based and machine-learning-based approaches for high performance temporal information extraction from clinical narratives. Copyright © 2013 Elsevier Inc. All rights reserved.

  2. Research of building information extraction and evaluation based on high-resolution remote-sensing imagery

    Cao, Qiong; Gu, Lingjia; Ren, Ruizhi; Wang, Lang

    2016-09-01

    Building extraction currently is important in the application of high-resolution remote sensing imagery. At present, quite a few algorithms are available for detecting building information, however, most of them still have some obvious disadvantages, such as the ignorance of spectral information, the contradiction between extraction rate and extraction accuracy. The purpose of this research is to develop an effective method to detect building information for Chinese GF-1 data. Firstly, the image preprocessing technique is used to normalize the image and image enhancement is used to highlight the useful information in the image. Secondly, multi-spectral information is analyzed. Subsequently, an improved morphological building index (IMBI) based on remote sensing imagery is proposed to get the candidate building objects. Furthermore, in order to refine building objects and further remove false objects, the post-processing (e.g., the shape features, the vegetation index and the water index) is employed. To validate the effectiveness of the proposed algorithm, the omission errors (OE), commission errors (CE), the overall accuracy (OA) and Kappa are used at final. The proposed method can not only effectively use spectral information and other basic features, but also avoid extracting excessive interference details from high-resolution remote sensing images. Compared to the original MBI algorithm, the proposed method reduces the OE by 33.14% .At the same time, the Kappa increase by 16.09%. In experiments, IMBI achieved satisfactory results and outperformed other algorithms in terms of both accuracies and visual inspection

  3. Information Extraction of High-Resolution Remotely Sensed Image Based on Multiresolution Segmentation

    Peng Shao

    2014-08-01

    Full Text Available The principle of multiresolution segmentation was represented in detail in this study, and the canny algorithm was applied for edge-detection of a remotely sensed image based on this principle. The target image was divided into regions based on object-oriented multiresolution segmentation and edge-detection. Furthermore, object hierarchy was created, and a series of features (water bodies, vegetation, roads, residential areas, bare land and other information were extracted by the spectral and geometrical features. The results indicate that the edge-detection has a positive effect on multiresolution segmentation, and overall accuracy of information extraction reaches to 94.6% by the confusion matrix.

  4. End-to-end information extraction without token-level supervision

    Palm, Rasmus Berg; Hovy, Dirk; Laws, Florian

    2017-01-01

    Most state-of-the-art information extraction approaches rely on token-level labels to find the areas of interest in text. Unfortunately, these labels are time-consuming and costly to create, and consequently, not available for many real-life IE tasks. To make matters worse, token-level labels...... and output text. We evaluate our model on the ATIS data set, MIT restaurant corpus and the MIT movie corpus and compare to neural baselines that do use token-level labels. We achieve competitive results, within a few percentage points of the baselines, showing the feasibility of E2E information extraction...

  5. Improving English Instruction through Neuro-Linguistic Programming

    Helm, David Jay

    2009-01-01

    This study examines the background information and numerous applications of neuro-linguistic programming as it applies to improving English instruction. In addition, the N.L.P. modalities of eye movement, the use of predicates, and posturing are discussed. Neuro-linguistic programming presents all students of English an opportunity to reach their…

  6. Extraction Method for Earthquake-Collapsed Building Information Based on High-Resolution Remote Sensing

    Chen, Peng; Wu, Jian; Liu, Yaolin; Wang, Jing

    2014-01-01

    At present, the extraction of earthquake disaster information from remote sensing data relies on visual interpretation. However, this technique cannot effectively and quickly obtain precise and efficient information for earthquake relief and emergency management. Collapsed buildings in the town of Zipingpu after the Wenchuan earthquake were used as a case study to validate two kinds of rapid extraction methods for earthquake-collapsed building information based on pixel-oriented and object-oriented theories. The pixel-oriented method is based on multi-layer regional segments that embody the core layers and segments of the object-oriented method. The key idea is to mask layer by layer all image information, including that on the collapsed buildings. Compared with traditional techniques, the pixel-oriented method is innovative because it allows considerably rapid computer processing. As for the object-oriented method, a multi-scale segment algorithm was applied to build a three-layer hierarchy. By analyzing the spectrum, texture, shape, location, and context of individual object classes in different layers, the fuzzy determined rule system was established for the extraction of earthquake-collapsed building information. We compared the two sets of results using three variables: precision assessment, visual effect, and principle. Both methods can extract earthquake-collapsed building information quickly and accurately. The object-oriented method successfully overcomes the pepper salt noise caused by the spectral diversity of high-resolution remote sensing data and solves the problem of same object, different spectrums and that of same spectrum, different objects. With an overall accuracy of 90.38%, the method achieves more scientific and accurate results compared with the pixel-oriented method (76.84%). The object-oriented image analysis method can be extensively applied in the extraction of earthquake disaster information based on high-resolution remote sensing

  7. Using text mining techniques to extract phenotypic information from the PhenoCHF corpus.

    Alnazzawi, Noha; Thompson, Paul; Batista-Navarro, Riza; Ananiadou, Sophia

    2015-01-01

    Phenotypic information locked away in unstructured narrative text presents significant barriers to information accessibility, both for clinical practitioners and for computerised applications used for clinical research purposes. Text mining (TM) techniques have previously been applied successfully to extract different types of information from text in the biomedical domain. They have the potential to be extended to allow the extraction of information relating to phenotypes from free text. To stimulate the development of TM systems that are able to extract phenotypic information from text, we have created a new corpus (PhenoCHF) that is annotated by domain experts with several types of phenotypic information relating to congestive heart failure. To ensure that systems developed using the corpus are robust to multiple text types, it integrates text from heterogeneous sources, i.e., electronic health records (EHRs) and scientific articles from the literature. We have developed several different phenotype extraction methods to demonstrate the utility of the corpus, and tested these methods on a further corpus, i.e., ShARe/CLEF 2013. Evaluation of our automated methods showed that PhenoCHF can facilitate the training of reliable phenotype extraction systems, which are robust to variations in text type. These results have been reinforced by evaluating our trained systems on the ShARe/CLEF corpus, which contains clinical records of various types. Like other studies within the biomedical domain, we found that solutions based on conditional random fields produced the best results, when coupled with a rich feature set. PhenoCHF is the first annotated corpus aimed at encoding detailed phenotypic information. The unique heterogeneous composition of the corpus has been shown to be advantageous in the training of systems that can accurately extract phenotypic information from a range of different text types. Although the scope of our annotation is currently limited to a single

  8. Terrain Extraction by Integrating Terrestrial Laser Scanner Data and Spectral Information

    Lau, C. L.; Halim, S.; Zulkepli, M.; Azwan, A. M.; Tang, W. L.; Chong, A. K.

    2015-10-01

    The extraction of true terrain points from unstructured laser point cloud data is an important process in order to produce an accurate digital terrain model (DTM). However, most of these spatial filtering methods just utilizing the geometrical data to discriminate the terrain points from nonterrain points. The point cloud filtering method also can be improved by using the spectral information available with some scanners. Therefore, the objective of this study is to investigate the effectiveness of using the three-channel (red, green and blue) of the colour image captured from built-in digital camera which is available in some Terrestrial Laser Scanner (TLS) for terrain extraction. In this study, the data acquisition was conducted at a mini replica landscape in Universiti Teknologi Malaysia (UTM), Skudai campus using Leica ScanStation C10. The spectral information of the coloured point clouds from selected sample classes are extracted for spectral analysis. The coloured point clouds which within the corresponding preset spectral threshold are identified as that specific feature point from the dataset. This process of terrain extraction is done through using developed Matlab coding. Result demonstrates that a higher spectral resolution passive image is required in order to improve the output. This is because low quality of the colour images captured by the sensor contributes to the low separability in spectral reflectance. In conclusion, this study shows that, spectral information is capable to be used as a parameter for terrain extraction.

  9. Information retrieval and terminology extraction in online resources for patients with diabetes.

    Seljan, Sanja; Baretić, Maja; Kucis, Vlasta

    2014-06-01

    Terminology use, as a mean for information retrieval or document indexing, plays an important role in health literacy. Specific types of users, i.e. patients with diabetes need access to various online resources (on foreign and/or native language) searching for information on self-education of basic diabetic knowledge, on self-care activities regarding importance of dietetic food, medications, physical exercises and on self-management of insulin pumps. Automatic extraction of corpus-based terminology from online texts, manuals or professional papers, can help in building terminology lists or list of "browsing phrases" useful in information retrieval or in document indexing. Specific terminology lists represent an intermediate step between free text search and controlled vocabulary, between user's demands and existing online resources in native and foreign language. The research aiming to detect the role of terminology in online resources, is conducted on English and Croatian manuals and Croatian online texts, and divided into three interrelated parts: i) comparison of professional and popular terminology use ii) evaluation of automatic statistically-based terminology extraction on English and Croatian texts iii) comparison and evaluation of extracted terminology performed on English manual using statistical and hybrid approaches. Extracted terminology candidates are evaluated by comparison with three types of reference lists: list created by professional medical person, list of highly professional vocabulary contained in MeSH and list created by non-medical persons, made as intersection of 15 lists. Results report on use of popular and professional terminology in online diabetes resources, on evaluation of automatically extracted terminology candidates in English and Croatian texts and on comparison of statistical and hybrid extraction methods in English text. Evaluation of automatic and semi-automatic terminology extraction methods is performed by recall

  10. OpenCV-Based Nanomanipulation Information Extraction and the Probe Operation in SEM

    Dongjie Li

    2015-02-01

    Full Text Available Aimed at the established telenanomanipulation system, the method of extracting location information and the strategies of probe operation were studied in this paper. First, the machine learning algorithm of OpenCV was used to extract location information from SEM images. Thus nanowires and probe in SEM images can be automatically tracked and the region of interest (ROI can be marked quickly. Then the location of nanowire and probe can be extracted from the ROI. To study the probe operation strategy, the Van der Waals force between probe and a nanowire was computed; thus relevant operating parameters can be obtained. With these operating parameters, the nanowire in 3D virtual environment can be preoperated and an optimal path of the probe can be obtained. The actual probe runs automatically under the telenanomanipulation system's control. Finally, experiments were carried out to verify the above methods, and results show the designed methods have achieved the expected effect.

  11. Methods to extract information on the atomic and molecular states from scientific abstracts

    Sasaki, Akira; Ueshima, Yutaka; Yamagiwa, Mitsuru; Murata, Masaki; Kanamaru, Toshiyuki; Shirado, Tamotsu; Isahara, Hitoshi

    2005-01-01

    We propose a new application of information technology to recognize and extract expressions of atomic and molecular states from electrical forms of scientific abstracts. Present results will help scientists to understand atomic states as well as the physics discussed in the articles. Combining with the internet search engines, it will make one possible to collect not only atomic and molecular data but broader scientific information over a wide range of research fields. (author)

  12. System and method for extracting physiological information from remotely detected electromagnetic radiation

    2016-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  13. System and method for extracting physiological information from remotely detected electromagnetic radiation

    2015-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  14. Network and Ensemble Enabled Entity Extraction in Informal Text (NEEEEIT) final report

    Kegelmeyer, Philip W. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Shead, Timothy M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Dunlavy, Daniel M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2013-09-01

    This SAND report summarizes the activities and outcomes of the Network and Ensemble Enabled Entity Extraction in Information Text (NEEEEIT) LDRD project, which addressed improving the accuracy of conditional random fields for named entity recognition through the use of ensemble methods.

  15. A construction scheme of web page comment information extraction system based on frequent subtree mining

    Zhang, Xiaowen; Chen, Bingfeng

    2017-08-01

    Based on the frequent sub-tree mining algorithm, this paper proposes a construction scheme of web page comment information extraction system based on frequent subtree mining, referred to as FSM system. The entire system architecture and the various modules to do a brief introduction, and then the core of the system to do a detailed description, and finally give the system prototype.

  16. EXTRACT

    Pafilis, Evangelos; Buttigieg, Pier Luigi; Ferrell, Barbra

    2016-01-01

    The microbial and molecular ecology research communities have made substantial progress on developing standards for annotating samples with environment metadata. However, sample manual annotation is a highly labor intensive process and requires familiarity with the terminologies used. We have the...... and text-mining-assisted curation revealed that EXTRACT speeds up annotation by 15-25% and helps curators to detect terms that would otherwise have been missed.Database URL: https://extract.hcmr.gr/......., organism, tissue and disease terms. The evaluators in the BioCreative V Interactive Annotation Task found the system to be intuitive, useful, well documented and sufficiently accurate to be helpful in spotting relevant text passages and extracting organism and environment terms. Comparison of fully manual...

  17. Semi-automatic building extraction in informal settlements from high-resolution satellite imagery

    Mayunga, Selassie David

    The extraction of man-made features from digital remotely sensed images is considered as an important step underpinning management of human settlements in any country. Man-made features and buildings in particular are required for varieties of applications such as urban planning, creation of geographical information systems (GIS) databases and Urban City models. The traditional man-made feature extraction methods are very expensive in terms of equipment, labour intensive, need well-trained personnel and cannot cope with changing environments, particularly in dense urban settlement areas. This research presents an approach for extracting buildings in dense informal settlement areas using high-resolution satellite imagery. The proposed system uses a novel strategy of extracting building by measuring a single point at the approximate centre of the building. The fine measurement of the building outlines is then effected using a modified snake model. The original snake model on which this framework is based, incorporates an external constraint energy term which is tailored to preserving the convergence properties of the snake model; its use to unstructured objects will negatively affect their actual shapes. The external constrained energy term was removed from the original snake model formulation, thereby, giving ability to cope with high variability of building shapes in informal settlement areas. The proposed building extraction system was tested on two areas, which have different situations. The first area was Tungi in Dar Es Salaam, Tanzania where three sites were tested. This area is characterized by informal settlements, which are illegally formulated within the city boundaries. The second area was Oromocto in New Brunswick, Canada where two sites were tested. Oromocto area is mostly flat and the buildings are constructed using similar materials. Qualitative and quantitative measures were employed to evaluate the accuracy of the results as well as the performance

  18. Peace linguistics for language teachers

    Francisco GOMES DE MATOS

    2014-12-01

    Full Text Available This text aims at presenting the concept of Peace Linguistics - origins and recent developments -- as being implemented in the author's ongoing work in that emerging branch of Applied Linguistics. Examples of applicational possibilities are given, with a focus on language teaching-learning and a Checklist is provided, of topics for suggested linguistic-educational research, centered on communicative peace.

  19. Linguistic Corpora and Language Teaching.

    Murison-Bowie, Simon

    1996-01-01

    Examines issues raised by corpus linguistics concerning the description of language. The article argues that it is necessary to start from correct descriptions of linguistic units and the contexts in which they occur. Corpus linguistics has joined with language teaching by sharing a recognition of the importance of a larger, schematic view of…

  20. The Routledge Applied Linguistics Reader

    Wei, Li, Ed.

    2011-01-01

    "The Routledge Applied Linguistics Reader" is an essential collection of readings for students of Applied Linguistics. Divided into five sections: Language Teaching and Learning, Second Language Acquisition, Applied Linguistics, Identity and Power and Language Use in Professional Contexts, the "Reader" takes a broad…

  1. Lexicography and Linguistic Creativity*

    rbr

    It could be argued that lexicography has little business with linguistic creativ- ...... The forms in which traditional proverbs are found can also vary greatly: many ... BoE has examples of the proverb every cloud has a silver lining but many more ...

  2. Variation and Linguistic Theory.

    Bailey, Charles-James N.

    This volume presents principles and models for describing language variation, and introduces a time-based, dynamic framework for linguistic description. The book first summarizes some of the problems of grammatical description encountered from Saussure through the present and then outlines possibilities for new descriptions of language which take…

  3. Untangling Linguistic Salience

    Boswijk, Vincent; Coler, Matt; Loerts, Hanneke; Hilton, Nanna

    2018-01-01

    The concept of linguistic salience is broadly used within sociolinguistics to account for processes as diverse as language change (Kerswill & Williams, 2002) and language acquisition (Ellis, 2016) in that salient forms are e.g. more likely to undergo change, or are often acquired earlier than other

  4. Guatemalan Linguistics Project

    Linguistic Reporter, 1974

    1974-01-01

    The general goals of the Guatemalan technical institution, the Proyecto Linguistico Francisco Marroquin, are to: create a national technical resource institution in linguistics and Mayan languages; enable Indians to influence programs for their communities; and stimulate the study of Mayan languages and their use as communication medium. (SW)

  5. Formal monkey linguistics

    Schlenker, Philippe; Chemla, Emmanuel; Schel, Anne M.; Fuller, James; Gautier, Jean Pierre; Kuhn, Jeremy; Veselinović, Dunja; Arnold, Kate; Cäsar, Cristiane; Keenan, Sumir; Lemasson, Alban; Ouattara, Karim; Ryder, Robin; Zuberbühler, Klaus

    2016-01-01

    We argue that rich data gathered in experimental primatology in the last 40 years can benefit from analytical methods used in contemporary linguistics. Focusing on the syntactic and especially semantic side, we suggest that these methods could help clarify five questions: (i) what morphology and

  6. Linguistic Corpora and Lexicography.

    Meijs, Willem

    1996-01-01

    Overviews the development of corpus linguistics, reviews the use of corpora in modern lexicography, and presents central issues in ongoing work aimed at broadening the scope of lexicographical use of corpus data. Focuses on how the field has developed in relation to the production of new monolingual English dictionaries by major British…

  7. Perspectives in Linguistics.

    Waterman, John T.

    Intended for the student of linguistics or the structural grammarian, who must develop an awareness of their intellectual heritage, the present work surveys the study of language in ancient times, the medieval and early modern periods, the nineteenth century, and the twentieth century to 1950. (This second edition includes additional material on…

  8. Gradual linguistic summaries

    Wilbik, A.M.; Kaymak, U.; Laurent, A.; Strauss, O.; Bouchon-Meunier, xx

    2014-01-01

    In this paper we propose a new type of protoform-based linguistic summary – the gradual summary. This new type of summaries aims in capturing the change over some time span. Such summaries can be useful in many domains, for instance in economics, e.g., "prices of X are getting smaller" in eldercare,

  9. Linguistics in Language Education

    Kumar, Rajesh; Yunus, Reva

    2014-01-01

    This article looks at the contribution of insights from theoretical linguistics to an understanding of language acquisition and the nature of language in terms of their potential benefit to language education. We examine the ideas of innateness and universal language faculty, as well as multilingualism and the language-society relationship. Modern…

  10. Linguistics and Literacy.

    Kindell, Gloria

    1983-01-01

    Discusses four general areas of linguistics studies that are particularly relevant to literacy issues: (1) discourse analysis, including text analysis, spoken and written language, and home and school discourse; (2) relationships between speech and writing, the distance between dialects and written norms, and developmental writing; (3)…

  11. Applied Linguistics in Europe

    de Bot, Kees

    2004-01-01

    In this contribution developments in Applied Linguistics in Europe are linked to major social changes that have taken place over the last decades. These include: The decline of the USSR and the end of the cold war; The development of the EEC and the EU and fading of borders; The economic growth of

  12. Sleepless in Seoul: Neoliberalism, English Fever, and Linguistic Insecurity among Korean Interpreters

    Cho, Jinhyun

    2015-01-01

    This article examines the socially constructed nature of significant linguistic insecurity with regard to the English language in Korean society as informed by neoliberalism. It specifically explores how linguistic insecurity leads to the pursuit of linguistic perfectionism under the popular discourse of neoliberal personhood. Participants are…

  13. Multiple Uses of Applied Linguistics Literature.

    Casanave, Christine Pearson

    2003-01-01

    Discusses ways that applied linguistics literature can be used in a multidisciplinary graduate-level English for academic purposes class. Focuses on three main uses: (1) providing students with information about issues in academic and professional writing; (2) helping them make comparisons of form and style with academic articles in their own…

  14. Stellenbosch Papers in Linguistics Plus: Editorial Policies

    ). The SPiL Plus series has two main aims. Firstly, it serves as a vehicle for the distribution of new and relatively inaccessible information in the field of modern linguistics. Secondly, it aims to stimulate critical discussion in Southern African ...

  15. Lexicography and Linguistic Creativity | Moon | Lexikos

    Abstract: Conventionally, dictionaries present information about institutionalized words, phrases, and senses of words; more creative formations and usages are generally ignored. Yet text and corpus data provide ample evidence of creativity in language, showing that it is part of ordinary linguistic behaviour and indeed ...

  16. Linguistic Prescription: Familiar Practices and New Perspectives.

    Finegan, Edward

    2003-01-01

    Reports on a question by a law student of whether a correction of "sneaked" to "snuck" suggests misinformation and misguided rigidity in the context of better information about current legal usage and a perennial tendency to linguistic prescription. Explores attitudes to current borrowings from English into Japanese and French…

  17. RESEARCH ON REMOTE SENSING GEOLOGICAL INFORMATION EXTRACTION BASED ON OBJECT ORIENTED CLASSIFICATION

    H. Gao

    2018-04-01

    Full Text Available The northern Tibet belongs to the Sub cold arid climate zone in the plateau. It is rarely visited by people. The geological working conditions are very poor. However, the stratum exposures are good and human interference is very small. Therefore, the research on the automatic classification and extraction of remote sensing geological information has typical significance and good application prospect. Based on the object-oriented classification in Northern Tibet, using the Worldview2 high-resolution remote sensing data, combined with the tectonic information and image enhancement, the lithological spectral features, shape features, spatial locations and topological relations of various geological information are excavated. By setting the threshold, based on the hierarchical classification, eight kinds of geological information were classified and extracted. Compared with the existing geological maps, the accuracy analysis shows that the overall accuracy reached 87.8561 %, indicating that the classification-oriented method is effective and feasible for this study area and provides a new idea for the automatic extraction of remote sensing geological information.

  18. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories.

    Yang, Wei; Ai, Tinghua; Lu, Wei

    2018-04-19

    Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT). First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS) traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction) by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  19. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories

    Wei Yang

    2018-04-01

    Full Text Available Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT. First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  20. YAdumper: extracting and translating large information volumes from relational databases to structured flat files.

    Fernández, José M; Valencia, Alfonso

    2004-10-12

    Downloading the information stored in relational databases into XML and other flat formats is a common task in bioinformatics. This periodical dumping of information requires considerable CPU time, disk and memory resources. YAdumper has been developed as a purpose-specific tool to deal with the integral structured information download of relational databases. YAdumper is a Java application that organizes database extraction following an XML template based on an external Document Type Declaration. Compared with other non-native alternatives, YAdumper substantially reduces memory requirements and considerably improves writing performance.

  1. The Extension of Quality Function Deployment Based on 2-Tuple Linguistic Representation Model for Product Design under Multigranularity Linguistic Environment

    Ming Li

    2012-01-01

    Full Text Available Quality function deployment (QFD is a customer-driven approach for product design and development. A QFD analysis process includes a series of subprocesses, such as determination of the importance of customer requirements (CRs, the correlation among engineering characteristics (ECs, and the relationship between CRs and ECs. Usually more than group of one decision makers are involved in the subprocesses to make the decision. In most decision making problems, they often provide their evaluation information in the linguistic form. Moreover, because of different knowledge, background, and discrimination ability, decision makers may express their linguistic preferences in multigranularity linguistic information. Therefore, an effective approach to deal with the multi-granularity linguistic information in QFD analysis process is highly needed. In this study, the QFD methodology is extended with 2-tuple linguistic representation model under multi-granularity linguistic environment. The extended QFD methodology can cope with multi-granularity linguistic evaluation information and avoid the loss of information. The applicability of the proposed approach is demonstrated with a numerical example.

  2. Linguistic Barriers and Bridges

    Thuesen, Frederik

    2016-01-01

    The influence of language on social capital in low-skill and ethnically diverse workplaces has thus far received very limited attention within the sociology of work. As the ethnically diverse workplace is an important social space for the construction of social relations bridging different social...... groups, the sociology of work needs to develop a better understanding of the way in which linguistic diversity influences the formation of social capital, i.e. resources such as the trust and reciprocity inherent in social relations in such workplaces. Drawing on theories about intergroup contact...... and intercultural communication, this article analyses interviews with 31 employees from two highly ethnically diverse Danish workplaces. The article shows how linguistic barriers such as different levels of majority language competence and their consequent misunderstandings breed mistrust and hostility, whilst...

  3. Extracting information from two-dimensional electrophoresis gels by partial least squares regression

    Jessen, Flemming; Lametsch, R.; Bendixen, E.

    2002-01-01

    of all proteins/spots in the gels. In the present study it is demonstrated how information can be extracted by multivariate data analysis. The strategy is based on partial least squares regression followed by variable selection to find proteins that individually or in combination with other proteins vary......Two-dimensional gel electrophoresis (2-DE) produces large amounts of data and extraction of relevant information from these data demands a cautious and time consuming process of spot pattern matching between gels. The classical approach of data analysis is to detect protein markers that appear...... or disappear depending on the experimental conditions. Such biomarkers are found by comparing the relative volumes of individual spots in the individual gels. Multivariate statistical analysis and modelling of 2-DE data for comparison and classification is an alternative approach utilising the combination...

  4. Multi-attribute Group Decision-Making with Incomplete Interval Linguistic Information%基于残缺语言区间信息的多属性群决策

    梁海明; 姜艳萍

    2011-01-01

    With respect to the multi-attribute group decision-making problem in which the attribute weights are linguistic variable and assessment information is incomplete interval linguistic information,a revised fuzzy interval evidential reasoning method was proposed.Firstly,the incomplete information was represented by random variables,and decision matrixes given by experts were combined into a credibility matrix according to the importance degrees of the experts.Then,the distributed assessment values of the alternatives were calculated by using the revised fuzzy interval evidential reasoning method.Further,fuzzy assessment values of all the alternatives were calculated to determine the alternative ranking result.Finally,a numerical example was given to illustrate the effectiveness of the proposed method.%针对属性权重为语言变量、评价信息为残缺语言区间信息的多属性群决策问题,提出了基于改进的模糊区间证据推理的分析方法.首先给出了残缺信息的随机变量表示方法,并根据专家在决策中的重要程度,将专家给出的决策矩阵组合成信任度矩阵,然后采用所提出的改进的模糊区间证据推理方法求得各方案的分布式评价值.计算各方案的模糊评价值,给出方案排序方法.最后给出了一个算例,证明了所提方法的有效性.

  5. From remote sensing data about information extraction for 3D geovisualization - Development of a workflow

    Tiede, D.

    2010-01-01

    With an increased availability of high (spatial) resolution remote sensing imagery since the late nineties, the need to develop operative workflows for the automated extraction, provision and communication of information from such data has grown. Monitoring requirements, aimed at the implementation of environmental or conservation targets, management of (environmental-) resources, and regional planning as well as international initiatives, especially the joint initiative of the European Commission and ESA (European Space Agency) for Global Monitoring for Environment and Security (GMES) play also a major part. This thesis addresses the development of an integrated workflow for the automated provision of information derived from remote sensing data. Considering applied data and fields of application, this work aims to design the workflow as generic as possible. Following research questions are discussed: What are the requirements of a workflow architecture that seamlessly links the individual workflow elements in a timely manner and secures accuracy of the extracted information effectively? How can the workflow retain its efficiency if mounds of data are processed? How can the workflow be improved with regards to automated object-based image analysis (OBIA)? Which recent developments could be of use? What are the limitations or which workarounds could be applied in order to generate relevant results? How can relevant information be prepared target-oriented and communicated effectively? How can the more recently developed freely available virtual globes be used for the delivery of conditioned information under consideration of the third dimension as an additional, explicit carrier of information? Based on case studies comprising different data sets and fields of application it is demonstrated how methods to extract and process information as well as to effectively communicate results can be improved and successfully combined within one workflow. It is shown that (1

  6. Addressing Risk Assessment for Patient Safety in Hospitals through Information Extraction in Medical Reports

    Proux, Denys; Segond, Frédérique; Gerbier, Solweig; Metzger, Marie Hélène

    Hospital Acquired Infections (HAI) is a real burden for doctors and risk surveillance experts. The impact on patients' health and related healthcare cost is very significant and a major concern even for rich countries. Furthermore required data to evaluate the threat is generally not available to experts and that prevents from fast reaction. However, recent advances in Computational Intelligence Techniques such as Information Extraction, Risk Patterns Detection in documents and Decision Support Systems allow now to address this problem.

  7. From Specific Information Extraction to Inferences: A Hierarchical Framework of Graph Comprehension

    2004-09-01

    The skill to interpret the information displayed in graphs is so important to have, the National Council of Teachers of Mathematics has created...guidelines to ensure that students learn these skills ( NCTM : Standards for Mathematics , 2003). These guidelines are based primarily on the extraction of...graphical perception. Human Computer Interaction, 8, 353-388. NCTM : Standards for Mathematics . (2003, 2003). Peebles, D., & Cheng, P. C.-H. (2002

  8. Extracting breathing rate information from a wearable reflectance pulse oximeter sensor.

    Johnston, W S; Mendelson, Y

    2004-01-01

    The integration of multiple vital physiological measurements could help combat medics and field commanders to better predict a soldier's health condition and enhance their ability to perform remote triage procedures. In this paper we demonstrate the feasibility of extracting accurate breathing rate information from a photoplethysmographic signal that was recorded by a reflectance pulse oximeter sensor mounted on the forehead and subsequently processed by a simple time domain filtering and frequency domain Fourier analysis.

  9. Mining of relations between proteins over biomedical scientific literature using a deep-linguistic approach.

    Rinaldi, Fabio; Schneider, Gerold; Kaljurand, Kaarel; Hess, Michael; Andronis, Christos; Konstandi, Ourania; Persidis, Andreas

    2007-02-01

    The amount of new discoveries (as published in the scientific literature) in the biomedical area is growing at an exponential rate. This growth makes it very difficult to filter the most relevant results, and thus the extraction of the core information becomes very expensive. Therefore, there is a growing interest in text processing approaches that can deliver selected information from scientific publications, which can limit the amount of human intervention normally needed to gather those results. This paper presents and evaluates an approach aimed at automating the process of extracting functional relations (e.g. interactions between genes and proteins) from scientific literature in the biomedical domain. The approach, using a novel dependency-based parser, is based on a complete syntactic analysis of the corpus. We have implemented a state-of-the-art text mining system for biomedical literature, based on a deep-linguistic, full-parsing approach. The results are validated on two different corpora: the manually annotated genomics information access (GENIA) corpus and the automatically annotated arabidopsis thaliana circadian rhythms (ATCR) corpus. We show how a deep-linguistic approach (contrary to common belief) can be used in a real world text mining application, offering high-precision relation extraction, while at the same time retaining a sufficient recall.

  10. Extraction of land cover change information from ENVISAT-ASAR data in Chengdu Plain

    Xu, Wenbo; Fan, Jinlong; Huang, Jianxi; Tian, Yichen; Zhang, Yong

    2006-10-01

    Land cover data are essential to most global change research objectives, including the assessment of current environmental conditions and the simulation of future environmental scenarios that ultimately lead to public policy development. Chinese Academy of Sciences generated a nationwide land cover database in order to carry out the quantification and spatial characterization of land use/cover changes (LUCC) in 1990s. In order to improve the reliability of the database, we will update the database anytime. But it is difficult to obtain remote sensing data to extract land cover change information in large-scale. It is hard to acquire optical remote sensing data in Chengdu plain, so the objective of this research was to evaluate multitemporal ENVISAT advanced synthetic aperture radar (ASAR) data for extracting land cover change information. Based on the fieldwork and the nationwide 1:100000 land cover database, the paper assesses several land cover changes in Chengdu plain, for example: crop to buildings, forest to buildings, and forest to bare land. The results show that ENVISAT ASAR data have great potential for the applications of extracting land cover change information.

  11. KneeTex: an ontology-driven system for information extraction from MRI reports.

    Spasić, Irena; Zhao, Bo; Jones, Christopher B; Button, Kate

    2015-01-01

    In the realm of knee pathology, magnetic resonance imaging (MRI) has the advantage of visualising all structures within the knee joint, which makes it a valuable tool for increasing diagnostic accuracy and planning surgical treatments. Therefore, clinical narratives found in MRI reports convey valuable diagnostic information. A range of studies have proven the feasibility of natural language processing for information extraction from clinical narratives. However, no study focused specifically on MRI reports in relation to knee pathology, possibly due to the complexity of knee anatomy and a wide range of conditions that may be associated with different anatomical entities. In this paper we describe KneeTex, an information extraction system that operates in this domain. As an ontology-driven information extraction system, KneeTex makes active use of an ontology to strongly guide and constrain text analysis. We used automatic term recognition to facilitate the development of a domain-specific ontology with sufficient detail and coverage for text mining applications. In combination with the ontology, high regularity of the sublanguage used in knee MRI reports allowed us to model its processing by a set of sophisticated lexico-semantic rules with minimal syntactic analysis. The main processing steps involve named entity recognition combined with coordination, enumeration, ambiguity and co-reference resolution, followed by text segmentation. Ontology-based semantic typing is then used to drive the template filling process. We adopted an existing ontology, TRAK (Taxonomy for RehAbilitation of Knee conditions), for use within KneeTex. The original TRAK ontology expanded from 1,292 concepts, 1,720 synonyms and 518 relationship instances to 1,621 concepts, 2,550 synonyms and 560 relationship instances. This provided KneeTex with a very fine-grained lexico-semantic knowledge base, which is highly attuned to the given sublanguage. Information extraction results were evaluated

  12. SAR matrices: automated extraction of information-rich SAR tables from large compound data sets.

    Wassermann, Anne Mai; Haebel, Peter; Weskamp, Nils; Bajorath, Jürgen

    2012-07-23

    We introduce the SAR matrix data structure that is designed to elucidate SAR patterns produced by groups of structurally related active compounds, which are extracted from large data sets. SAR matrices are systematically generated and sorted on the basis of SAR information content. Matrix generation is computationally efficient and enables processing of large compound sets. The matrix format is reminiscent of SAR tables, and SAR patterns revealed by different categories of matrices are easily interpretable. The structural organization underlying matrix formation is more flexible than standard R-group decomposition schemes. Hence, the resulting matrices capture SAR information in a comprehensive manner.

  13. Comparison of methods of extracting information for meta-analysis of observational studies in nutritional epidemiology

    Jong-Myon Bae

    2016-01-01

    Full Text Available OBJECTIVES: A common method for conducting a quantitative systematic review (QSR for observational studies related to nutritional epidemiology is the “highest versus lowest intake” method (HLM, in which only the information concerning the effect size (ES of the highest category of a food item is collected on the basis of its lowest category. However, in the interval collapsing method (ICM, a method suggested to enable a maximum utilization of all available information, the ES information is collected by collapsing all categories into a single category. This study aimed to compare the ES and summary effect size (SES between the HLM and ICM. METHODS: A QSR for evaluating the citrus fruit intake and risk of pancreatic cancer and calculating the SES by using the HLM was selected. The ES and SES were estimated by performing a meta-analysis using the fixed-effect model. The directionality and statistical significance of the ES and SES were used as criteria for determining the concordance between the HLM and ICM outcomes. RESULTS: No significant differences were observed in the directionality of SES extracted by using the HLM or ICM. The application of the ICM, which uses a broader information base, yielded more-consistent ES and SES, and narrower confidence intervals than the HLM. CONCLUSIONS: The ICM is advantageous over the HLM owing to its higher statistical accuracy in extracting information for QSR on nutritional epidemiology. The application of the ICM should hence be recommended for future studies.

  14. Sentence processing and grammaticality in functional linguistics

    Poulsen, Mads

    finding from research on sentence processing that sentences are processed incrementally. Empirical methods for establishing grammaticality status are discussed and applied in relation to non-WH extraction phenomena in Danish. In Chapter 2, I discuss the use of the notions of grammaticality......The dissertation presents a functional linguistic model of grammaticality and investigates methods for applying this notion in empirical work. The use of the notion of grammaticality in generative grammar has been criticized by functionalists (Harder, 1996; Lakoff & Johnson, 1999), but attempts...... grammaticality. It is concluded that the intuitions of linguists should in principle be considered hypotheses of grammaticality, and that such hypotheses need to be tested with independent data. Such data can for example take the form of corpus data or acceptability judgment experiments. It is furthermore argued...

  15. A Novel Integrated Approach for Green Supplier Selection with Interval-Valued Intuitionistic Uncertain Linguistic Information: A Case Study in the Agri-Food Industry

    Hua Shi

    2018-03-01

    Full Text Available With strengthening global consciousness of environmental protection, green supply chain management plays an increasingly important role in modern enterprise production operation management. A critical means to implement green supply chain management is incorporating environmental requirements into the supplier selection practices. In this paper, we put forward a novel integrated approach by using interval-valued intuitionistic uncertain linguistic sets (IVIULSs and grey relational analysis (GRA-technique for order preference by similarity to ideal solution (TOPSIS method for the evaluation and selection of green suppliers. First, various qualitative assessments of alternatives provided by decision makers are described by the IVIULSs. Then, the GRA-TOPSIS method is extended and employed to prioritize the alternative suppliers. The proposed model can handle the uncertainty and fuzziness of decision makers’ subjective evaluations more easily and get a more realistic and accurate ranking of green suppliers. Finally, an illustrative example in the agri-food industry is presented to verify the proposed green supplier selection model and demonstrate its practicality and effectiveness.

  16. Feature extraction and learning using context cue and Rényi entropy based mutual information

    Pan, Hong; Olsen, Søren Ingvor; Zhu, Yaping

    2015-01-01

    information. In particular, for feature extraction, we develop a new set of kernel descriptors−Context Kernel Descriptors (CKD), which enhance the original KDES by embedding the spatial context into the descriptors. Context cues contained in the context kernel enforce some degree of spatial consistency, thus...... improving the robustness of CKD. For feature learning and reduction, we propose a novel codebook learning method, based on a Rényi quadratic entropy based mutual information measure called Cauchy-Schwarz Quadratic Mutual Information (CSQMI), to learn a compact and discriminative CKD codebook. Projecting...... as the information about the underlying labels of the CKD using CSQMI. Thus the resulting codebook and reduced CKD are discriminative. We verify the effectiveness of our method on several public image benchmark datasets such as YaleB, Caltech-101 and CIFAR-10, as well as a challenging chicken feet dataset of our own...

  17. Method of extracting significant trouble information of nuclear power plants using probabilistic analysis technique

    Shimada, Yoshio; Miyazaki, Takamasa

    2005-01-01

    In order to analyze and evaluate large amounts of trouble information of overseas nuclear power plants, it is necessary to select information that is significant in terms of both safety and reliability. In this research, a method of efficiently and simply classifying degrees of importance of components in terms of safety and reliability while paying attention to root-cause components appearing in the information was developed. Regarding safety, the reactor core damage frequency (CDF), which is used in the probabilistic analysis of a reactor, was used. Regarding reliability, the automatic plant trip probability (APTP), which is used in the probabilistic analysis of automatic reactor trips, was used. These two aspects were reflected in the development of criteria for classifying degrees of importance of components. By applying these criteria, a simple method of extracting significant trouble information of overseas nuclear power plants was developed. (author)

  18. Automated concept-level information extraction to reduce the need for custom software and rules development.

    D'Avolio, Leonard W; Nguyen, Thien M; Goryachev, Sergey; Fiore, Louis D

    2011-01-01

    Despite at least 40 years of promising empirical performance, very few clinical natural language processing (NLP) or information extraction systems currently contribute to medical science or care. The authors address this gap by reducing the need for custom software and rules development with a graphical user interface-driven, highly generalizable approach to concept-level retrieval. A 'learn by example' approach combines features derived from open-source NLP pipelines with open-source machine learning classifiers to automatically and iteratively evaluate top-performing configurations. The Fourth i2b2/VA Shared Task Challenge's concept extraction task provided the data sets and metrics used to evaluate performance. Top F-measure scores for each of the tasks were medical problems (0.83), treatments (0.82), and tests (0.83). Recall lagged precision in all experiments. Precision was near or above 0.90 in all tasks. Discussion With no customization for the tasks and less than 5 min of end-user time to configure and launch each experiment, the average F-measure was 0.83, one point behind the mean F-measure of the 22 entrants in the competition. Strong precision scores indicate the potential of applying the approach for more specific clinical information extraction tasks. There was not one best configuration, supporting an iterative approach to model creation. Acceptable levels of performance can be achieved using fully automated and generalizable approaches to concept-level information extraction. The described implementation and related documentation is available for download.

  19. Linguistics and the digital humanities

    Jensen, Kim Ebensgaard

    2014-01-01

    Corpus linguistics has been closely intertwined with digital technology since the introduction of university computer mainframes in the 1960s. Making use of both digitized data in the form of the language corpus and computational methods of analysis involving concordancers and statistics software......, corpus linguistics arguably has a place in the digital humanities. Still, it remains obscure and figures only sporadically in the literature on the digital humanities. This article provides an overview of the main principles of corpus linguistics and the role of computer technology in relation to data...... and method and also offers a bird's-eye view of the history of corpus linguistics with a focus on its intimate relationship with digital technology and how digital technology has impacted the very core of corpus linguistics and shaped the identity of the corpus linguist. Ultimately, the article is oriented...

  20. The Feature Extraction Based on Texture Image Information for Emotion Sensing in Speech

    Kun-Ching Wang

    2014-09-01

    Full Text Available In this paper, we present a novel texture image feature for Emotion Sensing in Speech (ESS. This idea is based on the fact that the texture images carry emotion-related information. The feature extraction is derived from time-frequency representation of spectrogram images. First, we transform the spectrogram as a recognizable image. Next, we use a cubic curve to enhance the image contrast. Then, the texture image information (TII derived from the spectrogram image can be extracted by using Laws’ masks to characterize emotional state. In order to evaluate the effectiveness of the proposed emotion recognition in different languages, we use two open emotional databases including the Berlin Emotional Speech Database (EMO-DB and eNTERFACE corpus and one self-recorded database (KHUSC-EmoDB, to evaluate the performance cross-corpora. The results of the proposed ESS system are presented using support vector machine (SVM as a classifier. Experimental results show that the proposed TII-based feature extraction inspired by visual perception can provide significant classification for ESS systems. The two-dimensional (2-D TII feature can provide the discrimination between different emotions in visual expressions except for the conveyance pitch and formant tracks. In addition, the de-noising in 2-D images can be more easily completed than de-noising in 1-D speech.

  1. An Accurate Integral Method for Vibration Signal Based on Feature Information Extraction

    Yong Zhu

    2015-01-01

    Full Text Available After summarizing the advantages and disadvantages of current integral methods, a novel vibration signal integral method based on feature information extraction was proposed. This method took full advantage of the self-adaptive filter characteristic and waveform correction feature of ensemble empirical mode decomposition in dealing with nonlinear and nonstationary signals. This research merged the superiorities of kurtosis, mean square error, energy, and singular value decomposition on signal feature extraction. The values of the four indexes aforementioned were combined into a feature vector. Then, the connotative characteristic components in vibration signal were accurately extracted by Euclidean distance search, and the desired integral signals were precisely reconstructed. With this method, the interference problem of invalid signal such as trend item and noise which plague traditional methods is commendably solved. The great cumulative error from the traditional time-domain integral is effectively overcome. Moreover, the large low-frequency error from the traditional frequency-domain integral is successfully avoided. Comparing with the traditional integral methods, this method is outstanding at removing noise and retaining useful feature information and shows higher accuracy and superiority.

  2. A cascade of classifiers for extracting medication information from discharge summaries

    Halgrim Scott

    2011-07-01

    Full Text Available Abstract Background Extracting medication information from clinical records has many potential applications, and recently published research, systems, and competitions reflect an interest therein. Much of the early extraction work involved rules and lexicons, but more recently machine learning has been applied to the task. Methods We present a hybrid system consisting of two parts. The first part, field detection, uses a cascade of statistical classifiers to identify medication-related named entities. The second part uses simple heuristics to link those entities into medication events. Results The system achieved performance that is comparable to other approaches to the same task. This performance is further improved by adding features that reference external medication name lists. Conclusions This study demonstrates that our hybrid approach outperforms purely statistical or rule-based systems. The study also shows that a cascade of classifiers works better than a single classifier in extracting medication information. The system is available as is upon request from the first author.

  3. Three-dimensional information extraction from GaoFen-1 satellite images for landslide monitoring

    Wang, Shixin; Yang, Baolin; Zhou, Yi; Wang, Futao; Zhang, Rui; Zhao, Qing

    2018-05-01

    To more efficiently use GaoFen-1 (GF-1) satellite images for landslide emergency monitoring, a Digital Surface Model (DSM) can be generated from GF-1 across-track stereo image pairs to build a terrain dataset. This study proposes a landslide 3D information extraction method based on the terrain changes of slope objects. The slope objects are mergences of segmented image objects which have similar aspects; and the terrain changes are calculated from the post-disaster Digital Elevation Model (DEM) from GF-1 and the pre-disaster DEM from GDEM V2. A high mountain landslide that occurred in Wenchuan County, Sichuan Province is used to conduct a 3D information extraction test. The extracted total area of the landslide is 22.58 ha; the displaced earth volume is 652,100 m3; and the average sliding direction is 263.83°. The accuracies of them are 0.89, 0.87 and 0.95, respectively. Thus, the proposed method expands the application of GF-1 satellite images to the field of landslide emergency monitoring.

  4. DYNAMIC FEATURE SELECTION FOR WEB USER IDENTIFICATION ON LINGUISTIC AND STYLISTIC FEATURES OF ONLINE TEXTS

    A. A. Vorobeva

    2017-01-01

    Full Text Available The paper deals with identification and authentication of web users participating in the Internet information processes (based on features of online texts.In digital forensics web user identification based on various linguistic features can be used to discover identity of individuals, criminals or terrorists using the Internet to commit cybercrimes. Internet could be used as a tool in different types of cybercrimes (fraud and identity theft, harassment and anonymous threats, terrorist or extremist statements, distribution of illegal content and information warfare. Linguistic identification of web users is a kind of biometric identification, it can be used to narrow down the suspects, identify a criminal and prosecute him. Feature set includes various linguistic and stylistic features extracted from online texts. We propose dynamic feature selection for each web user identification task. Selection is based on calculating Manhattan distance to k-nearest neighbors (Relief-f algorithm. This approach improves the identification accuracy and minimizes the number of features. Experiments were carried out on several datasets with different level of class imbalance. Experiment results showed that features relevance varies in different set of web users (probable authors of some text; features selection for each set of web users improves identification accuracy by 4% at the average that is approximately 1% higher than with the use of static set of features. The proposed approach is most effective for a small number of training samples (messages per user.

  5. Learnability and linguistic performance

    Drozd, Kenneth

    2004-01-01

    of the human biological endowment for language in the form of a UNIVERSAL GRAMMAR (UG) (Chomsky, 1965). With respect to experimental design, C&T have strongly maintained that even young children know UG constraints but perform poorly in some experiments-due to the extralinguistic demands associated...... with experimental tasks, particularly those involved in presupposition accommodation and complex response planning. C&T specifically design their experiments to reduce the impact of extralinguistic demands on children's linguistic performance while at the same time providing felicitous environments for adultlike...... performance....

  6. Formal monkey linguistics

    Schlenker, Philippe; Chemla, Emmanuel; Schel, Anne M.; Fuller, James; Gautier, Jean-Pierre; Kuhn, Jeremy; Veselinović, Dunja; Arnold, Kate; Cäsar, Cristiane; Keenan, Sumir; Lemasson, Alban; Ouattara, Karim; Ryder, Robin; Zuberbühler, Klaus

    2016-01-01

    Zuberbühler: The research leading to these results received funding from the European Research Council under ERC grant ‘Prilang 283871’ and also from the Swiss National Science Foundation under grant ‘FN 310030_143359/1’. We argue that rich data gathered in experimental primatology in the last 40 years can benefit from analytical methods used in contemporary linguistics. Focusing on the syntactic and especially semantic side, we suggest that these methods could help clarify five questions:...

  7. Quantifying linguistic coordination

    Fusaroli, Riccardo; Tylén, Kristian

    task (Bahrami et al 2010, Fusaroli et al. 2012) we extend to linguistic coordination dynamical measures of recurrence employed in the analysis of sensorimotor coordination (such as heart-rate (Konvalinka et al 2011), postural sway (Shockley 2005) and eye-movements (Dale, Richardson and Kirkham 2012......). We employ nominal recurrence analysis (Orsucci et al 2005, Dale et al 2011) on the decision-making conversations between the participants. We report strong correlations between various indexes of recurrence and collective performance. We argue this method allows us to quantify the qualities...

  8. Linguistic summaries of categorical time series for septic shock patient data

    Almeida, R.J.; Lesot, M.-J.; Bouchon-Meunier, B.; Kaymak, U.; Moyse, G.

    2013-01-01

    Linguistic summarization is a data mining and knowledge discovery approach to extract patterns and sum up large volume of data into simple sentences. There is a large research in generating linguistic summaries which can be used to better understand and communicate about patterns, evolution and long

  9. THE EXTRACTION OF INDOOR BUILDING INFORMATION FROM BIM TO OGC INDOORGML

    T.-A. Teo

    2017-07-01

    Full Text Available Indoor Spatial Data Infrastructure (indoor-SDI is an important SDI for geosptial analysis and location-based services. Building Information Model (BIM has high degree of details in geometric and semantic information for building. This study proposed direct conversion schemes to extract indoor building information from BIM to OGC IndoorGML. The major steps of the research include (1 topological conversion from building model into indoor network model; and (2 generation of IndoorGML. The topological conversion is a major process of generating and mapping nodes and edges from IFC to indoorGML. Node represents every space (e.g. IfcSpace and objects (e.g. IfcDoor in the building while edge shows the relationships between nodes. According to the definition of IndoorGML, the topological model in the dual space is also represented as a set of nodes and edges. These definitions of IndoorGML are the same as in the indoor network. Therefore, we can extract the necessary data in the indoor network and easily convert them into IndoorGML based on IndoorGML Schema. The experiment utilized a real BIM model to examine the proposed method. The experimental results indicated that the 3D indoor model (i.e. IndoorGML model can be automatically imported from IFC model by the proposed procedure. In addition, the geometric and attribute of building elements are completely and correctly converted from BIM to indoor-SDI.

  10. Methods from Information Extraction from LIDAR Intensity Data and Multispectral LIDAR Technology

    Scaioni, M.; Höfle, B.; Baungarten Kersting, A. P.; Barazzetti, L.; Previtali, M.; Wujanz, D.

    2018-04-01

    LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on `Information Extraction from LiDAR Intensity Data' has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  11. The effect of informed consent on stress levels associated with extraction of impacted mandibular third molars.

    Casap, Nardy; Alterman, Michael; Sharon, Guy; Samuni, Yuval

    2008-05-01

    To evaluate the effect of informed consent on stress levels associated with removal of impacted mandibular third molars. A total of 60 patients scheduled for extraction of impacted mandibular third molars participated in this study. The patients were unaware of the study's objectives. Data from 20 patients established the baseline levels of electrodermal activity (EDA). The remaining 40 patients were randomly assigned into 2 equal groups receiving either a detailed document of informed consent, disclosing the possible risks involved with the surgery, or a simplified version. Pulse, blood pressure, and EDA were monitored before, during, and after completion of the consent document. Changes in EDA, but not in blood pressure, were measured on completion of either version of the consent document. A greater increase in EDA was associated with the detailed version of the consent document (P = .004). A similar concomitant increase (although nonsignificant) in pulse values was monitored on completion of both versions. Completion of overdisclosed document of informed consent is associated with changes in physiological parameters. The results suggest that overdetailed listing and disclosure before extraction of impacted mandibular third molars can increase patient stress.

  12. METHODS FROM INFORMATION EXTRACTION FROM LIDAR INTENSITY DATA AND MULTISPECTRAL LIDAR TECHNOLOGY

    M. Scaioni

    2018-04-01

    Full Text Available LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on ‘Information Extraction from LiDAR Intensity Data’ has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  13. About increasing informativity of diagnostic system of asynchronous electric motor by extracting additional information from values of consumed current parameter

    Zhukovskiy, Y.; Korolev, N.; Koteleva, N.

    2018-05-01

    This article is devoted to expanding the possibilities of assessing the technical state of the current consumption of asynchronous electric drives, as well as increasing the information capacity of diagnostic methods, in conditions of limited access to equipment and incompleteness of information. The method of spectral analysis of the electric drive current can be supplemented by an analysis of the components of the current of the Park's vector. The research of the hodograph evolution in the moment of appearance and development of defects was carried out using the example of current asymmetry in the phases of an induction motor. The result of the study is the new diagnostic parameters of the asynchronous electric drive. During the research, it was proved that the proposed diagnostic parameters allow determining the type and level of the defect. At the same time, there is no need to stop the equipment and taky it out of service for repair. Modern digital control and monitoring systems can use the proposed parameters based on the stator current of an electrical machine to improve the accuracy and reliability of obtaining diagnostic patterns and predicting their changes in order to improve the equipment maintenance systems. This approach can also be used in systems and objects where there are significant parasitic vibrations and unsteady loads. The extraction of useful information can be carried out in electric drive systems in the structure of which there is a power electric converter.

  14. Multi-Paradigm and Multi-Lingual Information Extraction as Support for Medical Web Labelling Authorities

    Martin Labsky

    2010-10-01

    Full Text Available Until recently, quality labelling of medical web content has been a pre-dominantly manual activity. However, the advances in automated text processing opened the way to computerised support of this activity. The core enabling technology is information extraction (IE. However, the heterogeneity of websites offering medical content imposes particular requirements on the IE techniques to be applied. In the paper we discuss these requirements and describe a multi-paradigm approach to IE addressing them. Experiments on multi-lingual data are reported. The research has been carried out within the EU MedIEQ project.

  15. Scholarly Information Extraction Is Going to Make a Quantum Leap with PubMed Central (PMC).

    Matthies, Franz; Hahn, Udo

    2017-01-01

    With the increasing availability of complete full texts (journal articles), rather than their surrogates (titles, abstracts), as resources for text analytics, entirely new opportunities arise for information extraction and text mining from scholarly publications. Yet, we gathered evidence that a range of problems are encountered for full-text processing when biomedical text analytics simply reuse existing NLP pipelines which were developed on the basis of abstracts (rather than full texts). We conducted experiments with four different relation extraction engines all of which were top performers in previous BioNLP Event Extraction Challenges. We found that abstract-trained engines loose up to 6.6% F-score points when run on full-text data. Hence, the reuse of existing abstract-based NLP software in a full-text scenario is considered harmful because of heavy performance losses. Given the current lack of annotated full-text resources to train on, our study quantifies the price paid for this short cut.

  16. Linguistics and the Literary Text.

    Ferrar, Madeleine

    1984-01-01

    Discusses the opposing viewpoints of the two most influential linguists of this century--Saussure and Chomsky--suggesting that while both are interested in form as opposed to substance, Saussure sees linguistics as a branch of semiotics and Chomsky sees it as part of cognitive psychology. Evaluates the relevance of these two viewpoints to the…

  17. New Conceptualizations of Linguistic Giftedness

    Biedron, Adriana; Pawlak, Miroslaw

    2016-01-01

    This state-of-the art paper focuses on the issue of linguistic giftedness, somewhat neglected in the second language acquisition (SLA) literature, attempting to reconceptualize, expand and update this concept in response to latest developments in the fields of psychology, linguistics and neurology. It first discusses contemporary perspectives on…

  18. Stellenbosch Papers in Linguistics: Contact

    Mailing Address. Editors SPiL. Department of General Linguistics University of Stellenbosch Private Bag X1 Matieland, 7602. Stellenbosch South Africa. Principal Contact. Dr Kate Huddlestone Journal Manager Department of General Linguistics. University of Stellenbosch. Private Bag X1. Matieland, 7602. Stellenbosch.

  19. Linguistic Theory and Actual Language.

    Segerdahl, Par

    1995-01-01

    Examines Noam Chomsky's (1957) discussion of "grammaticalness" and the role of linguistics in the "correct" way of speaking and writing. It is argued that the concern of linguistics with the tools of grammar has resulted in confusion, with the tools becoming mixed up with the actual language, thereby becoming the central…

  20. Linguistics and the TEFL Teacher.

    Fraser, Bruce

    This paper asserts the "unquestionable" relevance of linguistic insights in the training of and subsequent use by teachers of English as a foreign language. Although the author agrees with Chomsky's view that linguistics has nothing to offer the teacher in the form of specific proposals for language teaching methodology, he argues that linguistics…

  1. Machine Learning and Applied Linguistics

    Vajjala, Sowmya

    2018-01-01

    This entry introduces the topic of machine learning and provides an overview of its relevance for applied linguistics and language learning. The discussion will focus on giving an introduction to the methods and applications of machine learning in applied linguistics, and will provide references for further study.

  2. Conversation Analysis and Applied Linguistics.

    Schegloff, Emanuel A.; Koshik, Irene; Jacoby, Sally; Olsher, David

    2002-01-01

    Offers biographical guidance on several major areas of conversation-analytic work--turn-taking, repair, and word selection--and indicates past or potential points of contact with applied linguistics. Also discusses areas of applied linguistic work. (Author/VWL)

  3. Writing, Literacy, and Applied Linguistics.

    Leki, Ilona

    2000-01-01

    Discusses writing and literacy in the domain of applied linguistics. Focus is on needs analysis for literacy acquisition; second language learner identity; longitudinal studies as extensions of identity work; and applied linguistics contributions to second language literacy research. (Author/VWL)

  4. Literacy in Somali: Linguistic Consequences.

    Biber, Douglas; Hared, Mohamed

    1991-01-01

    Linguistic consequences of literacy in Somalia are examined in a review of the literature and through a study of five dimensions of variation among Somali registers and the expansion of linguistic variation in Somali resulting from the introduction of written registers. (36 references) (LB)

  5. Ontological problems of contemporary linguistics

    А В Бондаренко

    2009-03-01

    Full Text Available The article studies linguistic ontology problems such as evolution of essential-existential views of language, interrelation within Being-Language-Man triad, linguistics gnosiological principles, language essence localization, and «expression» as language metalinguistic unit as well as architectonics of language personality et alia.

  6. Concise Lexicon for Sign Linguistics

    dr. Jan Nijen Twilhaar; Dr. Beppie van den Bogaerde

    2016-01-01

    This extensive, well-researched and clearly formatted lexicon of a wide variety of linguistic terms is a long overdue. It is an extremely welcome addition to the bookshelves of sign language teachers, interpreters, linguists, learners and other sign language users, and of course of the Deaf

  7. Conceptual and Pragmatic Motivation as an Explanatory Concept in Linguistics

    Klaus-Uwe Panther

    2008-01-01

    The concept of motivation has been rejected by many formalist linguists as irrelevant, or at least, negligible. It isargued in this article that motivation deserves a place of honor in linguistic theorizing. After an introduction to variousperception verbs in English and Spanish are analyzed. These constructions provide evidence for the hypothesis thatlexicogrammatieal form is often motivated conceptually. Consequently, contrary to the dogma of autonomous syntax,semantic and pragmatic information has to be built into the formulation of lexicogrammatical regularities.

  8. Accurate facade feature extraction method for buildings from three-dimensional point cloud data considering structural information

    Wang, Yongzhi; Ma, Yuqing; Zhu, A.-xing; Zhao, Hui; Liao, Lixia

    2018-05-01

    Facade features represent segmentations of building surfaces and can serve as a building framework. Extracting facade features from three-dimensional (3D) point cloud data (3D PCD) is an efficient method for 3D building modeling. By combining the advantages of 3D PCD and two-dimensional optical images, this study describes the creation of a highly accurate building facade feature extraction method from 3D PCD with a focus on structural information. The new extraction method involves three major steps: image feature extraction, exploration of the mapping method between the image features and 3D PCD, and optimization of the initial 3D PCD facade features considering structural information. Results show that the new method can extract the 3D PCD facade features of buildings more accurately and continuously. The new method is validated using a case study. In addition, the effectiveness of the new method is demonstrated by comparing it with the range image-extraction method and the optical image-extraction method in the absence of structural information. The 3D PCD facade features extracted by the new method can be applied in many fields, such as 3D building modeling and building information modeling.

  9. Linguistic dating of biblical texts

    Young, Ian; Rezetko, Robert; Ehrensvärd, Martin Gustaf

    Since the beginning of critical scholarship biblical texts have been dated using linguistic evidence. In recent years this has become a controversial topic, especially with the publication of Ian Young (ed.), Biblical Hebrew: Studies in Chronology and Typology (2003). However, until now there has...... been no introduction and comprehensive study of the field. Volume 1 introduces the field of linguistic dating of biblical texts, particularly to intermediate and advanced students of biblical Hebrew who have a reasonable background in the language, having completed at least an introductory course...... in this volume are: What is it that makes Archaic Biblical Hebrew archaic , Early Biblical Hebrew early , and Late Biblical Hebrew late ? Does linguistic typology, i.e. different linguistic characteristics, convert easily and neatly into linguistic chronology, i.e. different historical origins? A large amount...

  10. The linguistic repudiation of Wundt.

    Nerlich, B; Clarke, D D

    1998-08-01

    Wilhelm Wundt's influence on the development of linguistics and psychology was pervasive. The foundations for this web of influence on the sciences of mind and language were laid down in Wundt's own research program, which was quite different from other attempts at founding a new psychology, as it was deeply rooted in German philosophy. This resulted in certain gaps in Wundt's conception of mind and language. These gaps provoked a double repudiation of Wundt's theories, by linguists and psychologists. The psychological repudiation has been studied by historians of psychology, and the linguistic repudiation has been studied by historians of linguistics. The intent of this article is to bring the linguistic repudiation to the attention of historians of psychology, especially the one outlined by two important figures in the history of psychology: Karl Buhler and George Mead.

  11. Developing an Approach to Prioritize River Restoration using Data Extracted from Flood Risk Information System Databases.

    Vimal, S.; Tarboton, D. G.; Band, L. E.; Duncan, J. M.; Lovette, J. P.; Corzo, G.; Miles, B.

    2015-12-01

    Prioritizing river restoration requires information on river geometry. In many states in the US detailed river geometry has been collected for floodplain mapping and is available in Flood Risk Information Systems (FRIS). In particular, North Carolina has, for its 100 Counties, developed a database of numerous HEC-RAS models which are available through its Flood Risk Information System (FRIS). These models that include over 260 variables were developed and updated by numerous contractors. They contain detailed surveyed or LiDAR derived cross-sections and modeled flood extents for different extreme event return periods. In this work, over 4700 HEC-RAS models' data was integrated and upscaled to utilize detailed cross-section information and 100-year modelled flood extent information to enable river restoration prioritization for the entire state of North Carolina. We developed procedures to extract geomorphic properties such as entrenchment ratio, incision ratio, etc. from these models. Entrenchment ratio quantifies the vertical containment of rivers and thereby their vulnerability to flooding and incision ratio quantifies the depth per unit width. A map of entrenchment ratio for the whole state was derived by linking these model results to a geodatabase. A ranking of highly entrenched counties enabling prioritization for flood allowance and mitigation was obtained. The results were shared through HydroShare and web maps developed for their visualization using Google Maps Engine API.

  12. Extracting Low-Frequency Information from Time Attenuation in Elastic Waveform Inversion

    Guo, Xuebao; Liu, Hong; Shi, Ying; Wang, Weihong

    2017-03-01

    Low-frequency information is crucial for recovering background velocity, but the lack of low-frequency information in field data makes inversion impractical without accurate initial models. Laplace-Fourier domain waveform inversion can recover a smooth model from real data without low-frequency information, which can be used for subsequent inversion as an ideal starting model. In general, it also starts with low frequencies and includes higher frequencies at later inversion stages, while the difference is that its ultralow frequency information comes from the Laplace-Fourier domain. Meanwhile, a direct implementation of the Laplace-transformed wavefield using frequency domain inversion is also very convenient. However, because broad frequency bands are often used in the pure time domain waveform inversion, it is difficult to extract the wavefields dominated by low frequencies in this case. In this paper, low-frequency components are constructed by introducing time attenuation into the recorded residuals, and the rest of the method is identical to the traditional time domain inversion. Time windowing and frequency filtering are also applied to mitigate the ambiguity of the inverse problem. Therefore, we can start at low frequencies and to move to higher frequencies. The experiment shows that the proposed method can achieve a good inversion result in the presence of a linear initial model and records without low-frequency information.

  13. Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images

    Koji Iwano

    2007-03-01

    Full Text Available This paper proposes an audio-visual speech recognition method using lip information extracted from side-face images as an attempt to increase noise robustness in mobile environments. Our proposed method assumes that lip images can be captured using a small camera installed in a handset. Two different kinds of lip features, lip-contour geometric features and lip-motion velocity features, are used individually or jointly, in combination with audio features. Phoneme HMMs modeling the audio and visual features are built based on the multistream HMM technique. Experiments conducted using Japanese connected digit speech contaminated with white noise in various SNR conditions show effectiveness of the proposed method. Recognition accuracy is improved by using the visual information in all SNR conditions. These visual features were confirmed to be effective even when the audio HMM was adapted to noise by the MLLR method.

  14. Approaching the largest ‘API’: extracting information from the Internet with Python

    Jonathan E. Germann

    2018-02-01

    Full Text Available This article explores the need for libraries to algorithmically access and manipulate the world’s largest API: the Internet. The billions of pages on the ‘Internet API’ (HTTP, HTML, CSS, XPath, DOM, etc. are easily accessible and manipulable. Libraries can assist in creating meaning through the datafication of information on the world wide web. Because most information is created for human consumption, some programming is required for automated extraction. Python is an easy-to-learn programming language with extensive packages and community support for web page automation. Four packages (Urllib, Selenium, BeautifulSoup, Scrapy in Python can automate almost any web page for all sized projects. An example warrant data project is explained to illustrate how well Python packages can manipulate web pages to create meaning through assembling custom datasets.

  15. DEVELOPMENT OF AUTOMATIC EXTRACTION METHOD FOR ROAD UPDATE INFORMATION BASED ON PUBLIC WORK ORDER OUTLOOK

    Sekimoto, Yoshihide; Nakajo, Satoru; Minami, Yoshitaka; Yamaguchi, Syohei; Yamada, Harutoshi; Fuse, Takashi

    Recently, disclosure of statistic data, representing financial effects or burden for public work, through each web site of national or local government, enables us to discuss macroscopic financial trends. However, it is still difficult to grasp a basic property nationwide how each spot was changed by public work. In this research, our research purpose is to collect road update information reasonably which various road managers provide, in order to realize efficient updating of various maps such as car navigation maps. In particular, we develop the system extracting public work concerned and registering summary including position information to database automatically from public work order outlook, released by each local government, combinating some web mining technologies. Finally, we collect and register several tens of thousands from web site all over Japan, and confirm the feasibility of our method.

  16. Hemispheric lateralization of linguistic prosody recognition in comparison to speech and speaker recognition.

    Kreitewolf, Jens; Friederici, Angela D; von Kriegstein, Katharina

    2014-11-15

    Hemispheric specialization for linguistic prosody is a controversial issue. While it is commonly assumed that linguistic prosody and emotional prosody are preferentially processed in the right hemisphere, neuropsychological work directly comparing processes of linguistic prosody and emotional prosody suggests a predominant role of the left hemisphere for linguistic prosody processing. Here, we used two functional magnetic resonance imaging (fMRI) experiments to clarify the role of left and right hemispheres in the neural processing of linguistic prosody. In the first experiment, we sought to confirm previous findings showing that linguistic prosody processing compared to other speech-related processes predominantly involves the right hemisphere. Unlike previous studies, we controlled for stimulus influences by employing a prosody and speech task using the same speech material. The second experiment was designed to investigate whether a left-hemispheric involvement in linguistic prosody processing is specific to contrasts between linguistic prosody and emotional prosody or whether it also occurs when linguistic prosody is contrasted against other non-linguistic processes (i.e., speaker recognition). Prosody and speaker tasks were performed on the same stimulus material. In both experiments, linguistic prosody processing was associated with activity in temporal, frontal, parietal and cerebellar regions. Activation in temporo-frontal regions showed differential lateralization depending on whether the control task required recognition of speech or speaker: recognition of linguistic prosody predominantly involved right temporo-frontal areas when it was contrasted against speech recognition; when contrasted against speaker recognition, recognition of linguistic prosody predominantly involved left temporo-frontal areas. The results show that linguistic prosody processing involves functions of both hemispheres and suggest that recognition of linguistic prosody is based on

  17. Geopositioning with a quadcopter: Extracted feature locations and predicted accuracy without a priori sensor attitude information

    Dolloff, John; Hottel, Bryant; Edwards, David; Theiss, Henry; Braun, Aaron

    2017-05-01

    This paper presents an overview of the Full Motion Video-Geopositioning Test Bed (FMV-GTB) developed to investigate algorithm performance and issues related to the registration of motion imagery and subsequent extraction of feature locations along with predicted accuracy. A case study is included corresponding to a video taken from a quadcopter. Registration of the corresponding video frames is performed without the benefit of a priori sensor attitude (pointing) information. In particular, tie points are automatically measured between adjacent frames using standard optical flow matching techniques from computer vision, an a priori estimate of sensor attitude is then computed based on supplied GPS sensor positions contained in the video metadata and a photogrammetric/search-based structure from motion algorithm, and then a Weighted Least Squares adjustment of all a priori metadata across the frames is performed. Extraction of absolute 3D feature locations, including their predicted accuracy based on the principles of rigorous error propagation, is then performed using a subset of the registered frames. Results are compared to known locations (check points) over a test site. Throughout this entire process, no external control information (e.g. surveyed points) is used other than for evaluation of solution errors and corresponding accuracy.

  18. Inexperienced clinicians can extract pathoanatomic information from MRI narrative reports with high reproducability for use in research/quality assurance

    Kent, Peter; Briggs, Andrew M; Albert, Hanne Birgit

    2011-01-01

    Background Although reproducibility in reading MRI images amongst radiologists and clinicians has been studied previously, no studies have examined the reproducibility of inexperienced clinicians in extracting pathoanatomic information from magnetic resonance imaging (MRI) narrative reports and t...

  19. [Extraction of buildings three-dimensional information from high-resolution satellite imagery based on Barista software].

    Zhang, Pei-feng; Hu, Yuan-man; He, Hong-shi

    2010-05-01

    The demand for accurate and up-to-date spatial information of urban buildings is becoming more and more important for urban planning, environmental protection, and other vocations. Today's commercial high-resolution satellite imagery offers the potential to extract the three-dimensional information of urban buildings. This paper extracted the three-dimensional information of urban buildings from QuickBird imagery, and validated the precision of the extraction based on Barista software. It was shown that the extraction of three-dimensional information of the buildings from high-resolution satellite imagery based on Barista software had the advantages of low professional level demand, powerful universality, simple operation, and high precision. One pixel level of point positioning and height determination accuracy could be achieved if the digital elevation model (DEM) and sensor orientation model had higher precision and the off-Nadir View Angle was relatively perfect.

  20. Overview of image processing tools to extract physical information from JET videos

    Craciunescu, T.; Murari, A.; Gelfusa, M.; Tiseanu, I.; Zoita, V.; EFDA Contributors, JET

    2014-11-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  1. Overview of image processing tools to extract physical information from JET videos

    Craciunescu, T; Tiseanu, I; Zoita, V; Murari, A; Gelfusa, M

    2014-01-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  2. Extraction and Analysis of Information Related to Research & Development Declared Under an Additional Protocol

    Idinger, J.; Labella, R.; Rialhe, A.; Teller, N.

    2015-01-01

    The additional protocol (AP) provides important tools to strengthen and improve the effectiveness and efficiency of the safeguards system. Safeguards are designed to verify that States comply with their international commitments not to use nuclear material or to engage in nuclear-related activities for the purpose of developing nuclear weapons or other nuclear explosive devices. Under an AP based on INFCIRC/540, a State must provide to the IAEA additional information about, and inspector access to, all parts of its nuclear fuel cycle. In addition, the State has to supply information about its nuclear fuel cycle-related research and development (R&D) activities. The majority of States declare their R&D activities under the AP Articles 2.a.(i), 2.a.(x), and 2.b.(i) as part of initial declarations and their annual updates under the AP. In order to verify consistency and completeness of information provided under the AP by States, the Agency has started to analyze declared R&D information by identifying interrelationships between States in different R&D areas relevant to safeguards. The paper outlines the quality of R&D information provided by States to the Agency, describes how the extraction and analysis of relevant declarations are currently carried out at the Agency and specifies what kinds of difficulties arise during evaluation in respect to cross-linking international projects and finding gaps in reporting. In addition, the paper tries to elaborate how the reporting quality of AP information with reference to R&D activities and the assessment process of R&D information could be improved. (author)

  3. Hesitant Fuzzy Linguistic Preference Utility Set and Its Application in Selection of Fire Rescue Plans

    Si, Guangsen; Xu, Zeshui

    2018-01-01

    Hesitant fuzzy linguistic term set provides an effective tool to represent uncertain decision information. However, the semantics corresponding to the linguistic terms in it cannot accurately reflect the decision-makers’ subjective cognition. In general, different decision-makers’ sensitivities towards the semantics are different. Such sensitivities can be represented by the cumulative prospect theory value function. Inspired by this, we propose a linguistic scale function to transform the semantics corresponding to linguistic terms into the linguistic preference values. Furthermore, we propose the hesitant fuzzy linguistic preference utility set, based on which, the decision-makers can flexibly express their distinct semantics and obtain the decision results that are consistent with their cognition. For calculations and comparisons over the hesitant fuzzy linguistic preference utility sets, we introduce some distance measures and comparison laws. Afterwards, to apply the hesitant fuzzy linguistic preference utility sets in emergency management, we develop a method to obtain objective weights of attributes and then propose a hesitant fuzzy linguistic preference utility-TOPSIS method to select the best fire rescue plan. Finally, the validity of the proposed method is verified by some comparisons of the method with other two representative methods including the hesitant fuzzy linguistic-TOPSIS method and the hesitant fuzzy linguistic-VIKOR method. PMID:29614019

  4. Zone analysis in biology articles as a basis for information extraction.

    Mizuta, Yoko; Korhonen, Anna; Mullen, Tony; Collier, Nigel

    2006-06-01

    In the field of biomedicine, an overwhelming amount of experimental data has become available as a result of the high throughput of research in this domain. The amount of results reported has now grown beyond the limits of what can be managed by manual means. This makes it increasingly difficult for the researchers in this area to keep up with the latest developments. Information extraction (IE) in the biological domain aims to provide an effective automatic means to dynamically manage the information contained in archived journal articles and abstract collections and thus help researchers in their work. However, while considerable advances have been made in certain areas of IE, pinpointing and organizing factual information (such as experimental results) remains a challenge. In this paper we propose tackling this task by incorporating into IE information about rhetorical zones, i.e. classification of spans of text in terms of argumentation and intellectual attribution. As the first step towards this goal, we introduce a scheme for annotating biological texts for rhetorical zones and provide a qualitative and quantitative analysis of the data annotated according to this scheme. We also discuss our preliminary research on automatic zone analysis, and its incorporation into our IE framework.

  5. Extract the Relational Information of Static Features and Motion Features for Human Activities Recognition in Videos

    Li Yao

    2016-01-01

    Full Text Available Both static features and motion features have shown promising performance in human activities recognition task. However, the information included in these features is insufficient for complex human activities. In this paper, we propose extracting relational information of static features and motion features for human activities recognition. The videos are represented by a classical Bag-of-Word (BoW model which is useful in many works. To get a compact and discriminative codebook with small dimension, we employ the divisive algorithm based on KL-divergence to reconstruct the codebook. After that, to further capture strong relational information, we construct a bipartite graph to model the relationship between words of different feature set. Then we use a k-way partition to create a new codebook in which similar words are getting together. With this new codebook, videos can be represented by a new BoW vector with strong relational information. Moreover, we propose a method to compute new clusters from the divisive algorithm’s projective function. We test our work on the several datasets and obtain very promising results.

  6. Functional categories in comparative linguistics

    Rijkhoff, Jan

    , Roger M. 1979. Linguistic knowledge and cultural knowledge: some doubts and speculation. American Anthropologist 81-1, 14-36. Levinson, Stephen C. 1997. From outer to inner space: linguistic categories and non-linguistic thinking. In J. Nuyts and E. Pederson (eds.), Language and Conceptualization, 13......). Furthermore certain ‘ontological categories’ are language-specific (Malt 1995). For example, speakers of Kalam (New Guinea) do not classify the cassowary as a bird, because they believe it has a mythical kinship relation with humans (Bulmer 1967).       In this talk I will discuss the role of functional...

  7. MedEx: a medication information extraction system for clinical narratives

    Stenner, Shane P; Doan, Son; Johnson, Kevin B; Waitman, Lemuel R; Denny, Joshua C

    2010-01-01

    Medication information is one of the most important types of clinical data in electronic medical records. It is critical for healthcare safety and quality, as well as for clinical research that uses electronic medical record data. However, medication data are often recorded in clinical notes as free-text. As such, they are not accessible to other computerized applications that rely on coded data. We describe a new natural language processing system (MedEx), which extracts medication information from clinical notes. MedEx was initially developed using discharge summaries. An evaluation using a data set of 50 discharge summaries showed it performed well on identifying not only drug names (F-measure 93.2%), but also signature information, such as strength, route, and frequency, with F-measures of 94.5%, 93.9%, and 96.0% respectively. We then applied MedEx unchanged to outpatient clinic visit notes. It performed similarly with F-measures over 90% on a set of 25 clinic visit notes. PMID:20064797

  8. Videomicroscopic extraction of specific information on cell proliferation and migration in vitro

    Debeir, Olivier; Megalizzi, Veronique; Warzee, Nadine; Kiss, Robert; Decaestecker, Christine

    2008-01-01

    In vitro cell imaging is a useful exploratory tool for cell behavior monitoring with a wide range of applications in cell biology and pharmacology. Combined with appropriate image analysis techniques, this approach has been shown to provide useful information on the detection and dynamic analysis of cell events. In this context, numerous efforts have been focused on cell migration analysis. In contrast, the cell division process has been the subject of fewer investigations. The present work focuses on this latter aspect and shows that, in complement to cell migration data, interesting information related to cell division can be extracted from phase-contrast time-lapse image series, in particular cell division duration, which is not provided by standard cell assays using endpoint analyses. We illustrate our approach by analyzing the effects induced by two sigma-1 receptor ligands (haloperidol and 4-IBP) on the behavior of two glioma cell lines using two in vitro cell models, i.e., the low-density individual cell model and the high-density scratch wound model. This illustration also shows that the data provided by our approach are suggestive as to the mechanism of action of compounds, and are thus capable of informing the appropriate selection of further time-consuming and more expensive biological evaluations required to elucidate a mechanism

  9. 5W1H Information Extraction with CNN-Bidirectional LSTM

    Nurdin, A.; Maulidevi, N. U.

    2018-03-01

    In this work, information about who, did what, when, where, why, and how on Indonesian news articles were extracted by combining Convolutional Neural Network and Bidirectional Long Short-Term Memory. Convolutional Neural Network can learn semantically meaningful representations of sentences. Bidirectional LSTM can analyze the relations among words in the sequence. We also use word embedding word2vec for word representation. By combining these algorithms, we obtained F-measure 0.808. Our experiments show that CNN-BLSTM outperforms other shallow methods, namely IBk, C4.5, and Naïve Bayes with the F-measure 0.655, 0.645, and 0.595, respectively.

  10. Metaproteomics: extracting and mining proteome information to characterize metabolic activities in microbial communities.

    Abraham, Paul E; Giannone, Richard J; Xiong, Weili; Hettich, Robert L

    2014-06-17

    Contemporary microbial ecology studies usually employ one or more "omics" approaches to investigate the structure and function of microbial communities. Among these, metaproteomics aims to characterize the metabolic activities of the microbial membership, providing a direct link between the genetic potential and functional metabolism. The successful deployment of metaproteomics research depends on the integration of high-quality experimental and bioinformatic techniques for uncovering the metabolic activities of a microbial community in a way that is complementary to other "meta-omic" approaches. The essential, quality-defining informatics steps in metaproteomics investigations are: (1) construction of the metagenome, (2) functional annotation of predicted protein-coding genes, (3) protein database searching, (4) protein inference, and (5) extraction of metabolic information. In this article, we provide an overview of current bioinformatic approaches and software implementations in metaproteome studies in order to highlight the key considerations needed for successful implementation of this powerful community-biology tool. Copyright © 2014 John Wiley & Sons, Inc.

  11. Developing a Process Model for the Forensic Extraction of Information from Desktop Search Applications

    Timothy Pavlic

    2008-03-01

    Full Text Available Desktop search applications can contain cached copies of files that were deleted from the file system. Forensic investigators see this as a potential source of evidence, as documents deleted by suspects may still exist in the cache. Whilst there have been attempts at recovering data collected by desktop search applications, there is no methodology governing the process, nor discussion on the most appropriate means to do so. This article seeks to address this issue by developing a process model that can be applied when developing an information extraction application for desktop search applications, discussing preferred methods and the limitations of each. This work represents a more structured approach than other forms of current research.

  12. An innovative method for extracting isotopic information from low-resolution gamma spectra

    Miko, D.; Estep, R.J.; Rawool-Sullivan, M.W.

    1998-01-01

    A method is described for the extraction of isotopic information from attenuated gamma ray spectra using the gross-count material basis set (GC-MBS) model. This method solves for the isotopic composition of an unknown mixture of isotopes attenuated through an absorber of unknown material. For binary isotopic combinations the problem is nonlinear in only one variable and is easily solved using standard line optimization techniques. Results are presented for NaI spectrum analyses of various binary combinations of enriched uranium, depleted uranium, low burnup Pu, 137 Cs, and 133 Ba attenuated through a suite of absorbers ranging in Z from polyethylene through lead. The GC-MBS method results are compared to those computed using ordinary response function fitting and with a simple net peak area method. The GC-MBS method was found to be significantly more accurate than the other methods over the range of absorbers and isotopic blends studied

  13. LANGUE AND PAROLE IN AMERICAN LINGUISTICS.

    LEVIN, SAMUEL R.

    THE PROBLEM OF THE NATURE OF LANGUAGE STRUCTURE IS CONSIDERED AND THE FORM WHICH ANY LINGUISTIC DESCRIPTION SHOULD TAKE. THE AUTHOR EXAMINES THE INFLUENCE OF THE SWISS LINGUIST, FERDINAND DE SAUSSURE, ON THE DEVELOPMENT OF AMERICAN LINGUISTICS. THE QUESTION OF "MENTALISM" IN LINGUISTICS IS REDUCED TO THE PROBLEM OF WHETHER LINGUISTIC…

  14. Teaching Hispanic Linguistics: Strategies to Engage Learners

    Knouse, Stephanie M.; Gupton, Timothy; Abreau, Laurel

    2015-01-01

    Even though many post-secondary institutions offer a variety of Hispanic linguistics classes (Hualde 2006; Lipski 2006), research on the pedagogy of Hispanic linguistics is an underdeveloped or non-existent area of the discipline. Courses in Hispanic linguistics can present not only linguistic challenges for non-native speakers of Spanish, but…

  15. LINGUISTICS AND SECOND LANGUAGE TEACHING: AN ...

    The relationship between linguistics and second language teaching has always been a controversial one. Many linguists have argued that linguistics has nothing to say to the teacher. Sampson (1980, p.10), for example, says: ·"1 do not believe that linguistics has any contribution to make to the teaching of English or the.

  16. Stellenbosch Papers in Linguistics Plus: Journal Sponsorship

    Publisher. Stellenbosch Papers in Linguistics (SPiL) is published by the Department of General Linguistics of Stellenbosch University. Department of General Linguistics, Stellenbosch University. Sources of Support. The Department of General Linguistics acknowledges the financial support provided by the Fonds ...

  17. EnvMine: A text-mining system for the automatic extraction of contextual information

    de Lorenzo Victor

    2010-06-01

    Full Text Available Abstract Background For ecological studies, it is crucial to count on adequate descriptions of the environments and samples being studied. Such a description must be done in terms of their physicochemical characteristics, allowing a direct comparison between different environments that would be difficult to do otherwise. Also the characterization must include the precise geographical location, to make possible the study of geographical distributions and biogeographical patterns. Currently, there is no schema for annotating these environmental features, and these data have to be extracted from textual sources (published articles. So far, this had to be performed by manual inspection of the corresponding documents. To facilitate this task, we have developed EnvMine, a set of text-mining tools devoted to retrieve contextual information (physicochemical variables and geographical locations from textual sources of any kind. Results EnvMine is capable of retrieving the physicochemical variables cited in the text, by means of the accurate identification of their associated units of measurement. In this task, the system achieves a recall (percentage of items retrieved of 92% with less than 1% error. Also a Bayesian classifier was tested for distinguishing parts of the text describing environmental characteristics from others dealing with, for instance, experimental settings. Regarding the identification of geographical locations, the system takes advantage of existing databases such as GeoNames to achieve 86% recall with 92% precision. The identification of a location includes also the determination of its exact coordinates (latitude and longitude, thus allowing the calculation of distance between the individual locations. Conclusion EnvMine is a very efficient method for extracting contextual information from different text sources, like published articles or web pages. This tool can help in determining the precise location and physicochemical

  18. Linguistic Features of English and Russian Dictionaries (A Comparative Study

    Robert Leščinskij

    2013-06-01

    Full Text Available The purpose of this study is to establish differences and similarities between linguistic characteristics of English and Russian dictionaries. Two dictionaries were selected for the study – electronic version of the 8th edition of Oxford Advanced Learner’s Dictionary (OALD and the online version of Ozhegov’s explanatory dictionary. The methods chosen for the study were descriptive, comparative and contrastive analysis. Linguistic characteristics of the dictionaries were analysed and compared. The research showed that both reference books provided different linguistic information on the headwords. OALD provided exhaustive phonetic information, which Ozhegov’s dictionary lacked. The two dictionaries provided different orthographic information. OALD disclosed semantic information via various tools available in the electronic version; these were unavailable in Ozhegov’s dictionary. Both dictionaries used similar stylistic labels.

  19. Linguistics: evolution and language change.

    Bowern, Claire

    2015-01-05

    Linguists have long identified sound changes that occur in parallel. Now novel research shows how Bayesian modeling can capture complex concerted changes, revealing how evolution of sounds proceeds. Copyright © 2015 Elsevier Ltd. All rights reserved.

  20. Stellenbosch Papers in Linguistics Plus

    Stellenbosch Papers in Linguistics Plus. Journal Home · ABOUT THIS JOURNAL · Advanced Search · Current Issue · Archives · Journal Home > Vol 42 (2013) >. Log in or Register to get access to full text downloads.

  1. Stellenbosch Papers in Linguistics Plus

    Stellenbosch Papers in Linguistics Plus. Journal Home · ABOUT THIS JOURNAL · Advanced Search · Current Issue · Archives · Journal Home > Vol 27 (1995) >. Log in or Register to get access to full text downloads.

  2. Gesture Modelling for Linguistic Purposes

    Olivrin, GJ

    2007-05-01

    Full Text Available The study of sign languages attempts to create a coherent model that binds the expressive nature of signs conveyed in gestures to a linguistic framework. Gesture modelling offers an alternative that provides device independence, scalability...

  3. Is Rorty a linguistic idealist?

    Marvan, Tomáš

    2011-01-01

    Roč. 21, č. 3 (2011), s. 272-279 ISSN 1210-3055 Institutional research plan: CEZ:AV0Z90090514 Keywords : Rorty * linguistic idealism * internal realism * intrinsic structure of reality * representation Subject RIV: AA - Philosophy ; Religion

  4. The Dutch Linguistic Intraoperative Protocol: a valid linguistic approach to awake brain surgery.

    De Witte, E; Satoer, D; Robert, E; Colle, H; Verheyen, S; Visch-Brink, E; Mariën, P

    2015-01-01

    Intraoperative direct electrical stimulation (DES) is increasingly used in patients operated on for tumours in eloquent areas. Although a positive impact of DES on postoperative linguistic outcome is generally advocated, information about the neurolinguistic methods applied in awake surgery is scarce. We developed for the first time a standardised Dutch linguistic test battery (measuring phonology, semantics, syntax) to reliably identify the critical language zones in detail. A normative study was carried out in a control group of 250 native Dutch-speaking healthy adults. In addition, the clinical application of the Dutch Linguistic Intraoperative Protocol (DuLIP) was demonstrated by means of anatomo-functional models and five case studies. A set of DuLIP tests was selected for each patient depending on the tumour location and degree of linguistic impairment. DuLIP is a valid test battery for pre-, intraoperative and postoperative language testing and facilitates intraoperative mapping of eloquent language regions that are variably located. Copyright © 2014 Elsevier Inc. All rights reserved.

  5. Extraction of prospecting information of uranium deposit based on high spatial resolution satellite data. Taking bashibulake region as an example

    Yang Xu; Liu Dechang; Zhang Jielin

    2008-01-01

    In this study, the signification and content of prospecting information of uranium deposit are expounded. Quickbird high spatial resolution satellite data are used to extract the prospecting information of uranium deposit in Bashibulake area in the north of Tarim Basin. By using the pertinent methods of image processing, the information of ore-bearing bed, ore-control structure and mineralized alteration have been extracted. The results show a high consistency with the field survey. The aim of this study is to explore practicability of high spatial resolution satellite data for prospecting minerals, and to broaden the thinking of prospectation at similar area. (authors)

  6. Linguistic Characteristics of Advertising English

    易高燕

    2010-01-01

    Advertising language takes form under the influence of linguistics,psychology and sociology,etc,and its way of choosing words and building sentences are quite different from normal English.And as a practical language,advertising English has its specific functions,and it has been distinguished from normal English as an independent language,and it has plentiful values.This paper aims to discuss some linguistic characteristics of advertising English.

  7. Translating Linguistic Jokes for Dubbing

    Elena ALEKSANDROVA

    2012-01-01

    Full Text Available This study has attempted to establish the possible ways of translating linguistic jokes whendubbing. The study is also intended to identify the most problematic cases of screen translation andthe factors which cause these problems. In order to support such an approach a corpus of 7American and British films has been compiled, including as many as 16 as their various dubbingtranslations into Russian. In the films, almost 12 instances of original linguistic jokes have beenidentified.

  8. Copyright Essentials for Linguists

    Paul Newman

    2007-06-01

    Full Text Available This paper addresses copyright issues that linguists confront in their capacity as users and creators of scholarly work. It is organized in a simple question-answer format. Questions 1-3 present the basics of U.S. copyright law, including the fundamental nature of copyright as a bundle of intellectual property rights and the role of registration. Questions 4-5 treat issues of copyright notice. Questions 6-8 explain licenses, especially Creative Commons licenses, and the function of an Author's Addendum. Questions 9-10 look at copyright in the context of online open access publishing. Question 11 discusses the concept of Fair Use. Question 12 analyzes the problem of what are called Orphan Works. Questions 13-19 explore issues of copyright ownership, including Work for Hire, joint authorship, and attribution. Questions 20-22 deal with copyright with specific reference to fieldwork situations and indigenous rights. The paper concludes with a brief presentation of key sources for further study and clarification.

  9. The new linguistic order

    Joshua A. Fishman

    2001-07-01

    Full Text Available The globalisation phenomenon that we are currently seeing has lead to major linguistic changes on a worldwide scale. English has become the leading international language, in economic and political spheres, and is becoming the language of high society and of the young. At the same time, however, regional languages are also making considerable headway, thanks to new social interaction and economic backing from their governments. In turn, and as a result of these two trends, there is impetus for feelings of belonging to local communities which see their language as a sign of their own authenticity, one that has to be defended against the phenomena of globalisation and regionalisation. We are thus heading towards a multilingual society, in which each language has its own, distinct social functions, even though it is inevitable that there will be conflict between the languages that come into contact. In this scenario, the author predicts a loss of hegemony for English, in favour of regional languages, and the future extinction of the least spoken minority languages.

  10. GEOLINGUISTICS: THE LINGUISTIC ATLAS OF PARANÁ

    Rosa Evangelina de Santana BELLI RODRIGUES

    2015-06-01

    Full Text Available The objective of this work is to analyze the methodology adopted by the Linguistic Atlas of Paraná – APLR (AGUILERA, 1990 and to describe its results in relation to other Brazilian atlas. To meet this objective, we first present the modifications, mainly methodological, under gone by Geolinguistics towards a more complete and in depth description of linguistic variation. The Pluridimensional Geolinguistics and Contractual model of Harald Thun (1998 and the Linguistics Atlas of Brazil – ALiB (CARDOSO et all, 2014, published in October, 2014, are presented. It was also necessary to describe, although briefly, the most traditional Geolinguistics research method, characteristic of the ALPR, before referring the text back to Aguilera’s Atlas. After discussing the criteria on which the ALPR was constructed, from choice of informers to the Geolinguistics charts that compose it, as well as its complementation by the ALPR II (ALTINO, 2007, it was possible to analyze the results and relate them to the hypotheses posed by the thesis which gave origin to it.

  11. Linguistic coding deficits in foreign language learners.

    Sparks, R; Ganschow, L; Pohlman, J

    1989-01-01

    As increasing numbers of colleges and universities require a foreign language for graduation in at least one of their degree programs, reports of students with difficulties in learning a second language are multiplying. Until recently, little research has been conducted to identify the nature of this problem. Recent attempts by the authors have focused upon subtle but ongoing language difficulties in these individuals as the source of their struggle to learn a foreign language. The present paper attempts to expand upon this concept by outlining a theoretical framework based upon a linguistic coding model that hypothesizes deficits in the processing of phonological, syntactic, and/or semantic information. Traditional psychoeducational assessment batteries of standardized intelligence and achievement tests generally are not sensitive to these linguistic coding deficits unless closely analyzed or, more often, used in conjunction with a more comprehensive language assessment battery. Students who have been waived from a foreign language requirement and their proposed type(s) of linguistic coding deficits are profiled. Tentative conclusions about the nature of these foreign language learning deficits are presented along with specific suggestions for tests to be used in psychoeducational evaluations.

  12. Linguistic Dating of Biblical Texts

    Ehrensvärd, Martin Gustaf

    2003-01-01

    For two centuries, scholars have pointed to consistent differences in the Hebrew of certain biblical texts and interpreted these differences as reflecting the date of composition of the texts. Until the 1980s, this was quite uncontroversial as the linguistic findings largely confirmed the chronol......For two centuries, scholars have pointed to consistent differences in the Hebrew of certain biblical texts and interpreted these differences as reflecting the date of composition of the texts. Until the 1980s, this was quite uncontroversial as the linguistic findings largely confirmed...... the chronology of the texts established by other means: the Hebrew of Genesis-2 Kings was judged to be early and that of Esther, Daniel, Ezra, Nehemiah, and Chronicles to be late. In the current debate where revisionists have questioned the traditional dating, linguistic arguments in the dating of texts have...... come more into focus. The study critically examines some linguistic arguments adduced to support the traditional position, and reviewing the arguments it points to weaknesses in the linguistic dating of EBH texts to pre-exilic times. When viewing the linguistic evidence in isolation it will be clear...

  13. Extracting chemical information from high-resolution Kβ X-ray emission spectroscopy

    Limandri, S.; Robledo, J.; Tirao, G.

    2018-06-01

    High-resolution X-ray emission spectroscopy allows studying the chemical environment of a wide variety of materials. Chemical information can be obtained by fitting the X-ray spectra and observing the behavior of some spectral features. Spectral changes can also be quantified by means of statistical parameters calculated by considering the spectrum as a probability distribution. Another possibility is to perform statistical multivariate analysis, such as principal component analysis. In this work the performance of these procedures for extracting chemical information in X-ray emission spectroscopy spectra for mixtures of Mn2+ and Mn4+ oxides are studied. A detail analysis of the parameters obtained, as well as the associated uncertainties is shown. The methodologies are also applied for Mn oxidation state characterization of double perovskite oxides Ba1+xLa1-xMnSbO6 (with 0 ≤ x ≤ 0.7). The results show that statistical parameters and multivariate analysis are the most suitable for the analysis of this kind of spectra.

  14. Information Extraction of Tourist Geological Resources Based on 3d Visualization Remote Sensing Image

    Wang, X.

    2018-04-01

    Tourism geological resources are of high value in admiration, scientific research and universal education, which need to be protected and rationally utilized. In the past, most of the remote sensing investigations of tourism geological resources used two-dimensional remote sensing interpretation method, which made it difficult for some geological heritages to be interpreted and led to the omission of some information. This aim of this paper is to assess the value of a method using the three-dimensional visual remote sensing image to extract information of geological heritages. skyline software system is applied to fuse the 0.36 m aerial images and 5m interval DEM to establish the digital earth model. Based on the three-dimensional shape, color tone, shadow, texture and other image features, the distribution of tourism geological resources in Shandong Province and the location of geological heritage sites were obtained, such as geological structure, DaiGu landform, granite landform, Volcanic landform, sandy landform, Waterscapes, etc. The results show that using this method for remote sensing interpretation is highly recognizable, making the interpretation more accurate and comprehensive.

  15. Information Management Processes for Extraction of Student Dropout Indicators in Courses in Distance Mode

    Renata Maria Abrantes Baracho

    2016-04-01

    Full Text Available This research addresses the use of information management processes in order to extract student dropout indicators in distance mode courses. Distance education in Brazil aims to facilitate access to information. The MEC (Ministry of Education announced, in the second semester of 2013, that the main obstacles faced by institutions offering courses in this mode were students dropping out and the resistance of both educators and students to this mode. The research used a mixed methodology, qualitative and quantitative, to obtain student dropout indicators. The factors found and validated in this research were: the lack of interest from students, insufficient training in the use of the virtual learning environment for students, structural problems in the schools that were chosen to offer the course, students without e-mail, incoherent answers to activities to the course, lack of knowledge on the part of the student when using the computer tool. The scenario considered was a course offered in distance mode called Aluno Integrado (Integrated Student

  16. Measuring nuclear reaction cross sections to extract information on neutrinoless double beta decay

    Cavallaro, M.; Cappuzzello, F.; Agodi, C.; Acosta, L.; Auerbach, N.; Bellone, J.; Bijker, R.; Bonanno, D.; Bongiovanni, D.; Borello-Lewin, T.; Boztosun, I.; Branchina, V.; Bussa, M. P.; Calabrese, S.; Calabretta, L.; Calanna, A.; Calvo, D.; Carbone, D.; Chávez Lomelí, E. R.; Coban, A.; Colonna, M.; D'Agostino, G.; De Geronimo, G.; Delaunay, F.; Deshmukh, N.; de Faria, P. N.; Ferraresi, C.; Ferreira, J. L.; Finocchiaro, P.; Fisichella, M.; Foti, A.; Gallo, G.; Garcia, U.; Giraudo, G.; Greco, V.; Hacisalihoglu, A.; Kotila, J.; Iazzi, F.; Introzzi, R.; Lanzalone, G.; Lavagno, A.; La Via, F.; Lay, J. A.; Lenske, H.; Linares, R.; Litrico, G.; Longhitano, F.; Lo Presti, D.; Lubian, J.; Medina, N.; Mendes, D. R.; Muoio, A.; Oliveira, J. R. B.; Pakou, A.; Pandola, L.; Petrascu, H.; Pinna, F.; Reito, S.; Rifuggiato, D.; Rodrigues, M. R. D.; Russo, A. D.; Russo, G.; Santagati, G.; Santopinto, E.; Sgouros, O.; Solakci, S. O.; Souliotis, G.; Soukeras, V.; Spatafora, A.; Torresi, D.; Tudisco, S.; Vsevolodovna, R. I. M.; Wheadon, R. J.; Yildirin, A.; Zagatto, V. A. B.

    2018-02-01

    Neutrinoless double beta decay (0vββ) is considered the best potential resource to access the absolute neutrino mass scale. Moreover, if observed, it will signal that neutrinos are their own anti-particles (Majorana particles). Presently, this physics case is one of the most important research “beyond Standard Model” and might guide the way towards a Grand Unified Theory of fundamental interactions. Since the 0vββ decay process involves nuclei, its analysis necessarily implies nuclear structure issues. In the NURE project, supported by a Starting Grant of the European Research Council (ERC), nuclear reactions of double charge-exchange (DCE) are used as a tool to extract information on the 0vββ Nuclear Matrix Elements. In DCE reactions and ββ decay indeed the initial and final nuclear states are the same and the transition operators have similar structure. Thus the measurement of the DCE absolute cross-sections can give crucial information on ββ matrix elements. In a wider view, the NUMEN international collaboration plans a major upgrade of the INFN-LNS facilities in the next years in order to increase the experimental production of nuclei of at least two orders of magnitude, thus making feasible a systematic study of all the cases of interest as candidates for 0vββ.

  17. Unsupervised Symbolization of Signal Time Series for Extraction of the Embedded Information

    Yue Li

    2017-03-01

    Full Text Available This paper formulates an unsupervised algorithm for symbolization of signal time series to capture the embedded dynamic behavior. The key idea is to convert time series of the digital signal into a string of (spatially discrete symbols from which the embedded dynamic information can be extracted in an unsupervised manner (i.e., no requirement for labeling of time series. The main challenges here are: (1 definition of the symbol assignment for the time series; (2 identification of the partitioning segment locations in the signal space of time series; and (3 construction of probabilistic finite-state automata (PFSA from the symbol strings that contain temporal patterns. The reported work addresses these challenges by maximizing the mutual information measures between symbol strings and PFSA states. The proposed symbolization method has been validated by numerical simulation as well as by experimentation in a laboratory environment. Performance of the proposed algorithm has been compared to that of two commonly used algorithms of time series partitioning.

  18. A methodology for the extraction of quantitative information from electron microscopy images at the atomic level

    Galindo, P L; Pizarro, J; Guerrero, E; Guerrero-Lebrero, M P; Scavello, G; Yáñez, A; Sales, D L; Herrera, M; Molina, S I; Núñez-Moraleda, B M; Maestre, J M

    2014-01-01

    In this paper we describe a methodology developed at the University of Cadiz (Spain) in the past few years for the extraction of quantitative information from electron microscopy images at the atomic level. This work is based on a coordinated and synergic activity of several research groups that have been working together over the last decade in two different and complementary fields: Materials Science and Computer Science. The aim of our joint research has been to develop innovative high-performance computing techniques and simulation methods in order to address computationally challenging problems in the analysis, modelling and simulation of materials at the atomic scale, providing significant advances with respect to existing techniques. The methodology involves several fundamental areas of research including the analysis of high resolution electron microscopy images, materials modelling, image simulation and 3D reconstruction using quantitative information from experimental images. These techniques for the analysis, modelling and simulation allow optimizing the control and functionality of devices developed using materials under study, and have been tested using data obtained from experimental samples

  19. Dual-wavelength phase-shifting digital holography selectively extracting wavelength information from wavelength-multiplexed holograms.

    Tahara, Tatsuki; Mori, Ryota; Kikunaga, Shuhei; Arai, Yasuhiko; Takaki, Yasuhiro

    2015-06-15

    Dual-wavelength phase-shifting digital holography that selectively extracts wavelength information from five wavelength-multiplexed holograms is presented. Specific phase shifts for respective wavelengths are introduced to remove the crosstalk components and extract only the object wave at the desired wavelength from the holograms. Object waves in multiple wavelengths are selectively extracted by utilizing 2π ambiguity and the subtraction procedures based on phase-shifting interferometry. Numerical results show the validity of the proposed technique. The proposed technique is also experimentally demonstrated.

  20. Information Extraction and Dependency on Open Government Data (ogd) for Environmental Monitoring

    Abdulmuttalib, Hussein

    2016-06-01

    Environmental monitoring practices support decision makers of different government / private institutions, besides environmentalists and planners among others. This support helps them act towards the sustainability of our environment, and also take efficient measures for protecting human beings in general, but it is difficult to explore useful information from 'OGD' and assure its quality for the purpose. On the other hand, Monitoring itself comprises detecting changes as happens, or within the mitigation period range, which means that any source of data, that is to be used for monitoring, should replicate the information related to the period of environmental monitoring, or otherwise it's considered almost useless or history. In this paper the assessment of information extraction and structuring from Open Government Data 'OGD', that can be useful to environmental monitoring is performed, looking into availability, usefulness to environmental monitoring of a certain type, checking its repetition period and dependences. The particular assessment is being performed on a small sample selected from OGD, bearing in mind the type of the environmental change monitored, such as the increase and concentrations of built up areas, and reduction of green areas, or monitoring the change of temperature in a specific area. The World Bank mentioned in its blog that Data is open if it satisfies both conditions of, being technically open, and legally open. The use of Open Data thus, is regulated by published terms of use, or an agreement which implies some conditions without violating the above mentioned two conditions. Within the scope of the paper I wish to share the experience of using some OGD for supporting an environmental monitoring work, that is performed to mitigate the production of carbon dioxide, by regulating energy consumption, and by properly designing the test area's landscapes, thus using Geodesign tactics, meanwhile wish to add to the results achieved by many

  1. How do we communicate stereotypes? Linguistic bases and inferential consequences

    Wigboldus, DHJ; Semin, GR; Spears, R

    The linguistic expectancy bias is defined as the tendency to describe expectancy-consistent information at a higher level of abstraction than expectancy-inconsistent information; The communicative consequences of this bias were examined in 3 experiments. Analyses of judgments that recipients made on

  2. Translation and linguistic validation of the Pediatric Patient-Reported Outcomes Measurement Information System measures into simplified Chinese using cognitive interviewing methodology.

    Liu, Yanyan; Hinds, Pamela S; Wang, Jichuan; Correia, Helena; Du, Shizheng; Ding, Jian; Gao, Wen Jun; Yuan, Changrong

    2013-01-01

    The Pediatric Patient-Reported Outcomes Measurement Information System (PROMIS) measures were developed using modern measurement theory and tested in a variety of settings to assess the quality of life, function, and symptoms of children and adolescents experiencing a chronic illness and its treatment. Developed in English, this set of measures had not been translated into Chinese. The objective of this study was to develop the Chinese version of the Pediatric PROMIS measures (C-Ped-PROMIS), specifically 8 short forms, and to pretest the translated measures in children and adolescents through cognitive interviewing methodology. The C-Ped-PROMIS was developed following the standard Functional Assessment of Chronic Illness Therapy Translation Methodology. Bilingual teams from the United States and China reviewed the translation to develop a provisional version, which was then pretested with cognitive interview by probing 10 native Chinese-speaking children aged 8 to 17 years in China. The translation was finalized by the bilingual teams. Most items, response options, and instructions were well understood by the children, and some revisions were made to address patient's comments during the cognitive interview. The results indicated that the C-Ped-PROMIS items were semantically and conceptually equivalent to the original. Children aged 8 to 17 years in China were able to comprehend these measures and express their experience and feelings about illness or their life. The C-Ped-PROMIS is available for psychometric validation. Future work will be directed at translating the rest of the item banks, calibrating them and creating a Chinese final version of the short forms.

  3. Probabilistic Linguistic Power Aggregation Operators for Multi-Criteria Group Decision Making

    Agbodah Kobina

    2017-12-01

    Full Text Available As an effective aggregation tool, power average (PA allows the input arguments being aggregated to support and reinforce each other, which provides more versatility in the information aggregation process. Under the probabilistic linguistic term environment, we deeply investigate the new power aggregation (PA operators for fusing the probabilistic linguistic term sets (PLTSs. In this paper, we firstly develop the probabilistic linguistic power average (PLPA, the weighted probabilistic linguistic power average (WPLPA operators, the probabilistic linguistic power geometric (PLPG and the weighted probabilistic linguistic power geometric (WPLPG operators. At the same time, we carefully analyze the properties of these new aggregation operators. With the aid of the WPLPA and WPLPG operators, we further design the approaches for the application of multi-criteria group decision-making (MCGDM with PLTSs. Finally, we use an illustrated example to expound our proposed methods and verify their performances.

  4. Linguistic and Psycho-Linguistic Principles of Linguadidactics (theoretical interpretation

    Liudmila Mauzienė

    2011-04-01

    Full Text Available This article considers linguadidactics being closely related to linguistics, psychology, psycholinguistics and didactics and applies their theoretical statements and regularities in its scientific studies. Methodology refers to linguistics which investigates the language as a teaching subject. Methodology is linked to psychology in two ways. First of all, it is based on psychology as the teaching process is an intellectual psychical act and its regularities are necessary to know. On the other hand, methodology applies rules of pedagogy that predicts ways of learning and development of language skills. The article emphasizes that sustainable work experience and analysis of scientific research show that teaching process is more effective if consistent patterns of linguistics and psychology are appropriately applied.

  5. Machine learning classification of surgical pathology reports and chunk recognition for information extraction noise reduction.

    Napolitano, Giulio; Marshall, Adele; Hamilton, Peter; Gavin, Anna T

    2016-06-01

    Machine learning techniques for the text mining of cancer-related clinical documents have not been sufficiently explored. Here some techniques are presented for the pre-processing of free-text breast cancer pathology reports, with the aim of facilitating the extraction of information relevant to cancer staging. The first technique was implemented using the freely available software RapidMiner to classify the reports according to their general layout: 'semi-structured' and 'unstructured'. The second technique was developed using the open source language engineering framework GATE and aimed at the prediction of chunks of the report text containing information pertaining to the cancer morphology, the tumour size, its hormone receptor status and the number of positive nodes. The classifiers were trained and tested respectively on sets of 635 and 163 manually classified or annotated reports, from the Northern Ireland Cancer Registry. The best result of 99.4% accuracy - which included only one semi-structured report predicted as unstructured - was produced by the layout classifier with the k nearest algorithm, using the binary term occurrence word vector type with stopword filter and pruning. For chunk recognition, the best results were found using the PAUM algorithm with the same parameters for all cases, except for the prediction of chunks containing cancer morphology. For semi-structured reports the performance ranged from 0.97 to 0.94 and from 0.92 to 0.83 in precision and recall, while for unstructured reports performance ranged from 0.91 to 0.64 and from 0.68 to 0.41 in precision and recall. Poor results were found when the classifier was trained on semi-structured reports but tested on unstructured. These results show that it is possible and beneficial to predict the layout of reports and that the accuracy of prediction of which segments of a report may contain certain information is sensitive to the report layout and the type of information sought. Copyright

  6. Linguistic Intuitions and Cognitive Penetrability

    Michael Devitt

    2014-12-01

    Full Text Available Metalinguistic intuitions play a very large evidential role in both linguistics and philosophy. Linguists think that these intuitions are products of underlying linguistic competence. I call this view “the voice of competence” (“VoC”. Although many philosophers seem to think that metalinguistic intuitions are a priori many may implicitly hold the more scientifically respectable VoC. According to VoC, I argue, these intuitions can be cognitively penetrated by the central processor. But, I have argued elsewhere, VoC is false. Instead, we should hold “the modest explanation” (“ME” according to which these intuitions are fairly unreflective empirical theory-laden central-processor responses to phenomena. On ME, no question of cognitive penetration arises. ME has great methodological significance for the study of language. Insofar as we rely on intuitions as evidence we should prefer those of linguists and philosophers because they are more expert. But, more importantly, we should be seeking other evidence in linguistic usage.

  7. Study of time-frequency characteristics of single snores: extracting new information for sleep apnea diagnosis

    Castillo Escario, Y.; Blanco Almazan, D.; Camara Vazquez, M.A.; Jane Campos, R.

    2016-07-01

    Obstructive sleep apnea (OSA) is a highly prevalent chronic disease, especially in elderly and obese population. Despite constituting a huge health and economic problem, most patients remain undiagnosed due to limitations in current strategies. Therefore, it is essential to find cost-effective diagnostic alternatives. One of these novel approaches is the analysis of acoustic snoring signals. Snoring is an early symptom of OSA which carries pathophysiological information of high diagnostic value. For this reason, the main objective of this work is to study the characteristics of single snores of different types, from healthy and OSA subjects. To do that, we analyzed snoring signals from previous databases and developed an experimental protocol to record simulated OSA-related sounds and characterize the response of two commercial tracheal microphones. Automatic programs for filtering, downsampling, event detection and time-frequency analysis were built in MATLAB. We found that time-frequency maps and spectral parameters (central, mean and peak frequency and energy in the 100-500 Hz band) allow distinguishing regular snores of healthy subjects from non-regular snores and snores of OSA subjects. Regarding the two commercial microphones, we found that one of them was a suitable snoring sensor, while the other had a too restricted frequency response. Future work shall include a higher number of episodes and subjects, but our study has contributed to show how important the differences between regular and non-regular snores can be for OSA diagnosis, and how much clinically relevant information can be extracted from time-frequency maps and spectral parameters of single snores. (Author)

  8. Quantum measurement information as a key to energy extraction from local vacuums

    Hotta, Masahiro

    2008-01-01

    In this paper, a protocol is proposed in which energy extraction from local vacuum states is possible by using quantum measurement information for the vacuum state of quantum fields. In the protocol, Alice, who stays at a spatial point, excites the ground state of the fields by a local measurement. Consequently, wave packets generated by Alice's measurement propagate the vacuum to spatial infinity. Let us assume that Bob stays away from Alice and fails to catch the excitation energy when the wave packets pass in front of him. Next Alice announces her local measurement result to Bob by classical communication. Bob performs a local unitary operation depending on the measurement result. In this process, positive energy is released from the fields to Bob's apparatus of the unitary operation. In the field systems, wave packets are generated with negative energy around Bob's location. Soon afterwards, the negative-energy wave packets begin to chase after the positive-energy wave packets generated by Alice and form loosely bound states.

  9. Oxygen octahedra picker: A software tool to extract quantitative information from STEM images

    Wang, Yi, E-mail: y.wang@fkf.mpg.de; Salzberger, Ute; Sigle, Wilfried; Eren Suyolcu, Y.; Aken, Peter A. van

    2016-09-15

    In perovskite oxide based materials and hetero-structures there are often strong correlations between oxygen octahedral distortions and functionality. Thus, atomistic understanding of the octahedral distortion, which requires accurate measurements of atomic column positions, will greatly help to engineer their properties. Here, we report the development of a software tool to extract quantitative information of the lattice and of BO{sub 6} octahedral distortions from STEM images. Center-of-mass and 2D Gaussian fitting methods are implemented to locate positions of individual atom columns. The precision of atomic column distance measurements is evaluated on both simulated and experimental images. The application of the software tool is demonstrated using practical examples. - Highlights: • We report a software tool for mapping atomic positions from HAADF and ABF images. • It enables quantification of both crystal lattice and oxygen octahedral distortions. • We test the measurement accuracy and precision on simulated and experimental images. • It works well for different orientations of perovskite structures and interfaces.

  10. Note on difference spectra for fast extraction of global image information.

    Van Wyk, BJ

    2007-06-01

    Full Text Available FOR FAST EXTRACTION OF GLOBAL IMAGE INFORMATION. B.J van Wyk* M.A. van Wyk* and F. van den Bergh** * c29c55c48c51c46c4bc03 c36c52c58c57c4bc03 c24c49c55c4cc46c44c51c03 c37c48c46c4bc51c4cc46c44c4fc03 c2cc51c56c57c4cc57c58c57c48c03 c4cc51c03 c28c4fc48c...46c57c55c52c51c4cc46c56c03 c0bc29cb6c36c24c37c2cc28c0cc03 c44c57c03 c57c4bc48c03 c37c56c4bc5ac44c51c48c03 c38c51c4cc59c48c55c56c4cc57c5cc03 c52c49c03 Technology, Private Bag X680, Pretoria 0001. ** Remote Sensing Research Group, Meraka Institute...

  11. COGNITIVE METAPHOR IN MODERN LINGUISTICS

    Antonina KARTASHOVA

    2010-11-01

    Full Text Available The article outlines the basic notions connected with cognitive metaphor which has lately undergone a thorough examination. The contribution made by linguists resulted in the rise of cognitive linguistics. This science regards metaphor not as a linguistic phenomenon but as a mental one that establishes connection between language and mind in the form of understanding new notions in terms of notions and categories known due to the previously gained experience. The interaction of new and previous experience can generate three main types of metaphors: structural metaphors which imply the structuring of target domain in terms of source domain, ontological metaphors which view abstract notions as concrete objects with clear outlines and orientational metaphors which represent the ways to fix the experience of spatial orientation. The classification of metaphors complemented with examples is presented below along with some controversial cases of determining the type of metaphor.

  12. Data Acquisition and Linguistic Resources

    Strassel, Stephanie; Christianson, Caitlin; McCary, John; Staderman, William; Olive, Joseph

    All human language technology demands substantial quantities of data for system training and development, plus stable benchmark data to measure ongoing progress. While creation of high quality linguistic resources is both costly and time consuming, such data has the potential to profoundly impact not just a single evaluation program but language technology research in general. GALE's challenging performance targets demand linguistic data on a scale and complexity never before encountered. Resources cover multiple languages (Arabic, Chinese, and English) and multiple genres -- both structured (newswire and broadcast news) and unstructured (web text, including blogs and newsgroups, and broadcast conversation). These resources include significant volumes of monolingual text and speech, parallel text, and transcribed audio combined with multiple layers of linguistic annotation, ranging from word aligned parallel text and Treebanks to rich semantic annotation.

  13. Analysis Methods for Extracting Knowledge from Large-Scale WiFi Monitoring to Inform Building Facility Planning

    Ruiz-Ruiz, Antonio; Blunck, Henrik; Prentow, Thor Siiger

    2014-01-01

    realistic data to inform facility planning. In this paper, we propose analysis methods to extract knowledge from large sets of network collected WiFi traces to better inform facility management and planning in large building complexes. The analysis methods, which build on a rich set of temporal and spatial......The optimization of logistics in large building com- plexes with many resources, such as hospitals, require realistic facility management and planning. Current planning practices rely foremost on manual observations or coarse unverified as- sumptions and therefore do not properly scale or provide....... Spatio-temporal visualization tools built on top of these methods enable planners to inspect and explore extracted information to inform facility-planning activities. To evaluate the methods, we present results for a large hospital complex covering more than 10 hectares. The evaluation is based on Wi...

  14. Social network size can influence linguistic malleability and the propagation of linguistic change.

    Lev-Ari, Shiri

    2018-07-01

    We learn language from our social environment, but the more sources we have, the less informative each source is, and therefore, the less weight we ascribe its input. According to this principle, people with larger social networks should give less weight to new incoming information, and should therefore be less susceptible to the influence of new speakers. This paper tests this prediction, and shows that speakers with smaller social networks indeed have more malleable linguistic representations. In particular, they are more likely to adjust their lexical boundary following exposure to a new speaker. Experiment 2 uses computational simulations to test whether this greater malleability could lead people with smaller social networks to be important for the propagation of linguistic change despite the fact that they interact with fewer people. The results indicate that when innovators were connected with people with smaller rather than larger social networks, the population exhibited greater and faster diffusion. Together these experiments show that the properties of people's social networks can influence individuals' learning and use as well as linguistic phenomena at the community level. Copyright © 2018 Elsevier B.V. All rights reserved.

  15. Mathematical Approaches to Cognitive Linguistics

    Chuluundorj Begz

    2013-05-01

    Full Text Available Cognitive linguistics, neuro-cognitive and psychological analysis of human verbal cognition present important area of multidisciplinary research. Mathematical methods and models have been introduced in number of publications with increasing attention to these theories. In this paper we have described some possible applications of mathematical methods to cognitive linguistics. Human verbal perception and verbal mapping deal with dissipative mental structures and symmetric/asymmetric relationships between objects of perception and deep (also surface structures of language. In that’s way methods of tensor analysis are ambitious candidate to be applied to analysis of human verbal thinking and mental space.

  16. Linguistics, human communication and psychiatry.

    Thomas, P; Fraser, W

    1994-11-01

    Psycholinguistics and sociolinguistics have extended our understanding of the abnormal communication seen in psychosis, as well as that of people with autism and Asperger's syndrome. Psycholinguistics has the potential to increase the explanatory power of cognitive and neuropsychological approaches to psychosis and new methods of assessment and therapy are now being developed, based on linguistic theory. A MEDLINE literature search was used. Of 205 relevant articles identified, 65 were selected for review. Greater familiarity with linguistic theory could improve psychiatrists' assessment skills and their understanding of the relevance of human communication to the new cognitive models of psychosis.

  17. Extraction as a source of additional information when concentrations in multicomponent systems are simultaneously determined

    Perkov, I.G.

    1988-01-01

    Using as an example photometric determination of Nd and Sm in their joint presence, the possibility to use the influence of extraction on analytic signal increase is considered. It is shown that interligand exchange in extracts in combination with simultaneous determination of concentrations can be used as a simple means increasing the accuracy of determination. 5 refs.; 2 figs.; 3 tabs

  18. On Norms and Linguistic Categories in Linguistic Diversity Management

    Marácz, L.

    2014-01-01

    Due to globalization there is an increase in the appearances of languages in the multilingual linguistic landscape in urban spaces. Commentators have described this state of affairs as super-, mega- or complex diversity. Mainstream sociolinguists have argued that languages have no fixed boundaries

  19. Linguistic Policies, Linguistic Planning, and Brazilian Sign Language in Brazil

    de Quadros, Ronice Muller

    2012-01-01

    This article explains the consolidation of Brazilian Sign Language in Brazil through a linguistic plan that arose from the Brazilian Sign Language Federal Law 10.436 of April 2002 and the subsequent Federal Decree 5695 of December 2005. Two concrete facts that emerged from this existing language plan are discussed: the implementation of bilingual…

  20. Linguistic Variability and Intellectual Development. Miami Linguistics Series No. 9.

    von Humboldt, Wilhelm

    Although this edition of Wilhelm von Humboldt's "Linguistic Variability and Intellectual Development" is based entirely on the original German edition, the translators (George C. Buck and Frithjof A. Raven) and the publisher have attempted to clarify certain aspects of this work for the modern-day reader. These features include the addition of…

  1. Validation and extraction of molecular-geometry information from small-molecule databases.

    Long, Fei; Nicholls, Robert A; Emsley, Paul; Graǽulis, Saulius; Merkys, Andrius; Vaitkus, Antanas; Murshudov, Garib N

    2017-02-01

    A freely available small-molecule structure database, the Crystallography Open Database (COD), is used for the extraction of molecular-geometry information on small-molecule compounds. The results are used for the generation of new ligand descriptions, which are subsequently used by macromolecular model-building and structure-refinement software. To increase the reliability of the derived data, and therefore the new ligand descriptions, the entries from this database were subjected to very strict validation. The selection criteria made sure that the crystal structures used to derive atom types, bond and angle classes are of sufficiently high quality. Any suspicious entries at a crystal or molecular level were removed from further consideration. The selection criteria included (i) the resolution of the data used for refinement (entries solved at 0.84 Å resolution or higher) and (ii) the structure-solution method (structures must be from a single-crystal experiment and all atoms of generated molecules must have full occupancies), as well as basic sanity checks such as (iii) consistency between the valences and the number of connections between atoms, (iv) acceptable bond-length deviations from the expected values and (v) detection of atomic collisions. The derived atom types and bond classes were then validated using high-order moment-based statistical techniques. The results of the statistical analyses were fed back to fine-tune the atom typing. The developed procedure was repeated four times, resulting in fine-grained atom typing, bond and angle classes. The procedure will be repeated in the future as and when new entries are deposited in the COD. The whole procedure can also be applied to any source of small-molecule structures, including the Cambridge Structural Database and the ZINC database.

  2. Extracting respiratory information from seismocardiogram signals acquired on the chest using a miniature accelerometer

    Pandia, Keya; Inan, Omer T; Kovacs, Gregory T A; Giovangrandi, Laurent

    2012-01-01

    Seismocardiography (SCG) is a non-invasive measurement of the vibrations of the chest caused by the heartbeat. SCG signals can be measured using a miniature accelerometer attached to the chest, and are thus well-suited for unobtrusive and long-term patient monitoring. Additionally, SCG contains information relating to both cardiovascular and respiratory systems. In this work, algorithms were developed for extracting three respiration-dependent features of the SCG signal: intensity modulation, timing interval changes within each heartbeat, and timing interval changes between successive heartbeats. Simultaneously with a reference respiration belt, SCG signals were measured from 20 healthy subjects and a respiration rate was estimated using each of the three SCG features and the reference signal. The agreement between each of the three accelerometer-derived respiration rate measurements was computed with respect to the respiration rate derived from the reference respiration belt. The respiration rate obtained from the intensity modulation in the SCG signal was found to be in closest agreement with the respiration rate obtained from the reference respiration belt: the bias was found to be 0.06 breaths per minute with a 95% confidence interval of −0.99 to 1.11 breaths per minute. The limits of agreement between the respiration rates estimated using SCG (intensity modulation) and the reference were within the clinically relevant ranges given in existing literature, demonstrating that SCG could be used for both cardiovascular and respiratory monitoring. Furthermore, phases of each of the three SCG parameters were investigated at four instances of a respiration cycle—start inspiration, peak inspiration, start expiration, and peak expiration—and during breath hold (apnea). The phases of the three SCG parameters observed during the respiration cycle were congruent with existing literature and physiologically expected trends. (paper)

  3. Extracting key information from historical data to quantify the transmission dynamics of smallpox

    Brockmann Stefan O

    2008-08-01

    Full Text Available Abstract Background Quantification of the transmission dynamics of smallpox is crucial for optimizing intervention strategies in the event of a bioterrorist attack. This article reviews basic methods and findings in mathematical and statistical studies of smallpox which estimate key transmission parameters from historical data. Main findings First, critically important aspects in extracting key information from historical data are briefly summarized. We mention different sources of heterogeneity and potential pitfalls in utilizing historical records. Second, we discuss how smallpox spreads in the absence of interventions and how the optimal timing of quarantine and isolation measures can be determined. Case studies demonstrate the following. (1 The upper confidence limit of the 99th percentile of the incubation period is 22.2 days, suggesting that quarantine should last 23 days. (2 The highest frequency (61.8% of secondary transmissions occurs 3–5 days after onset of fever so that infected individuals should be isolated before the appearance of rash. (3 The U-shaped age-specific case fatality implies a vulnerability of infants and elderly among non-immune individuals. Estimates of the transmission potential are subsequently reviewed, followed by an assessment of vaccination effects and of the expected effectiveness of interventions. Conclusion Current debates on bio-terrorism preparedness indicate that public health decision making must account for the complex interplay and balance between vaccination strategies and other public health measures (e.g. case isolation and contact tracing taking into account the frequency of adverse events to vaccination. In this review, we summarize what has already been clarified and point out needs to analyze previous smallpox outbreaks systematically.

  4. Social science and linguistic text analysis of nurses' records: a systematic review and critique.

    Buus, Niels; Hamilton, Bridget Elizabeth

    2016-03-01

    The two aims of the paper were to systematically review and critique social science and linguistic text analyses of nursing records in order to inform future research in this emerging area of research. Systematic searches in reference databases and in citation indexes identified 12 articles that included analyses of the social and linguistic features of records and recording. Two reviewers extracted data using established criteria for the evaluation of qualitative research papers. A common characteristic of nursing records was the economical use of language with local meanings that conveyed little information to the uninitiated reader. Records were dominated by technocratic-medical discourse focused on patients' bodies, and they depicted only very limited aspects of nursing practice. Nurses made moral evaluations in their categorisation of patients, which reflected detailed surveillance of patients' disturbing behaviour. The text analysis methods were rarely transparent in the articles, which could suggest research quality problems. For most articles, the significance of the findings was substantiated more by theoretical readings of the institutional settings than by the analysis of textual data. More probing empirical research of nurses' records and a wider range of theoretical perspectives has the potential to expose the situated meanings of nursing work in healthcare organisations. © 2015 John Wiley & Sons Ltd.

  5. Corpus Linguistics, Network Analysis and Co-occurrence Matrices Corpus Linguistics, Network Analysis and Co-occurrence Matrices

    Keith Stuart

    2009-12-01

    Full Text Available This article describes research undertaken in order to design a methodology for the reticular representation of knowledge of a specific discourse community. To achieve this goal, a representative corpus of the scientific production of the members of this discourse community (Universidad Politécnica de Valencia, UPV was created. The article presents the practical analysis (frequency, keyword, collocation and cluster analysis that was carried out in the initial phases of the study aimed at establishing the theoretical and practical background and framework for our matrix and network analysis of the scientific discourse of the UPV. In the methodology section, the processes that have allowed us to extract from the corpus the linguistic elements needed to develop co-occurrence matrices, as well as the computer tools used in the research, are described. From these co-occurrence matrices, semantic networks of subject and discipline knowledge were generated. Finally, based on the results obtained, we suggest that it may be viable to extract and to represent the intellectual capital of an academic institution using corpus linguistics methods in combination with the formulations of network theory.En este artículo describimos la investigación que se ha desarrollado en el diseño de una metodología para la representación reticular del conocimiento que se genera en el seno de una institución a partir de un corpus representativo de la producción científica de los integrantes de dicha comunidad discursiva, la Universidad Politécnica de Valencia.. Para ello, presentamos las acciones que se realizaron en las fases iniciales del estudio encaminadas a establecer el marco teórico y práctico en el que se inscribe nuestro análisis. En la sección de metodología se describen las herramientas informáticas utilizadas, así como los procesos que nos permitieron disponer de aquellos elementos presentes en el corpus, que nos llevarían al desarrollo de

  6. Clinical Linguistics--Retrospect and Prospect.

    Grunwell, Pamela

    In the past 20 years, linguistics has gained a prominent position in speech and language pathology in Britain, evolving into a new field, clinical linguistics. It includes three related areas of activity: training of speech pathologists/therapists; professional practice; and research. Linguistics and speech/language pathology have developed as…

  7. Quantitative Research in Systemic Functional Linguistics

    He, Qingshun

    2018-01-01

    The research of Systemic Functional Linguistics has been quite in-depth in both theory and practice. However, many linguists hold that Systemic Functional Linguistics has no hypothesis testing or experiments and its research is only qualitative. Analyses of the corpus, intelligent computing and language evolution on the ideological background of…

  8. Evaluating automatically annotated treebanks for linguistic research

    Bloem, J.; Bański, P.; Kupietz, M.; Lüngen, H.; Witt, A.; Barbaresi, A.; Biber, H.; Breiteneder, E.; Clematide, S.

    2016-01-01

    This study discusses evaluation methods for linguists to use when employing an automatically annotated treebank as a source of linguistic evidence. While treebanks are usually evaluated with a general measure over all the data, linguistic studies often focus on a particular construction or a group

  9. The Generic Style Rules for Linguistics

    Haspelmath, Martin

    2014-01-01

    The Generic Style Rules for Linguistics provide a style sheet that can be used by any linguistics journal or edited book, or for teaching purposes. They regulate aspects of text-structure style such as typographic highlighting, citation style, use of capitalization, and bibliographic style (based on the LSA's Unified Stylesheet for linguistics).

  10. Critical and Alternative Directions in Applied Linguistics

    Pennycook, Alastair

    2010-01-01

    Critical directions in applied linguistics can be understood in various ways. The term "critical" as it has been used in "critical applied linguistics," "critical discourse analysis," "critical literacy" and so forth, is now embedded as part of applied linguistic work, adding an overt focus on questions of power and inequality to discourse…

  11. Interdisciplinarity in pragmatics and linguistics

    Mey, Jacob L.

    2017-01-01

    At the Second International Conference ‘Zeichen und System der Sprache’ (Magdeburg, September 1964), a certain East German professor took the floor during a discussion of one of the linguistic presentations. He started his comments by saying: ‘Als Mathematiker weiß ich zwar von der Sache nichts...

  12. Fuzzy linguistic model for interpolation

    Abbasbandy, S.; Adabitabar Firozja, M.

    2007-01-01

    In this paper, a fuzzy method for interpolating of smooth curves was represented. We present a novel approach to interpolate real data by applying the universal approximation method. In proposed method, fuzzy linguistic model (FLM) applied as universal approximation for any nonlinear continuous function. Finally, we give some numerical examples and compare the proposed method with spline method

  13. Desiderata for Linguistic Software Design

    Garretson, Gregory

    2008-01-01

    This article presents a series of guidelines both for researchers in search of software to be used in linguistic analysis and for programmers designing such software. A description of the intended audience and the types of software under consideration and a review of some relevant literature are followed by a discussion of several important…

  14. Saussurean structuralism and cognitive linguistics

    Elffers, E.

    2012-01-01

    Cognitive linguistics (CL) is often regarded as a continuation of Saussurean structuralism. This paper explores the relationship between the two paradigms, focussing on the connection between semantics and views on the language-thought relationship. As it turns out, the similarity in this respect

  15. Pairing Linguistic and Music Intelligences

    DiEdwardo, MaryAnn Pasda

    2005-01-01

    This article describes how music in the language classroom setting can be a catalyst for developing reading, writing, and understanding skills. Studies suggest that pairing music and linguistic intelligences in the college classroom improves students' grades and abilities to compose theses statements for research papers in courses that emphasize…

  16. 140 CIRCULAR INTERACTION BETWEEN LINGUISTIC ...

    economy. Although a country or administrative district should have one or more official languages for obvious reasons, Nelde (1991) proposes that the ... circular interaction between linguistic departments and language departments. Finding an answer to' Plato's abovementioned problem entails that as many languages as ...

  17. Applied Linguistics Research on Asianness

    Kobayashi, Yoko

    2011-01-01

    As China is increasingly occupying the world's attention, its explosively expanding economical and political clout has also been felt in the applied linguistics domain, with the discussion on China's/Chinese language issues growing by leaps and bounds (e.g. China's English education policies, Chinese language classes in the West). Amid the world's…

  18. Applied Linguistics in the Philippines.

    Tucker, G. Richard

    This paper traces the three major developmental strands that converged to contribute to the definition of the applied linguistics field in the Philippines: the institution and capacity-building work supported by the Ford and Rockefeller Foundations; the forging of a vibrant consortium among three Filipino institutions of higher education to offer…

  19. Evaluation of needle trap micro-extraction and solid-phase micro-extraction: Obtaining comprehensive information on volatile emissions from in vitro cultures.

    Oertel, Peter; Bergmann, Andreas; Fischer, Sina; Trefz, Phillip; Küntzel, Anne; Reinhold, Petra; Köhler, Heike; Schubert, Jochen K; Miekisch, Wolfram

    2018-05-14

    Volatile organic compounds (VOCs) emitted from in vitro cultures may reveal information on species and metabolism. Owing to low nmol L -1 concentration ranges, pre-concentration techniques are required for gas chromatography-mass spectrometry (GC-MS) based analyses. This study was intended to compare the efficiency of established micro-extraction techniques - solid-phase micro-extraction (SPME) and needle-trap micro-extraction (NTME) - for the analysis of complex VOC patterns. For SPME, a 75 μm Carboxen®/polydimethylsiloxane fiber was used. The NTME needle was packed with divinylbenzene, Carbopack X and Carboxen 1000. The headspace was sampled bi-directionally. Seventy-two VOCs were calibrated by reference standard mixtures in the range of 0.041-62.24 nmol L -1 by means of GC-MS. Both pre-concentration methods were applied to profile VOCs from cultures of Mycobacterium avium ssp. paratuberculosis. Limits of detection ranged from 0.004 to 3.93 nmol L -1 (median = 0.030 nmol L -1 ) for NTME and from 0.001 to 5.684 nmol L -1 (median = 0.043 nmol L -1 ) for SPME. NTME showed advantages in assessing polar compounds such as alcohols. SPME showed advantages in reproducibility but disadvantages in sensitivity for N-containing compounds. Micro-extraction techniques such as SPME and NTME are well suited for trace VOC profiling over cultures if the limitations of each technique is taken into account. Copyright © 2018 John Wiley & Sons, Ltd.

  20. A COMPARATIVE ANALYSIS OF WEB INFORMATION EXTRACTION TECHNIQUES DEEP LEARNING vs. NAÏVE BAYES vs. BACK PROPAGATION NEURAL NETWORKS IN WEB DOCUMENT EXTRACTION

    J. Sharmila

    2016-01-01

    Full Text Available Web mining related exploration is getting the chance to be more essential these days in view of the reason that a lot of information is overseen through the web. Web utilization is expanding in an uncontrolled way. A particular framework is required for controlling such extensive measure of information in the web space. Web mining is ordered into three noteworthy divisions: Web content mining, web usage mining and web structure mining. Tak-Lam Wong has proposed a web content mining methodology in the exploration with the aid of Bayesian Networks (BN. In their methodology, they were learning on separating the web data and characteristic revelation in view of the Bayesian approach. Roused from their investigation, we mean to propose a web content mining methodology, in view of a Deep Learning Algorithm. The Deep Learning Algorithm gives the interest over BN on the basis that BN is not considered in any learning architecture planning like to propose system. The main objective of this investigation is web document extraction utilizing different grouping algorithm and investigation. This work extricates the data from the web URL. This work shows three classification algorithms, Deep Learning Algorithm, Bayesian Algorithm and BPNN Algorithm. Deep Learning is a capable arrangement of strategies for learning in neural system which is connected like computer vision, speech recognition, and natural language processing and biometrics framework. Deep Learning is one of the simple classification technique and which is utilized for subset of extensive field furthermore Deep Learning has less time for classification. Naive Bayes classifiers are a group of basic probabilistic classifiers in view of applying Bayes hypothesis with concrete independence assumptions between the features. At that point the BPNN algorithm is utilized for classification. Initially training and testing dataset contains more URL. We extract the content presently from the dataset. The

  1. Development, Underdevelopment and Poverty: A Linguistic Study of ...

    Development, Underdevelopment and Poverty: A Linguistic Study of Joseph Edoki's The African Dream. ... The PDF file you selected should load here if your Web browser has a PDF reader plug-in installed (for example, a recent version of Adobe Acrobat Reader). If you would like more information about how to print, save, ...

  2. Linguistic Theory in Practical Lexicography of African Languages

    Mev. R.B. Ruthven

    improve their work by using insights from theoretically-guided linguistic investigations. Our view .... The processes of compiling, editing and extending dictionaries as ... plinary nature of dictionaries, lexicographers should not only follow develop- ... as terminology, information technology, language teaching, translation, psy-.

  3. LINGUISTIC GAFFES IN THE NIGERIAN MEDIA NEWSCAST: AN ...

    Mitch

    It particularly explores cases of linguistic gaffes- wrong usage of .... of primary and secondary sources of information for its analysis. ... cybercrime in the sub region”. ... Lagos” (NTA Network News Broadcast, Sunday 12,June, 2005). .... keep tab on the trendiest of developments in the language in order not to lag behind or.

  4. Persian Linguistic and Rhetorical Pathology in The Telegram Communication Network

    M. Hedayat Mofidi

    2017-12-01

    Full Text Available Today, the mobile phone, with its wide range of features, is an inexpensive, easy to use, and most modern communication tool. One of the special applications of smart phones is providing spaces and facilities such as SMS services, virtual networks, and interactive groups that link people and groups from different linguistic and cultural background. In this paper, we tried to study the messages of the Farsi-speaking users of the Telegram communication network from the linguistics perspective. In this regard, the linguistics and rhetorical correspondence of short messages in Persian language was studied by the content-analysis method. The statistical population consisted of 150 messages which contained 327 sentences. These messages were randomly selected from 5 different groups. Persian language used in cyberspace differs from standard Persian language. The first step in correcting communication language errors in cyberspace is to inform the users about the type of errors.

  5. A Fuzzy Linguistic Methodology to Deal With Unbalanced Linguistic Term Sets

    Herrera, F.; Herrera-Viedma, Enrique; Martinez, L.

    2008-01-01

    Many real problems dealing with qualitative aspects use linguistic approaches to assess such aspects. In most of these problems, a uniform and symmetrical distribution of the linguistic term sets for linguistic modeling is assumed. However, there exist problems whose assessments need to be represented by means of unbalanced linguistic term sets, i.e., using term sets that are not uniformly and symmetrically distributed. The use of linguistic variables implies processes of computing with words...

  6. A Risk Assessment System with Automatic Extraction of Event Types

    Capet, Philippe; Delavallade, Thomas; Nakamura, Takuya; Sandor, Agnes; Tarsitano, Cedric; Voyatzi, Stavroula

    In this article we describe the joint effort of experts in linguistics, information extraction and risk assessment to integrate EventSpotter, an automatic event extraction engine, into ADAC, an automated early warning system. By detecting as early as possible weak signals of emerging risks ADAC provides a dynamic synthetic picture of situations involving risk. The ADAC system calculates risk on the basis of fuzzy logic rules operated on a template graph whose leaves are event types. EventSpotter is based on a general purpose natural language dependency parser, XIP, enhanced with domain-specific lexical resources (Lexicon-Grammar). Its role is to automatically feed the leaves with input data.

  7. OCCASIONAL ADNOMINAL IDIOM MODIFICATION - A COGNITIVE LINGUISTIC APPROACH

    Andreas Langlotz

    2006-06-01

    Full Text Available occasional Adnominal Idiom Modification - A Cognitive Linguistic Approach From a cognitive-linguistic perspective, this paper explores alternative types of adnoniinal modification in occasional variants of English verbal idioms. Being discussed against data extracted from the British National Corpiis (BNC, the model claims that in idioni-production idiomatic constructions are activated as complex linguistic schemas to code a context-specific target-conceptualisation. Adnominal pre- and postmodifications are one specific form of creative alteration to adapt the idiom for this purpose. Semantically, idiom-interna1 NPextension is not a uniforni process. It is necessary to distinguish two systematic types of adnominal modification: external and internal modification (Ernst 1981. While external NPmodification has adverbial function, ¡.e. it modifies the idiom as a unit, internal modification directly applies to the head-noun and thus depends on the degree of motivation and analysability of a given idiom. Following the cognitive-linguistic framework, these dimensions of idiom-transparency result from the language user's ability to remotivate the bipartite semantic structure by conceptual metaphors and metonymies.

  8. The Effects of Linguistic Labels Related to Abstract Scenes on Memory

    Kentaro Inomata

    2011-10-01

    Full Text Available Boundary extension is the false memory beyond the actual boundary of a picture scene. Gagnier (2011 suggested that a linguistic label has no effect on the magnitude of boundary extension. Although she controlled the timing of the presentation or information of the linguistic label, the information of stimulus was not changed. In the present study, the depiction of the main object was controlled in order to change the contextual information of a scene. In experiment, the 68 participants were shown 12 pictures. The stimulus consisted pictures that depicted the main object or did not depict the main object, and half of them were presented with linguistic description. Participants rated the object-less pictures more closely than the original pictures, when the former were presented with linguistic labels. However, when they were presented without linguistic labels, boundary extension did not occur. There was no effect of labels on the pictures that depicted the main objects. On the basis of these results, the linguistic label enhances the representation of the abstract scene like a homogeneous field or a wall. This finding suggests that boundary extension may be affected by not only visual information but also by other sensory information mediated by linguistic representation.

  9. Information extraction from dynamic PS-InSAR time series using machine learning

    van de Kerkhof, B.; Pankratius, V.; Chang, L.; van Swol, R.; Hanssen, R. F.

    2017-12-01

    Due to the increasing number of SAR satellites, with shorter repeat intervals and higher resolutions, SAR data volumes are exploding. Time series analyses of SAR data, i.e. Persistent Scatterer (PS) InSAR, enable the deformation monitoring of the built environment at an unprecedented scale, with hundreds of scatterers per km2, updated weekly. Potential hazards, e.g. due to failure of aging infrastructure, can be detected at an early stage. Yet, this requires the operational data processing of billions of measurement points, over hundreds of epochs, updating this data set dynamically as new data come in, and testing whether points (start to) behave in an anomalous way. Moreover, the quality of PS-InSAR measurements is ambiguous and heterogeneous, which will yield false positives and false negatives. Such analyses are numerically challenging. Here we extract relevant information from PS-InSAR time series using machine learning algorithms. We cluster (group together) time series with similar behaviour, even though they may not be spatially close, such that the results can be used for further analysis. First we reduce the dimensionality of the dataset in order to be able to cluster the data, since applying clustering techniques on high dimensional datasets often result in unsatisfying results. Our approach is to apply t-distributed Stochastic Neighbor Embedding (t-SNE), a machine learning algorithm for dimensionality reduction of high-dimensional data to a 2D or 3D map, and cluster this result using Density-Based Spatial Clustering of Applications with Noise (DBSCAN). The results show that we are able to detect and cluster time series with similar behaviour, which is the starting point for more extensive analysis into the underlying driving mechanisms. The results of the methods are compared to conventional hypothesis testing as well as a Self-Organising Map (SOM) approach. Hypothesis testing is robust and takes the stochastic nature of the observations into account

  10. Synthesis of High-Frequency Ground Motion Using Information Extracted from Low-Frequency Ground Motion

    Iwaki, A.; Fujiwara, H.

    2012-12-01

    Broadband ground motion computations of scenario earthquakes are often based on hybrid methods that are the combinations of deterministic approach in lower frequency band and stochastic approach in higher frequency band. Typical computation methods for low-frequency and high-frequency (LF and HF, respectively) ground motions are the numerical simulations, such as finite-difference and finite-element methods based on three-dimensional velocity structure model, and the stochastic Green's function method, respectively. In such hybrid methods, LF and HF wave fields are generated through two different methods that are completely independent of each other, and are combined at the matching frequency. However, LF and HF wave fields are essentially not independent as long as they are from the same event. In this study, we focus on the relation among acceleration envelopes at different frequency bands, and attempt to synthesize HF ground motion using the information extracted from LF ground motion, aiming to propose a new method for broad-band strong motion prediction. Our study area is Kanto area, Japan. We use the K-NET and KiK-net surface acceleration data and compute RMS envelope at four frequency bands: 0.5-1.0 Hz, 1.0-2.0 Hz, 2.0-4.0 Hz, .0-8.0 Hz, and 8.0-16.0 Hz. Taking the ratio of the envelopes of adjacent bands, we find that the envelope ratios have stable shapes at each site. The empirical envelope-ratio characteristics are combined with low-frequency envelope of the target earthquake to synthesize HF ground motion. We have applied the method to M5-class earthquakes and a M7 target earthquake that occurred in the vicinity of Kanto area, and successfully reproduced the observed HF ground motion of the target earthquake. The method can be applied to a broad band ground motion simulation for a scenario earthquake by combining numerically-computed low-frequency (~1 Hz) ground motion with the empirical envelope ratio characteristics to generate broadband ground motion

  11. Nature Disaster Risk Evaluation with a Group Decision Making Method Based on Incomplete Hesitant Fuzzy Linguistic Preference Relations

    Ming Tang

    2018-04-01

    Full Text Available Because the natural disaster system is a very comprehensive and large system, the disaster reduction scheme must rely on risk analysis. Experts’ knowledge and experiences play a critical role in disaster risk assessment. The hesitant fuzzy linguistic preference relation is an effective tool to express experts’ preference information when comparing pairwise alternatives. Owing to the lack of knowledge or a heavy workload, information may be missed in the hesitant fuzzy linguistic preference relation. Thus, an incomplete hesitant fuzzy linguistic preference relation is constructed. In this paper, we firstly discuss some properties of the additive consistent hesitant fuzzy linguistic preference relation. Next, the incomplete hesitant fuzzy linguistic preference relation, the normalized hesitant fuzzy linguistic preference relation, and the acceptable hesitant fuzzy linguistic preference relation are defined. Afterwards, three procedures to estimate the missing information are proposed. The first one deals with the situation in which there are only n − 1 known judgments involving all the alternatives; the second one is used to estimate the missing information of the hesitant fuzzy linguistic preference relation with more known judgments; while the third procedure is used to deal with ignorance situations in which there is at least one alternative with totally missing information. Furthermore, an algorithm for group decision making with incomplete hesitant fuzzy linguistic preference relations is given. Finally, we illustrate our model with a case study about flood disaster risk evaluation. A comparative analysis is presented to testify the advantage of our method.

  12. Nature Disaster Risk Evaluation with a Group Decision Making Method Based on Incomplete Hesitant Fuzzy Linguistic Preference Relations.

    Tang, Ming; Liao, Huchang; Li, Zongmin; Xu, Zeshui

    2018-04-13

    Because the natural disaster system is a very comprehensive and large system, the disaster reduction scheme must rely on risk analysis. Experts' knowledge and experiences play a critical role in disaster risk assessment. The hesitant fuzzy linguistic preference relation is an effective tool to express experts' preference information when comparing pairwise alternatives. Owing to the lack of knowledge or a heavy workload, information may be missed in the hesitant fuzzy linguistic preference relation. Thus, an incomplete hesitant fuzzy linguistic preference relation is constructed. In this paper, we firstly discuss some properties of the additive consistent hesitant fuzzy linguistic preference relation. Next, the incomplete hesitant fuzzy linguistic preference relation, the normalized hesitant fuzzy linguistic preference relation, and the acceptable hesitant fuzzy linguistic preference relation are defined. Afterwards, three procedures to estimate the missing information are proposed. The first one deals with the situation in which there are only n-1 known judgments involving all the alternatives; the second one is used to estimate the missing information of the hesitant fuzzy linguistic preference relation with more known judgments; while the third procedure is used to deal with ignorance situations in which there is at least one alternative with totally missing information. Furthermore, an algorithm for group decision making with incomplete hesitant fuzzy linguistic preference relations is given. Finally, we illustrate our model with a case study about flood disaster risk evaluation. A comparative analysis is presented to testify the advantage of our method.

  13. Clinical linguistics: conversational reflections.

    Crystal, David

    2013-04-01

    This is a report of the main points I made in an informal "conversation" with Paul Fletcher and the audience at the 14th ICPLA conference in Cork. The observations arose randomly, as part of an unstructured 1-h Q&A, so they do not provide a systematic account of the subject, but simply reflect the issues which were raised by the conference participants during that time.

  14. Non-linguistic Conditions for Causativization as a Linguistic Attractor.

    Nichols, Johanna

    2017-01-01

    An attractor, in complex systems theory, is any state that is more easily or more often entered or acquired than departed or lost; attractor states therefore accumulate more members than non-attractors, other things being equal. In the context of language evolution, linguistic attractors include sounds, forms, and grammatical structures that are prone to be selected when sociolinguistics and language contact make it possible for speakers to choose between competing forms. The reasons why an element is an attractor are linguistic (auditory salience, ease of processing, paradigm structure, etc.), but the factors that make selection possible and propagate selected items through the speech community are non-linguistic. This paper uses the consonants in personal pronouns to show what makes for an attractor and how selection and diffusion work, then presents a survey of several language families and areas showing that the derivational morphology of pairs of verbs like fear and frighten , or Turkish korkmak 'fear, be afraid' and korkutmak 'frighten, scare', or Finnish istua 'sit' and istutta 'seat (someone)', or Spanish sentarse 'sit down' and sentar 'seat (someone)' is susceptible to selection. Specifically, the Turkish and Finnish pattern, where 'seat' is derived from 'sit' by addition of a suffix-is an attractor and a favored target of selection. This selection occurs chiefly in sociolinguistic contexts of what is defined here as linguistic symbiosis, where languages mingle in speech, which in turn is favored by certain demographic, sociocultural, and environmental factors here termed frontier conditions. Evidence is surveyed from northern Eurasia, the Caucasus, North and Central America, and the Pacific and from both modern and ancient languages to raise the hypothesis that frontier conditions and symbiosis favor causativization.

  15. Non-linguistic Conditions for Causativization as a Linguistic Attractor

    Johanna Nichols; Johanna Nichols; Johanna Nichols

    2018-01-01

    An attractor, in complex systems theory, is any state that is more easily or more often entered or acquired than departed or lost; attractor states therefore accumulate more members than non-attractors, other things being equal. In the context of language evolution, linguistic attractors include sounds, forms, and grammatical structures that are prone to be selected when sociolinguistics and language contact make it possible for speakers to choose between competing forms. The reasons why an e...

  16. Non-linguistic Conditions for Causativization as a Linguistic Attractor

    Johanna Nichols

    2018-01-01

    Full Text Available An attractor, in complex systems theory, is any state that is more easily or more often entered or acquired than departed or lost; attractor states therefore accumulate more members than non-attractors, other things being equal. In the context of language evolution, linguistic attractors include sounds, forms, and grammatical structures that are prone to be selected when sociolinguistics and language contact make it possible for speakers to choose between competing forms. The reasons why an element is an attractor are linguistic (auditory salience, ease of processing, paradigm structure, etc., but the factors that make selection possible and propagate selected items through the speech community are non-linguistic. This paper uses the consonants in personal pronouns to show what makes for an attractor and how selection and diffusion work, then presents a survey of several language families and areas showing that the derivational morphology of pairs of verbs like fear and frighten, or Turkish korkmak ‘fear, be afraid’ and korkutmak ‘frighten, scare’, or Finnish istua ‘sit’ and istutta ‘seat (someone’, or Spanish sentarse ‘sit down’ and sentar ‘seat (someone’ is susceptible to selection. Specifically, the Turkish and Finnish pattern, where ‘seat’ is derived from ‘sit’ by addition of a suffix—is an attractor and a favored target of selection. This selection occurs chiefly in sociolinguistic contexts of what is defined here as linguistic symbiosis, where languages mingle in speech, which in turn is favored by certain demographic, sociocultural, and environmental factors here termed frontier conditions. Evidence is surveyed from northern Eurasia, the Caucasus, North and Central America, and the Pacific and from both modern and ancient languages to raise the hypothesis that frontier conditions and symbiosis favor causativization.

  17. TempoWordNet : une ressource lexicale pour l'extraction d'information temporelle

    Hasanuzzaman , Mohammed

    2016-01-01

    The ability to capture the time information conveyed in natural language, where that information is expressed either explicitly, or implicitly, or connotative, is essential to many natural language processing applications such as information retrieval, question answering, automatic summarization, targeted marketing, loan repayment forecasting, and understanding economic patterns. Associating word senses with temporal orientation to grasp the temporal information in language is relatively stra...

  18. Hemispheric association and dissociation of voice and speech information processing in stroke.

    Jones, Anna B; Farrall, Andrew J; Belin, Pascal; Pernet, Cyril R

    2015-10-01

    As we listen to someone speaking, we extract both linguistic and non-linguistic information. Knowing how these two sets of information are processed in the brain is fundamental for the general understanding of social communication, speech recognition and therapy of language impairments. We investigated the pattern of performances in phoneme versus gender categorization in left and right hemisphere stroke patients, and found an anatomo-functional dissociation in the right frontal cortex, establishing a new syndrome in voice discrimination abilities. In addition, phoneme and gender performances were most often associated than dissociated in the left hemisphere patients, suggesting a common neural underpinnings. Copyright © 2015 Elsevier Ltd. All rights reserved.

  19. A lattice-valued linguistic decision model for nuclear safeguards applications

    Ruan, D.; Liu, J.; Carchon, R.

    2001-01-01

    In this study, we focus our attention on decision making models to process uncertainty-based information directly without transforming them into any particular membership function, i.e., directly using linguistic information (linguistic values) instead of numbers (numerical values). By analyzing the feature of linguistic values ordered by their means of common usage, we argue that the set of linguistic values should be characterized by a lattice structure. We propose the lattice structure based on a logical algebraic structure i.e., lattice implication algebra. Finally, we obtain a multi-objective decision-making model by extending Yager's multi-objective model from the following aspects: (1) extension of linguistic information: from a set of linear ordered linguistic labels (values) to that of lattice-valued linguistic labels; (2) extension of the combination function M, which is used to combine the individual ratings with the weights of criteria. We propose an implication operation form of M. The implication operation can be drawn from lattice implication algebra. As an illustration, we will finally apply this decision model to the evaluation problem in safeguard relevant information. (orig.)

  20. The linguistic roots of natural pedagogy.

    Mattos, Otávio; Hinzen, Wolfram

    2015-01-01

    Natural pedagogy is a human-specific capacity that allows us to acquire cultural information from communication even before the emergence of the first words, encompassing three core elements: (i) a sensitivity to ostensive signals like eye contact that indicate to infants that they are being addressed through communication, (ii) a subsequent referential expectation (satisfied by the use of declarative gestures) and (iii) a biased interpretation of ostensive-referential communication as conveying relevant information about the referent's kind (Csibra and Gergely, 2006, 2009, 2011). Remarkably, the link between natural pedagogy and another human-specific capacity, namely language, has rarely been investigated in detail. We here argue that children's production and comprehension of declarative gestures around 10 months of age are in fact expressions of an evolving faculty of language. Through both declarative gestures and ostensive signals, infants can assign the roles of third, second, and first person, building the 'deictic space' that grounds both natural pedagogy and language use. Secondly, we argue that the emergence of two kinds of linguistic structures (i.e., proto-determiner phrases and proto-sentences) in the one-word period sheds light on the different kinds of information that children can acquire or convey at different stages of development (namely, generic knowledge about kinds and knowledge about particular events/actions/state of affairs, respectively). Furthermore, the development of nominal and temporal reference in speech allows children to cognize information in terms of spatial and temporal relations. In this way, natural pedagogy transpires as an inherent aspect of our faculty of language, rather than as an independent adaptation that pre-dates language in evolution or development (Csibra and Gergely, 2006). This hypothesis is further testable through predictions it makes on the different linguistic profiles of toddlers with developmental

  1. The linguistic roots of Natural Pedagogy

    Otávio eMattos

    2015-09-01

    Full Text Available Natural pedagogy is a human-specific capacity that allows us to acquire cultural information from communication even before the emergence of the first words, encompassing three core elements: (i a sensitivity to ostensive signals like eye contact that indicate to infants that they are being addressed through communication, (ii a subsequent referential expectation (satisfied by the use of declarative gestures and (iii a biased interpretation of ostensive-referential communication as conveying relevant information about the referent's kind (Csibra & Gergely, 2011, 2009, 2006. Remarkably, the link between natural pedagogy and another human-specific capacity, namely language, has rarely been investigated in detail. We here argue that children’s production and comprehension of declarative gestures around 10 months of age are in fact expressions of an evolving faculty of language. Through both declarative gestures and ostensive signals, infants can assign the roles of 3rd , 2nd and 1st person, building the ‘deictic space’ that grounds both natural pedagogy and language use. Secondly, we argue that the emergence of two kinds of linguistic structures (i.e. proto-determiner phrases and proto-sentences in the one-word period sheds light on the different kinds of information that children can acquire or convey at different stages of development (namely, generic knowledge about kinds and knowledge about particular events/actions/state of affairs, respectively. Furthermore, the development of nominal and temporal reference in speech allows children to cognize information in terms of spatial and temporal relations. In this way, natural pedagogy transpires as an inherent aspect of our faculty of language, rather than as an independent adaptation that pre-dates language in evolution or development (Csibra & Gergely, 2006. This hypothesis is further testable through predictions it makes on the different linguistic profiles of toddlers with developmental

  2. BioSimplify: an open source sentence simplification engine to improve recall in automatic biomedical information extraction

    Jonnalagadda, Siddhartha; Gonzalez, Graciela

    2011-01-01

    BioSimplify is an open source tool written in Java that introduces and facilitates the use of a novel model for sentence simplification tuned for automatic discourse analysis and information extraction (as opposed to sentence simplification for improving human readability). The model is based on a "shot-gun" approach that produces many different (simpler) versions of the original sentence by combining variants of its constituent elements. This tool is optimized for processing biomedical scien...

  3. The BEL information extraction workflow (BELIEF): evaluation in the BioCreative V BEL and IAT track

    Madan, Sumit; Hodapp, Sven; Senger, Philipp; Ansari, Sam; Szostak, Justyna; Hoeng, Julia; Peitsch, Manuel; Fluck, Juliane

    2016-01-01

    Network-based approaches have become extremely important in systems biology to achieve a better understanding of biological mechanisms. For network representation, the Biological Expression Language (BEL) is well designed to collate findings from the scientific literature into biological network models. To facilitate encoding and biocuration of such findings in BEL, a BEL Information Extraction Workflow (BELIEF) was developed. BELIEF provides a web-based curation interface, the BELIEF Dashboa...

  4. An Investigation of the Relationship Between Automated Machine Translation Evaluation Metrics and User Performance on an Information Extraction Task

    2007-01-01

    more reliable than BLEU and that it is easier to understand in terms familiar to NLP researchers. 19 2.2.3 METEOR Researchers at Carnegie Mellon...essential elements of infor- mation from output generated by three types of Arabic -English MT engines. The information extraction experiment was one of three...reviewing the task hierarchy and examining the MT output of several engines. A small, prior pilot experiment to evaluate Arabic -English MT engines for

  5. Non-linguistic learning in aphasia: Effects of training method and stimulus characteristics

    Vallila-Rohter, Sofia; Kiran, Swathi

    2013-01-01

    Purpose The purpose of the current study was to explore non-linguistic learning ability in patients with aphasia, examining the impact of stimulus typicality and feedback on success with learning. Method Eighteen patients with aphasia and eight healthy controls participated in this study. All participants completed four computerized, non-linguistic category-learning tasks. We probed learning ability under two methods of instruction: feedback-based (FB) and paired-associate (PA). We also examined the impact of task complexity on learning ability, comparing two stimulus conditions: typical (Typ) and atypical (Atyp). Performance was compared between groups and across conditions. Results Results demonstrated that healthy controls were able to successfully learn categories under all conditions. For our patients with aphasia, two patterns of performance arose. One subgroup of patients was able to maintain learning across task manipulations and conditions. The other subgroup of patients demonstrated a sensitivity to task complexity, learning successfully only in the typical training conditions. Conclusions Results support the hypothesis that impairments of general learning are present in aphasia. Some patients demonstrated the ability to extract category information under complex training conditions, while others learned only under conditions that were simplified and emphasized salient category features. Overall, the typical training condition facilitated learning for all participants. Findings have implications for therapy, which are discussed. PMID:23695914

  6. Comparison of Qinzhou bay wetland landscape information extraction by three methods

    X. Chang

    2014-04-01

    and OO is 219 km2, 193.70 km2, 217.40 km2 respectively. The result indicates that SC is in the f irst place, followed by OO approach, and the third DT method when used to extract Qingzhou Bay coastal wetland.

  7. Extracting topographic structure from digital elevation data for geographic information-system analysis

    Jenson, Susan K.; Domingue, Julia O.

    1988-01-01

    Software tools have been developed at the U.S. Geological Survey's EROS Data Center to extract topographic structure and to delineate watersheds and overland flow paths from digital elevation models. The tools are specialpurpose FORTRAN programs interfaced with general-purpose raster and vector spatial analysis and relational data base management packages.

  8. A primer in macromolecular linguistics.

    Searls, David B

    2013-03-01

    Polymeric macromolecules, when viewed abstractly as strings of symbols, can be treated in terms of formal language theory, providing a mathematical foundation for characterizing such strings both as collections and in terms of their individual structures. In addition this approach offers a framework for analysis of macromolecules by tools and conventions widely used in computational linguistics. This article introduces the ways that linguistics can be and has been applied to molecular biology, covering the relevant formal language theory at a relatively nontechnical level. Analogies between macromolecules and human natural language are used to provide intuitive insights into the relevance of grammars, parsing, and analysis of language complexity to biology. Copyright © 2012 Wiley Periodicals, Inc.

  9. Native Speakers in Linguistic Imperialism

    Phillipson, Robert

    2016-01-01

    An investigation of Native English Speaking Teachers’ performance in schemes in six Asian contexts, commissioned by the British Council, and undertaken by three British academics, is subjected to critical evaluation. Key issues for exploration are the issue of a monolingual approach to English le...... the economic and geopolitical agenda behind this English teaching business, there is clear evidence of linguistic imperialism in the functions of this global professional service. These activities serve to strengthen Western interests.......An investigation of Native English Speaking Teachers’ performance in schemes in six Asian contexts, commissioned by the British Council, and undertaken by three British academics, is subjected to critical evaluation. Key issues for exploration are the issue of a monolingual approach to English...... learning and teaching, and the inappropriate qualifications of those sent to education systems when they are unfamiliar with the learners’ languages, cultures, and pedagogical traditions. Whether the schemes involved constitute linguistic imperialismis analysed. Whereas the need for multilingual competence...

  10. Conversation Analysis in Applied Linguistics

    Kasper, Gabriele; Wagner, Johannes

    2014-01-01

    on applied CA, the application of basic CA's principles, methods, and findings to the study of social domains and practices that are interactionally constituted. We consider three strands—foundational, social problem oriented, and institutional applied CA—before turning to recent developments in CA research...... on learning and development. In conclusion, we address some emerging themes in the relationship of CA and applied linguistics, including the role of multilingualism, standard social science methods as research objects, CA's potential for direct social intervention, and increasing efforts to complement CA......For the last decade, conversation analysis (CA) has increasingly contributed to several established fields in applied linguistics. In this article, we will discuss its methodological contributions. The article distinguishes between basic and applied CA. Basic CA is a sociological endeavor concerned...

  11. Systematically extracting metal- and solvent-related occupational information from free-text responses to lifetime occupational history questionnaires.

    Friesen, Melissa C; Locke, Sarah J; Tornow, Carina; Chen, Yu-Cheng; Koh, Dong-Hee; Stewart, Patricia A; Purdue, Mark; Colt, Joanne S

    2014-06-01

    Lifetime occupational history (OH) questionnaires often use open-ended questions to capture detailed information about study participants' jobs. Exposure assessors use this information, along with responses to job- and industry-specific questionnaires, to assign exposure estimates on a job-by-job basis. An alternative approach is to use information from the OH responses and the job- and industry-specific questionnaires to develop programmable decision rules for assigning exposures. As a first step in this process, we developed a systematic approach to extract the free-text OH responses and convert them into standardized variables that represented exposure scenarios. Our study population comprised 2408 subjects, reporting 11991 jobs, from a case-control study of renal cell carcinoma. Each subject completed a lifetime OH questionnaire that included verbatim responses, for each job, to open-ended questions including job title, main tasks and activities (task), tools and equipment used (tools), and chemicals and materials handled (chemicals). Based on a review of the literature, we identified exposure scenarios (occupations, industries, tasks/tools/chemicals) expected to involve possible exposure to chlorinated solvents, trichloroethylene (TCE) in particular, lead, and cadmium. We then used a SAS macro to review the information reported by study participants to identify jobs associated with each exposure scenario; this was done using previously coded standardized occupation and industry classification codes, and a priori lists of associated key words and phrases related to possibly exposed tasks, tools, and chemicals. Exposure variables representing the occupation, industry, and task/tool/chemicals exposure scenarios were added to the work history records of the study respondents. Our identification of possibly TCE-exposed scenarios in the OH responses was compared to an expert's independently assigned probability ratings to evaluate whether we missed identifying

  12. Linguistic complex networks as a young field of quantitative linguistics. Comment on "Approaching human language with complex networks" by J. Cong and H. Liu

    Köhler, Reinhard

    2014-12-01

    We have long been used to the domination of qualitative methods in modern linguistics. Indeed, qualitative methods have advantages such as ease of use and wide applicability to many types of linguistic phenomena. However, this shall not overshadow the fact that a great part of human language is amenable to quantification. Moreover, qualitative methods may lead to over-simplification by employing the rigid yes/no scale. When variability and vagueness of human language must be taken into account, qualitative methods will prove inadequate and give way to quantitative methods [1, p. 11]. In addition to such advantages as exactness and precision, quantitative concepts and methods make it possible to find laws of human language which are just like those in natural sciences. These laws are fundamental elements of linguistic theories in the spirit of the philosophy of science [2,3]. Theorization effort of this type is what quantitative linguistics [1,4,5] is devoted to. The review of Cong and Liu [6] has provided an informative and insightful survey of linguistic complex networks as a young field of quantitative linguistics, including the basic concepts and measures, the major lines of research with linguistic motivation, and suggestions for future research.

  13. Linguistics, cognitive psychology, and the Now-or-Never bottleneck.

    Endress, Ansgar D; Katzir, Roni

    2016-01-01

    Christiansen & Chater (C&C)'s key premise is that "if linguistic information is not processed rapidly, that information is lost for good" (sect. 1, para. 1). From this "Now-or-Never bottleneck" (NNB), C&C derive "wide-reaching and fundamental implications for language processing, acquisition and change as well as for the structure of language itself" (sect. 2, para. 10). We question both the premise and the consequentiality of its purported implications.

  14. MIDAS. An algorithm for the extraction of modal information from experimentally determined transfer functions

    Durrans, R.F.

    1978-12-01

    In order to design reactor structures to withstand the large flow and acoustic forces present it is necessary to know something of their dynamic properties. In many cases these properties cannot be predicted theoretically and it is necessary to determine them experimentally. The algorithm MIDAS (Modal Identification for the Dynamic Analysis of Structures) which has been developed at B.N.L. for extracting these structural properties from experimental data is described. (author)

  15. A fuzzy linguistic interface for data bases in nuclear safety problems

    Lyapin, B.; Ryjov, A.; Moscow Univ.

    1994-01-01

    This work describes the idea of a fuzzy linguistic interface for large-scale data bases, allowing to effectively handle a large amount of information. This effect is reached by providing an opportunity to search information on the basis of generalised concepts, or in other words, linguistic descriptions. These concepts are formulated by the user in natural language, and modelled by fuzzy sets, defined on the universe of the significances of the characteristics of the data base objects

  16. Lexicography and Linguistic Creativity

    Rosamund Moon

    2011-10-01

    Full Text Available

    Abstract: Conventionally, dictionaries present information about institutionalized words,phrases, and senses of words; more creative formations and usages are generally ignored. Yet textand corpus data provide ample evidence of creativity in language, showing that it is part of ordinarylinguistic behaviour and indeed often systematic.This article looks at four specific types of lexical creativity in English: figurative meaning,word formation, idioms, and spelling. Focusing on selected examples, it discusses corpus evidenceand then treatment in (principally three recent monolingual dictionaries for learners of English. Itargues that, even taking into account the pedagogical function and limited scope of these dictionaries,more could be said about creative aspects of lexis, and the systematicity of creative usage.This would be of benefit and interest to dictionary users, and empower them.

    Keywords: AFFIXATION, CORPUS, CREATIVITY, DICTIONARY COVERAGE, ENGLISH,FIGURATIVE LANGUAGE, IDIOMS, LEXICOGRAPHY, NEOLOGISM, NORMATIVENESS,SPELLING, WORD FORMATION

    Opsomming: Leksikografie en taalkreatiwiteit. Normaalweg verskaf woordeboekeinligting oor geïnstitutionaliseerde woorde, frases en betekenisse van woorde; meer kreatiewe vormingeen gebruike word gewoonlik geïgnoreer. Tog bied teks- en korpusgegewens volop bewysevan kreatiwiteit in taal, wat toon dat dit deel van gewone taalkundige gedrag is en inderdaaddikwels sistematies.Hierdie artikel beskou vier spesifieke soorte leksikale kreatiwiteit in Engels: figuurlike betekenis,woordvorming, idiome, en spelling. Deur op uitgesoekte voorbeelde te fokus, bespreek ditkorpusbewyse en daarna behandeling in (hoofsaaklik drie resente eentalige woordeboeke viraanleerders van Engels. Dit voer aan dat, selfs al word die opvoedkundige funksie en beperkteomvang van hierdie woordeboeke in ag geneem, meer gesê sou kon word oor die kreatieweaspekte van leksis, en die sistematisiteit van kreatiewe gebruik

  17. Extracting Information about the Initial State from the Black Hole Radiation.

    Lochan, Kinjalk; Padmanabhan, T

    2016-02-05

    The crux of the black hole information paradox is related to the fact that the complete information about the initial state of a quantum field in a collapsing spacetime is not available to future asymptotic observers, belying the expectations from a unitary quantum theory. We study the imprints of the initial quantum state contained in a specific class of distortions of the black hole radiation and identify the classes of in states that can be partially or fully reconstructed from the information contained within. Even for the general in state, we can uncover some specific information. These results suggest that a classical collapse scenario ignores this richness of information in the resulting spectrum and a consistent quantum treatment of the entire collapse process might allow us to retrieve much more information from the spectrum of the final radiation.

  18. Point Cloud Classification of Tesserae from Terrestrial Laser Data Combined with Dense Image Matching for Archaeological Information Extraction

    Poux, F.; Neuville, R.; Billen, R.

    2017-08-01

    Reasoning from information extraction given by point cloud data mining allows contextual adaptation and fast decision making. However, to achieve this perceptive level, a point cloud must be semantically rich, retaining relevant information for the end user. This paper presents an automatic knowledge-based method for pre-processing multi-sensory data and classifying a hybrid point cloud from both terrestrial laser scanning and dense image matching. Using 18 features including sensor's biased data, each tessera in the high-density point cloud from the 3D captured complex mosaics of Germigny-des-prés (France) is segmented via a colour multi-scale abstraction-based featuring extracting connectivity. A 2D surface and outline polygon of each tessera is generated by a RANSAC plane extraction and convex hull fitting. Knowledge is then used to classify every tesserae based on their size, surface, shape, material properties and their neighbour's class. The detection and semantic enrichment method shows promising results of 94% correct semantization, a first step toward the creation of an archaeological smart point cloud.

  19. Bridging the Linguistic and Affective Gaps

    Westbrook, Peter Nils; Henriksen, Birgit

    2011-01-01

    feedback sessions. The study also includes a structured, in-depth interview with the informant, which yields very specific and rich data about how one lecturer feels about teaching in English, the informant’s own learning focus and the outcomes of a short language course. The aims of the study are fourfold......This paper reports on a small-scale case study which follows an experienced Danish university lecturer during a tailor-made, one-to-one language course to improve her English language skills for lecturing, consisting of a five-week cycle of observed English-medium lectures and subsequent language......, namely: 1) to describe the informant’s motivation for taking an English course; 2) to compare her affective and perceived linguistic needs with her objective needs; 3) to follow her own language focus areas during the course; and 4) to identify any subjective or objective gains she achieved from...

  20. Lung region extraction based on the model information and the inversed MIP method by using chest CT images

    Tomita, Toshihiro; Miguchi, Ryosuke; Okumura, Toshiaki; Yamamoto, Shinji; Matsumoto, Mitsuomi; Tateno, Yukio; Iinuma, Takeshi; Matsumoto, Toru.

    1997-01-01

    We developed a lung region extraction method based on the model information and the inversed MIP method in the Lung Cancer Screening CT (LSCT). Original model is composed of typical 3-D lung contour lines, a body axis, an apical point, and a convex hull. First, the body axis. the apical point, and the convex hull are automatically extracted from the input image Next, the model is properly transformed to fit to those of input image by the affine transformation. Using the same affine transformation coefficients, typical lung contour lines are also transferred, which correspond to rough contour lines of input image. Experimental results applied for 68 samples showed this method quite promising. (author)

  1. BioSimplify: an open source sentence simplification engine to improve recall in automatic biomedical information extraction.

    Jonnalagadda, Siddhartha; Gonzalez, Graciela

    2010-11-13

    BioSimplify is an open source tool written in Java that introduces and facilitates the use of a novel model for sentence simplification tuned for automatic discourse analysis and information extraction (as opposed to sentence simplification for improving human readability). The model is based on a "shot-gun" approach that produces many different (simpler) versions of the original sentence by combining variants of its constituent elements. This tool is optimized for processing biomedical scientific literature such as the abstracts indexed in PubMed. We tested our tool on its impact to the task of PPI extraction and it improved the f-score of the PPI tool by around 7%, with an improvement in recall of around 20%. The BioSimplify tool and test corpus can be downloaded from https://biosimplify.sourceforge.net.

  2. Unsupervised improvement of named entity extraction in short informal context using disambiguation clues

    Habib, Mena Badieh; van Keulen, Maurice

    2012-01-01

    Short context messages (like tweets and SMS’s) are a potentially rich source of continuously and instantly updated information. Shortness and informality of such messages are challenges for Natural Language Processing tasks. Most efforts done in this direction rely on machine learning techniques

  3. Automated Methods to Extract Patient New Information from Clinical Notes in Electronic Health Record Systems

    Zhang, Rui

    2013-01-01

    The widespread adoption of Electronic Health Record (EHR) has resulted in rapid text proliferation within clinical care. Clinicians' use of copying and pasting functions in EHR systems further compounds this by creating a large amount of redundant clinical information in clinical documents. A mixture of redundant information (especially outdated…

  4. Extracting principles for information management adaptability during crisis response : A dynamic capability view

    Bharosa, N.; Janssen, M.F.W.H.A.

    2010-01-01

    During crises, relief agency commanders have to make decisions in a complex and uncertain environment, requiring them to continuously adapt to unforeseen environmental changes. In the process of adaptation, the commanders depend on information management systems for information. Yet there are still

  5. Corpus linguistics and statistics with R introduction to quantitative methods in linguistics

    Desagulier, Guillaume

    2017-01-01

    This textbook examines empirical linguistics from a theoretical linguist’s perspective. It provides both a theoretical discussion of what quantitative corpus linguistics entails and detailed, hands-on, step-by-step instructions to implement the techniques in the field. The statistical methodology and R-based coding from this book teach readers the basic and then more advanced skills to work with large data sets in their linguistics research and studies. Massive data sets are now more than ever the basis for work that ranges from usage-based linguistics to the far reaches of applied linguistics. This book presents much of the methodology in a corpus-based approach. However, the corpus-based methods in this book are also essential components of recent developments in sociolinguistics, historical linguistics, computational linguistics, and psycholinguistics. Material from the book will also be appealing to researchers in digital humanities and the many non-linguistic fields that use textual data analysis and t...

  6. LINGUISTIC DIVERSITY AT PORTUGUESE TEXTBOOK: SOME CONSIDERATIONS

    Paula Gaida Winch

    2013-12-01

    Full Text Available It is analyzed how linguistic diversity is dealt with in a Portuguese textbook, where two chapters are designated to it. In these, it is pointed out that speaker ethnic origin can be manifested differently by: morphological changes; use of foreign expressions; accent in oral language. In synthesis, the linguistic diversity is dealt with through activities of identification and reproduction of linguistic varieties to be carried out by the students.

  7. English linguistic purism: history, development, criticism

    Grishechko Ovsanna Savvichna

    2015-12-01

    Full Text Available Linguistic purism as an area of linguistic analysis describes the practices of identification and acknowledgement of a certain language variety as more structurally advanced as compared to its other varieties. Linguistic protection is associated with preservation of some abstract, classical, conservative linguistic ideal and performs the regulatory function, above all. The puristic approach to the development of the English language has been subjected to heated debate for several centuries and is reflected in both scientific research and literary texts. Supporters of purification of the English language champion the idea of protection of “pure language”. The idea, however, is actively criticized by opponents.

  8. Extracting protein dynamics information from overlapped NMR signals using relaxation dispersion difference NMR spectroscopy.

    Konuma, Tsuyoshi; Harada, Erisa; Sugase, Kenji

    2015-12-01

    Protein dynamics plays important roles in many biological events, such as ligand binding and enzyme reactions. NMR is mostly used for investigating such protein dynamics in a site-specific manner. Recently, NMR has been actively applied to large proteins and intrinsically disordered proteins, which are attractive research targets. However, signal overlap, which is often observed for such proteins, hampers accurate analysis of NMR data. In this study, we have developed a new methodology called relaxation dispersion difference that can extract conformational exchange parameters from overlapped NMR signals measured using relaxation dispersion spectroscopy. In relaxation dispersion measurements, the signal intensities of fluctuating residues vary according to the Carr-Purcell-Meiboon-Gill pulsing interval, whereas those of non-fluctuating residues are constant. Therefore, subtraction of each relaxation dispersion spectrum from that with the highest signal intensities, measured at the shortest pulsing interval, leaves only the signals of the fluctuating residues. This is the principle of the relaxation dispersion difference method. This new method enabled us to extract exchange parameters from overlapped signals of heme oxygenase-1, which is a relatively large protein. The results indicate that the structural flexibility of a kink in the heme-binding site is important for efficient heme binding. Relaxation dispersion difference requires neither selectively labeled samples nor modification of pulse programs; thus it will have wide applications in protein dynamics analysis.

  9. Extracting protein dynamics information from overlapped NMR signals using relaxation dispersion difference NMR spectroscopy

    Konuma, Tsuyoshi [Icahn School of Medicine at Mount Sinai, Department of Structural and Chemical Biology (United States); Harada, Erisa [Suntory Foundation for Life Sciences, Bioorganic Research Institute (Japan); Sugase, Kenji, E-mail: sugase@sunbor.or.jp, E-mail: sugase@moleng.kyoto-u.ac.jp [Kyoto University, Department of Molecular Engineering, Graduate School of Engineering (Japan)

    2015-12-15

    Protein dynamics plays important roles in many biological events, such as ligand binding and enzyme reactions. NMR is mostly used for investigating such protein dynamics in a site-specific manner. Recently, NMR has been actively applied to large proteins and intrinsically disordered proteins, which are attractive research targets. However, signal overlap, which is often observed for such proteins, hampers accurate analysis of NMR data. In this study, we have developed a new methodology called relaxation dispersion difference that can extract conformational exchange parameters from overlapped NMR signals measured using relaxation dispersion spectroscopy. In relaxation dispersion measurements, the signal intensities of fluctuating residues vary according to the Carr-Purcell-Meiboon-Gill pulsing interval, whereas those of non-fluctuating residues are constant. Therefore, subtraction of each relaxation dispersion spectrum from that with the highest signal intensities, measured at the shortest pulsing interval, leaves only the signals of the fluctuating residues. This is the principle of the relaxation dispersion difference method. This new method enabled us to extract exchange parameters from overlapped signals of heme oxygenase-1, which is a relatively large protein. The results indicate that the structural flexibility of a kink in the heme-binding site is important for efficient heme binding. Relaxation dispersion difference requires neither selectively labeled samples nor modification of pulse programs; thus it will have wide applications in protein dynamics analysis.

  10. Data modelling in corpus linguistics: how low may we go?

    van Velzen, Marjolein H; Nanetti, Luca; de Deyn, Peter P

    2014-06-01

    Corpus linguistics allows researchers to process millions of words. However, the more words we analyse, i.e., the more data we acquire, the more urgent the call for correct data interpretation becomes. In recent years, a number of studies saw the light attempting to profile some prolific authors' linguistic decline, linking this decline to pathological conditions such as Alzheimer's Disease (AD). However, in line with the nature of the (literary) work that was analysed, numbers alone do not suffice to 'tell the story'. The one and only objective of using statistical methods for the analysis of research data is to tell a story--what happened, when, and how. In the present study we describe a computerised but individualised approach to linguistic analysis--we propose a unifying approach, with firm grounds in Information Theory, that, independently from the specific parameter being investigated, guarantees to produce a robust model of the temporal dynamics of an author's linguistic richness over his or her lifetime. We applied this methodology to six renowned authors with an active writing life of four decades or more: Iris Murdoch, Gerard Reve, Hugo Claus, Agatha Christie, P.D. James, and Harry Mulisch. The first three were diagnosed with probable Alzheimer Disease, confirmed post-mortem for Iris Murdoch; this same condition was hypothesized for Agatha Christie. Our analysis reveals different evolutive patterns of lexical richness, in turn plausibly correlated with the authors' different conditions. Copyright © 2013 Elsevier Ltd. All rights reserved.

  11. On Functional Potential of Interrogative Structures in Academic Linguistic Discourse

    Sergey Trofimovich Nefedov

    2015-11-01

    Full Text Available The paper deals with the language of scientific communication in the field of linguistics, namely, with the functional potential of the interrogative structures in the form of direct and embedded questions. From a pragmatic perspective the interrogative structures does not seem to be compatible with the contexts of scientific interaction: scientists do not ask for information, but they offer their own solutions for the problem situations. This is reflected in the extremely low frequency of questions in academic research articles and monographs. Their text frequency works out a little over 2 % of the total number of the text predications and that of direct questions is about 1 %. Therefore, their place in verbalization of scientific knowledge in linguistics is metaphorically characterized in this article as «interrogative prohibition» by analogy with the «prohibitions» of Harald Weinrich who introduced several notions point to rarely used linguistic units: «prohibition of authorization» – das «Ich-Verbot», «narrative prohibition» – das «Erzähl-Verbot»; «prohibition of metaphors»– das «MetaphernVerbot». In its turn, low frequency makes the analyzed structures an effective tool to formulate the discussed problems, enables further argumentation, integrate the current text into overall linguistic discourse, control the development of the argumentation in scientific text and finally to draw the recipient's attention to a crucial or unexpected argument.

  12. Wavelet analysis of molecular dynamics: Efficient extraction of time-frequency information in ultrafast optical processes

    Prior, Javier; Castro, Enrique; Chin, Alex W.; Almeida, Javier; Huelga, Susana F.; Plenio, Martin B.

    2013-01-01

    New experimental techniques based on nonlinear ultrafast spectroscopies have been developed over the last few years, and have been demonstrated to provide powerful probes of quantum dynamics in different types of molecular aggregates, including both natural and artificial light harvesting complexes. Fourier transform-based spectroscopies have been particularly successful, yet “complete” spectral information normally necessitates the loss of all information on the temporal sequence of events in a signal. This information though is particularly important in transient or multi-stage processes, in which the spectral decomposition of the data evolves in time. By going through several examples of ultrafast quantum dynamics, we demonstrate that the use of wavelets provide an efficient and accurate way to simultaneously acquire both temporal and frequency information about a signal, and argue that this greatly aids the elucidation and interpretation of physical process responsible for non-stationary spectroscopic features, such as those encountered in coherent excitonic energy transport

  13. Extracting information from an ensemble of GCMs to reliably assess future global runoff change

    Sperna Weiland, F.C.; Beek, L.P.H. van; Weerts, A.H.; Bierkens, M.F.P.

    2011-01-01

    Future runoff projections derived from different global climate models (GCMs) show large differences. Therefore, within this study the, information from multiple GCMs has been combined to better assess hydrological changes. For projections of precipitation and temperature the Reliability ensemble

  14. The Grammar of Linguistic Semiotics

    Durst-Andersen, Per

    2009-01-01

    The paper presents a new typology of linguistic signs primarily based on Peirce’s sign conception. It is demonstrated that the fundamental simple sign, the symbolic nominal lexeme, has an arbitrary relationship to its object in order to make it omnipotent, that is, open to various possible...... objects (ensured by nouns) and situations (ensured by the verb)--the latter corresponding to Peirce's rhematic sign-- and in addition to the level of assertion--corresponding to Peirce's dicentic sign-- there is a third level at which verbal categories collaborate in order to make a deduction, abduction...... or induction-- corresponding to Peirce's argumentative signs....

  15. Investigation of the Impact of Extracting and Exchanging Health Information by Using Internet and Social Networks.

    Pistolis, John; Zimeras, Stelios; Chardalias, Kostas; Roupa, Zoe; Fildisis, George; Diomidous, Marianna

    2016-06-01

    Social networks (1) have been embedded in our daily life for a long time. They constitute a powerful tool used nowadays for both searching and exchanging information on different issues by using Internet searching engines (Google, Bing, etc.) and Social Networks (Facebook, Twitter etc.). In this paper, are presented the results of a research based on the frequency and the type of the usage of the Internet and the Social Networks by the general public and the health professionals. The objectives of the research were focused on the investigation of the frequency of seeking and meticulously searching for health information in the social media by both individuals and health practitioners. The exchanging of information is a procedure that involves the issues of reliability and quality of information. In this research, by using advanced statistical techniques an effort is made to investigate the participant's profile in using social networks for searching and exchanging information on health issues. Based on the answers 93 % of the people, use the Internet to find information on health-subjects. Considering principal component analysis, the most important health subjects were nutrition (0.719 %), respiratory issues (0.79 %), cardiological issues (0.777%), psychological issues (0.667%) and total (73.8%). The research results, based on different statistical techniques revealed that the 61.2% of the males and 56.4% of the females intended to use the social networks for searching medical information. Based on the principal components analysis, the most important sources that the participants mentioned, were the use of the Internet and social networks for exchanging information on health issues. These sources proved to be of paramount importance to the participants of the study. The same holds for nursing, medical and administrative staff in hospitals.

  16. Forensic Linguistics: The Linguistic Analyst and Expert Witness of Language Evidence in Criminal Trials.

    Jordan, Sherilynn Nidever

    Forensic linguistics (FL) provides consultation to lawyers through the analysis of language evidence during the pre-trial investigation. Evidence commonly analyzed by linguists in criminal cases includes transcripts of police interviews and language crimes (such as bribery) and anonymous or questioned texts. Forensic linguistic testimony is rarely…

  17. Measuring Linguistic Empathy: An Experimental Approach to Connecting Linguistic and Social Psychological Notions of Empathy

    Kann, Trevor

    2017-01-01

    This dissertation investigated the relationship between Linguistic Empathy and Psychological Empathy by implementing a psycholinguistic experiment that measured a person's acceptability ratings of sentences with violations of Linguistic Empathy and correlating them with a measure of the person's Psychological Empathy. Linguistic Empathy…

  18. Right Lateral Cerebellum Represents Linguistic Predictability.

    Lesage, Elise; Hansen, Peter C; Miall, R Chris

    2017-06-28

    Mounting evidence indicates that posterolateral portions of the cerebellum (right Crus I/II) contribute to language processing, but the nature of this role remains unclear. Based on a well-supported theory of cerebellar motor function, which ascribes to the cerebellum a role in short-term prediction through internal modeling, we hypothesize that right cerebellar Crus I/II supports prediction of upcoming sentence content. We tested this hypothesis using event-related fMRI in male and female human subjects by manipulating the predictability of written sentences. Our design controlled for motor planning and execution, as well as for linguistic features and working memory load; it also allowed separation of the prediction interval from the presentation of the final sentence item. In addition, three further fMRI tasks captured semantic, phonological, and orthographic processing to shed light on the nature of the information processed. As hypothesized, activity in right posterolateral cerebellum correlated with the predictability of the upcoming target word. This cerebellar region also responded to prediction error during the outcome of the trial. Further, this region was engaged in phonological, but not semantic or orthographic, processing. This is the first imaging study to demonstrate a right cerebellar contribution in language comprehension independently from motor, cognitive, and linguistic confounds. These results complement our work using other methodologies showing cerebellar engagement in linguistic prediction and suggest that internal modeling of phonological representations aids language production and comprehension. SIGNIFICANCE STATEMENT The cerebellum is traditionally seen as a motor structure that allows for smooth movement by predicting upcoming signals. However, the cerebellum is also consistently implicated in nonmotor functions such as language and working memory. Using fMRI, we identify a cerebellar area that is active when words are predicted and

  19. Amplitude extraction in pseudoscalar-meson photoproduction: towards a situation of complete information

    Nys, Jannes; Vrancx, Tom; Ryckebusch, Jan

    2015-01-01

    A complete set for pseudoscalar-meson photoproduction is a minimum set of observables from which one can determine the underlying reaction amplitudes unambiguously. The complete sets considered in this work involve single- and double-polarization observables. It is argued that for extracting amplitudes from data, the transversity representation of the reaction amplitudes offers advantages over alternate representations. It is shown that with the available single-polarization data for the p(γ,K + )Λ reaction, the energy and angular dependence of the moduli of the normalized transversity amplitudes in the resonance region can be determined to a fair accuracy. Determining the relative phases of the amplitudes from double-polarization observables is far less evident. (paper)

  20. The Analysis of Tree Species Distribution Information Extraction and Landscape Pattern Based on Remote Sensing Images

    Yi Zeng

    2017-08-01

    Full Text Available The forest ecosystem is the largest land vegetation type, which plays the role of unreplacement with its unique value. And in the landscape scale, the research on forest landscape pattern has become the current hot spot, wherein the study of forest canopy structure is very important. They determines the process and the strength of forests energy flow, which influences the adjustments of ecosystem for climate and species diversity to some extent. The extraction of influencing factors of canopy structure and the analysis of the vegetation distribution pattern are especially important. To solve the problems, remote sensing technology, which is superior to other technical means because of its fine timeliness and large-scale monitoring, is applied to the study. Taking Lingkong Mountain as the study area, the paper uses the remote sensing image to analyze the forest distribution pattern and obtains the spatial characteristics of canopy structure distribution, and DEM data are as the basic data to extract the influencing factors of canopy structure. In this paper, pattern of trees distribution is further analyzed by using terrain parameters, spatial analysis tools and surface processes quantitative simulation. The Hydrological Analysis tool is used to build distributed hydrological model, and corresponding algorithm is applied to determine surface water flow path, rivers network and basin boundary. Results show that forest vegetation distribution of dominant tree species present plaque on the landscape scale and their distribution have spatial heterogeneity which is related to terrain factors closely. After the overlay analysis of aspect, slope and forest distribution pattern respectively, the most suitable area for stand growth and the better living condition are obtained.

  1. Linking attentional processes and conceptual problem solving: visual cues facilitate the automaticity of extracting relevant information from diagrams.

    Rouinfar, Amy; Agra, Elise; Larson, Adam M; Rebello, N Sanjay; Loschky, Lester C

    2014-01-01

    This study investigated links between visual attention processes and conceptual problem solving. This was done by overlaying visual cues on conceptual physics problem diagrams to direct participants' attention to relevant areas to facilitate problem solving. Participants (N = 80) individually worked through four problem sets, each containing a diagram, while their eye movements were recorded. Each diagram contained regions that were relevant to solving the problem correctly and separate regions related to common incorrect responses. Problem sets contained an initial problem, six isomorphic training problems, and a transfer problem. The cued condition saw visual cues overlaid on the training problems. Participants' verbal responses were used to determine their accuracy. This study produced two major findings. First, short duration visual cues which draw attention to solution-relevant information and aid in the organizing and integrating of it, facilitate both immediate problem solving and generalization of that ability to new problems. Thus, visual cues can facilitate re-representing a problem and overcoming impasse, enabling a correct solution. Importantly, these cueing effects on problem solving did not involve the solvers' attention necessarily embodying the solution to the problem, but were instead caused by solvers attending to and integrating relevant information in the problems into a solution path. Second, this study demonstrates that when such cues are used across multiple problems, solvers can automatize the extraction of problem-relevant information extraction. These results suggest that low-level attentional selection processes provide a necessary gateway for relevant information to be used in problem solving, but are generally not sufficient for correct problem solving. Instead, factors that lead a solver to an impasse and to organize and integrate problem information also greatly facilitate arriving at correct solutions.

  2. The Application of Chinese High-Spatial Remote Sensing Satellite Image in Land Law Enforcement Information Extraction

    Wang, N.; Yang, R.

    2018-04-01

    Chinese high -resolution (HR) remote sensing satellites have made huge leap in the past decade. Commercial satellite datasets, such as GF-1, GF-2 and ZY-3 images, the panchromatic images (PAN) resolution of them are 2 m, 1 m and 2.1 m and the multispectral images (MS) resolution are 8 m, 4 m, 5.8 m respectively have been emerged in recent years. Chinese HR satellite imagery has been free downloaded for public welfare purposes using. Local government began to employ more professional technician to improve traditional land management technology. This paper focused on analysing the actual requirements of the applications in government land law enforcement in Guangxi Autonomous Region. 66 counties in Guangxi Autonomous Region were selected for illegal land utilization spot extraction with fusion Chinese HR images. The procedure contains: A. Defines illegal land utilization spot type. B. Data collection, GF-1, GF-2, and ZY-3 datasets were acquired in the first half year of 2016 and other auxiliary data were collected in 2015. C. Batch process, HR images were collected for batch preprocessing through ENVI/IDL tool. D. Illegal land utilization spot extraction by visual interpretation. E. Obtaining attribute data with ArcGIS Geoprocessor (GP) model. F. Thematic mapping and surveying. Through analysing 42 counties results, law enforcement officials found 1092 illegal land using spots and 16 suspicious illegal mining spots. The results show that Chinese HR satellite images have great potential for feature information extraction and the processing procedure appears robust.

  3. Implementation of generalized quantum measurements: Superadditive quantum coding, accessible information extraction, and classical capacity limit

    Takeoka, Masahiro; Fujiwara, Mikio; Mizuno, Jun; Sasaki, Masahide

    2004-01-01

    Quantum-information theory predicts that when the transmission resource is doubled in quantum channels, the amount of information transmitted can be increased more than twice by quantum-channel coding technique, whereas the increase is at most twice in classical information theory. This remarkable feature, the superadditive quantum-coding gain, can be implemented by appropriate choices of code words and corresponding quantum decoding which requires a collective quantum measurement. Recently, an experimental demonstration was reported [M. Fujiwara et al., Phys. Rev. Lett. 90, 167906 (2003)]. The purpose of this paper is to describe our experiment in detail. Particularly, a design strategy of quantum-collective decoding in physical quantum circuits is emphasized. We also address the practical implication of the gain on communication performance by introducing the quantum-classical hybrid coding scheme. We show how the superadditive quantum-coding gain, even in a small code length, can boost the communication performance of conventional coding techniques

  4. An Extended TOPSIS Method for Multiple Attribute Decision Making based on Interval Neutrosophic Uncertain Linguistic Variables

    Said Broumi

    2015-03-01

    Full Text Available The interval neutrosophic uncertain linguistic variables can easily express the indeterminate and inconsistent information in real world, and TOPSIS is a very effective decision making method more and more extensive applications. In this paper, we will extend the TOPSIS method to deal with the interval neutrosophic uncertain linguistic information, and propose an extended TOPSIS method to solve the multiple attribute decision making problems in which the attribute value takes the form of the interval neutrosophic uncertain linguistic variables and attribute weight is unknown. Firstly, the operational rules and properties for the interval neutrosophic variables are introduced. Then the distance between two interval neutrosophic uncertain linguistic variables is proposed and the attribute weight is calculated by the maximizing deviation method, and the closeness coefficients to the ideal solution for each alternatives. Finally, an illustrative example is given to illustrate the decision making steps and the effectiveness of the proposed method.

  5. Linguistic spatial classifications of event domains in narratives of crime

    Blake Stephen Howald

    2010-07-01

    Full Text Available Structurally, formal definitions of the linguistic narrative minimally require two temporally linked past-time events. The role of space in this definition, based on spatial language indicating where events occur, is considered optional and non-structural. However, based on narratives with a high frequency of spatial language, recent research has questioned this perspective, suggesting that space is more critical than may be readily apparent. Through an analysis of spatially rich serial criminal narratives, it will be demonstrated that spatial information qualitatively varies relative to narrative events. In particular, statistical classifiers in a supervised machine learning task achieve a 90% accuracy in predicting Pre-Crime, Crime, and Post-Crime events based on spatial (and temporal information. Overall, these results suggest a deeper spatial organization of discourse, which not only provides practical event resolution possibilities, but also challenges traditional formal linguistic definitions of narrative.

  6. Extraction of basic roadway information for non-state roads in Florida : [summary].

    2015-07-01

    The Florida Department of Transportation (FDOT) maintains a map of all the roads in Florida, : containing over one and a half million road links. For planning purposes, a wide variety : of information, such as stop lights, signage, lane number, and s...

  7. Extracting additional risk managers information from a risk assessment of Listeria monocytogenes in deli meats

    Pérez-Rodríguez, F.; Asselt, van E.D.; García-Gimeno, R.M.; Zurera, G.; Zwietering, M.H.

    2007-01-01

    The risk assessment study of Listeria monocytogenes in ready-to-eat foods conducted by the U.S. Food and Drug Administration is an example of an extensive quantitative microbiological risk assessment that could be used by risk analysts and other scientists to obtain information and by managers and

  8. Synthetic aperture radar ship discrimination, generation and latent variable extraction using information maximizing generative adversarial networks

    Schwegmann, Colin P

    2017-07-01

    Full Text Available such as Synthetic Aperture Radar imagery. To aid in the creation of improved machine learning-based ship detection and discrimination methods this paper applies a type of neural network known as an Information Maximizing Generative Adversarial Network. Generative...

  9. Multidisciplinary Approach in Teaching Foreign Languages to Information Security Professionals

    N. M. Nikiforova

    2012-12-01

    Full Text Available The program of teaching foreign languages to information security professionals is aimed at unifying linguistic, extra linguistic and professional information distributed in the contents of the course.

  10. Youth Culture, Language Endangerment and Linguistic Survivance

    Wyman, Leisy

    2012-01-01

    Detailing a decade of life and language use in a remote Alaskan Yup'ik community, Youth Culture, Language Endangerment and Linguistic Survivance provides rare insight into young people's language brokering and Indigenous people's contemporary linguistic ecologies. This book examines how two consecutive groups of youth in a Yup'ik village…

  11. MODERN LINGUISTICS, ITS DEVELOPMENT AND SCOPE.

    LEVIN, SAMUEL R.

    THE DEVELOPMENT OF MODERN LINGUISTICS STARTED WITH JONES' DISCOVERY IN 1786 THAT SANSKRIT IS CLOSELY RELATED TO THE CLASSICAL, GERMANIC, AND CELTIC LANGUAGES, AND HAS ADVANCED TO INCLUDE THE APPLICATION OF COMPUTERS IN LANGUAGE ANALYSIS. THE HIGHLIGHTS OF LINGUISTIC RESEARCH HAVE BEEN DE SAUSSURE'S DISTINCTION BETWEEN THE DIACHRONIC AND THE…

  12. What can literature do for linguistics?

    Nørgaard, Nina

    2007-01-01

      Through analyses of selected passages from James Joyce's Ulysses, this article demonstrates how the challenging of the boundaries between linguistics and literary studies can be more than a one-way process aimed at uncovering linguistic patterns of literary texts. The theoretical basis...

  13. Statistical Measures for Usage-Based Linguistics

    Gries, Stefan Th.; Ellis, Nick C.

    2015-01-01

    The advent of usage-/exemplar-based approaches has resulted in a major change in the theoretical landscape of linguistics, but also in the range of methodologies that are brought to bear on the study of language acquisition/learning, structure, and use. In particular, methods from corpus linguistics are now frequently used to study distributional…

  14. Exploring Linguistic Identity in Young Multilingual Learners

    Dressler, Roswita

    2014-01-01

    This article explores the linguistic identity of young multilingual learners through the use of a Language Portrait Silhouette. Examples from a research study of children aged 6-8 years in a German bilingual program in Canada provide teachers with an understanding that linguistic identity comprises expertise, affiliation, and inheritance. This…

  15. Applied Linguistics: The Challenge of Theory

    McNamara, Tim

    2015-01-01

    Language has featured prominently in contemporary social theory, but the relevance of this fact to the concerns of Applied Linguistics, with its necessary orientation to practical issues of language in context, represents an ongoing challenge. This article supports the need for a greater engagement with theory in Applied Linguistics. It considers…

  16. Political Liberalism, Linguistic Diversity and Equal Treatment

    Bonotti, Matteo

    2017-01-01

    This article explores the implications of John Rawls' political liberalism for linguistic diversity and language policy, by focusing on the following question: what kind(s) of equality between speakers of different languages and with different linguistic identities should the state guarantee under political liberalism? The article makes three…

  17. Using the Linguistic Landscape to Bridge Languages

    Mari, Vanessa

    2018-01-01

    In this article Vanessa Mari describes how she uses the linguistic landscape to bridge two or more languages with students learning English. The linguistic landscape is defined by Landry and Bourhis (1997, 25) as "the language of public road signs, advertising billboards, street names, place names, commercial shop signs, and public signs on…

  18. Linguistic Recycling and the Open Community.

    Dasgupta, Probal

    2001-01-01

    Examines linguistic recycling in the context of domestic Esperanto use. Argues that word-meaning recycling reflects the same fundamental principles as sentential recursion, and that a linguistics theoretically sensitive to these principles strengthens practical efforts towards the social goal of an open speech community. (Author/VWL)

  19. Stellenbosch Papers in Linguistics: Journal Sponsorship

    Publisher. Stellenbosch Papers in Linguistics (SPiL) is published by the Department of General Linguistics of Stellenbosch University. Publisher contact person: Mrs Christine Smit. Email: linguis@sun.ac.za. Phone: 021 808 2052. Fax: 021 808 2009. Mailing address: Private Bag X1, Matieland, 7602. Department of General ...

  20. Are Prospective English Teachers Linguistically Intelligent?

    Tezel, Kadir Vefa

    2017-01-01

    Language is normally associated with linguistic capabilities of individuals. In the theory of multiple intelligences, language is considered to be related primarily to linguistic intelligence. Using the theory of Multiple Intelligences as its starting point, this descriptive survey study investigated to what extent prospective English teachers'…

  1. A General Overview of Motivation in Linguistics

    王航

    2014-01-01

    In recent years, the term of motivation in linguistics study has aroused the interests of scholars. Different studies of mo -tivation have been produced by different scholars. In this paper, the writer organizes the recent studies on motivation in linguistics. the paper is divided into three parts, the introduction of the term motivation, different types of motivation, and theories of moti -vation.

  2. Ghana Journal of Linguistics: Editorial Policies

    Focus and Scope. The Ghana Journal of Linguistics is a peer-reviewed scholarly journal appearing twice a year, published by the Linguistics Association of Ghana. Beginning with Volume 2 (2013) it is published in electronic format only, open access, at www.ajol.info. However print-on-demand copies can be made ...

  3. Child Participant Roles in Applied Linguistics Research

    Pinter, Annamaria

    2014-01-01

    Children's status as research participants in applied linguistics has been largely overlooked even though unique methodological and ethical concerns arise in projects where children, rather than adults, are involved. This article examines the role of children as research participants in applied linguistics and discusses the limitations of…

  4. The Transition from Animal to Linguistic Communication

    Smit, Harry

    2016-01-01

    Darwin's theory predicts that linguistic behavior gradually evolved out of animal forms of communication (signaling). However, this prediction is confronted by the conceptual problem that there is an essential difference between signaling and linguistic behavior: using words is a normative practice.

  5. Applied Linguistics in Its Disciplinary Context

    Liddicoat, Anthony J.

    2010-01-01

    Australia's current attempt to develop a process to evaluate the quality of research (Excellence in Research for Australia--ERA) places a central emphasis on the disciplinary organisation of academic work. This disciplinary focus poses particular problems for Applied Linguistics in Australia. This paper will examine Applied Linguistics in relation…

  6. Plenary Speeches: Applied Linguists without Borders

    Tarone, Elaine

    2013-01-01

    Until 1989, the American Association for Applied Linguistics (AAAL) could have been viewed as an interest group of the Linguistics Society of America (LSA); AAAL met in two designated meeting rooms as a subsection of the LSA conference. In 1991, I was asked to organize the first independent meeting of AAAL in New York City, with the help of…

  7. Non-linguistic analysis of call center conversations

    Kopparapu, Sunil Kumar

    2014-01-01

    The book focuses on the part of the audio conversation not related to language such as speaking rate (in terms of number of syllables per unit time) and emotion centric features. This text examines using non-linguistics features to infer information from phone calls to call centers. The author analyzes 'how' the conversation happens and not 'what' the conversation is about by audio signal processing and analysis.

  8. International Meeting on Languages, Applied Linguistics and Translation

    Guerra, Luis

    2012-01-01

    This meeting aims at providing an overview of the current theory and practice, exploring new directions and emerging trends, sharing good practice, and exchanging information regarding foreign languages, applied linguistics and translation. The meeting is an excellent opportunity for the presentation of current or previous language learning and translation projects funded by the European Commission or by other sources. Proposals are invited for one or more of the following topics, in any of t...

  9. Linguistic Characteristics of Commercial and Social Advertising Slogans

    Pavel Skorupa; Tatjana Dubovičienė

    2015-01-01

    The current paper presents the analysis of linguistic characteristics of commercial and social advertising slogans. There is no uniform definition of the advertising slogan in the scientific literature, therefore, an attempt to provide the definition of a slogan in the context of marketing communication was made. One of the main functions of both social and commercial advertising is to provide information to the target audience and make it act in a way desired by the advertisers. As language,...

  10. Tracking Anglicisms in Domains by the Corpus-Linguistic Method

    Mousten, Birthe; Laursen, Anne Lise

    2015-01-01

    Lay investors and semi-professionals lean on professional stock bloggers and stock analysts for advice on stock investments; semi-professionals and professionals write about investments globally, and stock information has to be available in many local markets. Using the correct terminology......’s critical sense is not enough to make the right choices. Our corpus-linguistic tool can be a help in this specialized field....

  11. You had me at "Hello": Rapid extraction of dialect information from spoken words.

    Scharinger, Mathias; Monahan, Philip J; Idsardi, William J

    2011-06-15

    Research on the neuronal underpinnings of speaker identity recognition has identified voice-selective areas in the human brain with evolutionary homologues in non-human primates who have comparable areas for processing species-specific calls. Most studies have focused on estimating the extent and location of these areas. In contrast, relatively few experiments have investigated the time-course of speaker identity, and in particular, dialect processing and identification by electro- or neuromagnetic means. We show here that dialect extraction occurs speaker-independently, pre-attentively and categorically. We used Standard American English and African-American English exemplars of 'Hello' in a magnetoencephalographic (MEG) Mismatch Negativity (MMN) experiment. The MMN as an automatic change detection response of the brain reflected dialect differences that were not entirely reducible to acoustic differences between the pronunciations of 'Hello'. Source analyses of the M100, an auditory evoked response to the vowels suggested additional processing in voice-selective areas whenever a dialect change was detected. These findings are not only relevant for the cognitive neuroscience of language, but also for the social sciences concerned with dialect and race perception. Copyright © 2011 Elsevier Inc. All rights reserved.

  12. Extraction of indirectly captured information for use in a comparison of offline pH measurement technologies.

    Ritchie, Elspeth K; Martin, Elaine B; Racher, Andy; Jaques, Colin

    2017-06-10

    Understanding the causes of discrepancies in pH readings of a sample can allow more robust pH control strategies to be implemented. It was found that 59.4% of differences between two offline pH measurement technologies for an historical dataset lay outside an expected instrument error range of ±0.02pH. A new variable, Osmo Res , was created using multiple linear regression (MLR) to extract information indirectly captured in the recorded measurements for osmolality. Principal component analysis and time series analysis were used to validate the expansion of the historical dataset with the new variable Osmo Res . MLR was used to identify variables strongly correlated (p<0.05) with differences in pH readings by the two offline pH measurement technologies. These included concentrations of specific chemicals (e.g. glucose) and Osmo Res, indicating culture medium and bolus feed additions as possible causes of discrepancies between the offline pH measurement technologies. Temperature was also identified as statistically significant. It is suggested that this was a result of differences in pH-temperature compensations employed by the pH measurement technologies. In summary, a method for extracting indirectly captured information has been demonstrated, and it has been shown that competing pH measurement technologies were not necessarily interchangeable at the desired level of control (±0.02pH). Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Flexible and fast: linguistic shortcut affects both shallow and deep conceptual processing.

    Connell, Louise; Lynott, Dermot

    2013-06-01

    Previous research has shown that people use linguistic distributional information during conceptual processing, and that it is especially useful for shallow tasks and rapid responding. Using two conceptual combination tasks, we showed that this linguistic shortcut extends to the processing of novel stimuli, is used in both successful and unsuccessful conceptual processing, and is evident in both shallow and deep conceptual tasks. Specifically, as predicted by the ECCo theory of conceptual combination, people use the linguistic shortcut as a "quick-and-dirty" guide to whether the concepts are likely to combine into a coherent conceptual representation, in both shallow sensibility judgment and deep interpretation generation tasks. Linguistic distributional frequency predicts both the likelihood and the time course of rejecting a novel word compound as nonsensical or uninterpretable. However, it predicts the time course of successful processing only in shallow sensibility judgment, because the deeper conceptual process of interpretation generation does not allow the linguistic shortcut to suffice. Furthermore, the effects of linguistic distributional frequency are independent of any effects of conventional word frequency. We discuss the utility of the linguistic shortcut as a cognitive triage mechanism that can optimize processing in a limited-resource conceptual system.

  14. Combining shallow and deep processing for a robust, fast, deep-linguistic dependency parser

    Schneider, G

    2004-01-01

    This paper describes Pro3Gres, a fast, robust, broad-coverage parser that delivers deep-linguistic grammatical relation structures as output, which are closer to predicate-argument structures and more informative than pure constituency structures. The parser stays as shallow as is possible for each task, combining shallow and deep-linguistic methods by integrating chunking and by expressing the majority of long-distance dependencies in a context-free way. It combines statistical and rule-base...

  15. The Use of Non-linguistic Data in a Terminology and Knowledge Bank

    Madsen, Bodil Nistrup

    ‘symbol’, non-verbal form’ and ‘non-linguistic form’ – are they synonymous designations of one data category or do they designate diff erent data categories? In the presentation we will discuss defi nitions from e.g. ISOcat, ISO 704:2009 and the DanTermBank taxonomy of terminological data categories......, and we will present some thoughts about the relevance of non-linguistic information in a national term bank....

  16. Extracting 3d Semantic Information from Video Surveillance System Using Deep Learning

    Zhang, J. S.; Cao, J.; Mao, B.; Shen, D. Q.

    2018-04-01

    At present, intelligent video analysis technology has been widely used in various fields. Object tracking is one of the important part of intelligent video surveillance, but the traditional target tracking technology based on the pixel coordinate system in images still exists some unavoidable problems. Target tracking based on pixel can't reflect the real position information of targets, and it is difficult to track objects across scenes. Based on the analysis of Zhengyou Zhang's camera calibration method, this paper presents a method of target tracking based on the target's space coordinate system after converting the 2-D coordinate of the target into 3-D coordinate. It can be seen from the experimental results: Our method can restore the real position change information of targets well, and can also accurately get the trajectory of the target in space.

  17. Information extracting and processing with diffraction enhanced imaging of X-ray

    Chen Bo; Chinese Academy of Science, Beijing; Chen Chunchong; Jiang Fan; Chen Jie; Ming Hai; Shu Hang; Zhu Peiping; Wang Junyue; Yuan Qingxi; Wu Ziyu

    2006-01-01

    X-ray imaging at high energies has been used for many years in many fields. Conventional X-ray imaging is based on the different absorption within a sample. It is difficult to distinguish different tissues of a biological sample because of their small difference in absorption. The authors use the diffraction enhanced imaging (DEI) method. The authors took images of absorption, extinction, scattering and refractivity. In the end, the authors presented pictures of high resolution with all these information combined. (authors)

  18. Effective use of latent semantic indexing and computational linguistics in biological and biomedical applications.

    Chen, Hongyu; Martin, Bronwen; Daimon, Caitlin M; Maudsley, Stuart

    2013-01-01

    Text mining is rapidly becoming an essential technique for the annotation and analysis of large biological data sets. Biomedical literature currently increases at a rate of several thousand papers per week, making automated information retrieval methods the only feasible method of managing this expanding corpus. With the increasing prevalence of open-access journals and constant growth of publicly-available repositories of biomedical literature, literature mining has become much more effective with respect to the extraction of biomedically-relevant data. In recent years, text mining of popular databases such as MEDLINE has evolved from basic term-searches to more sophisticated natural language processing techniques, indexing and retrieval methods, structural analysis and integration of literature with associated metadata. In this review, we will focus on Latent Semantic Indexing (LSI), a computational linguistics technique increasingly used for a variety of biological purposes. It is noted for its ability to consistently outperform benchmark Boolean text searches and co-occurrence models at information retrieval and its power to extract indirect relationships within a data set. LSI has been used successfully to formulate new hypotheses, generate novel connections from existing data, and validate empirical data.

  19. Clinical linguistics: its past, present and future.

    Perkins, Michael R

    2011-11-01

    Historiography is a growing area of research within the discipline of linguistics, but so far the subfield of clinical linguistics has received virtually no systematic attention. This article attempts to rectify this by tracing the development of the discipline from its pre-scientific days up to the present time. As part of this, I include the results of a survey of articles published in Clinical Linguistics & Phonetics between 1987 and 2008 which shows, for example, a consistent primary focus on phonetics and phonology at the expense of grammar, semantics and pragmatics. I also trace the gradual broadening of the discipline from its roots in structural linguistics to its current reciprocal relationship with speech and language pathology and a range of other academic disciplines. Finally, I consider the scope of clinical linguistic research in 2011 and assess how the discipline seems likely develop in the future.

  20. Analysis of linguistic terms of variables representing the wave of arterial diameter variation in radial arteries using fuzzy entropies

    Nuno Almirantearena, F; Introzzi, A; Clara, F; Burillo Lopez, P

    2007-01-01

    In this work we use 53 Arterial Diameter Variation (ADV) waves extracted from radial artery of normotense males, along with the values of variables that represent the ADV wave, obtained by means of multivariate analysis. Then, we specify the linguistic variables and the linguistic terms. The variables are fuzzified using triangular and trapezoidal fuzzy numbers. We analyze the fuzziness of the linguistic terms by applying discrete and continuous fuzzy entropies. Finally, we infer which variable presents the greatest disorder associated to the loss of arterial elasticity in radial artery

  1. Architecture and data processing alternatives for the TSE computer. Volume 2: Extraction of topological information from an image by the Tse computer

    Jones, J. R.; Bodenheimer, R. E.

    1976-01-01

    A simple programmable Tse processor organization and arithmetic operations necessary for extraction of the desired topological information are described. Hardware additions to this organization are discussed along with trade-offs peculiar to the tse computing concept. An improved organization is presented along with the complementary software for the various arithmetic operations. The performance of the two organizations is compared in terms of speed, power, and cost. Software routines developed to extract the desired information from an image are included.

  2. Electronic processing of informed consents in a global pharmaceutical company environment.

    Vishnyakova, Dina; Gobeill, Julien; Oezdemir-Zaech, Fatma; Kreim, Olivier; Vachon, Therese; Clade, Thierry; Haenning, Xavier; Mikhailov, Dmitri; Ruch, Patrick

    2014-01-01

    We present an electronic capture tool to process informed consents, which are mandatory recorded when running a clinical trial. This tool aims at the extraction of information expressing the duration of the consent given by the patient to authorize the exploitation of biomarker-related information collected during clinical trials. The system integrates a language detection module (LDM) to route a document into the appropriate information extraction module (IEM). The IEM is based on language-specific sets of linguistic rules for the identification of relevant textual facts. The achieved accuracy of both the LDM and IEM is 99%. The architecture of the system is described in detail.

  3. What do professional forecasters' stock market expectations tell us about herding, information extraction and beauty contests?

    Rangvid, Jesper; Schmeling, M.; Schrimpf, A.

    2013-01-01

    We study how professional forecasters form equity market expectations based on a new micro-level dataset which includes rich cross-sectional information about individual characteristics. We focus on testing whether agents rely on the beliefs of others, i.e., consensus expectations, when forming...... their own forecast. We find strong evidence that the average of all forecasters' beliefs influences an individual's own forecast. This effect is stronger for young and less experienced forecasters as well as forecasters whose pay depends more on performance relative to a benchmark. Further tests indicate...

  4. CLASSIFICATION OF INFORMAL SETTLEMENTS THROUGH THE INTEGRATION OF 2D AND 3D FEATURES EXTRACTED FROM UAV DATA

    C. M. Gevaert

    2016-06-01

    Full Text Available Unmanned Aerial Vehicles (UAVs are capable of providing very high resolution and up-to-date information to support informal settlement upgrading projects. In order to provide accurate basemaps, urban scene understanding through the identification and classification of buildings and terrain is imperative. However, common characteristics of informal settlements such as small, irregular buildings with heterogeneous roof material and large presence of clutter challenge state-of-the-art algorithms. Especially the dense buildings and steeply sloped terrain cause difficulties in identifying elevated objects. This work investigates how 2D radiometric and textural features, 2.5D topographic features, and 3D geometric features obtained from UAV imagery can be integrated to obtain a high classification accuracy in challenging classification problems for the analysis of informal settlements. It compares the utility of pixel-based and segment-based features obtained from an orthomosaic and DSM with point-based and segment-based features extracted from the point cloud to classify an unplanned settlement in Kigali, Rwanda. Findings show that the integration of 2D and 3D features leads to higher classification accuracies.

  5. How Linguistic Metaphor Scaffolds Reasoning.

    Thibodeau, Paul H; Hendricks, Rose K; Boroditsky, Lera

    2017-11-01

    Language helps people communicate and think. Precise and accurate language would seem best suited to achieve these goals. But a close look at the way people actually talk reveals an abundance of apparent imprecision in the form of metaphor: ideas are 'light bulbs', crime is a 'virus', and cancer is an 'enemy' in a 'war'. In this article, we review recent evidence that metaphoric language can facilitate communication and shape thinking even though it is literally false. We first discuss recent experiments showing that linguistic metaphor can guide thought and behavior. Then we explore the conditions under which metaphors are most influential. Throughout, we highlight theoretical and practical implications, as well as key challenges and opportunities for future research. Copyright © 2017 Elsevier Ltd. All rights reserved.

  6. Swearing, Euphemisms, and Linguistic Relativity

    Bowers, Jeffrey S.; Pleydell-Pearce, Christopher W.

    2011-01-01

    Participants read aloud swear words, euphemisms of the swear words, and neutral stimuli while their autonomic activity was measured by electrodermal activity. The key finding was that autonomic responses to swear words were larger than to euphemisms and neutral stimuli. It is argued that the heightened response to swear words reflects a form of verbal conditioning in which the phonological form of the word is directly associated with an affective response. Euphemisms are effective because they replace the trigger (the offending word form) by another word form that expresses a similar idea. That is, word forms exert some control on affect and cognition in turn. We relate these findings to the linguistic relativity hypothesis, and suggest a simple mechanistic account of how language may influence thinking in this context. PMID:21799832

  7. A method to extract quantitative information in analyzer-based x-ray phase contrast imaging

    Pagot, E.; Cloetens, P.; Fiedler, S.; Bravin, A.; Coan, P.; Baruchel, J.; Haertwig, J.; Thomlinson, W.

    2003-01-01

    Analyzer-based imaging is a powerful phase-sensitive technique that generates improved contrast compared to standard absorption radiography. Combining numerically two images taken on either side at ±1/2 of the full width at half-maximum (FWHM) of the rocking curve provides images of 'pure refraction' and of 'apparent absorption'. In this study, a similar approach is made by combining symmetrical images with respect to the peak of the analyzer rocking curve but at general positions, ±α·FWHM. These two approaches do not consider the ultrasmall angle scattering produced by the object independently, which can lead to inconsistent results. An accurate way to separately retrieve the quantitative information intrinsic to the object is proposed. It is based on a statistical analysis of the local rocking curve, and allows one to overcome the problems encountered using the previous approaches

  8. Breast cancer and quality of life: medical information extraction from health forums.

    Opitz, Thomas; Aze, Jérome; Bringay, Sandra; Joutard, Cyrille; Lavergne, Christian; Mollevi, Caroline

    2014-01-01

    Internet health forums are a rich textual resource with content generated through free exchanges among patients and, in certain cases, health professionals. We tackle the problem of retrieving clinically relevant information from such forums, with relevant topics being defined from clinical auto-questionnaires. Texts in forums are largely unstructured and noisy, calling for adapted preprocessing and query methods. We minimize the number of false negatives in queries by using a synonym tool to achieve query expansion of initial topic keywords. To avoid false positives, we propose a new measure based on a statistical comparison of frequent co-occurrences in a large reference corpus (Web) to keep only relevant expansions. Our work is motivated by a study of breast cancer patients' health-related quality of life (QoL). We consider topics defined from a breast-cancer specific QoL-questionnaire. We quantify and structure occurrences in posts of a specialized French forum and outline important future developments.

  9. EXTRACTION OF BENTHIC COVER INFORMATION FROM VIDEO TOWS AND PHOTOGRAPHS USING OBJECT-BASED IMAGE ANALYSIS

    M. T. L. Estomata

    2012-07-01

    Full Text Available Mapping benthic cover in deep waters comprises a very small proportion of studies in the field of research. Majority of benthic cover mapping makes use of satellite images and usually, classification is carried out only for shallow waters. To map the seafloor in optically deep waters, underwater videos and photos are needed. Some researchers have applied this method on underwater photos, but made use of different classification methods such as: Neural Networks, and rapid classification via down sampling. In this study, accurate bathymetric data obtained using a multi-beam echo sounder (MBES was attempted to be used as complementary data with the underwater photographs. Due to the absence of a motion reference unit (MRU, which applies correction to the data gathered by the MBES, accuracy of the said depth data was compromised. Nevertheless, even with the absence of accurate bathymetric data, object-based image analysis (OBIA, which used rule sets based on information such as shape, size, area, relative distance, and spectral information, was still applied. Compared to pixel-based classifications, OBIA was able to classify more specific benthic cover types other than coral and sand, such as rubble and fish. Through the use of rule sets on area, less than or equal to 700 pixels for fish and between 700 to 10,000 pixels for rubble, as well as standard deviation values to distinguish texture, fish and rubble were identified. OBIA produced benthic cover maps that had higher overall accuracy, 93.78±0.85%, as compared to pixel-based methods that had an average accuracy of only 87.30±6.11% (p-value = 0.0001, α = 0.05.

  10. Information Extraction and Interpretation Analysis of Mineral Potential Targets Based on ETM+ Data and GIS technology: A Case Study of Copper and Gold Mineralization in Burma

    Wenhui, Du; Yongqing, Chen; Nana, Guo; Yinglong, Hao; Pengfei, Zhao; Gongwen, Wang

    2014-01-01

    Mineralization-alteration and structure information extraction plays important roles in mineral resource prospecting and assessment using remote sensing data and the Geographical Information System (GIS) technology. Choosing copper and gold mines in Burma as example, the authors adopt band ratio, threshold segmentation and principal component analysis (PCA) to extract the hydroxyl alteration information using ETM+ remote sensing images. Digital elevation model (DEM) (30m spatial resolution) and ETM+ data was used to extract linear and circular faults that are associated with copper and gold mineralization. Combining geological data and the above information, the weights of evidence method and the C-A fractal model was used to integrate and identify the ore-forming favourable zones in this area. Research results show that the high grade potential targets are located with the known copper and gold deposits, and the integrated information can be used to the next exploration for the mineral resource decision-making

  11. The BEL information extraction workflow (BELIEF): evaluation in the BioCreative V BEL and IAT track.

    Madan, Sumit; Hodapp, Sven; Senger, Philipp; Ansari, Sam; Szostak, Justyna; Hoeng, Julia; Peitsch, Manuel; Fluck, Juliane

    2016-01-01

    Network-based approaches have become extremely important in systems biology to achieve a better understanding of biological mechanisms. For network representation, the Biological Expression Language (BEL) is well designed to collate findings from the scientific literature into biological network models. To facilitate encoding and biocuration of such findings in BEL, a BEL Information Extraction Workflow (BELIEF) was developed. BELIEF provides a web-based curation interface, the BELIEF Dashboard, that incorporates text mining techniques to support the biocurator in the generation of BEL networks. The underlying UIMA-based text mining pipeline (BELIEF Pipeline) uses several named entity recognition processes and relationship extraction methods to detect concepts and BEL relationships in literature. The BELIEF Dashboard allows easy curation of the automatically generated BEL statements and their context annotations. Resulting BEL statements and their context annotations can be syntactically and semantically verified to ensure consistency in the BEL network. In summary, the workflow supports experts in different stages of systems biology network building. Based on the BioCreative V BEL track evaluation, we show that the BELIEF Pipeline automatically extracts relationships with an F-score of 36.4% and fully correct statements can be obtained with an F-score of 30.8%. Participation in the BioCreative V Interactive task (IAT) track with BELIEF revealed a systems usability scale (SUS) of 67. Considering the complexity of the task for new users-learning BEL, working with a completely new interface, and performing complex curation-a score so close to the overall SUS average highlights the usability of BELIEF.Database URL: BELIEF is available at http://www.scaiview.com/belief/. © The Author(s) 2016. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  12. Automated Trait Extraction using ClearEarth, a Natural Language Processing System for Text Mining in Natural Sciences

    Thessen,Anne; Preciado,Jenette; Jain,Payoj; Martin,James; Palmer,Martha; Bhat,Riyaz

    2018-01-01

    The cTAKES package (using the ClearTK Natural Language Processing toolkit Bethard et al. 2014, http://cleartk.github.io/cleartk/) has been successfully used to automatically read clinical notes in the medical field (Albright et al. 2013, Styler et al. 2014). It is used on a daily basis to automatically process clinical notes and extract relevant information by dozens of medical institutions. ClearEarth is a collaborative project that brings together computational linguistics and domain scient...

  13. Measuring the diffusion of linguistic change.

    Nerbonne, John

    2010-12-12

    We examine situations in which linguistic changes have probably been propagated via normal contact as opposed to via conquest, recent settlement and large-scale migration. We proceed then from two simplifying assumptions: first, that all linguistic variation is the result of either diffusion or independent innovation, and, second, that we may operationalize social contact as geographical distance. It is clear that both of these assumptions are imperfect, but they allow us to examine diffusion via the distribution of linguistic variation as a function of geographical distance. Several studies in quantitative linguistics have examined this relation, starting with Séguy (Séguy 1971 Rev. Linguist. Romane 35, 335-357), and virtually all report a sublinear growth in aggregate linguistic variation as a function of geographical distance. The literature from dialectology and historical linguistics has mostly traced the diffusion of individual features, however, so that it is sensible to ask what sort of dynamic in the diffusion of individual features is compatible with Séguy's curve. We examine some simulations of diffusion in an effort to shed light on this question.

  14. A Linguistic Multigranular Sensory Evaluation Model for Olive Oil

    Luis Martinez

    2008-06-01

    Full Text Available Evaluation is a process that analyzes elements in order to achieve different objectives such as quality inspection, marketing and other fields in industrial companies. This paper focuses on sensory evaluation where the evaluated items are assessed by a panel of experts according to the knowledge acquired via human senses. In these evaluation processes the information provided by the experts implies uncertainty, vagueness and imprecision. The use of the Fuzzy Linguistic Approach 32 has provided successful results modelling such a type of information. In sensory evaluation it may happen that the panel of experts have more or less degree knowledge of about the evaluated items or indicators. So, it seems suitable that each expert could express their preferences in different linguistic term sets based on their own knowledge. In this paper, we present a sensory evaluation model that manages multigranular linguistic evaluation framework based on a decision analysis scheme. This model will be applied to the sensory evaluation process of Olive Oil.

  15. "Voices of the people": linguistic research among Germany's prisoners of war during World War I.

    Kaplan, Judith

    2013-01-01

    This paper investigates the history of the Royal Prussian Phonographic Commission, a body that collected and archived linguistic, ethnographic, and anthropological data from prisoners-of-war (POWs) in Germany during World War I. Recent literature has analyzed the significance of this research for the rise of conservative physical anthropology. Taking a complementary approach, the essay charts new territory in seeking to understand how the prison-camp studies informed philology and linguistics specifically. I argue that recognizing philological commitments of the Phonographic Commission is essential to comprehending the project contextually. My approach reveals that linguists accommodated material and contemporary evidence to older text-based research models, sustaining dynamic theories of language. Through a case study based on the Iranian philologist F. C. Andreas (1846-1930), the paper ultimately argues that linguistics merits greater recognition in the historiography of the behavioral sciences. © 2013 Wiley Periodicals, Inc.

  16. Keeleliste elulugude uurimisvõimalusi: Dagmar Normeti mitmekeelne lapsepõlv Eestis. Possibilities of Research on Linguistic Biographies: Dagmar Normet, a Multilingual Childhood in Estonia

    Anna Verschik

    2012-04-01

    Full Text Available Recently the investigation of linguistic biographies has become popular among linguists for several reasons. Instead of studying formally-oriented, traditional approaches to second language acquisition and language learning, such research focuses on an individual’s conceptualisation of languages, language acquisition and living with and among multiple languages. Linguistic biographies can be either oral or written narratives, elicited by a researcher or produced by individuals. This includes language-learning memoirs as well. As some studies have demonstrated, a closer look at a linguistic history of a particular individual helps to discover new aspects that generally remain unnoticed in formally-oriented studies, such as the speaker’s personal attitudes, emotions attached to his/ her languages, self-expression in different languages, and instances of multilingual speech (for example, cross-linguistic influence, code-switching, etc.. However, a multilingual person’s narratives, either in written or oral form, should be treated with caution. It has been demonstrated in recent studies that grounded theory approach (i.e., coding and establishing emergent categories and content analysis alone cannot present a full picture of a linguistic biography. As Pavlenko (2007 argues, at least three kinds of reality should be considered: subject reality (how the narrator sees his/her life with multiple languages, text reality (that is, how the text of narration is structured, in what order events are presented and life reality (biographical facts. As in fieldwork in general, a researcher should be prepared to face discrepancies between the picture presented by the informant and other types of reality. From a methodological point of view, an informant should be interviewed several times in his/her different languages or, at the very least; a researcher should be familiar with the languages. In this sense, the European tradition of linguistic biographies

  17. Eodataservice.org: Big Data Platform to Enable Multi-disciplinary Information Extraction from Geospatial Data

    Natali, S.; Mantovani, S.; Barboni, D.; Hogan, P.

    2017-12-01

    In 1999, US Vice-President Al Gore outlined the concept of `Digital Earth' as a multi-resolution, three-dimensional representation of the planet to find, visualise and make sense of vast amounts of geo- referenced information on physical and social environments, allowing to navigate through space and time, accessing historical and forecast data to support scientists, policy-makers, and any other user. The eodataservice platform (http://eodataservice.org/) implements the Digital Earth Concept: eodatasevice is a cross-domain platform that makes available a large set of multi-year global environmental collections allowing data discovery, visualization, combination, processing and download. It implements a "virtual datacube" approach where data stored on distributed data centers are made available via standardized OGC-compliant interfaces. Dedicated web-based Graphic User Interfaces (based on the ESA-NASA WebWorldWind technology) as well as web-based notebooks (e.g. Jupyter notebook), deskop GIS tools and command line interfaces can be used to access and manipulate the data. The platform can be fully customized on users' needs. So far eodataservice has been used for the following thematic applications: High resolution satellite data distribution Land surface monitoring using SAR surface deformation data Atmosphere, ocean and climate applications Climate-health applications Urban Environment monitoring Safeguard of cultural heritage sites Support to farmers and (re)-insurances in the agriculturés field In the current work, the EO Data Service concept is presented as key enabling technology; furthermore various examples are provided to demonstrate the high level of interdisciplinarity of the platform.

  18. Applied linguistics - a science of culture?

    Benke, Gertraud

    2003-01-01

    Full Text Available In this article, the status of applied linguistics as discipline is questioned and problems of establishing it - and other newly formed scientific enterprises like cultural science - as disciplines are discussed. This discussion is contextualized using the author's own experience as applied linguist working in (the institutional structure of Austria. Secondly, applied linguistics is presented as complementing cultural science, with both exploring at times the same phenomena albeit under different perspectives and focussing on different levels of experience. Two examples of research involving such a joint interest with different foci are discussed.

  19. Educational Linguistics and College English Syllabus Design

    LIU Ji-xin

    2016-01-01

    The direct application of linguistic theories to syllabus design gives rise to frequent change of syllabus type in the histo-ry of syllabus development, which makes language teachers feel difficult to adapt to, to adopt and to implement. The recognition and popularization of the new-born discipline educational linguistics servers as a method to ease the situation, especially in the college English syllabus design in China. The development and application of the fruitful achievements in educational linguis-tics is bound to provide us with a more scientific approach to syllabus design in the future.

  20. Linguistic fire and human cognitive powers

    Cowley, Stephen

    2012-01-01

    To view language as a cultural tool challenges much of what claims to be linguistic science while opening up a new people-centred linguistics. On this view, how we speak, think and act depends on, not just brains (or minds), but also cultural traditions. Yet, Everett is conservative: like others...... theory, bodily dynamics themselves act as cues to meaning. Linguistic exostructures resemble tools that constrain how people concert acting-perceiving bodies. The result is unending renewal of verbal structures: like artefacts and institutions, they function to sustain a species-specific cultural ecology...

  1. Collective Variables in Apphed Linguistics Research

    ヘンスリー, ジョール; HENSLEY, Joel

    2011-01-01

    This paper focuses on the key dynamic(al)systems theory concept of collective variables as it relates to developmental research in applied linguistics. Dynamic(al) systems theory is becoming prevalent in linguistic research and in the past two decades has jumped to the forefront of cutting edge in the field. One key concept in dynamic(al) systems theory is that of collective variables. In order to help properly orient this concept in the field of applied linguistics, this paper discusses the ...

  2. A METHOD OF EXTRACTING SHORELINE BASED ON SEMANTIC INFORMATION USING DUAL-LENGTH LiDAR DATA

    C. Yao

    2017-09-01

    Full Text Available Shoreline is a spatial varying separation between water and land. By utilizing dual-wavelength LiDAR point data together with semantic information that shoreline often appears beyond water surface profile and is observable on the beach, the paper generates the shoreline and the details are as follows: (1 Gain the water surface profile: first we obtain water surface by roughly selecting water points based on several features of water body, then apply least square fitting method to get the whole water trend surface. Then we get the ground surface connecting the under -water surface by both TIN progressive filtering method and surface interpolation method. After that, we have two fitting surfaces intersected to get water surface profile of the island. (2 Gain the sandy beach: we grid all points and select the water surface profile grids points as seeds, then extract sandy beach points based on eight-neighborhood method and features, then we get all sandy beaches. (3 Get the island shoreline: first we get the sandy beach shoreline based on intensity information, then we get a threshold value to distinguish wet area and dry area, therefore we get the shoreline of several sandy beaches. In some extent, the shoreline has the same height values within a small area, by using all the sandy shoreline points to fit a plane P, and the intersection line of the ground surface and the shoreline plane P can be regarded as the island shoreline. By comparing with the surveying shoreline, the results show that the proposed method can successfully extract shoreline.

  3. Improving Climate Communication through Comprehensive Linguistic Analyses Using Computational Tools

    Gann, T. M.; Matlock, T.

    2014-12-01

    An important lesson on climate communication research is that there is no single way to reach out and inform the public. Different groups conceptualize climate issues in different ways and different groups have different values and assumptions. This variability makes it extremely difficult to effectively and objectively communicate climate information. One of the main challenges is the following: How do we acquire a better understanding of how values and assumptions vary across groups, including political groups? A necessary starting point is to pay close attention to the linguistic content of messages used across current popular media sources. Careful analyses of that information—including how it is realized in language for conservative and progressive media—may ultimately help climate scientists, government agency officials, journalists and others develop more effective messages. Past research has looked at partisan media coverage of climate change, but little attention has been given to the fine-grained linguistic content of such media. And when researchers have done detailed linguistic analyses, they have relied primarily on hand-coding, an approach that is costly, labor intensive, and time-consuming. Our project, building on recent work on partisan news media (Gann & Matlock, 2014; under review) uses high dimensional semantic analyses and other methods of automated classification techniques from the field of natural language processing to quantify how climate issues are characterized in media sources that differ according to political orientation. In addition to discussing varied linguistic patterns, we share new methods for improving climate communication for varied stakeholders, and for developing better assessments of their effectiveness.

  4. Linguistic measures of chemical diversity and the "keywords" of molecular collections.

    Woźniak, Michał; Wołos, Agnieszka; Modrzyk, Urszula; Górski, Rafał L; Winkowski, Jan; Bajczyk, Michał; Szymkuć, Sara; Grzybowski, Bartosz A; Eder, Maciej

    2018-05-15

    Computerized linguistic analyses have proven of immense value in comparing and searching through large text collections ("corpora"), including those deposited on the Internet - indeed, it would nowadays be hard to imagine browsing the Web without, for instance, search algorithms extracting most appropriate keywords from documents. This paper describes how such corpus-linguistic concepts can be extended to chemistry based on characteristic "chemical words" that span more than traditional functional groups and, instead, look at common structural fragments molecules share. Using these words, it is possible to quantify the diversity of chemical collections/databases in new ways and to define molecular "keywords" by which such collections are best characterized and annotated.

  5. Circular Interaction Between Linguistic Departments And Language ...

    Stellenbosch Papers in Linguistics Plus. Journal Home · ABOUT THIS JOURNAL · Advanced Search · Current Issue · Archives · Journal Home > Vol 21 (1992) >. Log in or Register to get access to full text downloads.

  6. Linguistik und Didaktik (Linguistics and Didactics)

    Mollay, Karl

    1974-01-01

    Briefly summarizes the papers presented at the 10th annual convention of the German Language Institute in Mannheim. The relationship between linguistic research and its applicability in the area of language instruction is discussed. (Text is in German.) (DS)

  7. Glossematik und Linguistik (Glossematics and Linguistics)

    Hoger, Alfons

    1974-01-01

    Provides a short summary on the background, current development and future perspectives of the glossematic theory of language and linguistics, as developed by Hjelmslev and those associated with him (Loosely called "the Danish school"). (Text is in German.) (DS)

  8. Zweiter Linguistischer Orientierungskurs (Second Linguistic Orientation Course)

    Gosewitz, Uta; Wiegand, Herbert Ernst

    1973-01-01

    Report on the Second Linguistic Orientation Course sponsored by the Institut fur deutsche Sprache (Institute for the German Language) and funded by the Volkswagen Foundation; held at Mannheim, West Germany, February 21-March 3, 1972. (RS)

  9. LINGUISTICS AND SECOND LANGUAGE TEACHING: AN ...

    methods, the cognitive code method and the cognitive anti-method, emerged, both drawing on .... sciences; he must have some knowledge of linguistics. ... much as the nature of the organising power that is capable of handling such data.

  10. Design and Practice: Enacting Functional Linguistics.

    Martin, James R.

    2000-01-01

    Draws on experience with a transdisciplinary literacy project in writing development at the secondary level to address the sub-field of "writing-literacy," writing as a linguist working across an applied versus theoretical frontier. (Author/VWL)

  11. The Unbalanced Linguistic Aggregation Operator in Group Decision Making

    Li Zou

    2012-01-01

    Full Text Available Many linguistic aggregation methods have been proposed and applied in the linguistic decision-making problems. In practice, experts need to assess a number of values in a side of reference domain higher than in the other one; that is, experts use unbalanced linguistic values to express their evaluation for problems. In this paper, we propose a new linguistic aggregation operator to deal with unbalanced linguistic values in group decision making, we adopt 2-tuple representation model of linguistic values and linguistic hierarchies to express unbalanced linguistic values, and moreover, we present the unbalanced linguistic ordered weighted geometric operator to aggregate unbalanced linguistic evaluation values; a comparison example is given to show the advantage of our method.

  12. FORENSIC LINGUISTICS: AUTOMATIC WEB AUTHOR IDENTIFICATION

    A. A. Vorobeva

    2016-03-01

    Full Text Available Internet is anonymous, this allows posting under a false name, on behalf of others or simply anonymous. Thus, individuals, criminal or terrorist organizations can use Internet for criminal purposes; they hide their identity to avoid the prosecuting. Existing approaches and algorithms for author identification of web-posts on Russian language are not effective. The development of proven methods, technics and tools for author identification is extremely important and challenging task. In this work the algorithm and software for authorship identification of web-posts was developed. During the study the effectiveness of several classification and feature selection algorithms were tested. The algorithm includes some important steps: 1 Feature extraction; 2 Features discretization; 3 Feature selection with the most effective Relief-f algorithm (to find the best feature set with the most discriminating power for each set of candidate authors and maximize accuracy of author identification; 4 Author identification on model based on Random Forest algorithm. Random Forest and Relief-f algorithms are used to identify the author of a short text on Russian language for the first time. The important step of author attribution is data preprocessing - discretization of continuous features; earlier it was not applied to improve the efficiency of author identification. The software outputs top q authors with maximum probabilities of authorship. This approach is helpful for manual analysis in forensic linguistics, when developed tool is used to narrow the set of candidate authors. For experiments on 10 candidate authors, real author appeared in to top 3 in 90.02% cases, on first place real author appeared in 70.5% of cases.

  13. An audit of the reliability of influenza vaccination and medical information extracted from eHealth records in general practice.

    Regan, Annette K; Gibbs, Robyn A; Effler, Paul V

    2018-05-31

    To evaluate the reliability of information in general practice (GP) electronic health records (EHRs), 2100 adult patients were randomly selected for interview regarding the presence of specific medical conditions and recent influenza vaccination. Agreement between self-report and data extracted from EHRs was compared using Cohen's kappa coefficient (k) and interpreted in accordance with Altman's Kappa Benchmarking criteria; 377 (18%) patients declined participation, and 608 (29%) could not be contacted. Of 1115 (53%) remaining, 856 (77%) were active patients (≥3 visits to the GP practice in the last two years) who provided complete information for analysis. Although a higher proportion of patients self-reported being vaccinated or having a medical condition compared to the EHR (50.7% vs 36.9%, and 39.4% vs 30.3%, respectively), there was "good" agreement between self-report and EHR for both vaccination status (κ = 0.67) and medical conditions (κ = 0.66). These findings suggest EHR may be useful for public health surveillance. Crown Copyright © 2018. Published by Elsevier Ltd. All rights reserved.

  14. Ling An: Linguistic analysis of NPP instructions

    Karlsson, F.; Salo, L. (Helsingfors Univ., Institutionen foer allmaen spraakvetenskap (Finland)); Wahlstroem, B. (VTT (Finland))

    2008-07-15

    The project consists of two sub-projects, 1) to find out whether the available linguistic method SWECG (Swedish Constraint Grammar) might be used for analyzing the safety manuals for Forsmark nuclear power plant, and 2) to find out whether it is possible to create a working system based on the SWECG method. The conclusion of the project is that an applicable linguistic analysis system may be realized by the company Lingsoft Inc., Aabo, Finland. (ln)

  15. A Python Library for Historical Comparative Linguistics

    Moran , Steven; List , Johann-Mattis

    2012-01-01

    Awarded best paper award; International audience; In this talk we will discuss a European Research Council funded collaborative effort to build a Python library for undertaking academic research in historical-comparative linguistics. Our aim of implementing quantitative methods, specifically in Python, is to transform historical-comparative linguistics from a primarily handcrafted scientific scholarly endeavor, performed by individual researchers, into a quantitative and collaborative field o...

  16. Ling An: LINGUISTIC ANALYSIS OF NPP INSTRUCTIONS

    Karlsson, F.; Salo, L.; Wahlstroem, B.

    2008-07-01

    The project consists of two sub-projects, 1) to find out whether the available linguistic method SWECG (Swedish Constraint Grammar) might be used for analyzing the safety manuals for Forsmark nuclear power plant, and 2) to find out whether it is possible to create a working system based on the SWECG method. The conclusion of the project is that an applicable linguistic analysis system may be realized by the company Lingsoft Inc., Aabo, Finland. (ln)

  17. Legal Linguistics as a Mutual Arena for Cooperation: Recent Developments in the Field of Applied Linguistics and Law

    Engberg, Jan

    2013-01-01

    This article reports on some of the recent projects and individual works in the field of Legal Linguistics as examples of cooperation between Applied Linguistics and law. The article starts by discussing relevant prototypical concepts of Legal Linguistics. Legal Linguistics scrutinizes interactions between human beings in the framework of legal…

  18. Formal linguistics as a cue to demographic history.

    Longobardi, Giuseppe; Ceolin, Andrea; Ecay, Aaron; Ghirotto, Silvia; Guardiano, Cristina; Irimia, Monica-Alexandrina; Michelioudakis, Dimitris; Radkevich, Nina; Pettener, Davide; Luiselli, Donata; Barbujani, Guido

    2016-06-20

    Beyond its theoretical success, the development of molecular genetics has brought about the possibility of extraordinary progress in the study of classification and in the inference of the evolutionary history of many species and populations. A major step forward was represented by the availability of extremely large sets of molecular data suited to quantitative and computational treatments. In this paper, we argue that even in cognitive sciences, purely theoretical progress in a discipline such as linguistics may have analogous impact. Thus, exactly on the model of molecular biology, we propose to unify two traditionally unrelated lines of linguistic investigation: 1) the formal study of syntactic variation (parameter theory) in the biolinguistic program; 2) the reconstruction of relatedness among languages (phylogenetic taxonomy). The results of our linguistic analysis have thus been plotted against data from population genetics and the correlations have turned out to be largely significant: given a non-trivial set of languages/populations, the description of their variation provided by the comparison of systematic parametric analysis and molecular anthropology informatively recapitulates their history and relationships. As a result, we can claim that the reality of some parametric model of the language faculty and language acquisition/transmission (more broadly of generative grammar) receives strong and original support from its historical heuristic power. Then, on these grounds, we can begin testing Darwin's prediction that, when properly generated, the trees of human populations and of their languages should eventually turn out to be significantly parallel.

  19. SCHOOL LINGUISTIC CREATIVITY BASED ON SCIENTIFIC GEOGRAPHICAL TEXTS

    VIORICA BLÎNDĂ

    2012-01-01

    Full Text Available The analysis and observation of the natural environment and of the social and economic one, observing phenomena, objects, beings, and geographical events are at the basis of producing geographical scientific texts. The symbols of iconotexts and cartotexts are another source of inspiration for linguistic interpretation. The linguistic creations that we selected for our study are the scientific analysis, the commentary, the characterization, the parallel, the synthesis, epitomizing and abstracting, the scientific communication, the essay, and the scientific description. The representations on maps, photos, graphics and profiles are translated into verbal or written expression in order to render geographical scientific information from diagrams and images through diverse discursive procedures. Through school linguistic creations, teachers develop their students’ observation spirit, in a written and oral form, their geographical thinking through metaphors, they develop and stimulate their students’ imagination and fantasy, their cognitive, reflexive and affective sensitivity, their abilities to express themselves, to present and argument in a scientific way according to different criteria (sufficiency, demonstrative reasoning, lineal reasoning, pros and cons, giving examples, inferential deduction through using truth tables, etc.. Trough description, students give names and define geographical objects and beings (plants, animals, and people according to their form and aspect, they explain toponyms and appellatives, they classify and make hierarchies, they define their identity through processes of differentiation, emblematizing, personification, location in time and space.

  20. DEVELOPING LINGUISTIC SKILLS AND ABILITIES IN EMP STUDENTS

    Nataša Milosavljević

    2012-03-01

    Full Text Available English for medical purposes falls within the category of discipline-specific language learning. It is characterized by specific linguistic features and requires specific study that is possible to carry out using specially designed programmes. Reading skill is one way of learning strategy of English for medical purposes. Most often, students use this skill in order to obtain information for some particular topic from the area of medical science, or because they need some sort of instruction in order to carry out certain task. Vocabulary acquisition plays a very important role in EMP teaching. It is achieved best through learning vocabulary in context, rather than in isolation. One of the most important questions in EMP teaching is related to what grammatical constructions should be analysed and emphasised. Some types of these activities would include, for example, asking students to find examples of one particular structure in the text or fill in the blanks with the missing forms such as tenses, passive forms, prepositions, etc. Speaking skill represents productive skill the aim of which is communication. In order for foreign language communication to be suucessful, a student should know linguistic and cultural features of native speakers and follow certain rules and conventions that are not easy to define. It is of crucial importance that during teaching process students develop strategies and techniques that will help them use linguistic structures fluently, apply language in different situations, take part in discussions, and use acquired vocabulary in accurate and precise way.