WorldWideScience

Sample records for extracting relevant information

  1. Linking attentional processes and conceptual problem solving: visual cues facilitate the automaticity of extracting relevant information from diagrams.

    Science.gov (United States)

    Rouinfar, Amy; Agra, Elise; Larson, Adam M; Rebello, N Sanjay; Loschky, Lester C

    2014-01-01

    This study investigated links between visual attention processes and conceptual problem solving. This was done by overlaying visual cues on conceptual physics problem diagrams to direct participants' attention to relevant areas to facilitate problem solving. Participants (N = 80) individually worked through four problem sets, each containing a diagram, while their eye movements were recorded. Each diagram contained regions that were relevant to solving the problem correctly and separate regions related to common incorrect responses. Problem sets contained an initial problem, six isomorphic training problems, and a transfer problem. The cued condition saw visual cues overlaid on the training problems. Participants' verbal responses were used to determine their accuracy. This study produced two major findings. First, short duration visual cues which draw attention to solution-relevant information and aid in the organizing and integrating of it, facilitate both immediate problem solving and generalization of that ability to new problems. Thus, visual cues can facilitate re-representing a problem and overcoming impasse, enabling a correct solution. Importantly, these cueing effects on problem solving did not involve the solvers' attention necessarily embodying the solution to the problem, but were instead caused by solvers attending to and integrating relevant information in the problems into a solution path. Second, this study demonstrates that when such cues are used across multiple problems, solvers can automatize the extraction of problem-relevant information extraction. These results suggest that low-level attentional selection processes provide a necessary gateway for relevant information to be used in problem solving, but are generally not sufficient for correct problem solving. Instead, factors that lead a solver to an impasse and to organize and integrate problem information also greatly facilitate arriving at correct solutions.

  2. Has Financial Statement Information become Less Relevant?

    DEFF Research Database (Denmark)

    Thinggaard, Frank; Damkier, Jesper

    This paper presents insights into the question of whether accounting information based on the EU’s Accounting Directives has become less value-relevant to investors over time. The study is based on a research design first used by Francis and Schipper (1999), where value-relevance is measured......? The sample is based on non-financial companies listed on the Copenhagen Stock Exchange in the period 1984-2002. Our analyses show that all the applied accounting measures are value-relevant as investment strategies based on the information earn positive market-adjusted returns in our sample period....... The results provide some indication of a decline in the value-relevance of earnings information in the 1984-2001 period, and mixed, but not statistically reliable, evidence for accounting measures where book value information and asset values are also extracted from financial statements. The results seem...

  3. System for selecting relevant information for decision support.

    Science.gov (United States)

    Kalina, Jan; Seidl, Libor; Zvára, Karel; Grünfeldová, Hana; Slovák, Dalibor; Zvárová, Jana

    2013-01-01

    We implemented a prototype of a decision support system called SIR which has a form of a web-based classification service for diagnostic decision support. The system has the ability to select the most relevant variables and to learn a classification rule, which is guaranteed to be suitable also for high-dimensional measurements. The classification system can be useful for clinicians in primary care to support their decision-making tasks with relevant information extracted from any available clinical study. The implemented prototype was tested on a sample of patients in a cardiological study and performs an information extraction from a high-dimensional set containing both clinical and gene expression data.

  4. Extraction of CT dose information from DICOM metadata: automated Matlab-based approach.

    Science.gov (United States)

    Dave, Jaydev K; Gingold, Eric L

    2013-01-01

    The purpose of this study was to extract exposure parameters and dose-relevant indexes of CT examinations from information embedded in DICOM metadata. DICOM dose report files were identified and retrieved from a PACS. An automated software program was used to extract from these files information from the structured elements in the DICOM metadata relevant to exposure. Extracting information from DICOM metadata eliminated potential errors inherent in techniques based on optical character recognition, yielding 100% accuracy.

  5. NAMED ENTITY RECOGNITION FROM BIOMEDICAL TEXT -AN INFORMATION EXTRACTION TASK

    Directory of Open Access Journals (Sweden)

    N. Kanya

    2016-07-01

    Full Text Available Biomedical Text Mining targets the Extraction of significant information from biomedical archives. Bio TM encompasses Information Retrieval (IR and Information Extraction (IE. The Information Retrieval will retrieve the relevant Biomedical Literature documents from the various Repositories like PubMed, MedLine etc., based on a search query. The IR Process ends up with the generation of corpus with the relevant document retrieved from the Publication databases based on the query. The IE task includes the process of Preprocessing of the document, Named Entity Recognition (NER from the documents and Relationship Extraction. This process includes Natural Language Processing, Data Mining techniques and machine Language algorithm. The preprocessing task includes tokenization, stop word Removal, shallow parsing, and Parts-Of-Speech tagging. NER phase involves recognition of well-defined objects such as genes, proteins or cell-lines etc. This process leads to the next phase that is extraction of relationships (IE. The work was based on machine learning algorithm Conditional Random Field (CRF.

  6. Extracting Information from Multimedia Meeting Collections

    OpenAIRE

    Gatica-Perez, Daniel; Zhang, Dong; Bengio, Samy

    2005-01-01

    Multimedia meeting collections, composed of unedited audio and video streams, handwritten notes, slides, and electronic documents that jointly constitute a raw record of complex human interaction processes in the workplace, have attracted interest due to the increasing feasibility of recording them in large quantities, by the opportunities for information access and retrieval applications derived from the automatic extraction of relevant meeting information, and by the challenges that the ext...

  7. Multi-Filter String Matching and Human-Centric Entity Matching for Information Extraction

    Science.gov (United States)

    Sun, Chong

    2012-01-01

    More and more information is being generated in text documents, such as Web pages, emails and blogs. To effectively manage this unstructured information, one broadly used approach includes locating relevant content in documents, extracting structured information and integrating the extracted information for querying, mining or further analysis. In…

  8. Fine-grained information extraction from German transthoracic echocardiography reports.

    Science.gov (United States)

    Toepfer, Martin; Corovic, Hamo; Fette, Georg; Klügl, Peter; Störk, Stefan; Puppe, Frank

    2015-11-12

    Information extraction techniques that get structured representations out of unstructured data make a large amount of clinically relevant information about patients accessible for semantic applications. These methods typically rely on standardized terminologies that guide this process. Many languages and clinical domains, however, lack appropriate resources and tools, as well as evaluations of their applications, especially if detailed conceptualizations of the domain are required. For instance, German transthoracic echocardiography reports have not been targeted sufficiently before, despite of their importance for clinical trials. This work therefore aimed at development and evaluation of an information extraction component with a fine-grained terminology that enables to recognize almost all relevant information stated in German transthoracic echocardiography reports at the University Hospital of Würzburg. A domain expert validated and iteratively refined an automatically inferred base terminology. The terminology was used by an ontology-driven information extraction system that outputs attribute value pairs. The final component has been mapped to the central elements of a standardized terminology, and it has been evaluated according to documents with different layouts. The final system achieved state-of-the-art precision (micro average.996) and recall (micro average.961) on 100 test documents that represent more than 90 % of all reports. In particular, principal aspects as defined in a standardized external terminology were recognized with f 1=.989 (micro average) and f 1=.963 (macro average). As a result of keyword matching and restraint concept extraction, the system obtained high precision also on unstructured or exceptionally short documents, and documents with uncommon layout. The developed terminology and the proposed information extraction system allow to extract fine-grained information from German semi-structured transthoracic echocardiography reports

  9. Advanced applications of natural language processing for performing information extraction

    CERN Document Server

    Rodrigues, Mário

    2015-01-01

    This book explains how can be created information extraction (IE) applications that are able to tap the vast amount of relevant information available in natural language sources: Internet pages, official documents such as laws and regulations, books and newspapers, and social web. Readers are introduced to the problem of IE and its current challenges and limitations, supported with examples. The book discusses the need to fill the gap between documents, data, and people, and provides a broad overview of the technology supporting IE. The authors present a generic architecture for developing systems that are able to learn how to extract relevant information from natural language documents, and illustrate how to implement working systems using state-of-the-art and freely available software tools. The book also discusses concrete applications illustrating IE uses.   ·         Provides an overview of state-of-the-art technology in information extraction (IE), discussing achievements and limitations for t...

  10. Post-processing of Deep Web Information Extraction Based on Domain Ontology

    Directory of Open Access Journals (Sweden)

    PENG, T.

    2013-11-01

    Full Text Available Many methods are utilized to extract and process query results in deep Web, which rely on the different structures of Web pages and various designing modes of databases. However, some semantic meanings and relations are ignored. So, in this paper, we present an approach for post-processing deep Web query results based on domain ontology which can utilize the semantic meanings and relations. A block identification model (BIM based on node similarity is defined to extract data blocks that are relevant to specific domain after reducing noisy nodes. Feature vector of domain books is obtained by result set extraction model (RSEM based on vector space model (VSM. RSEM, in combination with BIM, builds the domain ontology on books which can not only remove the limit of Web page structures when extracting data information, but also make use of semantic meanings of domain ontology. After extracting basic information of Web pages, a ranking algorithm is adopted to offer an ordered list of data records to users. Experimental results show that BIM and RSEM extract data blocks and build domain ontology accurately. In addition, relevant data records and basic information are extracted and ranked. The performances precision and recall show that our proposed method is feasible and efficient.

  11. Information extraction system

    Science.gov (United States)

    Lemmond, Tracy D; Hanley, William G; Guensche, Joseph Wendell; Perry, Nathan C; Nitao, John J; Kidwell, Paul Brandon; Boakye, Kofi Agyeman; Glaser, Ron E; Prenger, Ryan James

    2014-05-13

    An information extraction system and methods of operating the system are provided. In particular, an information extraction system for performing meta-extraction of named entities of people, organizations, and locations as well as relationships and events from text documents are described herein.

  12. Information Needs/Relevance

    OpenAIRE

    Wildemuth, Barbara M.

    2009-01-01

    A user's interaction with a DL is often initiated as the result of the user experiencing an information need of some kind. Aspects of that experience and how it might affect the user's interactions with the DL are discussed in this module. In addition, users continuously make decisions about and evaluations of the materials retrieved from a DL, relative to their information needs. Relevance judgments, and their relationship to the user's information needs, are discussed in this module. Draft

  13. Is Information Still Relevant?

    Science.gov (United States)

    Ma, Lia

    2013-01-01

    Introduction: The term "information" in information science does not share the characteristics of those of a nomenclature: it does not bear a generally accepted definition and it does not serve as the bases and assumptions for research studies. As the data deluge has arrived, is the concept of information still relevant for information…

  14. Relevance: An Interdisciplinary and Information Science Perspective

    Directory of Open Access Journals (Sweden)

    Howard Greisdorf

    2000-01-01

    Full Text Available Although relevance has represented a key concept in the field of information science for evaluating information retrieval effectiveness, the broader context established by interdisciplinary frameworks could provide greater depth and breadth to on-going research in the field. This work provides an overview of the nature of relevance in the field of information science with a cursory view of how cross-disciplinary approaches to relevance could represent avenues for further investigation into the evaluative characteristics of relevance as a means for enhanced understanding of human information behavior.

  15. Multimedia Information Extraction

    CERN Document Server

    Maybury, Mark T

    2012-01-01

    The advent of increasingly large consumer collections of audio (e.g., iTunes), imagery (e.g., Flickr), and video (e.g., YouTube) is driving a need not only for multimedia retrieval but also information extraction from and across media. Furthermore, industrial and government collections fuel requirements for stock media access, media preservation, broadcast news retrieval, identity management, and video surveillance.  While significant advances have been made in language processing for information extraction from unstructured multilingual text and extraction of objects from imagery and vid

  16. Research on Crowdsourcing Emergency Information Extraction of Based on Events' Frame

    Science.gov (United States)

    Yang, Bo; Wang, Jizhou; Ma, Weijun; Mao, Xi

    2018-01-01

    At present, the common information extraction method cannot extract the structured emergency event information accurately; the general information retrieval tool cannot completely identify the emergency geographic information; these ways also do not have an accurate assessment of these results of distilling. So, this paper proposes an emergency information collection technology based on event framework. This technique is to solve the problem of emergency information picking. It mainly includes emergency information extraction model (EIEM), complete address recognition method (CARM) and the accuracy evaluation model of emergency information (AEMEI). EIEM can be structured to extract emergency information and complements the lack of network data acquisition in emergency mapping. CARM uses a hierarchical model and the shortest path algorithm and allows the toponomy pieces to be joined as a full address. AEMEI analyzes the results of the emergency event and summarizes the advantages and disadvantages of the event framework. Experiments show that event frame technology can solve the problem of emergency information drawing and provides reference cases for other applications. When the emergency disaster is about to occur, the relevant departments query emergency's data that has occurred in the past. They can make arrangements ahead of schedule which defense and reducing disaster. The technology decreases the number of casualties and property damage in the country and world. This is of great significance to the state and society.

  17. A Compositional Relevance Model for Adaptive Information Retrieval

    Science.gov (United States)

    Mathe, Nathalie; Chen, James; Lu, Henry, Jr. (Technical Monitor)

    1994-01-01

    There is a growing need for rapid and effective access to information in large electronic documentation systems. Access can be facilitated if information relevant in the current problem solving context can be automatically supplied to the user. This includes information relevant to particular user profiles, tasks being performed, and problems being solved. However most of this knowledge on contextual relevance is not found within the contents of documents, and current hypermedia tools do not provide any easy mechanism to let users add this knowledge to their documents. We propose a compositional relevance network to automatically acquire the context in which previous information was found relevant. The model records information on the relevance of references based on user feedback for specific queries and contexts. It also generalizes such information to derive relevant references for similar queries and contexts. This model lets users filter information by context of relevance, build personalized views of documents over time, and share their views with other users. It also applies to any type of multimedia information. Compared to other approaches, it is less costly and doesn't require any a priori statistical computation, nor an extended training period. It is currently being implemented into the Computer Integrated Documentation system which enables integration of various technical documents in a hypertext framework.

  18. Challenges in Managing Information Extraction

    Science.gov (United States)

    Shen, Warren H.

    2009-01-01

    This dissertation studies information extraction (IE), the problem of extracting structured information from unstructured data. Example IE tasks include extracting person names from news articles, product information from e-commerce Web pages, street addresses from emails, and names of emerging music bands from blogs. IE is all increasingly…

  19. 46 CFR 560.5 - Receipt of relevant information.

    Science.gov (United States)

    2010-10-01

    ... 46 Shipping 9 2010-10-01 2010-10-01 false Receipt of relevant information. 560.5 Section 560.5... FOREIGN PORTS § 560.5 Receipt of relevant information. (a) In making its decision on matters arising under... submissions should be supported by affidavits of fact and memorandum of law. Relevant information may include...

  20. Improving information extraction using a probability-based approach

    DEFF Research Database (Denmark)

    Kim, S.; Ahmed, Saeema; Wallace, K.

    2007-01-01

    Information plays a crucial role during the entire life-cycle of a product. It has been shown that engineers frequently consult colleagues to obtain the information they require to solve problems. However, the industrial world is now more transient and key personnel move to other companies...... or retire. It is becoming essential to retrieve vital information from archived product documents, if it is available. There is, therefore, great interest in ways of extracting relevant and sharable information from documents. A keyword-based search is commonly used, but studies have shown...... the recall, while maintaining the high precision, a learning approach that makes identification decisions based on a probability model, rather than simply looking up the presence of the pre-defined variations, looks promising. This paper presents the results of developing such a probability-based entity...

  1. Signal Enhancement as Minimization of Relevant Information Loss

    OpenAIRE

    Geiger, Bernhard C.; Kubin, Gernot

    2012-01-01

    We introduce the notion of relevant information loss for the purpose of casting the signal enhancement problem in information-theoretic terms. We show that many algorithms from machine learning can be reformulated using relevant information loss, which allows their application to the aforementioned problem. As a particular example we analyze principle component analysis for dimensionality reduction, discuss its optimality, and show that the relevant information loss can indeed vanish if the r...

  2. Natural brain-information interfaces: Recommending information by relevance inferred from human brain signals

    Science.gov (United States)

    Eugster, Manuel J. A.; Ruotsalo, Tuukka; Spapé, Michiel M.; Barral, Oswald; Ravaja, Niklas; Jacucci, Giulio; Kaski, Samuel

    2016-01-01

    Finding relevant information from large document collections such as the World Wide Web is a common task in our daily lives. Estimation of a user’s interest or search intention is necessary to recommend and retrieve relevant information from these collections. We introduce a brain-information interface used for recommending information by relevance inferred directly from brain signals. In experiments, participants were asked to read Wikipedia documents about a selection of topics while their EEG was recorded. Based on the prediction of word relevance, the individual’s search intent was modeled and successfully used for retrieving new relevant documents from the whole English Wikipedia corpus. The results show that the users’ interests toward digital content can be modeled from the brain signals evoked by reading. The introduced brain-relevance paradigm enables the recommendation of information without any explicit user interaction and may be applied across diverse information-intensive applications. PMID:27929077

  3. 49 CFR 556.9 - Public inspection of relevant information.

    Science.gov (United States)

    2010-10-01

    ... 49 Transportation 6 2010-10-01 2010-10-01 false Public inspection of relevant information. 556.9... NONCOMPLIANCE § 556.9 Public inspection of relevant information. Information relevant to a petition under this... Administration, 400 Seventh Street, SW., Washington, DC 20590. Copies of available information may be obtained in...

  4. Software Helps Retrieve Information Relevant to the User

    Science.gov (United States)

    Mathe, Natalie; Chen, James

    2003-01-01

    The Adaptive Indexing and Retrieval Agent (ARNIE) is a code library, designed to be used by an application program, that assists human users in retrieving desired information in a hypertext setting. Using ARNIE, the program implements a computational model for interactively learning what information each human user considers relevant in context. The model, called a "relevance network," incrementally adapts retrieved information to users individual profiles on the basis of feedback from the users regarding specific queries. The model also generalizes such knowledge for subsequent derivation of relevant references for similar queries and profiles, thereby, assisting users in filtering information by relevance. ARNIE thus enables users to categorize and share information of interest in various contexts. ARNIE encodes the relevance and structure of information in a neural network dynamically configured with a genetic algorithm. ARNIE maintains an internal database, wherein it saves associations, and from which it returns associated items in response to a query. A C++ compiler for a platform on which ARNIE will be utilized is necessary for creating the ARNIE library but is not necessary for the execution of the software.

  5. The Development of Relevance in Information Retrieval

    Directory of Open Access Journals (Sweden)

    Mu-hsuan Huang

    1997-12-01

    Full Text Available This article attempts to investigate the notion of relevance in information retrieval. It discusses various definitions for relevance from historical viewpoints and the characteristics of relevance judgments. Also, it introduces empirical results of important related researches.[Article content in Chinese

  6. Extracting of implicit information in English advertising texts with phonetic and lexical-morphological means

    Directory of Open Access Journals (Sweden)

    Traikovskaya Natalya Petrovna

    2015-12-01

    Full Text Available The article deals with phonetic and lexical-morphological language means participating in the process of extracting implicit information in English-speaking advertising texts for men and women. The functioning of phonetic means of the English language is not the basis for implication of information in advertising texts. Lexical and morphological means play the role of markers of relevant information, playing the role of the activator ofimplicit information in the texts of advertising.

  7. Evaluating automatic attentional capture by self-relevant information.

    Science.gov (United States)

    Ocampo, Brenda; Kahan, Todd A

    2016-01-01

    Our everyday decisions and memories are inadvertently influenced by self-relevant information. For example, we are faster and more accurate at making perceptual judgments about stimuli associated with ourselves, such as our own face or name, as compared with familiar non-self-relevant stimuli. Humphreys and Sui propose a "self-attention network" to account for these effects, wherein self-relevant stimuli automatically capture our attention and subsequently enhance the perceptual processing of self-relevant information. We propose that the masked priming paradigm and continuous flash suppression represent two ways to experimentally examine these controversial claims.

  8. Evolutionary relevance facilitates visual information processing.

    Science.gov (United States)

    Jackson, Russell E; Calvillo, Dusti P

    2013-11-03

    Visual search of the environment is a fundamental human behavior that perceptual load affects powerfully. Previously investigated means for overcoming the inhibitions of high perceptual load, however, generalize poorly to real-world human behavior. We hypothesized that humans would process evolutionarily relevant stimuli more efficiently than evolutionarily novel stimuli, and evolutionary relevance would mitigate the repercussions of high perceptual load during visual search. Animacy is a significant component to evolutionary relevance of visual stimuli because perceiving animate entities is time-sensitive in ways that pose significant evolutionary consequences. Participants completing a visual search task located evolutionarily relevant and animate objects fastest and with the least impact of high perceptual load. Evolutionarily novel and inanimate objects were located slowest and with the highest impact of perceptual load. Evolutionary relevance may importantly affect everyday visual information processing.

  9. Information extraction from multi-institutional radiology reports.

    Science.gov (United States)

    Hassanpour, Saeed; Langlotz, Curtis P

    2016-01-01

    The radiology report is the most important source of clinical imaging information. It documents critical information about the patient's health and the radiologist's interpretation of medical findings. It also communicates information to the referring physicians and records that information for future clinical and research use. Although efforts to structure some radiology report information through predefined templates are beginning to bear fruit, a large portion of radiology report information is entered in free text. The free text format is a major obstacle for rapid extraction and subsequent use of information by clinicians, researchers, and healthcare information systems. This difficulty is due to the ambiguity and subtlety of natural language, complexity of described images, and variations among different radiologists and healthcare organizations. As a result, radiology reports are used only once by the clinician who ordered the study and rarely are used again for research and data mining. In this work, machine learning techniques and a large multi-institutional radiology report repository are used to extract the semantics of the radiology report and overcome the barriers to the re-use of radiology report information in clinical research and other healthcare applications. We describe a machine learning system to annotate radiology reports and extract report contents according to an information model. This information model covers the majority of clinically significant contents in radiology reports and is applicable to a wide variety of radiology study types. Our automated approach uses discriminative sequence classifiers for named-entity recognition to extract and organize clinically significant terms and phrases consistent with the information model. We evaluated our information extraction system on 150 radiology reports from three major healthcare organizations and compared its results to a commonly used non-machine learning information extraction method. We

  10. Extracting information from multiplex networks

    Science.gov (United States)

    Iacovacci, Jacopo; Bianconi, Ginestra

    2016-06-01

    Multiplex networks are generalized network structures that are able to describe networks in which the same set of nodes are connected by links that have different connotations. Multiplex networks are ubiquitous since they describe social, financial, engineering, and biological networks as well. Extending our ability to analyze complex networks to multiplex network structures increases greatly the level of information that is possible to extract from big data. For these reasons, characterizing the centrality of nodes in multiplex networks and finding new ways to solve challenging inference problems defined on multiplex networks are fundamental questions of network science. In this paper, we discuss the relevance of the Multiplex PageRank algorithm for measuring the centrality of nodes in multilayer networks and we characterize the utility of the recently introduced indicator function Θ ˜ S for describing their mesoscale organization and community structure. As working examples for studying these measures, we consider three multiplex network datasets coming for social science.

  11. Evolutionary Relevance Facilitates Visual Information Processing

    Directory of Open Access Journals (Sweden)

    Russell E. Jackson

    2013-07-01

    Full Text Available Visual search of the environment is a fundamental human behavior that perceptual load affects powerfully. Previously investigated means for overcoming the inhibitions of high perceptual load, however, generalize poorly to real-world human behavior. We hypothesized that humans would process evolutionarily relevant stimuli more efficiently than evolutionarily novel stimuli, and evolutionary relevance would mitigate the repercussions of high perceptual load during visual search. Animacy is a significant component to evolutionary relevance of visual stimuli because perceiving animate entities is time-sensitive in ways that pose significant evolutionary consequences. Participants completing a visual search task located evolutionarily relevant and animate objects fastest and with the least impact of high perceptual load. Evolutionarily novel and inanimate objects were located slowest and with the highest impact of perceptual load. Evolutionary relevance may importantly affect everyday visual information processing.

  12. OpenCV-Based Nanomanipulation Information Extraction and the Probe Operation in SEM

    Directory of Open Access Journals (Sweden)

    Dongjie Li

    2015-02-01

    Full Text Available Aimed at the established telenanomanipulation system, the method of extracting location information and the strategies of probe operation were studied in this paper. First, the machine learning algorithm of OpenCV was used to extract location information from SEM images. Thus nanowires and probe in SEM images can be automatically tracked and the region of interest (ROI can be marked quickly. Then the location of nanowire and probe can be extracted from the ROI. To study the probe operation strategy, the Van der Waals force between probe and a nanowire was computed; thus relevant operating parameters can be obtained. With these operating parameters, the nanowire in 3D virtual environment can be preoperated and an optimal path of the probe can be obtained. The actual probe runs automatically under the telenanomanipulation system's control. Finally, experiments were carried out to verify the above methods, and results show the designed methods have achieved the expected effect.

  13. Extracting the relevant delays in time series modelling

    DEFF Research Database (Denmark)

    Goutte, Cyril

    1997-01-01

    selection, and more precisely stepwise forward selection. The method is compared to other forward selection schemes, as well as to a nonparametric tests aimed at estimating the embedding dimension of time series. The final application extends these results to the efficient estimation of FIR filters on some......In this contribution, we suggest a convenient way to use generalisation error to extract the relevant delays from a time-varying process, i.e. the delays that lead to the best prediction performance. We design a generalisation-based algorithm that takes its inspiration from traditional variable...

  14. From remote sensing data about information extraction for 3D geovisualization - Development of a workflow

    International Nuclear Information System (INIS)

    Tiede, D.

    2010-01-01

    With an increased availability of high (spatial) resolution remote sensing imagery since the late nineties, the need to develop operative workflows for the automated extraction, provision and communication of information from such data has grown. Monitoring requirements, aimed at the implementation of environmental or conservation targets, management of (environmental-) resources, and regional planning as well as international initiatives, especially the joint initiative of the European Commission and ESA (European Space Agency) for Global Monitoring for Environment and Security (GMES) play also a major part. This thesis addresses the development of an integrated workflow for the automated provision of information derived from remote sensing data. Considering applied data and fields of application, this work aims to design the workflow as generic as possible. Following research questions are discussed: What are the requirements of a workflow architecture that seamlessly links the individual workflow elements in a timely manner and secures accuracy of the extracted information effectively? How can the workflow retain its efficiency if mounds of data are processed? How can the workflow be improved with regards to automated object-based image analysis (OBIA)? Which recent developments could be of use? What are the limitations or which workarounds could be applied in order to generate relevant results? How can relevant information be prepared target-oriented and communicated effectively? How can the more recently developed freely available virtual globes be used for the delivery of conditioned information under consideration of the third dimension as an additional, explicit carrier of information? Based on case studies comprising different data sets and fields of application it is demonstrated how methods to extract and process information as well as to effectively communicate results can be improved and successfully combined within one workflow. It is shown that (1

  15. Types of Lexicographical Information Needs and their Relevance for Information Science

    Directory of Open Access Journals (Sweden)

    Bergenholtz, Henning

    2017-09-01

    Full Text Available In some situations, you need information in order to solve a problem that has occurred. In information science, user needs are often described through very specific examples rather than through a classification of situation types in which information needs occur. Furthermore, information science often describes general human needs, typically with a reference to Maslow's classification of needs (1954, instead of actual information needs. Lexicography has also focused on information needs, but has developed a more abstract classification of types of information needs, though (until more recent research into lexicographical functions with a particular interest in linguistic uncertainties and the lack of knowledge and skills in relation to one or several languages. In this article, we suggest a classification of information needs in which a tripartition has been made according to the different types of situations: communicative needs, cognitive needs, and operative needs. This is a classification that is relevant and useful in general in our modern information society and therefore also relevant for information science, including lexicography.

  16. Extracting the Behaviorally Relevant Stimulus: Unique Neural Representation of Farnesol, a Component of the Recruitment Pheromone of Bombus terrestris.

    Directory of Open Access Journals (Sweden)

    Martin F Strube-Bloss

    Full Text Available To trigger innate behavior, sensory neural networks are pre-tuned to extract biologically relevant stimuli. Many male-female or insect-plant interactions depend on this phenomenon. Especially communication among individuals within social groups depends on innate behaviors. One example is the efficient recruitment of nest mates by successful bumblebee foragers. Returning foragers release a recruitment pheromone in the nest while they perform a 'dance' behavior to activate unemployed nest mates. A major component of this pheromone is the sesquiterpenoid farnesol. How farnesol is processed and perceived by the olfactory system, has not yet been identified. It is much likely that processing farnesol involves an innate mechanism for the extraction of relevant information to trigger a fast and reliable behavioral response. To test this hypothesis, we used population response analyses of 100 antennal lobe (AL neurons recorded in alive bumblebee workers under repeated stimulation with four behaviorally different, but chemically related odorants (geraniol, citronellol, citronellal and farnesol. The analysis identified a unique neural representation of the recruitment pheromone component compared to the other odorants that are predominantly emitted by flowers. The farnesol induced population activity in the AL allowed a reliable separation of farnesol from all other chemically related odor stimuli we tested. We conclude that the farnesol induced population activity may reflect a predetermined representation within the AL-neural network allowing efficient and fast extraction of a behaviorally relevant stimulus. Furthermore, the results show that population response analyses of multiple single AL-units may provide a powerful tool to identify distinct representations of behaviorally relevant odors.

  17. Testing the idea of privileged awareness of self-relevant information.

    Science.gov (United States)

    Stein, Timo; Siebold, Alisha; van Zoest, Wieske

    2016-03-01

    Self-relevant information is prioritized in processing. Some have suggested the mechanism driving this advantage is akin to the automatic prioritization of physically salient stimuli in information processing (Humphreys & Sui, 2015). Here we investigate whether self-relevant information is prioritized for awareness under continuous flash suppression (CFS), as has been found for physical salience. Gabor patches with different orientations were first associated with the labels You or Other. Participants were more accurate in matching the self-relevant association, replicating previous findings of self-prioritization. However, breakthrough into awareness from CFS did not differ between self- and other-associated Gabors. These findings demonstrate that self-relevant information has no privileged access to awareness. Rather than modulating the initial visual processes that precede and lead to awareness, the advantage of self-relevant information may better be characterized as prioritization at later processing stages. (c) 2016 APA, all rights reserved).

  18. Extracting information from two-dimensional electrophoresis gels by partial least squares regression

    DEFF Research Database (Denmark)

    Jessen, Flemming; Lametsch, R.; Bendixen, E.

    2002-01-01

    of all proteins/spots in the gels. In the present study it is demonstrated how information can be extracted by multivariate data analysis. The strategy is based on partial least squares regression followed by variable selection to find proteins that individually or in combination with other proteins vary......Two-dimensional gel electrophoresis (2-DE) produces large amounts of data and extraction of relevant information from these data demands a cautious and time consuming process of spot pattern matching between gels. The classical approach of data analysis is to detect protein markers that appear...... or disappear depending on the experimental conditions. Such biomarkers are found by comparing the relative volumes of individual spots in the individual gels. Multivariate statistical analysis and modelling of 2-DE data for comparison and classification is an alternative approach utilising the combination...

  19. Value Relevance of Accounting Information in the United Arab Emirates

    Directory of Open Access Journals (Sweden)

    Jamal Barzegari Khanagha

    2011-01-01

    Full Text Available This paper examines the value relevance of accounting information in per and post-periods of International Financial Reporting Standards implementation using the regression and portfolio approaches for sample of the UAE companies. The results obtained from a combination of regression and portfolio approaches, show accounting information is value relevant in UAE stock market. A comparison of the results for the periods before and after adoption, based on both regression and portfolio approaches, shows a decline in value relevance of accounting information after the reform in accounting standards. It could be interpreted to mean that following to IFRS in UAE didn’t improve value relevancy of accounting information. However, results based on and portfolio approach shows that cash flows’ incremental information content increased for the post-IFRS period.

  20. Extracting useful information from images

    DEFF Research Database (Denmark)

    Kucheryavskiy, Sergey

    2011-01-01

    The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic and heter......The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic...

  1. 50 CFR 424.13 - Sources of information and relevant data.

    Science.gov (United States)

    2010-10-01

    ... 50 Wildlife and Fisheries 7 2010-10-01 2010-10-01 false Sources of information and relevant data... Sources of information and relevant data. When considering any revision of the lists, the Secretary shall..., administrative reports, maps or other graphic materials, information received from experts on the subject, and...

  2. Alpha power gates relevant information during working memory updating.

    Science.gov (United States)

    Manza, Peter; Hau, Chui Luen Vera; Leung, Hoi-Chung

    2014-04-23

    Human working memory (WM) is inherently limited, so we must filter out irrelevant information in our environment or our mind while retaining limited important relevant contents. Previous work suggests that neural oscillations in the alpha band (8-14 Hz) play an important role in inhibiting incoming distracting information during attention and selective encoding tasks. However, whether alpha power is involved in inhibiting no-longer-relevant content or in representing relevant WM content is still debated. To clarify this issue, we manipulated the amount of relevant/irrelevant information using a task requiring spatial WM updating while measuring neural oscillatory activity via EEG and localized current sources across the scalp using a surface Laplacian transform. An initial memory set of two, four, or six spatial locations was to be memorized over a delay until an updating cue was presented indicating that only one or three locations remained relevant for a subsequent recognition test. Alpha amplitude varied with memory maintenance and updating demands among a cluster of left frontocentral electrodes. Greater postcue alpha power was associated with the high relevant load conditions (six and four dots cued to reduce to three relevant) relative to the lower load conditions (four and two dots reduced to one). Across subjects, this difference in alpha power was correlated with condition differences in performance accuracy. In contrast, no significant effects of irrelevant load were observed. These findings demonstrate that, during WM updating, alpha power reflects maintenance of relevant memory contents rather than suppression of no-longer-relevant memory traces.

  3. Extraction of Information of Audio-Visual Contents

    Directory of Open Access Journals (Sweden)

    Carlos Aguilar

    2011-10-01

    Full Text Available In this article we show how it is possible to use Channel Theory (Barwise and Seligman, 1997 for modeling the process of information extraction realized by audiences of audio-visual contents. To do this, we rely on the concepts pro- posed by Channel Theory and, especially, its treatment of representational systems. We then show how the information that an agent is capable of extracting from the content depends on the number of channels he is able to establish between the content and the set of classifications he is able to discriminate. The agent can endeavor the extraction of information through these channels from the totality of content; however, we discuss the advantages of extracting from its constituents in order to obtain a greater number of informational items that represent it. After showing how the extraction process is endeavored for each channel, we propose a method of representation of all the informative values an agent can obtain from a content using a matrix constituted by the channels the agent is able to establish on the content (source classifications, and the ones he can understand as individual (destination classifications. We finally show how this representation allows reflecting the evolution of the informative items through the evolution of audio-visual content.

  4. 76 FR 34075 - Request for Information (RFI) To Identify and Obtain Relevant Information From Public or Private...

    Science.gov (United States)

    2011-06-10

    ... Relevant Information From Public or Private Entities With an Interest in Biovigilance; Extension AGENCY... and obtain relevant information regarding the possible development of a public-private partnership... Identify and Obtain Relevant Information from Public or Private Entities with an Interest in Biovigilance...

  5. Scenario Customization for Information Extraction

    National Research Council Canada - National Science Library

    Yangarber, Roman

    2001-01-01

    Information Extraction (IE) is an emerging NLP technology, whose function is to process unstructured, natural language text, to locate specific pieces of information, or facts, in the text, and to use these facts to fill a database...

  6. Can we replace curation with information extraction software?

    Science.gov (United States)

    Karp, Peter D

    2016-01-01

    Can we use programs for automated or semi-automated information extraction from scientific texts as practical alternatives to professional curation? I show that error rates of current information extraction programs are too high to replace professional curation today. Furthermore, current IEP programs extract single narrow slivers of information, such as individual protein interactions; they cannot extract the large breadth of information extracted by professional curators for databases such as EcoCyc. They also cannot arbitrate among conflicting statements in the literature as curators can. Therefore, funding agencies should not hobble the curation efforts of existing databases on the assumption that a problem that has stymied Artificial Intelligence researchers for more than 60 years will be solved tomorrow. Semi-automated extraction techniques appear to have significantly more potential based on a review of recent tools that enhance curator productivity. But a full cost-benefit analysis for these tools is lacking. Without such analysis it is possible to expend significant effort developing information-extraction tools that automate small parts of the overall curation workflow without achieving a significant decrease in curation costs.Database URL. © The Author(s) 2016. Published by Oxford University Press.

  7. Transductive Pattern Learning for Information Extraction

    National Research Council Canada - National Science Library

    McLernon, Brian; Kushmerick, Nicholas

    2006-01-01

    .... We present TPLEX, a semi-supervised learning algorithm for information extraction that can acquire extraction patterns from a small amount of labelled text in conjunction with a large amount of unlabelled text...

  8. CellBase, a comprehensive collection of RESTful web services for retrieving relevant biological information from heterogeneous sources.

    Science.gov (United States)

    Bleda, Marta; Tarraga, Joaquin; de Maria, Alejandro; Salavert, Francisco; Garcia-Alonso, Luz; Celma, Matilde; Martin, Ainoha; Dopazo, Joaquin; Medina, Ignacio

    2012-07-01

    During the past years, the advances in high-throughput technologies have produced an unprecedented growth in the number and size of repositories and databases storing relevant biological data. Today, there is more biological information than ever but, unfortunately, the current status of many of these repositories is far from being optimal. Some of the most common problems are that the information is spread out in many small databases; frequently there are different standards among repositories and some databases are no longer supported or they contain too specific and unconnected information. In addition, data size is increasingly becoming an obstacle when accessing or storing biological data. All these issues make very difficult to extract and integrate information from different sources, to analyze experiments or to access and query this information in a programmatic way. CellBase provides a solution to the growing necessity of integration by easing the access to biological data. CellBase implements a set of RESTful web services that query a centralized database containing the most relevant biological data sources. The database is hosted in our servers and is regularly updated. CellBase documentation can be found at http://docs.bioinfo.cipf.es/projects/cellbase.

  9. Support Vector Machines: Relevance Feedback and Information Retrieval.

    Science.gov (United States)

    Drucker, Harris; Shahrary, Behzad; Gibbon, David C.

    2002-01-01

    Compares support vector machines (SVMs) to Rocchio, Ide regular and Ide dec-hi algorithms in information retrieval (IR) of text documents using relevancy feedback. If the preliminary search is so poor that one has to search through many documents to find at least one relevant document, then SVM is preferred. Includes nine tables. (Contains 24…

  10. Optical Aperture Synthesis Object's Information Extracting Based on Wavelet Denoising

    International Nuclear Information System (INIS)

    Fan, W J; Lu, Y

    2006-01-01

    Wavelet denoising is studied to improve OAS(optical aperture synthesis) object's Fourier information extracting. Translation invariance wavelet denoising based on Donoho wavelet soft threshold denoising is researched to remove Pseudo-Gibbs in wavelet soft threshold image. OAS object's information extracting based on translation invariance wavelet denoising is studied. The study shows that wavelet threshold denoising can improve the precision and the repetition of object's information extracting from interferogram, and the translation invariance wavelet denoising information extracting is better than soft threshold wavelet denoising information extracting

  11. KneeTex: an ontology-driven system for information extraction from MRI reports.

    Science.gov (United States)

    Spasić, Irena; Zhao, Bo; Jones, Christopher B; Button, Kate

    2015-01-01

    on a test set of 100 MRI reports. A gold standard consisted of 1,259 filled template records with the following slots: finding, finding qualifier, negation, certainty, anatomy and anatomy qualifier. KneeTex extracted information with precision of 98.00 %, recall of 97.63 % and F-measure of 97.81 %, the values of which are in line with human-like performance. KneeTex is an open-source, stand-alone application for information extraction from narrative reports that describe an MRI scan of the knee. Given an MRI report as input, the system outputs the corresponding clinical findings in the form of JavaScript Object Notation objects. The extracted information is mapped onto TRAK, an ontology that formally models knowledge relevant for the rehabilitation of knee conditions. As a result, formally structured and coded information allows for complex searches to be conducted efficiently over the original MRI reports, thereby effectively supporting epidemiologic studies of knee conditions.

  12. A User-Centered Approach to Adaptive Hypertext Based on an Information Relevance Model

    Science.gov (United States)

    Mathe, Nathalie; Chen, James

    1994-01-01

    Rapid and effective to information in large electronic documentation systems can be facilitated if information relevant in an individual user's content can be automatically supplied to this user. However most of this knowledge on contextual relevance is not found within the contents of documents, it is rather established incrementally by users during information access. We propose a new model for interactively learning contextual relevance during information retrieval, and incrementally adapting retrieved information to individual user profiles. The model, called a relevance network, records the relevance of references based on user feedback for specific queries and user profiles. It also generalizes such knowledge to later derive relevant references for similar queries and profiles. The relevance network lets users filter information by context of relevance. Compared to other approaches, it does not require any prior knowledge nor training. More importantly, our approach to adaptivity is user-centered. It facilitates acceptance and understanding by users by giving them shared control over the adaptation without disturbing their primary task. Users easily control when to adapt and when to use the adapted system. Lastly, the model is independent of the particular application used to access information, and supports sharing of adaptations among users.

  13. Extracting Behaviorally Relevant Traits from Natural Stimuli: Benefits of Combinatorial Representations at the Accessory Olfactory Bulb.

    Directory of Open Access Journals (Sweden)

    Anat Kahan

    2016-03-01

    Full Text Available For many animals, chemosensation is essential for guiding social behavior. However, because multiple factors can modulate levels of individual chemical cues, deriving information about other individuals via natural chemical stimuli involves considerable challenges. How social information is extracted despite these sources of variability is poorly understood. The vomeronasal system provides an excellent opportunity to study this topic due to its role in detecting socially relevant traits. Here, we focus on two such traits: a female mouse's strain and reproductive state. In particular, we measure stimulus-induced neuronal activity in the accessory olfactory bulb (AOB in response to various dilutions of urine, vaginal secretions, and saliva, from estrus and non-estrus female mice from two different strains. We first show that all tested secretions provide information about a female's receptivity and genotype. Next, we investigate how these traits can be decoded from neuronal activity despite multiple sources of variability. We show that individual neurons are limited in their capacity to allow trait classification across multiple sources of variability. However, simple linear classifiers sampling neuronal activity from small neuronal ensembles can provide a substantial improvement over that attained with individual units. Furthermore, we show that some traits are more efficiently detected than others, and that particular secretions may be optimized for conveying information about specific traits. Across all tested stimulus sources, discrimination between strains is more accurate than discrimination of receptivity, and detection of receptivity is more accurate with vaginal secretions than with urine. Our findings highlight the challenges of chemosensory processing of natural stimuli, and suggest that downstream readout stages decode multiple behaviorally relevant traits by sampling information from distinct but overlapping populations of AOB neurons.

  14. A content relevance model for social media health information.

    Science.gov (United States)

    Prybutok, Gayle Linda; Koh, Chang; Prybutok, Victor R

    2014-04-01

    Consumer health informatics includes the development and implementation of Internet-based systems to deliver health risk management information and health intervention applications to the public. The application of consumer health informatics to educational and interventional efforts such as smoking reduction and cessation has garnered attention from both consumers and health researchers in recent years. Scientists believe that smoking avoidance or cessation before the age of 30 years can prevent more than 90% of smoking-related cancers and that individuals who stop smoking fare as well in preventing cancer as those who never start. The goal of this study was to determine factors that were most highly correlated with content relevance for health information provided on the Internet for a study group of 18- to 30-year-old college students. Data analysis showed that the opportunity for convenient entertainment, social interaction, health information-seeking behavior, time spent surfing on the Internet, the importance of available activities on the Internet (particularly e-mail), and perceived site relevance for Internet-based sources of health information were significantly correlated with content relevance for 18- to 30-year-old college students, an educated subset of this population segment.

  15. Cause Information Extraction from Financial Articles Concerning Business Performance

    Science.gov (United States)

    Sakai, Hiroyuki; Masuyama, Shigeru

    We propose a method of extracting cause information from Japanese financial articles concerning business performance. Our method acquires cause informtion, e. g. “_??__??__??__??__??__??__??__??__??__??_ (zidousya no uriage ga koutyou: Sales of cars were good)”. Cause information is useful for investors in selecting companies to invest. Our method extracts cause information as a form of causal expression by using statistical information and initial clue expressions automatically. Our method can extract causal expressions without predetermined patterns or complex rules given by hand, and is expected to be applied to other tasks for acquiring phrases that have a particular meaning not limited to cause information. We compared our method with our previous one originally proposed for extracting phrases concerning traffic accident causes and experimental results showed that our new method outperforms our previous one.

  16. Sample-based XPath Ranking for Web Information Extraction

    NARCIS (Netherlands)

    Jundt, Oliver; van Keulen, Maurice

    Web information extraction typically relies on a wrapper, i.e., program code or a configuration that specifies how to extract some information from web pages at a specific website. Manually creating and maintaining wrappers is a cumbersome and error-prone task. It may even be prohibitive as some

  17. Ontology-Based Information Extraction for Business Intelligence

    Science.gov (United States)

    Saggion, Horacio; Funk, Adam; Maynard, Diana; Bontcheva, Kalina

    Business Intelligence (BI) requires the acquisition and aggregation of key pieces of knowledge from multiple sources in order to provide valuable information to customers or feed statistical BI models and tools. The massive amount of information available to business analysts makes information extraction and other natural language processing tools key enablers for the acquisition and use of that semantic information. We describe the application of ontology-based extraction and merging in the context of a practical e-business application for the EU MUSING Project where the goal is to gather international company intelligence and country/region information. The results of our experiments so far are very promising and we are now in the process of building a complete end-to-end solution.

  18. Cogito ergo video: Task-relevant information is involuntarily boosted into awareness.

    Science.gov (United States)

    Gayet, Surya; Brascamp, Jan W; Van der Stigchel, Stefan; Paffen, Chris L E

    2015-01-01

    Only part of the visual information that impinges on our retinae reaches visual awareness. In a series of three experiments, we investigated how the task relevance of incoming visual information affects its access to visual awareness. On each trial, participants were instructed to memorize one of two presented hues, drawn from different color categories (e.g., red and green), for later recall. During the retention interval, participants were presented with a differently colored grating in each eye such as to elicit binocular rivalry. A grating matched either the task-relevant (memorized) color category or the task-irrelevant (nonmemorized) color category. We found that the rivalrous stimulus that matched the task-relevant color category tended to dominate awareness over the rivalrous stimulus that matched the task-irrelevant color category. This effect of task relevance persisted when participants reported the orientation of the rivalrous stimuli, even though in this case color information was completely irrelevant for the task of reporting perceptual dominance during rivalry. When participants memorized the shape of a colored stimulus, however, its color category did not affect predominance of rivalrous stimuli during retention. Taken together, these results indicate that the selection of task-relevant information is under volitional control but that visual input that matches this information is boosted into awareness irrespective of whether this is useful for the observer.

  19. Bootstrapping agency: How control-relevant information affects motivation.

    Science.gov (United States)

    Karsh, Noam; Eitam, Baruch; Mark, Ilya; Higgins, E Tory

    2016-10-01

    How does information about one's control over the environment (e.g., having an own-action effect) influence motivation? The control-based response selection framework was proposed to predict and explain such findings. Its key tenant is that control relevant information modulates both the frequency and speed of responses by determining whether a perceptual event is an outcome of one's actions or not. To test this framework empirically, the current study examines whether and how temporal and spatial contiguity/predictability-previously established as being important for one's sense of agency-modulate motivation from control. In 5 experiments, participants responded to a cue, potentially triggering a perceptual effect. Temporal (Experiments 1a-c) and spatial (Experiments 2a and b) contiguity/predictability between actions and their potential effects were experimentally manipulated. The influence of these control-relevant factors was measured, both indirectly (through their effect on explicit judgments of agency) and directly on response time and response frequency. The pattern of results was highly consistent with the control-based response selection framework in suggesting that control relevant information reliably modulates the impact of "having an effect" on different levels of action selection. We discuss the implications of this study for the notion of motivation from control and for the empirical work on the sense of agency. (PsycINFO Database Record (c) 2016 APA, all rights reserved).

  20. Fuzzy Mutual Information Based min-Redundancy and Max-Relevance Heterogeneous Feature Selection

    Directory of Open Access Journals (Sweden)

    Daren Yu

    2011-08-01

    Full Text Available Feature selection is an important preprocessing step in pattern classification and machine learning, and mutual information is widely used to measure relevance between features and decision. However, it is difficult to directly calculate relevance between continuous or fuzzy features using mutual information. In this paper we introduce the fuzzy information entropy and fuzzy mutual information for computing relevance between numerical or fuzzy features and decision. The relationship between fuzzy information entropy and differential entropy is also discussed. Moreover, we combine fuzzy mutual information with qmin-Redundancy-Max-Relevanceq, qMax-Dependencyq and min-Redundancy-Max-Dependencyq algorithms. The performance and stability of the proposed algorithms are tested on benchmark data sets. Experimental results show the proposed algorithms are effective and stable.

  1. A Two-Step Resume Information Extraction Algorithm

    Directory of Open Access Journals (Sweden)

    Jie Chen

    2018-01-01

    Full Text Available With the rapid growth of Internet-based recruiting, there are a great number of personal resumes among recruiting systems. To gain more attention from the recruiters, most resumes are written in diverse formats, including varying font size, font colour, and table cells. However, the diversity of format is harmful to data mining, such as resume information extraction, automatic job matching, and candidates ranking. Supervised methods and rule-based methods have been proposed to extract facts from resumes, but they strongly rely on hierarchical structure information and large amounts of labelled data, which are hard to collect in reality. In this paper, we propose a two-step resume information extraction approach. In the first step, raw text of resume is identified as different resume blocks. To achieve the goal, we design a novel feature, Writing Style, to model sentence syntax information. Besides word index and punctuation index, word lexical attribute and prediction results of classifiers are included in Writing Style. In the second step, multiple classifiers are employed to identify different attributes of fact information in resumes. Experimental results on a real-world dataset show that the algorithm is feasible and effective.

  2. Extraction of Pluvial Flood Relevant Volunteered Geographic Information (VGI by Deep Learning from User Generated Texts and Photos

    Directory of Open Access Journals (Sweden)

    Yu Feng

    2018-01-01

    Full Text Available In recent years, pluvial floods caused by extreme rainfall events have occurred frequently. Especially in urban areas, they lead to serious damages and endanger the citizens’ safety. Therefore, real-time information about such events is desirable. With the increasing popularity of social media platforms, such as Twitter or Instagram, information provided by voluntary users becomes a valuable source for emergency response. Many applications have been built for disaster detection and flood mapping using crowdsourcing. Most of the applications so far have merely used keyword filtering or classical language processing methods to identify disaster relevant documents based on user generated texts. As the reliability of social media information is often under criticism, the precision of information retrieval plays a significant role for further analyses. Thus, in this paper, high quality eyewitnesses of rainfall and flooding events are retrieved from social media by applying deep learning approaches on user generated texts and photos. Subsequently, events are detected through spatiotemporal clustering and visualized together with these high quality eyewitnesses in a web map application. Analyses and case studies are conducted during flooding events in Paris, London and Berlin.

  3. The Agent of extracting Internet Information with Lead Order

    Science.gov (United States)

    Mo, Zan; Huang, Chuliang; Liu, Aijun

    In order to carry out e-commerce better, advanced technologies to access business information are in need urgently. An agent is described to deal with the problems of extracting internet information that caused by the non-standard and skimble-scamble structure of Chinese websites. The agent designed includes three modules which respond to the process of extracting information separately. A method of HTTP tree and a kind of Lead algorithm is proposed to generate a lead order, with which the required web can be retrieved easily. How to transform the extracted information structuralized with natural language is also discussed.

  4. Screening of plant extracts for antimicrobial activity against bacteria and yeasts with dermatological relevance.

    Science.gov (United States)

    Weckesser, S; Engel, K; Simon-Haarhaus, B; Wittmer, A; Pelz, K; Schempp, C M

    2007-08-01

    There is cumulative resistance against antibiotics of many bacteria. Therefore, the development of new antiseptics and antimicrobial agents for the treatment of skin infections is of increasing interest. We have screened six plant extracts and isolated compounds for antimicrobial effects on bacteria and yeasts with dermatological relevance. The following plant extracts have been tested: Gentiana lutea, Harpagophytum procumbens, Boswellia serrata (dry extracts), Usnea barbata, Rosmarinus officinalis and Salvia officinalis (supercritical carbon dioxide [CO2] extracts). Additionally, the following characteristic plant substances were tested: usnic acid, carnosol, carnosic acid, ursolic acid, oleanolic acid, harpagoside, boswellic acid and gentiopicroside. The extracts and compounds were tested against 29 aerobic and anaerobic bacteria and yeasts in the agar dilution test. U. barbata-extract and usnic acid were the most active compounds, especially in anaerobic bacteria. Usnea CO2-extract effectively inhibited the growth of several Gram-positive bacteria like Staphylococcus aureus (including methicillin-resistant strains - MRSA), Propionibacterium acnes and Corynebacterium species. Growth of the dimorphic yeast Malassezia furfur was also inhibited by Usnea-extract. Besides the Usnea-extract, Rosmarinus-, Salvia-, Boswellia- and Harpagophytum-extracts proved to be effective against a panel of bacteria. It is concluded that due to their antimicrobial effects some of the plant extracts may be used for the topical treatment of skin disorders like acne vulgaris and seborrhoic eczema.

  5. Fuzzy Information Retrieval Using Genetic Algorithms and Relevance Feedback.

    Science.gov (United States)

    Petry, Frederick E.; And Others

    1993-01-01

    Describes an approach that combines concepts from information retrieval, fuzzy set theory, and genetic programing to improve weighted Boolean query formulation via relevance feedback. Highlights include background on information retrieval systems; genetic algorithms; subproblem formulation; and preliminary results based on a testbed. (Contains 12…

  6. Disclosure of Non-Financial Information: Relevant to Financial Analysts?

    OpenAIRE

    ORENS, Raf; LYBAERT, Nadine

    2013-01-01

    The decline in the relevance of financial statement information to value firms leads to calls from organizational stakeholders to convey non-financial information in order to be able to judge firms' financial performance and value. This literature review aims to report extant literature findings on the use of corporate non-financial information by sell-side financial analysts, the information intermediaries between corporate management and investors. Prior studies highlight that financial ana...

  7. Waveform shape analysis: extraction of physiologically relevant information from Doppler recordings.

    Science.gov (United States)

    Ramsay, M M; Broughton Pipkin, F; Rubin, P C; Skidmore, R

    1994-05-01

    1. Doppler recordings were made from the brachial artery of healthy female subjects during a series of manoeuvres which altered the pressure-flow characteristics of the vessel. 2. Changes were induced in the peripheral circulation of the forearm by the application of heat or ice-packs. A sphygmomanometer cuff was used to create graded occlusion of the vessel above and below the point of measurement. Recordings were also made whilst the subjects performed a standardized Valsalva manoeuvre. 3. The Doppler recordings were analysed both with the standard waveform indices (systolic/diastolic ratio, pulsatility index and resistance index) and by the method of Laplace transform analysis. 4. The waveform parameters obtained by Laplace transform analysis distinguished the different changes in flow conditions; they thus had direct physiological relevance, unlike the standard waveform indices.

  8. Information sharing during diagnostic assessments: what is relevant for parents?

    Science.gov (United States)

    Klein, Sheryl; Wynn, Kerry; Ray, Lynne; Demeriez, Lori; LaBerge, Patricia; Pei, Jacqueline; St Pierre, Cherie

    2011-05-01

    ABSTRACT This descriptive qualitative study facilitates the application of family-centered care within a tertiary care interdisciplinary neurodevelopmental diagnostic assessment clinic by furthering an understanding of parent perceptions of the relevance of diagnostic information provision. An interdisciplinary assessment team completed an open-ended questionnaire to describe parent information provision. Parents from 9 families completed in-depth parent interviews following clinic attendance to discuss perceptions of information received. Interviews were audiotaped, transcribed, and coded by related themes. Parents did not perceive the information in the way professionals expected. Parents acknowledged receipt of comprehensive information relevant to the diagnosis but indicated that not all their needs were met. During the interviews, parents described the assessment process, preassessment information, and "steps in their journey." They noted that a strength-based approach and a focus on parental competency would support their coping efforts. Results underscore the need for professionals to be attentive to parents' individualized needs.

  9. Impact of Non Accounting Information on The Value Relevance of Accounting Information: The Case of Jordan

    Directory of Open Access Journals (Sweden)

    DHIAA SHAMKI

    2013-07-01

    Full Text Available The paper presents empirical evidence about the impact of firm’s shareholders number as non accounting information on the value relevance of its earnings and book value of equity as accounting information for Jordanian industrial firms for the period from 1993 to 2002. Employing the return regression analysis and using shareholders number in two proxies namely local and foreign shareholders number, the findings of the study are fourfold. First, Individual earnings are value relevant while book value is irrelevant. Second, combining earnings with book value leads both of them to be irrelevant. Third, extending local shareholders number has significant impact on the value relevance of individual and combined earnings. Forth, extending foreign shareholders number has significant impact on the value relevance of individual book value and combined earnings. Since studies on the value relevance of these variables have neglected Jordan (and the Middle Eastern region, the study is the first especially in Jordan that tries to fill this gap by examiningthe impact of shareholders numbers on the value relevance of earnings and book valueto indicate firm value.

  10. Extractive Summarisation of Medical Documents

    OpenAIRE

    Abeed Sarker; Diego Molla; Cecile Paris

    2012-01-01

    Background Evidence Based Medicine (EBM) practice requires practitioners to extract evidence from published medical research when answering clinical queries. Due to the time-consuming nature of this practice, there is a strong motivation for systems that can automatically summarise medical documents and help practitioners find relevant information. Aim The aim of this work is to propose an automatic query-focused, extractive summarisation approach that selects informative sentences from medic...

  11. The Relevance of Information and Communication Technologies in ...

    African Journals Online (AJOL)

    The Relevance of Information and Communication Technologies in Libraries Services ... Technologies in Libraries Services and Librarianship Profession in the 21th Century ... This paper therefore, examines the importance of ICT in librarianship as a ... for Authors · for Policy Makers · about Open Access · Journal Quality.

  12. The relevance of accounting information enclosed in performance indicators

    Directory of Open Access Journals (Sweden)

    Mihaela-Cristina Onica

    2012-12-01

    Full Text Available This research study is analyzing the relevance of accounting information reflected through the elaboration of firm performance variables and administration because of the necessity of performance to be administrated. The subject of the theme is enclosed in current development of accounting norms at national , european, (Directives and international levels (IAS/IFRS. The analyised topic is based upon the capabilty of accounting to generate information , throug synthesis calculus being settled the nature , the characteristics and the informational valences of financial performance of an organization. The accounting infromation is base for performing the decison process. The rol of accounting in insurring the relevance and comparability of information increased significantly, being already indispensable. A real solution for communication misunderstanig elimination emerged, as result of diputes in perception and interpretation of economic information, as results for the national speciffic norms.The economic communication is demanding for firm not only in its expression but in thinking and in the process of method conceptualization of organization and administration. A detailed financial situation analysis, which are employing annual financial analysis procedures, underling the performance and risks influencing factors, are considering one starting point for addressing the issue. The introduced variables are insuring a whole vision of firm activity and an appropriate strategy for results significance.

  13. Electronic patient records in action: Transforming information into professionally relevant knowledge.

    Science.gov (United States)

    Winman, Thomas; Rystedt, Hans

    2011-03-01

    The implementation of generic models for organizing information in complex institutions like those in healthcare creates a gap between standardization and the need for locally relevant knowledge. The present study addresses how this gap can be bridged by focusing on the practical work of healthcare staff in transforming information in EPRs into knowledge that is useful for everyday work. Video recording of shift handovers on a rehabilitation ward serves as the empirical case. The results show how extensive selections and reorganizations of information in EPRs are carried out in order to transform information into professionally relevant accounts. We argue that knowledge about the institutional obligations and professional ways of construing information are fundamental for these transitions. The findings point to the need to consider the role of professional knowledge inherent in unpacking information in efforts to develop information systems intended to bridge between institutional and professional boundaries in healthcare. © The Author(s) 2011.

  14. Extraction and Analysis of Information Related to Research & Development Declared Under an Additional Protocol

    International Nuclear Information System (INIS)

    Idinger, J.; Labella, R.; Rialhe, A.; Teller, N.

    2015-01-01

    The additional protocol (AP) provides important tools to strengthen and improve the effectiveness and efficiency of the safeguards system. Safeguards are designed to verify that States comply with their international commitments not to use nuclear material or to engage in nuclear-related activities for the purpose of developing nuclear weapons or other nuclear explosive devices. Under an AP based on INFCIRC/540, a State must provide to the IAEA additional information about, and inspector access to, all parts of its nuclear fuel cycle. In addition, the State has to supply information about its nuclear fuel cycle-related research and development (R&D) activities. The majority of States declare their R&D activities under the AP Articles 2.a.(i), 2.a.(x), and 2.b.(i) as part of initial declarations and their annual updates under the AP. In order to verify consistency and completeness of information provided under the AP by States, the Agency has started to analyze declared R&D information by identifying interrelationships between States in different R&D areas relevant to safeguards. The paper outlines the quality of R&D information provided by States to the Agency, describes how the extraction and analysis of relevant declarations are currently carried out at the Agency and specifies what kinds of difficulties arise during evaluation in respect to cross-linking international projects and finding gaps in reporting. In addition, the paper tries to elaborate how the reporting quality of AP information with reference to R&D activities and the assessment process of R&D information could be improved. (author)

  15. Task-relevant information is prioritized in spatiotemporal contextual cueing.

    Science.gov (United States)

    Higuchi, Yoko; Ueda, Yoshiyuki; Ogawa, Hirokazu; Saiki, Jun

    2016-11-01

    Implicit learning of visual contexts facilitates search performance-a phenomenon known as contextual cueing; however, little is known about contextual cueing under situations in which multidimensional regularities exist simultaneously. In everyday vision, different information, such as object identity and location, appears simultaneously and interacts with each other. We tested the hypothesis that, in contextual cueing, when multiple regularities are present, the regularities that are most relevant to our behavioral goals would be prioritized. Previous studies of contextual cueing have commonly used the visual search paradigm. However, this paradigm is not suitable for directing participants' attention to a particular regularity. Therefore, we developed a new paradigm, the "spatiotemporal contextual cueing paradigm," and manipulated task-relevant and task-irrelevant regularities. In four experiments, we demonstrated that task-relevant regularities were more responsible for search facilitation than task-irrelevant regularities. This finding suggests our visual behavior is focused on regularities that are relevant to our current goal.

  16. Iterative Filtering of Retrieved Information to Increase Relevance

    Directory of Open Access Journals (Sweden)

    Robert Zeidman

    2007-12-01

    Full Text Available Efforts have been underway for years to find more effective ways to retrieve information from large knowledge domains. This effort is now being driven particularly by the Internet and the vast amount of information that is available to unsophisticated users. In the early days of the Internet, some effort involved allowing users to enter Boolean equations of search terms into search engines, for example, rather than just a list of keywords. More recently, effort has focused on understanding a user's desires from past search histories in order to narrow searches. Also there has been much effort to improve the ranking of results based on some measure of relevancy. This paper discusses using iterative filtering of retrieved information to focus in on useful information. This work was done for finding source code correlation and the author extends his findings to Internet searching and e-commerce. The paper presents specific information about a particular filtering application and then generalizes it to other forms of information retrieval.

  17. On the meniscus formation and the negative hydrogen ion extraction from ITER neutral beam injection relevant ion source

    Science.gov (United States)

    Mochalskyy, S.; Wünderlich, D.; Ruf, B.; Fantz, U.; Franzen, P.; Minea, T.

    2014-10-01

    The development of a large area (Asource,ITER = 0.9 × 2 m2) hydrogen negative ion (NI) source constitutes a crucial step in construction of the neutral beam injectors of the international fusion reactor ITER. To understand the plasma behaviour in the boundary layer close to the extraction system the 3D PIC MCC code ONIX is exploited. Direct cross checked analysis of the simulation and experimental results from the ITER-relevant BATMAN source testbed with a smaller area (Asource,BATMAN ≈ 0.32 × 0.59 m2) has been conducted for a low perveance beam, but for a full set of plasma parameters available. ONIX has been partially benchmarked by comparison to the results obtained using the commercial particle tracing code for positive ion extraction KOBRA3D. Very good agreement has been found in terms of meniscus position and its shape for simulations of different plasma densities. The influence of the initial plasma composition on the final meniscus structure was then investigated for NIs. As expected from the Child-Langmuir law, the results show that not only does the extraction potential play a crucial role on the meniscus formation, but also the initial plasma density and its electronegativity. For the given parameters, the calculated meniscus locates a few mm downstream of the plasma grid aperture provoking a direct NI extraction. Most of the surface produced NIs do not reach the plasma bulk, but move directly towards the extraction grid guided by the extraction field. Even for artificially increased electronegativity of the bulk plasma the extracted NI current from this region is low. This observation indicates a high relevance of the direct NI extraction. These calculations show that the extracted NI current from the bulk region is low even if a complete ion-ion plasma is assumed, meaning that direct extraction from surface produced ions should be present in order to obtain sufficiently high extracted NI current density. The calculated extracted currents, both ions

  18. Semantic Information Extraction of Lanes Based on Onboard Camera Videos

    Science.gov (United States)

    Tang, L.; Deng, T.; Ren, C.

    2018-04-01

    In the field of autonomous driving, semantic information of lanes is very important. This paper proposes a method of automatic detection of lanes and extraction of semantic information from onboard camera videos. The proposed method firstly detects the edges of lanes by the grayscale gradient direction, and improves the Probabilistic Hough transform to fit them; then, it uses the vanishing point principle to calculate the lane geometrical position, and uses lane characteristics to extract lane semantic information by the classification of decision trees. In the experiment, 216 road video images captured by a camera mounted onboard a moving vehicle were used to detect lanes and extract lane semantic information. The results show that the proposed method can accurately identify lane semantics from video images.

  19. Online drug databases: a new method to assess and compare inclusion of clinically relevant information.

    Science.gov (United States)

    Silva, Cristina; Fresco, Paula; Monteiro, Joaquim; Rama, Ana Cristina Ribeiro

    2013-08-01

    Evidence-Based Practice requires health care decisions to be based on the best available evidence. The model "Information Mastery" proposes that clinicians should use sources of information that have previously evaluated relevance and validity, provided at the point of care. Drug databases (DB) allow easy and fast access to information and have the benefit of more frequent content updates. Relevant information, in the context of drug therapy, is that which supports safe and effective use of medicines. Accordingly, the European Guideline on the Summary of Product Characteristics (EG-SmPC) was used as a standard to evaluate the inclusion of relevant information contents in DB. To develop and test a method to evaluate relevancy of DB contents, by assessing the inclusion of information items deemed relevant for effective and safe drug use. Hierarchical organisation and selection of the principles defined in the EGSmPC; definition of criteria to assess inclusion of selected information items; creation of a categorisation and quantification system that allows score calculation; calculation of relative differences (RD) of scores for comparison with an "ideal" database, defined as the one that achieves the best quantification possible for each of the information items; pilot test on a sample of 9 drug databases, using 10 drugs frequently associated in literature with morbidity-mortality and also being widely consumed in Portugal. Main outcome measure Calculate individual and global scores for clinically relevant information items of drug monographs in databases, using the categorisation and quantification system created. A--Method development: selection of sections, subsections, relevant information items and corresponding requisites; system to categorise and quantify their inclusion; score and RD calculation procedure. B--Pilot test: calculated scores for the 9 databases; globally, all databases evaluated significantly differed from the "ideal" database; some DB performed

  20. Point Cloud Classification of Tesserae from Terrestrial Laser Data Combined with Dense Image Matching for Archaeological Information Extraction

    Science.gov (United States)

    Poux, F.; Neuville, R.; Billen, R.

    2017-08-01

    Reasoning from information extraction given by point cloud data mining allows contextual adaptation and fast decision making. However, to achieve this perceptive level, a point cloud must be semantically rich, retaining relevant information for the end user. This paper presents an automatic knowledge-based method for pre-processing multi-sensory data and classifying a hybrid point cloud from both terrestrial laser scanning and dense image matching. Using 18 features including sensor's biased data, each tessera in the high-density point cloud from the 3D captured complex mosaics of Germigny-des-prés (France) is segmented via a colour multi-scale abstraction-based featuring extracting connectivity. A 2D surface and outline polygon of each tessera is generated by a RANSAC plane extraction and convex hull fitting. Knowledge is then used to classify every tesserae based on their size, surface, shape, material properties and their neighbour's class. The detection and semantic enrichment method shows promising results of 94% correct semantization, a first step toward the creation of an archaeological smart point cloud.

  1. Communicating stereotype-relevant information: is factual information subject to the same communication biases as fictional information?

    Science.gov (United States)

    Goodman, Ruth L; Webb, Thomas L; Stewart, Andrew J

    2009-07-01

    Factual information is more frequently read and discussed than fictional information. However, research on the role of communication in shaping stereotypes has focused almost exclusively on fictional narratives. In Experiments 1 and 2 a newspaper article containing information about heroin users was communicated along chains of 4 people. No stereotype-consistency bias was observed. Instead, a greater proportion of stereotype-inconsistent information was communicated than was stereotype-consistent or -neutral information. Three further experiments investigated explanations for the difference between the communication of fictional and factual information. Experiment 3 ruled out the possibility that participants' beliefs about the validity of the information could influence the way that it is communicated. Experiments 4 and 5 divided information into concrete (a specific event or fact) or abstract (opinion). A stereotype-consistency bias emerged only for abstract information. In summary, linguistic abstraction moderates whether stereotype-consistency biases emerge in the communication of stereotype-relevant factual information.

  2. Integrating Information Extraction Agents into a Tourism Recommender System

    Science.gov (United States)

    Esparcia, Sergio; Sánchez-Anguix, Víctor; Argente, Estefanía; García-Fornes, Ana; Julián, Vicente

    Recommender systems face some problems. On the one hand information needs to be maintained updated, which can result in a costly task if it is not performed automatically. On the other hand, it may be interesting to include third party services in the recommendation since they improve its quality. In this paper, we present an add-on for the Social-Net Tourism Recommender System that uses information extraction and natural language processing techniques in order to automatically extract and classify information from the Web. Its goal is to maintain the system updated and obtain information about third party services that are not offered by service providers inside the system.

  3. System for Selection of Relevant Information for Decision Support

    Czech Academy of Sciences Publication Activity Database

    Kalina, Jan; Seidl, L.; Zvára, K.; Grünfeldová, H.; Slovák, Dalibor; Zvárová, Jana

    2013-01-01

    Roč. 1, č. 1 (2013), s. 46-46 ISSN 1805-8698. [EFMI 2013 Special Topic Conference. 17.04.2013-19.04.2013, Prague] Institutional support: RVO:67985807 Keywords : decision support system * web-service * information extraction * high-dimension * gene expressions Subject RIV: IN - Informatics, Computer Science

  4. Why relevance theory is relevant for lexicography

    DEFF Research Database (Denmark)

    Bothma, Theo; Tarp, Sven

    2014-01-01

    This article starts by providing a brief summary of relevance theory in information science in relation to the function theory of lexicography, explaining the different types of relevance, viz. objective system relevance and the subjective types of relevance, i.e. topical, cognitive, situational...... that is very important for lexicography as well as for information science, viz. functional relevance. Since all lexicographic work is ultimately aimed at satisfying users’ information needs, the article then discusses why the lexicographer should take note of all these types of relevance when planning a new...... dictionary project, identifying new tasks and responsibilities of the modern lexicographer. The article furthermore discusses how relevance theory impacts on teaching dictionary culture and reference skills. By integrating insights from lexicography and information science, the article contributes to new...

  5. Optimal Information Extraction of Laser Scanning Dataset by Scale-Adaptive Reduction

    Science.gov (United States)

    Zang, Y.; Yang, B.

    2018-04-01

    3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  6. OPTIMAL INFORMATION EXTRACTION OF LASER SCANNING DATASET BY SCALE-ADAPTIVE REDUCTION

    Directory of Open Access Journals (Sweden)

    Y. Zang

    2018-04-01

    Full Text Available 3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  7. Age differences in attention toward decision-relevant information: education matters.

    Science.gov (United States)

    Xing, Cai; Isaacowitz, Derek

    2011-01-01

    Previous studies suggested that older adults are more likely to engage in heuristic decision-making than young adults. This study used eye tracking technique to examine young adults' and highly educated older adults' attention toward two types of decision-relevant information: heuristic cue vs. factual cues. Surprisingly, highly educated older adults showed the reversed age pattern-they looked more toward factual cues than did young adults. This age difference disappeared after controlling for educational level. Additionally, education correlated with attentional pattern to decision-relevant information. We interpret this finding as an indication of the power of education: education may modify what are thought to be "typical" age differences in decision-making, and education may influence young and older people's decision-making via different paths.

  8. Knowledge Dictionary for Information Extraction on the Arabic Text Data

    Directory of Open Access Journals (Sweden)

    Wahyu Jauharis Saputra

    2013-04-01

    Full Text Available Information extraction is an early stage of a process of textual data analysis. Information extraction is required to get information from textual data that can be used for process analysis, such as classification and categorization. A textual data is strongly influenced by the language. Arabic is gaining a significant attention in many studies because Arabic language is very different from others, and in contrast to other languages, tools and research on the Arabic language is still lacking. The information extracted using the knowledge dictionary is a concept of expression. A knowledge dictionary is usually constructed manually by an expert and this would take a long time and is specific to a problem only. This paper proposed a method for automatically building a knowledge dictionary. Dictionary knowledge is formed by classifying sentences having the same concept, assuming that they will have a high similarity value. The concept that has been extracted can be used as features for subsequent computational process such as classification or categorization. Dataset used in this paper was the Arabic text dataset. Extraction result was tested by using a decision tree classification engine and the highest precision value obtained was 71.0% while the highest recall value was 75.0%. 

  9. Phase synchronization of delta and theta oscillations increase during the detection of relevant lexical information

    Directory of Open Access Journals (Sweden)

    Enzo eBrunetti

    2013-06-01

    Full Text Available During monitoring of the discourse, the detection of the relevance of incoming lexical information could be critical for its incorporation to update mental representations in memory. Because, in these situations, the relevance for lexical information is defined by abstract rules that are maintained in memory, results critical to understand how an abstract level of knowledge maintained in mind mediates the detection of the lower-level semantic information. In the present study, we propose that neuronal oscillations participate in the detection of relevant lexical information, based on ‘kept in mind’ rules deriving from more abstract semantic information. We tested our hypothesis using an experimental paradigm that restricted the detection of relevance to inferences based on explicit information, thus controlling for ambiguities derived from implicit aspects. We used a categorization task, in which the semantic relevance was previously defined based on the congruency between a kept in mind category (abstract knowledge, and the lexical-semantic information presented. Our results show that during the detection of the relevant lexical information, phase synchronization of neuronal oscillations selectively increases in delta and theta frequency bands during the interval of semantic analysis. These increments were independent of the semantic category maintained in memory, had a temporal profile specific for each subject, and were mainly induced, as they had no effect on the evoked mean global field power. Also, recruitment of an increased number of pairs of electrodes was a robust observation during the detection of semantic contingent words. These results are consistent with the notion that the detection of relevant lexical information based on a particular semantic rule, could be mediated by increasing the global phase synchronization of neuronal oscillations, which may contribute to the recruitment of an extended number of cortical regions.

  10. Phase synchronization of delta and theta oscillations increase during the detection of relevant lexical information.

    Science.gov (United States)

    Brunetti, Enzo; Maldonado, Pedro E; Aboitiz, Francisco

    2013-01-01

    During monitoring of the discourse, the detection of the relevance of incoming lexical information could be critical for its incorporation to update mental representations in memory. Because, in these situations, the relevance for lexical information is defined by abstract rules that are maintained in memory, a central aspect to elucidate is how an abstract level of knowledge maintained in mind mediates the detection of the lower-level semantic information. In the present study, we propose that neuronal oscillations participate in the detection of relevant lexical information, based on "kept in mind" rules deriving from more abstract semantic information. We tested our hypothesis using an experimental paradigm that restricted the detection of relevance to inferences based on explicit information, thus controlling for ambiguities derived from implicit aspects. We used a categorization task, in which the semantic relevance was previously defined based on the congruency between a kept in mind category (abstract knowledge), and the lexical semantic information presented. Our results show that during the detection of the relevant lexical information, phase synchronization of neuronal oscillations selectively increases in delta and theta frequency bands during the interval of semantic analysis. These increments occurred irrespective of the semantic category maintained in memory, had a temporal profile specific for each subject, and were mainly induced, as they had no effect on the evoked mean global field power. Also, recruitment of an increased number of pairs of electrodes was a robust observation during the detection of semantic contingent words. These results are consistent with the notion that the detection of relevant lexical information based on a particular semantic rule, could be mediated by increasing the global phase synchronization of neuronal oscillations, which may contribute to the recruitment of an extended number of cortical regions.

  11. Detecting clinically relevant new information in clinical notes across specialties and settings.

    Science.gov (United States)

    Zhang, Rui; Pakhomov, Serguei V S; Arsoniadis, Elliot G; Lee, Janet T; Wang, Yan; Melton, Genevieve B

    2017-07-05

    Automated methods for identifying clinically relevant new versus redundant information in electronic health record (EHR) clinical notes is useful for clinicians and researchers involved in patient care and clinical research, respectively. We evaluated methods to automatically identify clinically relevant new information in clinical notes, and compared the quantity of redundant information across specialties and clinical settings. Statistical language models augmented with semantic similarity measures were evaluated as a means to detect and quantify clinically relevant new and redundant information over longitudinal clinical notes for a given patient. A corpus of 591 progress notes over 40 inpatient admissions was annotated for new information longitudinally by physicians to generate a reference standard. Note redundancy between various specialties was evaluated on 71,021 outpatient notes and 64,695 inpatient notes from 500 solid organ transplant patients (April 2015 through August 2015). Our best method achieved at best performance of 0.87 recall, 0.62 precision, and 0.72 F-measure. Addition of semantic similarity metrics compared to baseline improved recall but otherwise resulted in similar performance. While outpatient and inpatient notes had relatively similar levels of high redundancy (61% and 68%, respectively), redundancy differed by author specialty with mean redundancy of 75%, 66%, 57%, and 55% observed in pediatric, internal medicine, psychiatry and surgical notes, respectively. Automated techniques with statistical language models for detecting redundant versus clinically relevant new information in clinical notes do not improve with the addition of semantic similarity measures. While levels of redundancy seem relatively similar in the inpatient and ambulatory settings in the Fairview Health Services, clinical note redundancy appears to vary significantly with different medical specialties.

  12. On the meniscus formation and the negative hydrogen ion extraction from ITER neutral beam injection relevant ion source

    International Nuclear Information System (INIS)

    Mochalskyy, S; Wünderlich, D; Ruf, B; Fantz, U; Franzen, P; Minea, T

    2014-01-01

    The development of a large area (A source,ITER  = 0.9 × 2 m 2 ) hydrogen negative ion (NI) source constitutes a crucial step in construction of the neutral beam injectors of the international fusion reactor ITER. To understand the plasma behaviour in the boundary layer close to the extraction system the 3D PIC MCC code ONIX is exploited. Direct cross checked analysis of the simulation and experimental results from the ITER-relevant BATMAN source testbed with a smaller area (A source,BATMAN  ≈ 0.32 × 0.59 m 2 ) has been conducted for a low perveance beam, but for a full set of plasma parameters available. ONIX has been partially benchmarked by comparison to the results obtained using the commercial particle tracing code for positive ion extraction KOBRA3D. Very good agreement has been found in terms of meniscus position and its shape for simulations of different plasma densities. The influence of the initial plasma composition on the final meniscus structure was then investigated for NIs. As expected from the Child–Langmuir law, the results show that not only does the extraction potential play a crucial role on the meniscus formation, but also the initial plasma density and its electronegativity. For the given parameters, the calculated meniscus locates a few mm downstream of the plasma grid aperture provoking a direct NI extraction. Most of the surface produced NIs do not reach the plasma bulk, but move directly towards the extraction grid guided by the extraction field. Even for artificially increased electronegativity of the bulk plasma the extracted NI current from this region is low. This observation indicates a high relevance of the direct NI extraction. These calculations show that the extracted NI current from the bulk region is low even if a complete ion–ion plasma is assumed, meaning that direct extraction from surface produced ions should be present in order to obtain sufficiently high extracted NI current density. The calculated

  13. A rapid extraction of landslide disaster information research based on GF-1 image

    Science.gov (United States)

    Wang, Sai; Xu, Suning; Peng, Ling; Wang, Zhiyi; Wang, Na

    2015-08-01

    In recent years, the landslide disasters occurred frequently because of the seismic activity. It brings great harm to people's life. It has caused high attention of the state and the extensive concern of society. In the field of geological disaster, landslide information extraction based on remote sensing has been controversial, but high resolution remote sensing image can improve the accuracy of information extraction effectively with its rich texture and geometry information. Therefore, it is feasible to extract the information of earthquake- triggered landslides with serious surface damage and large scale. Taking the Wenchuan county as the study area, this paper uses multi-scale segmentation method to extract the landslide image object through domestic GF-1 images and DEM data, which uses the estimation of scale parameter tool to determine the optimal segmentation scale; After analyzing the characteristics of landslide high-resolution image comprehensively and selecting spectrum feature, texture feature, geometric features and landform characteristics of the image, we can establish the extracting rules to extract landslide disaster information. The extraction results show that there are 20 landslide whose total area is 521279.31 .Compared with visual interpretation results, the extraction accuracy is 72.22%. This study indicates its efficient and feasible to extract earthquake landslide disaster information based on high resolution remote sensing and it provides important technical support for post-disaster emergency investigation and disaster assessment.

  14. Extraction and derivatization of chemical weapons convention relevant aminoalcohols on magnetic cation-exchange resins.

    Science.gov (United States)

    Singh, Varoon; Garg, Prabhat; Chinthakindi, Sridhar; Tak, Vijay; Dubey, Devendra Kumar

    2014-02-14

    Analysis and identification of nitrogen containing aminoalcohols is an integral part of the verification analysis of chemical weapons convention (CWC). This study was aimed to develop extraction and derivatization of aminoalcohols of CWC relevance by using magnetic dispersive solid-phase extraction (MDSPE) in combination with on-resin derivatization (ORD). For this purpose, sulfonated magnetic cation-exchange resins (SMRs) were prepared using magnetite nanoparticles as core, styrene and divinylbenzene as polymer coat and sulfonic acid as acidic cation exchanger. SMRs were successfully employed as extractant for targeted basic analytes. Adsorbed analytes were derivatized with hexamethyldisilazane (HMDS) on the surface of extractant. Derivatized (silylated) compounds were analyzed by GC-MS in SIM and full scan mode. The linearity of the method ranged from 5 to 200ngmL(-1). The LOD and LOQ ranged from 2 to 6ngmL(-1) and 5 to 19ngmL(-1) respectively. The relative standard deviation for intra-day repeatability and inter-day intermediate precision ranged from 5.1% to 6.6% and 0.2% to 7.6% respectively. Recoveries of analytes from spiked water samples from different sources varied from 28.4% to 89.3%. Copyright © 2014 Elsevier B.V. All rights reserved.

  15. The relevance of music information representation metadata from the perspective of expert users

    Directory of Open Access Journals (Sweden)

    Camila Monteiro de Barros

    Full Text Available The general goal of this research was to verify which metadata elements of music information representation are relevant for its retrieval from the perspective of expert music users. Based on a bibliographical research, a comprehensive metadata set of music information representation was developed and transformed into a questionnaire for data collection, which was applied to students and professors of the Graduate Program in Music at the Federal University of Rio Grande do Sul. The results show that the most relevant information for expert music users is related to identification and authorship responsibilities. The respondents from Composition and Interpretative Practice areas agree with these results, while the respondents from Musicology/Ethnomusicology and Music Education areas also consider the metadata related to the historical context of composition relevant.

  16. Information Extraction with Character-level Neural Networks and Free Noisy Supervision

    OpenAIRE

    Meerkamp, Philipp; Zhou, Zhengyi

    2016-01-01

    We present an architecture for information extraction from text that augments an existing parser with a character-level neural network. The network is trained using a measure of consistency of extracted data with existing databases as a form of noisy supervision. Our architecture combines the ability of constraint-based information extraction systems to easily incorporate domain knowledge and constraints with the ability of deep neural networks to leverage large amounts of data to learn compl...

  17. Unsupervised information extraction by text segmentation

    CERN Document Server

    Cortez, Eli

    2013-01-01

    A new unsupervised approach to the problem of Information Extraction by Text Segmentation (IETS) is proposed, implemented and evaluated herein. The authors' approach relies on information available on pre-existing data to learn how to associate segments in the input string with attributes of a given domain relying on a very effective set of content-based features. The effectiveness of the content-based features is also exploited to directly learn from test data structure-based features, with no previous human-driven training, a feature unique to the presented approach. Based on the approach, a

  18. the effect of current and relevant information sources on the use

    African Journals Online (AJOL)

    Admin

    reported similar findings at Yaba College of. Technology, Lagos. However, in a ... values. In other words, current information sources resulted in the use of the library. Jam (1992) identified lack of relevant information sources to be one of the problems facing library users and has ... Bachelor's degree holders. That those with.

  19. Task-Relevant Information Modulates Primary Motor Cortex Activity Before Movement Onset.

    Science.gov (United States)

    Calderon, Cristian B; Van Opstal, Filip; Peigneux, Philippe; Verguts, Tom; Gevers, Wim

    2018-01-01

    Monkey neurophysiology research supports the affordance competition hypothesis (ACH) proposing that cognitive information useful for action selection is integrated in sensorimotor areas. In this view, action selection would emerge from the simultaneous representation of competing action plans, in parallel biased by relevant task factors. This biased competition would take place up to primary motor cortex (M1). Although ACH is plausible in environments affording choices between actions, its relevance for human decision making is less clear. To address this issue, we designed an functional magnetic resonance imaging (fMRI) experiment modeled after monkey neurophysiology studies in which human participants processed cues conveying predictive information about upcoming button presses. Our results demonstrate that, as predicted by the ACH, predictive information (i.e., the relevant task factor) biases activity of primary motor regions. Specifically, first, activity before movement onset in contralateral M1 increases as the competition is biased in favor of a specific button press relative to activity in ipsilateral M1. Second, motor regions were more tightly coupled with fronto-parietal regions when competition between potential actions was high, again suggesting that motor regions are also part of the biased competition network. Our findings support the idea that action planning dynamics as proposed in the ACH are valid both in human and non-human primates.

  20. Relevant Information and Informed Consent in Research: In Defense of the Subjective Standard of Disclosure.

    Science.gov (United States)

    Dranseika, Vilius; Piasecki, Jan; Waligora, Marcin

    2017-02-01

    In this article, we seek to contribute to the debate on the requirement of disclosure in the context of informed consent for research. We defend the subjective standard of disclosure and describe ways to implement this standard in research practice. We claim that the researcher should make an effort to find out what kinds of information are likely to be relevant for those consenting to research. This invites researchers to take empirical survey information seriously, attempt to understand the cultural context, talk to patients to be better able to understand what can be potentially different concerns and interests prevalent in the target population. The subjective standard of disclosure should be seen as a moral ideal that perhaps can never be perfectly implemented but still can and should be used as a normative ideal guiding research practice. In the light of these discussions, we call for more empirical research on what considerations are likely to be perceived as relevant by potential research participants recruited from different socio-economic and cultural groups.

  1. ¹H-NMR simultaneous identification of health-relevant compounds in propolis extracts.

    Science.gov (United States)

    Bertelli, Davide; Papotti, Giulia; Bortolotti, Laura; Marcazzan, Gian Luigi; Plessi, Maria

    2012-01-01

    Propolis is a resinous substance collected by bees from exudates of different plants that is rich in well-known health-relevant phenolic compounds such as flavonoids and phenolic acids. Propolis extracts are very complex matrices difficult to study. Different analytical methods are usable to analyse propolis extracts and to obtain chemical fingerprint but to our knowledge NMR has not previously been used for this purpose. This study aims to demonstrate that it is possible to use ¹H-NMR for the simultaneous recognition of phenolic compounds in complex matrices, such as propolis extracts, using appropriate tools for spectra pre-treatment and analysis. In this work 12 typical phenolic propolis compounds (apigenin, chrysin, galangin, kaempferol, quercetin, naringenin, pinocembrin, pinostrobin, caffeic acid, cinnamic acid, p-coumaric acid and ferulic acid) were considered as reference compounds and their presence in samples was verified by HPLC-MS. A simple ¹H-NMR sequence was used to obtain spectra of samples. Spectra were pre-treated by using an appropriate tool for spectra alignment and analysed by using software for the study of spectra originated from complex matrices. Sixty-five propolis samples were used to test the proposed identification procedure. Ten out of 12 considered compounds were identified as statistically significant in most of the samples. This work suggests that it is possible to efficiently use ¹H-NMR, coupled with appropriate spectral analytical tools, for the simultaneous detection of phenolic compounds in complex matrices. Copyright © 2011 John Wiley & Sons, Ltd.

  2. 77 FR 42339 - Improving Contracting Officers' Access to Relevant Integrity Information

    Science.gov (United States)

    2012-07-18

    ... contracting officers' access to relevant information about contractor business ethics in the Federal Awardee... ability to evaluate the business ethics of prospective contractors and protect the Government from...

  3. An Effective Approach to Biomedical Information Extraction with Limited Training Data

    Science.gov (United States)

    Jonnalagadda, Siddhartha

    2011-01-01

    In the current millennium, extensive use of computers and the internet caused an exponential increase in information. Few research areas are as important as information extraction, which primarily involves extracting concepts and the relations between them from free text. Limitations in the size of training data, lack of lexicons and lack of…

  4. The Common Body of Knowledge: A Framework to Promote Relevant Information Security Research

    Directory of Open Access Journals (Sweden)

    Kenneth J. Knapp

    2007-03-01

    Full Text Available This study proposes using an established common body of knowledge (CBK as one means of organizing information security literature.  Consistent with calls for more relevant information systems (IS research, this industry-developed framework can motivate future research towards topics that are important to the security practitioner.  In this review, forty-eight articles from ten IS journals from 1995 to 2004 are selected and cross-referenced to the ten domains of the information security CBK.  Further, we distinguish articles as empirical research, frameworks, or tutorials.  Generally, this study identified a need for additional empirical research in every CBK domain including topics related to legal aspects of information security.  Specifically, this study identified a need for additional IS security research relating to applications development, physical security, operations security, and business continuity.  The CBK framework is inherently practitioner oriented and using it will promote relevancy by steering IS research towards topics important to practitioners.  This is important considering the frequent calls by prominent information systems scholars for more relevant research.  Few research frameworks have emerged from the literature that specifically classify the diversity of security threats and range of problems that businesses today face.  With the recent surge of interest in security, the need for a comprehensive framework that also promotes relevant research can be of great value.

  5. MedTime: a temporal information extraction system for clinical narratives.

    Science.gov (United States)

    Lin, Yu-Kai; Chen, Hsinchun; Brown, Randall A

    2013-12-01

    Temporal information extraction from clinical narratives is of critical importance to many clinical applications. We participated in the EVENT/TIMEX3 track of the 2012 i2b2 clinical temporal relations challenge, and presented our temporal information extraction system, MedTime. MedTime comprises a cascade of rule-based and machine-learning pattern recognition procedures. It achieved a micro-averaged f-measure of 0.88 in both the recognitions of clinical events and temporal expressions. We proposed and evaluated three time normalization strategies to normalize relative time expressions in clinical texts. The accuracy was 0.68 in normalizing temporal expressions of dates, times, durations, and frequencies. This study demonstrates and evaluates the integration of rule-based and machine-learning-based approaches for high performance temporal information extraction from clinical narratives. Copyright © 2013 Elsevier Inc. All rights reserved.

  6. a Statistical Texture Feature for Building Collapse Information Extraction of SAR Image

    Science.gov (United States)

    Li, L.; Yang, H.; Chen, Q.; Liu, X.

    2018-04-01

    Synthetic Aperture Radar (SAR) has become one of the most important ways to extract post-disaster collapsed building information, due to its extreme versatility and almost all-weather, day-and-night working capability, etc. In view of the fact that the inherent statistical distribution of speckle in SAR images is not used to extract collapsed building information, this paper proposed a novel texture feature of statistical models of SAR images to extract the collapsed buildings. In the proposed feature, the texture parameter of G0 distribution from SAR images is used to reflect the uniformity of the target to extract the collapsed building. This feature not only considers the statistical distribution of SAR images, providing more accurate description of the object texture, but also is applied to extract collapsed building information of single-, dual- or full-polarization SAR data. The RADARSAT-2 data of Yushu earthquake which acquired on April 21, 2010 is used to present and analyze the performance of the proposed method. In addition, the applicability of this feature to SAR data with different polarizations is also analysed, which provides decision support for the data selection of collapsed building information extraction.

  7. AUTOMATIC SUMMARIZATION OF WEB FORUMS AS SOURCES OF PROFESSIONALLY SIGNIFICANT INFORMATION

    Directory of Open Access Journals (Sweden)

    K. I. Buraya

    2016-07-01

    Full Text Available Subject of Research.The competitive advantage of a modern specialist is the widest possible coverage of informationsources useful from the point of view of obtaining and acquisition of relevant professionally significant information. Among these sources professional web forums occupy a significant place. The paperconsiders the problem of automaticforum text summarization, i.e. identification ofthose fragments that contain professionally relevant information. Method.The research is based on statistical analysis of texts of forums by means of machine learning. Six web forums were selected for research considering aspects of technologies of various subject domains as their subject-matter. The marking of forums was carried out by an expert way. Using various methods of machine learning the models were designed reflecting functional communication between the estimated characteristics of PSI extraction quality and signs of posts. The cumulative NDCG metrics and its dispersion were used for an assessment of quality of models.Main Results. We have shown that an important role in an assessment of PSI extraction efficiency is played by requestcontext. The contexts of requestshave been selected,characteristic of PSI extraction, reflecting various interpretations of information needs of users, designated by terms relevance and informational content. The scales for their estimates have been designed corresponding to worldwide approaches. We have experimentally confirmed that results of the summarization of forums carried out by experts manually significantly depend on requestcontext. We have shown that in the general assessment of PSI extraction efficiency relevance is rather well described by a linear combination of features, and the informational content assessment already requires their nonlinear combination. At the same time at a relevance assessment the leading role is played by the features connected with keywords, and at an informational content

  8. DKIE: Open Source Information Extraction for Danish

    DEFF Research Database (Denmark)

    Derczynski, Leon; Field, Camilla Vilhelmsen; Bøgh, Kenneth Sejdenfaden

    2014-01-01

    Danish is a major Scandinavian language spoken daily by around six million people. However, it lacks a unified, open set of NLP tools. This demonstration will introduce DKIE, an extensible open-source toolkit for processing Danish text. We implement an information extraction architecture for Danish...

  9. Transliteration normalization for Information Extraction and Machine Translation

    Directory of Open Access Journals (Sweden)

    Yuval Marton

    2014-12-01

    Full Text Available Foreign name transliterations typically include multiple spelling variants. These variants cause data sparseness and inconsistency problems, increase the Out-of-Vocabulary (OOV rate, and present challenges for Machine Translation, Information Extraction and other natural language processing (NLP tasks. This work aims to identify and cluster name spelling variants using a Statistical Machine Translation method: word alignment. The variants are identified by being aligned to the same “pivot” name in another language (the source-language in Machine Translation settings. Based on word-to-word translation and transliteration probabilities, as well as the string edit distance metric, names with similar spellings in the target language are clustered and then normalized to a canonical form. With this approach, tens of thousands of high-precision name transliteration spelling variants are extracted from sentence-aligned bilingual corpora in Arabic and English (in both languages. When these normalized name spelling variants are applied to Information Extraction tasks, improvements over strong baseline systems are observed. When applied to Machine Translation tasks, a large improvement potential is shown.

  10. End-to-end information extraction without token-level supervision

    DEFF Research Database (Denmark)

    Palm, Rasmus Berg; Hovy, Dirk; Laws, Florian

    2017-01-01

    Most state-of-the-art information extraction approaches rely on token-level labels to find the areas of interest in text. Unfortunately, these labels are time-consuming and costly to create, and consequently, not available for many real-life IE tasks. To make matters worse, token-level labels...... and output text. We evaluate our model on the ATIS data set, MIT restaurant corpus and the MIT movie corpus and compare to neural baselines that do use token-level labels. We achieve competitive results, within a few percentage points of the baselines, showing the feasibility of E2E information extraction...

  11. THE RELEVANCE OF ECONOMIC INFORMATION IN ANALYZING THE ECONOMIC PERFORMANCE

    Directory of Open Access Journals (Sweden)

    PATRUTA MIRCEA IOAN

    2017-12-01

    Full Text Available The performance analysis is based on an informational system, which provides financial information in various formatsand with various applicabilities.We intend to formulate a set of important caracteristics of financial information along with identifying a set of relevant financial rates and indicatorsused to appreciate the performance level of a company. Economic performance can be interpreted in different ways at each level of analysis. Generally, it refers to economic growth, increased productivity and profitability. The growth of labor productivity or increased production per worker is a measure of efficient use of resources in value creation.

  12. Pengaruh Participation Budgeting, Information Asimetry dan Job Relevant Information terhadap Budget Slack pada Institusi Pendidikan (Studi pada Institusi Pendidikan Universitas Kristen Maranatha)

    OpenAIRE

    K. S., Christine Dwi; Agustina, Lidya

    2010-01-01

    The purpose of this research is to analyze and examine the hyphothesis effect of participation budgeting on job relevant information and  information asimetry as a moderating variable , and effect of participation budgeting and information asimetry on budget slack and job relevant information as mediating varible. The respondent of this research are 30 structural staf of program and ministry in Maranatha Christian University who have participated to make  budgeting. This method that...

  13. Mining knowledge from text repositories using information extraction ...

    Indian Academy of Sciences (India)

    Information extraction (IE); text mining; text repositories; knowledge discovery from .... general purpose English words. However ... of precision and recall, as extensive experimentation is required due to lack of public tagged corpora. 4. Mining ...

  14. Towards an information extraction and knowledge formation framework based on Shannon entropy

    Directory of Open Access Journals (Sweden)

    Iliescu Dragoș

    2017-01-01

    Full Text Available Information quantity subject is approached in this paperwork, considering the specific domain of nonconforming product management as information source. This work represents a case study. Raw data were gathered from a heavy industrial works company, information extraction and knowledge formation being considered herein. Involved method for information quantity estimation is based on Shannon entropy formula. Information and entropy spectrum are decomposed and analysed for extraction of specific information and knowledge-that formation. The result of the entropy analysis point out the information needed to be acquired by the involved organisation, this being presented as a specific knowledge type.

  15. Age Differences in Attention toward Decision-Relevant Information: Education Matters

    Science.gov (United States)

    Xing, Cai; Isaacowitz, Derek

    2011-01-01

    Previous studies suggested that older adults are more likely to engage in heuristic decision-making than young adults. This study used eye tracking technique to examine young adults' and highly educated older adults' attention toward two types of decision-relevant information: heuristic cue vs. factual cues. Surprisingly, highly educated older…

  16. The pricing relevance of insider information; Die Preiserheblichkeit von Insiderinformationen

    Energy Technology Data Exchange (ETDEWEB)

    Kruse, Dominik

    2011-07-01

    The publication attempts to describe the so far discussion concerning the feature of pricing relevance and to develop it further with the aid of new research approaches. First, a theoretical outline is presented of the elementary regulation problem of insider trading, its historical development, and the regulation goals of the WpHG. This is followed by an analysis of the concrete specifications of the law. In view of the exemplarity of US law, a country with long experience in regulation of the capital market, the materiality doctrine of US insider law is gone into in some detail. The goals and development of the doctrine are reviewed in the light of court rulings. The third part outlines the requirements of German law in order to forecast the pricing relevance of insider information, while the final part presents a critical review of the current regulations on pricing relevance. (orig./RHM)

  17. Spatially Compact Neural Clusters in the Dorsal Striatum Encode Locomotion Relevant Information.

    Science.gov (United States)

    Barbera, Giovanni; Liang, Bo; Zhang, Lifeng; Gerfen, Charles R; Culurciello, Eugenio; Chen, Rong; Li, Yun; Lin, Da-Ting

    2016-10-05

    An influential striatal model postulates that neural activities in the striatal direct and indirect pathways promote and inhibit movement, respectively. Normal behavior requires coordinated activity in the direct pathway to facilitate intended locomotion and indirect pathway to inhibit unwanted locomotion. In this striatal model, neuronal population activity is assumed to encode locomotion relevant information. Here, we propose a novel encoding mechanism for the dorsal striatum. We identified spatially compact neural clusters in both the direct and indirect pathways. Detailed characterization revealed similar cluster organization between the direct and indirect pathways, and cluster activities from both pathways were correlated with mouse locomotion velocities. Using machine-learning algorithms, cluster activities could be used to decode locomotion relevant behavioral states and locomotion velocity. We propose that neural clusters in the dorsal striatum encode locomotion relevant information and that coordinated activities of direct and indirect pathway neural clusters are required for normal striatal controlled behavior. VIDEO ABSTRACT. Published by Elsevier Inc.

  18. Changing Zaire to Congo: the fate of no-longer relevant mnemonic information.

    Science.gov (United States)

    Eriksson, Johan; Stiernstedt, Mikael; Öhlund, Maria; Nyberg, Lars

    2014-11-01

    In an ever-changing world there is constant pressure on revising long-term memory, such when people or countries change name. What happens to the old, pre-existing information? One possibility is that old associations gradually are weakened and eventually lost. Alternatively, old and no longer relevant information may still be an integral part of memory traces. To test the hypothesis that old mnemonic information still becomes activated when people correctly retrieve new, currently relevant information, brain activity was measured with fMRI while participants performed a cued-retrieval task. Paired associates (symbol-sound and symbol-face pairs) were first learned during two days. Half of the associations were then updated during the next two days, followed by fMRI scanning on day 5 and also 18 months later. As expected, retrieval reactivated sensory cortex related to the most recently learned association (visual cortex for symbol-face pairs, auditory cortex for symbol-sound pairs). Critically, retrieval also reactivated sensory cortex related to the no-longer relevant associate. Eighteen months later, only non-updated symbol-face associations were intact. Intriguingly, a subset of the updated associations was now treated as though the original association had taken over, in that memory performance was significantly worse than chance and that activity in sensory cortex for the original but not the updated associate correlated (negatively) with performance. Moreover, the degree of "residual" reactivation during day 5 inversely predicted memory performance 18 months later. Thus, updating of long-term memory involves adding new information to already existing networks, in which old information can stay resilient for a long time. Copyright © 2014. Published by Elsevier Inc.

  19. Tagline: Information Extraction for Semi-Structured Text Elements in Medical Progress Notes

    Science.gov (United States)

    Finch, Dezon Kile

    2012-01-01

    Text analysis has become an important research activity in the Department of Veterans Affairs (VA). Statistical text mining and natural language processing have been shown to be very effective for extracting useful information from medical documents. However, neither of these techniques is effective at extracting the information stored in…

  20. Mars Target Encyclopedia: Information Extraction for Planetary Science

    Science.gov (United States)

    Wagstaff, K. L.; Francis, R.; Gowda, T.; Lu, Y.; Riloff, E.; Singh, K.

    2017-06-01

    Mars surface targets / and published compositions / Seek and ye will find. We used text mining methods to extract information from LPSC abstracts about the composition of Mars surface targets. Users can search by element, mineral, or target.

  1. How Long Should Routine EEG Be Recorded to Get Relevant Information?

    Science.gov (United States)

    Doudoux, Hannah; Skaare, Kristina; Geay, Thomas; Kahane, Philippe; Bosson, Jean L; Sabourdy, Cécile; Vercueil, Laurent

    2017-03-01

    The optimal duration of routine EEG (rEEG) has not been determined on a clinical basis. This study aims to determine the time required to obtain relevant information during rEEG with respect to the clinical request. All rEEGs performed over 3 months in unselected patients older than 14 years in an academic hospital were analyzed retrospectively. The latency required to obtain relevant information was determined for each rEEG by 2 independent readers blinded to the clinical data. EEG final diagnoses and latencies were analyzed with respect to the main clinical requests: subacute cognitive impairment, spells, transient focal neurologic manifestation or patients referred by epileptologists. From 430 rEEGs performed in the targeted period, 364 were analyzed: 92% of the pathological rEEGs were provided within the first 10 minutes of recording. Slowing background activity was diagnosed from the beginning, whereas interictal epileptiform discharges were recorded over time. Moreover, the time elapsed to demonstrate a pattern differed significantly in the clinical groups: in patients with subacute cognitive impairment, EEG abnormalities appeared within the first 10 minutes, whereas in the other groups, data could be provided over time. Patients with subacute cognitive impairment differed from those in the other groups significantly in the elapsed time required to obtain relevant information during rEEG, suggesting that 10-minute EEG recordings could be sufficient, arguing in favor of individualized rEEG. However, this conclusion does not apply to intensive care unit patients.

  2. Readability, relevance and quality of the information in Spanish on the Web for patients with rheumatoid arthritis.

    Science.gov (United States)

    Castillo-Ortiz, Jose Dionisio; Valdivia-Nuno, Jose de Jesus; Ramirez-Gomez, Andrea; Garagarza-Mariscal, Heber; Gallegos-Rios, Carlos; Flores-Hernandez, Gabriel; Hernandez-Sanchez, Luis; Brambila-Barba, Victor; Castaneda-Sanchez, Jose Juan; Barajas-Ochoa, Zalathiel; Suarez-Rico, Angel; Sanchez-Gonzalez, Jorge Manuel; Ramos-Remus, Cesar

    Education is a major health determinant and one of the main independent outcome predictors in rheumatoid arthritis (RA). The use of the Internet by patients has grown exponentially in the last decade. To assess the characteristics, legibility and quality of the information available in Spanish in the Internet regarding to rheumatoid arthritis. The search was performed in Google using the phrase rheumatoid arthritis. Information from the first 30 pages was evaluated according to a pre-established format (relevance, scope, authorship, type of publication and financial objective). The quality and legibility of the pages were assessed using two validated tools, DISCERN and INFLESZ respectively. Data extraction was performed by senior medical students and evaluation was achieved by consensus. The Google search returned 323 hits but only 63% were considered relevant; 80% of them were information sites (71% discussed exclusively RA, 44% conventional treatment and 12% alternative therapies) and 12.5% had a primary financial interest. 60% of the sites were created by nonprofit organizations and 15% by medical associations. Web sites posted by medical institutions from the United States of America were better positioned in Spanish (Arthritis Foundation 4th position and American College of Rheumatology 10th position) than web sites posted by Spanish speaking countries. There is a risk of disinformation for patients with RA that use the Internet. We identified a window of opportunity for rheumatology medical institutions from Spanish-speaking countries to have a more prominent societal involvement in the education of their patients with RA. Copyright © 2016 Elsevier España, S.L.U. and Sociedad Española de Reumatología y Colegio Mexicano de Reumatología. All rights reserved.

  3. Optimum detection for extracting maximum information from symmetric qubit sets

    International Nuclear Information System (INIS)

    Mizuno, Jun; Fujiwara, Mikio; Sasaki, Masahide; Akiba, Makoto; Kawanishi, Tetsuya; Barnett, Stephen M.

    2002-01-01

    We demonstrate a class of optimum detection strategies for extracting the maximum information from sets of equiprobable real symmetric qubit states of a single photon. These optimum strategies have been predicted by Sasaki et al. [Phys. Rev. A 59, 3325 (1999)]. The peculiar aspect is that the detections with at least three outputs suffice for optimum extraction of information regardless of the number of signal elements. The cases of ternary (or trine), quinary, and septenary polarization signals are studied where a standard von Neumann detection (a projection onto a binary orthogonal basis) fails to access the maximum information. Our experiments demonstrate that it is possible with present technologies to attain about 96% of the theoretical limit

  4. Study on methods and techniques of aeroradiometric weak information extraction for sandstone-hosted uranium deposits based on GIS

    International Nuclear Information System (INIS)

    Han Shaoyang; Ke Dan; Hou Huiqun

    2005-01-01

    The weak information extraction is one of the important research contents in the current sandstone-type uranium prospecting in China. This paper introduces the connotation of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information and establishes some effective mathematic models for weak information extraction. Models for weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are completed in known uranium mineralized areas. Research results prove that the prospective areas of sandstone-type uranium deposits can be rapidly delineated by extracting aeroradiometric weak information. (authors)

  5. Using text mining techniques to extract phenotypic information from the PhenoCHF corpus.

    Science.gov (United States)

    Alnazzawi, Noha; Thompson, Paul; Batista-Navarro, Riza; Ananiadou, Sophia

    2015-01-01

    Phenotypic information locked away in unstructured narrative text presents significant barriers to information accessibility, both for clinical practitioners and for computerised applications used for clinical research purposes. Text mining (TM) techniques have previously been applied successfully to extract different types of information from text in the biomedical domain. They have the potential to be extended to allow the extraction of information relating to phenotypes from free text. To stimulate the development of TM systems that are able to extract phenotypic information from text, we have created a new corpus (PhenoCHF) that is annotated by domain experts with several types of phenotypic information relating to congestive heart failure. To ensure that systems developed using the corpus are robust to multiple text types, it integrates text from heterogeneous sources, i.e., electronic health records (EHRs) and scientific articles from the literature. We have developed several different phenotype extraction methods to demonstrate the utility of the corpus, and tested these methods on a further corpus, i.e., ShARe/CLEF 2013. Evaluation of our automated methods showed that PhenoCHF can facilitate the training of reliable phenotype extraction systems, which are robust to variations in text type. These results have been reinforced by evaluating our trained systems on the ShARe/CLEF corpus, which contains clinical records of various types. Like other studies within the biomedical domain, we found that solutions based on conditional random fields produced the best results, when coupled with a rich feature set. PhenoCHF is the first annotated corpus aimed at encoding detailed phenotypic information. The unique heterogeneous composition of the corpus has been shown to be advantageous in the training of systems that can accurately extract phenotypic information from a range of different text types. Although the scope of our annotation is currently limited to a single

  6. Earlier saccades to task-relevant targets irrespective of relative gain between peripheral and foveal information.

    Science.gov (United States)

    Wolf, Christian; Schütz, Alexander C

    2017-06-01

    Saccades bring objects of interest onto the fovea for high-acuity processing. Saccades to rewarded targets show shorter latencies that correlate negatively with expected motivational value. Shorter latencies are also observed when the saccade target is relevant for a perceptual discrimination task. Here we tested whether saccade preparation is equally influenced by informational value as it is by motivational value. We defined informational value as the probability that information is task-relevant times the ratio between postsaccadic foveal and presaccadic peripheral discriminability. Using a gaze-contingent display, we independently manipulated peripheral and foveal discriminability of the saccade target. Latencies of saccades with perceptual task were reduced by 36 ms in general, but they were not modulated by the information saccades provide (Experiments 1 and 2). However, latencies showed a clear negative linear correlation with the probability that the target is task-relevant (Experiment 3). We replicated that the facilitation by a perceptual task is spatially specific and not due to generally heightened arousal (Experiment 4). Finally, the facilitation only emerged when the perceptual task is in the visual but not in the auditory modality (Experiment 5). Taken together, these results suggest that saccade latencies are not equally modulated by informational value as by motivational value. The facilitation by a perceptual task only arises when task-relevant visual information is foveated, irrespective of whether the foveation is useful or not.

  7. The temporal-relevance temporal-uncertainty model of prospective duration judgment.

    Science.gov (United States)

    Zakay, Dan

    2015-12-15

    A model aimed at explaining prospective duration judgments in real life settings (as well as in the laboratory) is presented. The model is based on the assumption that situational meaning is continuously being extracted by humans' perceptual and cognitive information processing systems. Time is one of the important dimensions of situational meaning. Based on the situational meaning, a value for Temporal Relevance is set. Temporal Relevance reflects the importance of temporal aspects for enabling adaptive behavior in a specific moment in time. When Temporal Relevance is above a certain threshold a prospective duration judgment process is evoked automatically. In addition, a search for relevant temporal information is taking place and its outcomes determine the level of Temporal Uncertainty which reflects the degree of knowledge one has regarding temporal aspects of the task to be performed. The levels of Temporal Relevance and Temporal Uncertainty determine the amount of attentional resources allocated for timing by the executive system. The merit of the model is in connecting timing processes with the ongoing general information processing stream. The model rests on findings in various domains which indicate that cognitive-relevance and self-relevance are powerful determinants of resource allocation policy. The feasibility of the model is demonstrated by analyzing various temporal phenomena. Suggestions for further empirical validation of the model are presented. Copyright © 2015 Elsevier Inc. All rights reserved.

  8. Information Extraction, Data Integration, and Uncertain Data Management: The State of The Art

    NARCIS (Netherlands)

    Habib, Mena Badieh; van Keulen, Maurice

    2011-01-01

    Information Extraction, data Integration, and uncertain data management are different areas of research that got vast focus in the last two decades. Many researches tackled those areas of research individually. However, information extraction systems should have integrated with data integration

  9. Information extraction from muon radiography data

    International Nuclear Information System (INIS)

    Borozdin, K.N.; Asaki, T.J.; Chartrand, R.; Hengartner, N.W.; Hogan, G.E.; Morris, C.L.; Priedhorsky, W.C.; Schirato, R.C.; Schultz, L.J.; Sottile, M.J.; Vixie, K.R.; Wohlberg, B.E.; Blanpied, G.

    2004-01-01

    Scattering muon radiography was proposed recently as a technique of detection and 3-d imaging for dense high-Z objects. High-energy cosmic ray muons are deflected in matter in the process of multiple Coulomb scattering. By measuring the deflection angles we are able to reconstruct the configuration of high-Z material in the object. We discuss the methods for information extraction from muon radiography data. Tomographic methods widely used in medical images have been applied to a specific muon radiography information source. Alternative simple technique based on the counting of high-scattered muons in the voxels seems to be efficient in many simulated scenes. SVM-based classifiers and clustering algorithms may allow detection of compact high-Z object without full image reconstruction. The efficiency of muon radiography can be increased using additional informational sources, such as momentum estimation, stopping power measurement, and detection of muonic atom emission.

  10. Proactive Support of Internet Browsing when Searching for Relevant Health Information.

    Science.gov (United States)

    Rurik, Clas; Zowalla, Richard; Wiesner, Martin; Pfeifer, Daniel

    2015-01-01

    Many people use the Internet as one of the primary sources of health information. This is due to the high volume and easy access of freely available information regarding diseases, diagnoses and treatments. However, users may find it difficult to retrieve information which is easily understandable and does not require a deep medical background. In this paper, we present a new kind of Web browser add-on, in order to proactively support users when searching for relevant health information. Our add-on not only visualizes the understandability of displayed medical text but also provides further recommendations of Web pages which hold similar content but are potentially easier to comprehend.

  11. Overview of image processing tools to extract physical information from JET videos

    Science.gov (United States)

    Craciunescu, T.; Murari, A.; Gelfusa, M.; Tiseanu, I.; Zoita, V.; EFDA Contributors, JET

    2014-11-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  12. Overview of image processing tools to extract physical information from JET videos

    International Nuclear Information System (INIS)

    Craciunescu, T; Tiseanu, I; Zoita, V; Murari, A; Gelfusa, M

    2014-01-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  13. Towards a realistic 3D simulation of the extraction region in ITER NBI relevant ion source

    Science.gov (United States)

    Mochalskyy, S.; Wünderlich, D.; Fantz, U.; Franzen, P.; Minea, T.

    2015-03-01

    The development of negative ion (NI) sources for ITER is strongly accompanied by modelling activities. The ONIX code addresses the physics of formation and extraction of negative hydrogen ions at caesiated sources as well as the amount of co-extracted electrons. In order to be closer to the experimental conditions the code has been improved. It includes now the bias potential applied to first grid (plasma grid) of the extraction system, and the presence of Cs+ ions in the plasma. The simulation results show that such aspects play an important role for the formation of an ion-ion plasma in the boundary region by reducing the depth of the negative potential well in vicinity to the plasma grid that limits the extraction of the NIs produced at the Cs covered plasma grid surface. The influence of the initial temperature of the surface produced NI and its emission rate on the NI density in the bulk plasma that in turn affects the beam formation region was analysed. The formation of the plasma meniscus, the boundary between the plasma and the beam, was investigated for the extraction potentials of 5 and 10 kV. At the smaller extraction potential the meniscus moves closer to the plasma grid but as in the case of 10 kV the deepest meniscus bend point is still outside of the aperture. Finally, a plasma containing the same amount of NI and electrons (nH- =ne =1017 m-3) , representing good source conditioning, was simulated. It is shown that at such conditions the extracted NI current can reach values of ˜32 mA cm-2 using ITER-relevant extraction potential of 10 kV and ˜19 mA cm-2 at 5 kV. These results are in good agreement with experimental measurements performed at the small scale ITER prototype source at the test facility BATMAN.

  14. Performance of multi-aperture grid extraction systems for an ITER-relevant RF-driven negative hydrogen ion source

    Science.gov (United States)

    Franzen, P.; Gutser, R.; Fantz, U.; Kraus, W.; Falter, H.; Fröschle, M.; Heinemann, B.; McNeely, P.; Nocentini, R.; Riedl, R.; Stäbler, A.; Wünderlich, D.

    2011-07-01

    The ITER neutral beam system requires a negative hydrogen ion beam of 48 A with an energy of 0.87 MeV, and a negative deuterium beam of 40 A with an energy of 1 MeV. The beam is extracted from a large ion source of dimension 1.9 × 0.9 m2 by an acceleration system consisting of seven grids with 1280 apertures each. Currently, apertures with a diameter of 14 mm in the first grid are foreseen. In 2007, the IPP RF source was chosen as the ITER reference source due to its reduced maintenance compared with arc-driven sources and the successful development at the BATMAN test facility of being equipped with the small IPP prototype RF source ( {\\sim}\\frac{1}{8} of the area of the ITER NBI source). These results, however, were obtained with an extraction system with 8 mm diameter apertures. This paper reports on the comparison of the source performance at BATMAN of an ITER-relevant extraction system equipped with chamfered apertures with a 14 mm diameter and 8 mm diameter aperture extraction system. The most important result is that there is almost no difference in the achieved current density—being consistent with ion trajectory calculations—and the amount of co-extracted electrons. Furthermore, some aspects of the beam optics of both extraction systems are discussed.

  15. Performance of multi-aperture grid extraction systems for an ITER-relevant RF-driven negative hydrogen ion source

    International Nuclear Information System (INIS)

    Franzen, P.; Gutser, R.; Fantz, U.; Kraus, W.; Falter, H.; Froeschle, M.; Heinemann, B.; McNeely, P.; Nocentini, R.; Riedl, R.; Staebler, A.; Wuenderlich, D.

    2011-01-01

    The ITER neutral beam system requires a negative hydrogen ion beam of 48 A with an energy of 0.87 MeV, and a negative deuterium beam of 40 A with an energy of 1 MeV. The beam is extracted from a large ion source of dimension 1.9 x 0.9 m 2 by an acceleration system consisting of seven grids with 1280 apertures each. Currently, apertures with a diameter of 14 mm in the first grid are foreseen. In 2007, the IPP RF source was chosen as the ITER reference source due to its reduced maintenance compared with arc-driven sources and the successful development at the BATMAN test facility of being equipped with the small IPP prototype RF source ( ∼ 1/8 of the area of the ITER NBI source). These results, however, were obtained with an extraction system with 8 mm diameter apertures. This paper reports on the comparison of the source performance at BATMAN of an ITER-relevant extraction system equipped with chamfered apertures with a 14 mm diameter and 8 mm diameter aperture extraction system. The most important result is that there is almost no difference in the achieved current density-being consistent with ion trajectory calculations-and the amount of co-extracted electrons. Furthermore, some aspects of the beam optics of both extraction systems are discussed.

  16. Prioritising the relevant information for learning and decision making within orbital and ventromedial prefrontal cortex.

    Science.gov (United States)

    Walton, Mark E; Chau, Bolton K H; Kennerley, Steven W

    2015-02-01

    Our environment and internal states are frequently complex, ambiguous and dynamic, meaning we need to have selection mechanisms to ensure we are basing our decisions on currently relevant information. Here, we review evidence that orbitofrontal (OFC) and ventromedial prefrontal cortex (VMPFC) play conserved, critical but distinct roles in this process. While OFC may use specific sensory associations to enhance task-relevant information, particularly in the context of learning, VMPFC plays a role in ensuring irrelevant information does not impinge on the decision in hand.

  17. A Parallel Relational Database Management System Approach to Relevance Feedback in Information Retrieval.

    Science.gov (United States)

    Lundquist, Carol; Frieder, Ophir; Holmes, David O.; Grossman, David

    1999-01-01

    Describes a scalable, parallel, relational database-drive information retrieval engine. To support portability across a wide range of execution environments, all algorithms adhere to the SQL-92 standard. By incorporating relevance feedback algorithms, accuracy is enhanced over prior database-driven information retrieval efforts. Presents…

  18. Enhancing biomedical text summarization using semantic relation extraction.

    Directory of Open Access Journals (Sweden)

    Yue Shang

    Full Text Available Automatic text summarization for a biomedical concept can help researchers to get the key points of a certain topic from large amount of biomedical literature efficiently. In this paper, we present a method for generating text summary for a given biomedical concept, e.g., H1N1 disease, from multiple documents based on semantic relation extraction. Our approach includes three stages: 1 We extract semantic relations in each sentence using the semantic knowledge representation tool SemRep. 2 We develop a relation-level retrieval method to select the relations most relevant to each query concept and visualize them in a graphic representation. 3 For relations in the relevant set, we extract informative sentences that can interpret them from the document collection to generate text summary using an information retrieval based method. Our major focus in this work is to investigate the contribution of semantic relation extraction to the task of biomedical text summarization. The experimental results on summarization for a set of diseases show that the introduction of semantic knowledge improves the performance and our results are better than the MEAD system, a well-known tool for text summarization.

  19. Enhancing biomedical text summarization using semantic relation extraction.

    Science.gov (United States)

    Shang, Yue; Li, Yanpeng; Lin, Hongfei; Yang, Zhihao

    2011-01-01

    Automatic text summarization for a biomedical concept can help researchers to get the key points of a certain topic from large amount of biomedical literature efficiently. In this paper, we present a method for generating text summary for a given biomedical concept, e.g., H1N1 disease, from multiple documents based on semantic relation extraction. Our approach includes three stages: 1) We extract semantic relations in each sentence using the semantic knowledge representation tool SemRep. 2) We develop a relation-level retrieval method to select the relations most relevant to each query concept and visualize them in a graphic representation. 3) For relations in the relevant set, we extract informative sentences that can interpret them from the document collection to generate text summary using an information retrieval based method. Our major focus in this work is to investigate the contribution of semantic relation extraction to the task of biomedical text summarization. The experimental results on summarization for a set of diseases show that the introduction of semantic knowledge improves the performance and our results are better than the MEAD system, a well-known tool for text summarization.

  20. Recognition techniques for extracting information from semistructured documents

    Science.gov (United States)

    Della Ventura, Anna; Gagliardi, Isabella; Zonta, Bruna

    2000-12-01

    Archives of optical documents are more and more massively employed, the demand driven also by the new norms sanctioning the legal value of digital documents, provided they are stored on supports that are physically unalterable. On the supply side there is now a vast and technologically advanced market, where optical memories have solved the problem of the duration and permanence of data at costs comparable to those for magnetic memories. The remaining bottleneck in these systems is the indexing. The indexing of documents with a variable structure, while still not completely automated, can be machine supported to a large degree with evident advantages both in the organization of the work, and in extracting information, providing data that is much more detailed and potentially significant for the user. We present here a system for the automatic registration of correspondence to and from a public office. The system is based on a general methodology for the extraction, indexing, archiving, and retrieval of significant information from semi-structured documents. This information, in our prototype application, is distributed among the database fields of sender, addressee, subject, date, and body of the document.

  1. Video coding and decoding devices and methods preserving ppg relevant information

    NARCIS (Netherlands)

    2013-01-01

    The present invention relates to a video encoding device (10) for encoding video data and a corresponding video decoding device, wherein during decoding PPG relevant information shall be preserved. For this purpose the video coding device (10) comprises a first encoder (20) for encoding input video

  2. Acoustic fine structure may encode biologically relevant information for zebra finches.

    Science.gov (United States)

    Prior, Nora H; Smith, Edward; Lawson, Shelby; Ball, Gregory F; Dooling, Robert J

    2018-04-18

    The ability to discriminate changes in the fine structure of complex sounds is well developed in birds. However, the precise limit of this discrimination ability and how it is used in the context of natural communication remains unclear. Here we describe natural variability in acoustic fine structure of male and female zebra finch calls. Results from psychoacoustic experiments demonstrate that zebra finches are able to discriminate extremely small differences in fine structure, which are on the order of the variation in acoustic fine structure that is present in their vocal signals. Results from signal analysis methods also suggest that acoustic fine structure may carry information that distinguishes between biologically relevant categories including sex, call type and individual identity. Combined, our results are consistent with the hypothesis that zebra finches can encode biologically relevant information within the fine structure of their calls. This study provides a foundation for our understanding of how acoustic fine structure may be involved in animal communication.

  3. A Planetary Defense Gateway for Smart Discovery of relevant Information for Decision Support

    Science.gov (United States)

    Bambacus, Myra; Yang, Chaowei Phil; Leung, Ronald Y.; Barbee, Brent; Nuth, Joseph A.; Seery, Bernard; Jiang, Yongyao; Qin, Han; Li, Yun; Yu, Manzhu; hide

    2017-01-01

    A Planetary Defense Gateway for Smart Discovery of relevant Information for Decision Support presentation discussing background, framework architecture, current results, ongoing research, conclusions.

  4. Disseminating relevant health information to underserved audiences: implications of the Digital Divide Pilot Projects.

    Science.gov (United States)

    Kreps, Gary L

    2005-10-01

    This paper examines the influence of the digital divide on disparities in health outcomes for vulnerable populations, identifying implications for medical and public libraries. The paper describes the results of the Digital Divide Pilot Projects demonstration research programs funded by the National Cancer Institute to test new strategies for disseminating relevant health information to underserved and at-risk audiences. The Digital Divide Pilot Projects field-tested innovative systemic strategies for helping underserved populations access and utilize relevant health information to make informed health-related decisions about seeking appropriate health care and support, resisting avoidable and significant health risks, and promoting their own health. The paper builds on the Digital Divide Pilot Projects by identifying implications for developing health communication strategies that libraries can adopt to provide digital health information to vulnerable populations.

  5. Kernel-Based Relevance Analysis with Enhanced Interpretability for Detection of Brain Activity Patterns

    Directory of Open Access Journals (Sweden)

    Andres M. Alvarez-Meza

    2017-10-01

    Full Text Available We introduce Enhanced Kernel-based Relevance Analysis (EKRA that aims to support the automatic identification of brain activity patterns using electroencephalographic recordings. EKRA is a data-driven strategy that incorporates two kernel functions to take advantage of the available joint information, associating neural responses to a given stimulus condition. Regarding this, a Centered Kernel Alignment functional is adjusted to learning the linear projection that best discriminates the input feature set, optimizing the required free parameters automatically. Our approach is carried out in two scenarios: (i feature selection by computing a relevance vector from extracted neural features to facilitating the physiological interpretation of a given brain activity task, and (ii enhanced feature selection to perform an additional transformation of relevant features aiming to improve the overall identification accuracy. Accordingly, we provide an alternative feature relevance analysis strategy that allows improving the system performance while favoring the data interpretability. For the validation purpose, EKRA is tested in two well-known tasks of brain activity: motor imagery discrimination and epileptic seizure detection. The obtained results show that the EKRA approach estimates a relevant representation space extracted from the provided supervised information, emphasizing the salient input features. As a result, our proposal outperforms the state-of-the-art methods regarding brain activity discrimination accuracy with the benefit of enhanced physiological interpretation about the task at hand.

  6. Video coding and decoding devices and methods preserving PPG relevant information

    NARCIS (Netherlands)

    2015-01-01

    The present invention relates to a video encoding device (10, 10', 10") and method for encoding video data and to a corresponding video decoding device (60, 60') and method. To preserve PPG relevant information after encoding without requiring a large amount of additional data for the video encoder

  7. Video coding and decoding devices and methods preserving ppg relevant information

    NARCIS (Netherlands)

    2013-01-01

    The present invention relates to a video encoding device (10, 10', 10'') and method for encoding video data and to a corresponding video decoding device (60, 60') and method. To preserve PPG relevant information after encoding without requiring a large amount of additional data for the video encoder

  8. Perceived relevance and information needs regarding food topics and preferred information sources among Dutch adults: results of a quantitative consumer study

    NARCIS (Netherlands)

    Dillen, van S.M.E.; Hiddink, G.J.; Koelen, M.A.; Graaf, de C.; Woerkum, van C.M.J.

    2004-01-01

    Objective: For more effective nutrition communication, it is crucial to identify sources from which consumers seek information. Our purpose was to assess perceived relevance and information needs regarding food topics, and preferred information sources by means of quantitative consumer research.

  9. RESEARCH ON REMOTE SENSING GEOLOGICAL INFORMATION EXTRACTION BASED ON OBJECT ORIENTED CLASSIFICATION

    Directory of Open Access Journals (Sweden)

    H. Gao

    2018-04-01

    Full Text Available The northern Tibet belongs to the Sub cold arid climate zone in the plateau. It is rarely visited by people. The geological working conditions are very poor. However, the stratum exposures are good and human interference is very small. Therefore, the research on the automatic classification and extraction of remote sensing geological information has typical significance and good application prospect. Based on the object-oriented classification in Northern Tibet, using the Worldview2 high-resolution remote sensing data, combined with the tectonic information and image enhancement, the lithological spectral features, shape features, spatial locations and topological relations of various geological information are excavated. By setting the threshold, based on the hierarchical classification, eight kinds of geological information were classified and extracted. Compared with the existing geological maps, the accuracy analysis shows that the overall accuracy reached 87.8561 %, indicating that the classification-oriented method is effective and feasible for this study area and provides a new idea for the automatic extraction of remote sensing geological information.

  10. Terrain Extraction by Integrating Terrestrial Laser Scanner Data and Spectral Information

    Science.gov (United States)

    Lau, C. L.; Halim, S.; Zulkepli, M.; Azwan, A. M.; Tang, W. L.; Chong, A. K.

    2015-10-01

    The extraction of true terrain points from unstructured laser point cloud data is an important process in order to produce an accurate digital terrain model (DTM). However, most of these spatial filtering methods just utilizing the geometrical data to discriminate the terrain points from nonterrain points. The point cloud filtering method also can be improved by using the spectral information available with some scanners. Therefore, the objective of this study is to investigate the effectiveness of using the three-channel (red, green and blue) of the colour image captured from built-in digital camera which is available in some Terrestrial Laser Scanner (TLS) for terrain extraction. In this study, the data acquisition was conducted at a mini replica landscape in Universiti Teknologi Malaysia (UTM), Skudai campus using Leica ScanStation C10. The spectral information of the coloured point clouds from selected sample classes are extracted for spectral analysis. The coloured point clouds which within the corresponding preset spectral threshold are identified as that specific feature point from the dataset. This process of terrain extraction is done through using developed Matlab coding. Result demonstrates that a higher spectral resolution passive image is required in order to improve the output. This is because low quality of the colour images captured by the sensor contributes to the low separability in spectral reflectance. In conclusion, this study shows that, spectral information is capable to be used as a parameter for terrain extraction.

  11. Extracting Semantic Information from Visual Data: A Survey

    Directory of Open Access Journals (Sweden)

    Qiang Liu

    2016-03-01

    Full Text Available The traditional environment maps built by mobile robots include both metric ones and topological ones. These maps are navigation-oriented and not adequate for service robots to interact with or serve human users who normally rely on the conceptual knowledge or semantic contents of the environment. Therefore, the construction of semantic maps becomes necessary for building an effective human-robot interface for service robots. This paper reviews recent research and development in the field of visual-based semantic mapping. The main focus is placed on how to extract semantic information from visual data in terms of feature extraction, object/place recognition and semantic representation methods.

  12. Selecting Relevant Information for Medical Decision Support with Application in Cardiology

    Czech Academy of Sciences Publication Activity Database

    Kalina, Jan; Seidl, L.; Grünfeldová, H.; Slovák, Dalibor; Zvárová, Jana

    2013-01-01

    Roč. 9, č. 1 (2013), s. 2-6 ISSN 1801-5603 R&D Projects: GA MŠk(CZ) 1M06014 Institutional support: RVO:67985807 Keywords : decision support system * web-service * information extraction * high dimension * gene expressions Subject RIV: IN - Informatics, Computer Science http://www.ejbi.org/img/ejbi/2013/1/Kalina_en.pdf

  13. Disseminating relevant health information to underserved audiences: implications of the Digital Divide Pilot Projects*

    Science.gov (United States)

    Kreps, Gary L.

    2005-01-01

    Objective: This paper examines the influence of the digital divide on disparities in health outcomes for vulnerable populations, identifying implications for medical and public libraries. Method: The paper describes the results of the Digital Divide Pilot Projects demonstration research programs funded by the National Cancer Institute to test new strategies for disseminating relevant health information to underserved and at-risk audiences. Results: The Digital Divide Pilot Projects field-tested innovative systemic strategies for helping underserved populations access and utilize relevant health information to make informed health-related decisions about seeking appropriate health care and support, resisting avoidable and significant health risks, and promoting their own health. Implications: The paper builds on the Digital Divide Pilot Projects by identifying implications for developing health communication strategies that libraries can adopt to provide digital health information to vulnerable populations. PMID:16239960

  14. Towards a realistic 3D simulation of the extraction region in ITER NBI relevant ion source

    International Nuclear Information System (INIS)

    Mochalskyy, S.; Wünderlich, D.; Fantz, U.; Franzen, P.; Minea, T.

    2015-01-01

    The development of negative ion (NI) sources for ITER is strongly accompanied by modelling activities. The ONIX code addresses the physics of formation and extraction of negative hydrogen ions at caesiated sources as well as the amount of co-extracted electrons. In order to be closer to the experimental conditions the code has been improved. It includes now the bias potential applied to first grid (plasma grid) of the extraction system, and the presence of Cs + ions in the plasma. The simulation results show that such aspects play an important role for the formation of an ion–ion plasma in the boundary region by reducing the depth of the negative potential well in vicinity to the plasma grid that limits the extraction of the NIs produced at the Cs covered plasma grid surface. The influence of the initial temperature of the surface produced NI and its emission rate on the NI density in the bulk plasma that in turn affects the beam formation region was analysed. The formation of the plasma meniscus, the boundary between the plasma and the beam, was investigated for the extraction potentials of 5 and 10 kV. At the smaller extraction potential the meniscus moves closer to the plasma grid but as in the case of 10 kV the deepest meniscus bend point is still outside of the aperture. Finally, a plasma containing the same amount of NI and electrons (n H − =n e =10 17 m −3 ), representing good source conditioning, was simulated. It is shown that at such conditions the extracted NI current can reach values of ∼32 mA cm −2 using ITER-relevant extraction potential of 10 kV and ∼19 mA cm −2 at 5 kV. These results are in good agreement with experimental measurements performed at the small scale ITER prototype source at the test facility BATMAN. (paper)

  15. Effect of the Adoption of IFRS on the Information Relevance of Accounting Profits in Brazil

    Directory of Open Access Journals (Sweden)

    Mateus Alexandre Costa dos Santos

    2014-12-01

    Full Text Available This study aimed to assess the effect of adopting the International Financial Reporting Standards (IFRS in Brazil on the information relevance of accounting profits of publicly traded companies. International studies have shown that the adoption of IFRS improves the quality of accounting information compared with domestic accounting standards. Concurrent evidence is sparse in Brazil. Information relevance is understood herein as a multidimensional attribute that is closely related to the quality and usefulness of the information conveyed by accounting profits. The associative capacity and information timeliness of accounting profits in relation to share prices were examined. Furthermore, the level of conditional conservatism present in accounting profits was also analyzed because according to Basu (1997, this aspect is related to timeliness. The study used pooled regressions and panel data models to analyze the quarterly accounting profits of 246 companies between the first quarter of 1999 and the first quarter of 2013, resulting in 9,558 quarter-company observations. The results indicated that the adoption of IFRS in Brazil (1 increased the associative capacity of accounting profits; (2 reduced information timeliness to non-significant levels; and (3 had no effect on conditional conservatism. The joint analysis of the empirical evidence from the present study conclusively precludes stating that the adoption of IFRS in Brazil contributed to an increase the information relevance of accounting profits of publicly traded companies.

  16. Perceived Relevance of an Introductory Information Systems Course to Prospective Business Students

    Directory of Open Access Journals (Sweden)

    Irene Govender

    2013-12-01

    Full Text Available The study is designed to examine students’ perceptions of the introductory Information Systems (IS course. It was an exploratory study in which 67 students participated. A quantitative approach was followed making use of questionnaires for the collection of data. Using the theory of reasoned action as a framework, the study explores the factors that influence non-IS major students’ perceived relevance of the IS introductory course. The analysis of collected data included descriptive and inferential statistics. Using multiple regression analysis, the results suggest that overall, the independent variables, relevance of the content, previous IT knowledge, relevance for professional practice, IT preference in courses and peers’ influence may account for 72% of the explanatory power for the dependent variable, perceived relevance of the IS course. In addition, the results have shown some strong predictors (IT preference and peers’ influence that influence students’ perceived relevance of the IS course. Practical work was found to be a strong mediating variable toward positive perceptions of IS. The results of this study suggest that students do indeed perceive the introductory IS course to be relevant and match their professional needs, but more practical work would enhance their learning. Implications for theory and practice are discussed as a result of the behavioural intention to perceive the IS course to be relevant and eventually to recruit more IS students.

  17. Extracting the information backbone in online system.

    Directory of Open Access Journals (Sweden)

    Qian-Ming Zhang

    Full Text Available Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  18. Extracting the information backbone in online system.

    Science.gov (United States)

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  19. Extracting the Information Backbone in Online System

    Science.gov (United States)

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such “less can be more” feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency. PMID:23690946

  20. Behavioral and Event-Related-Potential Correlates of Processing Congruent and Incongruent Self-Relevant Information

    Science.gov (United States)

    Clark, Sheri L.

    2013-01-01

    People want to be viewed by others as they view themselves. Being confronted with self-relevant information that is either congruent or incongruent with one's self-view has been shown to differentially affect subsequent behavior, memory for the information, and evaluation of the source of the information. However, no research has examined…

  1. Suppression of no-longer relevant information in Working Memory: An alpha-power related mechanism?

    Science.gov (United States)

    Poch, Claudia; Valdivia, María; Capilla, Almudena; Hinojosa, José Antonio; Campo, Pablo

    2018-03-27

    Selective attention can enhance Working Memory (WM) performance by selecting relevant information, while preventing distracting items from encoding or from further maintenance. Alpha oscillatory modulations are a correlate of visuospatial attention. Specifically, an enhancement of alpha power is observed in the ipsilateral posterior cortex to the locus of attention, along with a suppression in the contralateral hemisphere. An influential model proposes that the alpha enhancement is functionally related to the suppression of information. However, whether ipsilateral alpha power represents a mechanism through which no longer relevant WM representations are inhibited has yet not been explored. Here we examined whether the amount of distractors to be suppressed during WM maintenance is functionally related to alpha power lateralized activity. We measure EEG activity while participants (N = 36) performed a retro-cue task in which the WM load was varied across the relevant/irrelevant post-cue hemifield. We found that alpha activity was lateralized respect to the locus of attention, but did not track post-cue irrelevant load. Additionally, non-lateralized alpha activity increased with post-cue relevant load. We propose that alpha lateralization associated to retro-cuing might be related to a general orienting mechanism toward relevant representation. Copyright © 2018 Elsevier B.V. All rights reserved.

  2. Task relevance of emotional information affects anxiety-linked attention bias in visual search.

    Science.gov (United States)

    Dodd, Helen F; Vogt, Julia; Turkileri, Nilgun; Notebaert, Lies

    2017-01-01

    Task relevance affects emotional attention in healthy individuals. Here, we investigate whether the association between anxiety and attention bias is affected by the task relevance of emotion during an attention task. Participants completed two visual search tasks. In the emotion-irrelevant task, participants were asked to indicate whether a discrepant face in a crowd of neutral, middle-aged faces was old or young. Irrelevant to the task, target faces displayed angry, happy, or neutral expressions. In the emotion-relevant task, participants were asked to indicate whether a discrepant face in a crowd of middle-aged neutral faces was happy or angry (target faces also varied in age). Trait anxiety was not associated with attention in the emotion-relevant task. However, in the emotion-irrelevant task, trait anxiety was associated with a bias for angry over happy faces. These findings demonstrate that the task relevance of emotional information affects conclusions about the presence of an anxiety-linked attention bias. Copyright © 2016 Elsevier B.V. All rights reserved.

  3. Collecting Safeguards Relevant Trade Information: The IAEA Procurement Outreach Programme

    International Nuclear Information System (INIS)

    Schot, P.; El Gebaly, A.; Tarvainen, M.

    2010-01-01

    The increasing awareness of activities of transnational procurement networks to covertly acquire sensitive nuclear related dual use equipment prompted an evolution of safeguards methodologies. One of the responses to this challenge by the Department of Safeguards in the IAEA was to establish the Trade and Technology Unit (TTA) in November 2004 to analyse and report on these covert nuclear related trade activities. To obtain information relevant to this analysis, TTA is engaging States that might be willing to provide this information to the Secretariat on a voluntary basis. This paper will give an overview of current activities, sum up the results achieved and discuss suggestions to further improve this programme made by Member States. (author)

  4. Extraction Method for Earthquake-Collapsed Building Information Based on High-Resolution Remote Sensing

    International Nuclear Information System (INIS)

    Chen, Peng; Wu, Jian; Liu, Yaolin; Wang, Jing

    2014-01-01

    At present, the extraction of earthquake disaster information from remote sensing data relies on visual interpretation. However, this technique cannot effectively and quickly obtain precise and efficient information for earthquake relief and emergency management. Collapsed buildings in the town of Zipingpu after the Wenchuan earthquake were used as a case study to validate two kinds of rapid extraction methods for earthquake-collapsed building information based on pixel-oriented and object-oriented theories. The pixel-oriented method is based on multi-layer regional segments that embody the core layers and segments of the object-oriented method. The key idea is to mask layer by layer all image information, including that on the collapsed buildings. Compared with traditional techniques, the pixel-oriented method is innovative because it allows considerably rapid computer processing. As for the object-oriented method, a multi-scale segment algorithm was applied to build a three-layer hierarchy. By analyzing the spectrum, texture, shape, location, and context of individual object classes in different layers, the fuzzy determined rule system was established for the extraction of earthquake-collapsed building information. We compared the two sets of results using three variables: precision assessment, visual effect, and principle. Both methods can extract earthquake-collapsed building information quickly and accurately. The object-oriented method successfully overcomes the pepper salt noise caused by the spectral diversity of high-resolution remote sensing data and solves the problem of same object, different spectrums and that of same spectrum, different objects. With an overall accuracy of 90.38%, the method achieves more scientific and accurate results compared with the pixel-oriented method (76.84%). The object-oriented image analysis method can be extensively applied in the extraction of earthquake disaster information based on high-resolution remote sensing

  5. The impact of intangibles on the value relevance of accounting information: Evidence from French companies

    Directory of Open Access Journals (Sweden)

    Bilal Kimouche

    2016-03-01

    Full Text Available Purpose: The paper aims to explore whether intangible items that recognised in financial statements are value-relevant to investors in the French context, and whether these items affect the value relevance of accounting information. Design/methodology/approach: Empirical data were collected from a sample of French listed companies, over the nine-year period of 2005 to 2013. Starting of Ohlson’s (1995 model, the correlation analysis and the linear multiple regressions have been applied. Findings: We find that intangibles and traditional accounting measures as a whole are value relevant. However, the amortization and impairment charges of intangibles and cash flows do not affect the market values of French companies, unlike other variables, which affect positively and substantially the market values. Also goodwill and book values are more associated with market values than intangible assets and earnings respectively. Finally, we find that intangibles have improved the value relevance of accounting information. Practical implications: French legislators must give more interest for intangibles, in order to enrich the financial statements content and increasing the pertinence of accounting information. Auditors must give more attention for intangibles’ examination process, in order to certify the amounts related to intangibles in financial statements, and hence enrich their reliability, what provides adequacy guarantees for investors to use them in decision making. Originality/value: The paper used recently available financial data, and proposed an improvement concerning the measure of incremental value relevance of intangibles items.

  6. Semi-automatic building extraction in informal settlements from high-resolution satellite imagery

    Science.gov (United States)

    Mayunga, Selassie David

    The extraction of man-made features from digital remotely sensed images is considered as an important step underpinning management of human settlements in any country. Man-made features and buildings in particular are required for varieties of applications such as urban planning, creation of geographical information systems (GIS) databases and Urban City models. The traditional man-made feature extraction methods are very expensive in terms of equipment, labour intensive, need well-trained personnel and cannot cope with changing environments, particularly in dense urban settlement areas. This research presents an approach for extracting buildings in dense informal settlement areas using high-resolution satellite imagery. The proposed system uses a novel strategy of extracting building by measuring a single point at the approximate centre of the building. The fine measurement of the building outlines is then effected using a modified snake model. The original snake model on which this framework is based, incorporates an external constraint energy term which is tailored to preserving the convergence properties of the snake model; its use to unstructured objects will negatively affect their actual shapes. The external constrained energy term was removed from the original snake model formulation, thereby, giving ability to cope with high variability of building shapes in informal settlement areas. The proposed building extraction system was tested on two areas, which have different situations. The first area was Tungi in Dar Es Salaam, Tanzania where three sites were tested. This area is characterized by informal settlements, which are illegally formulated within the city boundaries. The second area was Oromocto in New Brunswick, Canada where two sites were tested. Oromocto area is mostly flat and the buildings are constructed using similar materials. Qualitative and quantitative measures were employed to evaluate the accuracy of the results as well as the performance

  7. In the Dark: Young Men's Stories of Sexual Initiation in the Absence of Relevant Sexual Health Information

    Science.gov (United States)

    Kubicek, Katrina; Beyer, William J.; Weiss, George; Iverson, Ellen; Kipke, Michele D.

    2010-01-01

    A growing body of research has investigated the effectiveness of abstinence-only sexual education. There remains a dearth of research on the relevant sexual health information available to young men who have sex with men (YMSM). Drawing on a mixed-methods study with 526 YMSM, this study explores how and where YMSM receive relevant information on…

  8. Feature-selective Attention in Frontoparietal Cortex: Multivoxel Codes Adjust to Prioritize Task-relevant Information.

    Science.gov (United States)

    Jackson, Jade; Rich, Anina N; Williams, Mark A; Woolgar, Alexandra

    2017-02-01

    Human cognition is characterized by astounding flexibility, enabling us to select appropriate information according to the objectives of our current task. A circuit of frontal and parietal brain regions, often referred to as the frontoparietal attention network or multiple-demand (MD) regions, are believed to play a fundamental role in this flexibility. There is evidence that these regions dynamically adjust their responses to selectively process information that is currently relevant for behavior, as proposed by the "adaptive coding hypothesis" [Duncan, J. An adaptive coding model of neural function in prefrontal cortex. Nature Reviews Neuroscience, 2, 820-829, 2001]. Could this provide a neural mechanism for feature-selective attention, the process by which we preferentially process one feature of a stimulus over another? We used multivariate pattern analysis of fMRI data during a perceptually challenging categorization task to investigate whether the representation of visual object features in the MD regions flexibly adjusts according to task relevance. Participants were trained to categorize visually similar novel objects along two orthogonal stimulus dimensions (length/orientation) and performed short alternating blocks in which only one of these dimensions was relevant. We found that multivoxel patterns of activation in the MD regions encoded the task-relevant distinctions more strongly than the task-irrelevant distinctions: The MD regions discriminated between stimuli of different lengths when length was relevant and between the same objects according to orientation when orientation was relevant. The data suggest a flexible neural system that adjusts its representation of visual objects to preferentially encode stimulus features that are currently relevant for behavior, providing a neural mechanism for feature-selective attention.

  9. Accurate facade feature extraction method for buildings from three-dimensional point cloud data considering structural information

    Science.gov (United States)

    Wang, Yongzhi; Ma, Yuqing; Zhu, A.-xing; Zhao, Hui; Liao, Lixia

    2018-05-01

    Facade features represent segmentations of building surfaces and can serve as a building framework. Extracting facade features from three-dimensional (3D) point cloud data (3D PCD) is an efficient method for 3D building modeling. By combining the advantages of 3D PCD and two-dimensional optical images, this study describes the creation of a highly accurate building facade feature extraction method from 3D PCD with a focus on structural information. The new extraction method involves three major steps: image feature extraction, exploration of the mapping method between the image features and 3D PCD, and optimization of the initial 3D PCD facade features considering structural information. Results show that the new method can extract the 3D PCD facade features of buildings more accurately and continuously. The new method is validated using a case study. In addition, the effectiveness of the new method is demonstrated by comparing it with the range image-extraction method and the optical image-extraction method in the absence of structural information. The 3D PCD facade features extracted by the new method can be applied in many fields, such as 3D building modeling and building information modeling.

  10. Information analysis of iris biometrics for the needs of cryptology key extraction

    Directory of Open Access Journals (Sweden)

    Adamović Saša

    2013-01-01

    Full Text Available The paper presents a rigorous analysis of iris biometric information for the synthesis of an optimized system for the extraction of a high quality cryptology key. Estimations of local entropy and mutual information were identified as segments of the iris most suitable for this purpose. In order to optimize parameters, corresponding wavelets were transformed, in order to obtain the highest possible entropy and mutual information lower in the transformation domain, which set frameworks for the synthesis of systems for the extraction of truly random sequences of iris biometrics, without compromising authentication properties. [Projekat Ministarstva nauke Republike Srbije, br. TR32054 i br. III44006

  11. Analysis of the Relevance of Information Content of the Value Added Statement in the Brazilian Capital Markets

    Directory of Open Access Journals (Sweden)

    Márcio André Veras Machado

    2015-04-01

    Full Text Available The usefulness of financial statements depends, fundamentally, on the degree of relevance of the information they disclose to users. Thus, studies that measure the relevance of accounting information to the users of financial statements are of some importance. One line of research within this subject is in ascertaining the relevance and importance of accounting information for the capital markets: if a particular item of accounting information is minimally reflected in the price of a share, it is because this information has relevance, at least at a certain level of significance, for investors and analysts of the capital markets. This present study aims to analyze the relevance, in the Brazilian capital markets, of the information content of the Value Added Statement (or VAS - referred to in Brazil as the Demonstração do Valor Adicionado, or DVA. It analyzed the ratio between stock price and Wealth created per share (WCPS, using linear regressions, for the period 2005-2011, for non-financial listed companies included in Melhores & Maiores ('Biggest & Best', an annual listing published by Exame Magazine in Brazil. As a secondary objective, this article seeks to establish whether WCPS represents a better indication of a company's result than Net profit per share (in this study, referred to as NPPS. The empirical evidence that was found supports the concept that the VAS has relevant information content, because it shows a capacity to explain a variation in the share price of the companies studied. Additionally, the relationship between WCPS and the stock price was shown to be significant, even after the inclusion of the control variables Stockholders' equity per share (which we abbreviate in this study to SEPS and NPPS. Finally, the evidence found indicates that the market reacts more to WCPS (Wealth created per share than to NPPS. Thus, the results obtained give some indication that, for the Brazilian capital markets, WCPS may be a better proxy

  12. Information retrieval and terminology extraction in online resources for patients with diabetes.

    Science.gov (United States)

    Seljan, Sanja; Baretić, Maja; Kucis, Vlasta

    2014-06-01

    Terminology use, as a mean for information retrieval or document indexing, plays an important role in health literacy. Specific types of users, i.e. patients with diabetes need access to various online resources (on foreign and/or native language) searching for information on self-education of basic diabetic knowledge, on self-care activities regarding importance of dietetic food, medications, physical exercises and on self-management of insulin pumps. Automatic extraction of corpus-based terminology from online texts, manuals or professional papers, can help in building terminology lists or list of "browsing phrases" useful in information retrieval or in document indexing. Specific terminology lists represent an intermediate step between free text search and controlled vocabulary, between user's demands and existing online resources in native and foreign language. The research aiming to detect the role of terminology in online resources, is conducted on English and Croatian manuals and Croatian online texts, and divided into three interrelated parts: i) comparison of professional and popular terminology use ii) evaluation of automatic statistically-based terminology extraction on English and Croatian texts iii) comparison and evaluation of extracted terminology performed on English manual using statistical and hybrid approaches. Extracted terminology candidates are evaluated by comparison with three types of reference lists: list created by professional medical person, list of highly professional vocabulary contained in MeSH and list created by non-medical persons, made as intersection of 15 lists. Results report on use of popular and professional terminology in online diabetes resources, on evaluation of automatically extracted terminology candidates in English and Croatian texts and on comparison of statistical and hybrid extraction methods in English text. Evaluation of automatic and semi-automatic terminology extraction methods is performed by recall

  13. 40 CFR 86.1862-04 - Maintenance of records and submittal of information relevant to compliance with fleet average NOX...

    Science.gov (United States)

    2010-07-01

    ... of information relevant to compliance with fleet average NOX standards. 86.1862-04 Section 86.1862-04...-Cycle Heavy-Duty Vehicles § 86.1862-04 Maintenance of records and submittal of information relevant to..., maintain, and retain the following information in adequately organized and indexed records for each model...

  14. [Extraction of buildings three-dimensional information from high-resolution satellite imagery based on Barista software].

    Science.gov (United States)

    Zhang, Pei-feng; Hu, Yuan-man; He, Hong-shi

    2010-05-01

    The demand for accurate and up-to-date spatial information of urban buildings is becoming more and more important for urban planning, environmental protection, and other vocations. Today's commercial high-resolution satellite imagery offers the potential to extract the three-dimensional information of urban buildings. This paper extracted the three-dimensional information of urban buildings from QuickBird imagery, and validated the precision of the extraction based on Barista software. It was shown that the extraction of three-dimensional information of the buildings from high-resolution satellite imagery based on Barista software had the advantages of low professional level demand, powerful universality, simple operation, and high precision. One pixel level of point positioning and height determination accuracy could be achieved if the digital elevation model (DEM) and sensor orientation model had higher precision and the off-Nadir View Angle was relatively perfect.

  15. Information Extraction of High Resolution Remote Sensing Images Based on the Calculation of Optimal Segmentation Parameters

    Science.gov (United States)

    Zhu, Hongchun; Cai, Lijie; Liu, Haiying; Huang, Wei

    2016-01-01

    Multi-scale image segmentation and the selection of optimal segmentation parameters are the key processes in the object-oriented information extraction of high-resolution remote sensing images. The accuracy of remote sensing special subject information depends on this extraction. On the basis of WorldView-2 high-resolution data, the optimal segmentation parameters methodof object-oriented image segmentation and high-resolution image information extraction, the following processes were conducted in this study. Firstly, the best combination of the bands and weights was determined for the information extraction of high-resolution remote sensing image. An improved weighted mean-variance method was proposed andused to calculatethe optimal segmentation scale. Thereafter, the best shape factor parameter and compact factor parameters were computed with the use of the control variables and the combination of the heterogeneity and homogeneity indexes. Different types of image segmentation parameters were obtained according to the surface features. The high-resolution remote sensing images were multi-scale segmented with the optimal segmentation parameters. Ahierarchical network structure was established by setting the information extraction rules to achieve object-oriented information extraction. This study presents an effective and practical method that can explain expert input judgment by reproducible quantitative measurements. Furthermore the results of this procedure may be incorporated into a classification scheme. PMID:27362762

  16. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories.

    Science.gov (United States)

    Yang, Wei; Ai, Tinghua; Lu, Wei

    2018-04-19

    Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT). First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS) traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction) by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  17. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories

    Directory of Open Access Journals (Sweden)

    Wei Yang

    2018-04-01

    Full Text Available Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT. First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  18. Strategic relevance and accountability expectations: new perspectives for health care information technology design.

    Science.gov (United States)

    Tan, J K; Modrow, R E

    1999-05-01

    In this article, we discuss the traditional systems analysis perspective on end-user information requirements analysis and extend it to merge with the new accountability expectations perspective to guide the future planning and design of health organization information systems. Underlying the strategic relevance of health care information technology (HCIT) are three critical questions: (1) What is the ideal HCIT model for the health organization in terms of achieving strategic expertise and competitive advantage? Specifically, how does this model link industry performance standards with organizational performance and accountability expectations? (2) How should the limitations of past HCIT models be reconciled to the benefits presented by the superior arrangement of the ideal model in the context of changing accountability expectations? (3) How should alternative HCIT solutions be evaluated in light of evidence-based accountability and organizational performance benchmarking? Insights into these questions will ensure that health care managers, HCIT practitioners and researchers can continue to focus on the most critical issues in harnessing today's fast-paced changing technologies for evolving strategically relevant, performance-based health organization systems.

  19. A cascade of classifiers for extracting medication information from discharge summaries

    Directory of Open Access Journals (Sweden)

    Halgrim Scott

    2011-07-01

    Full Text Available Abstract Background Extracting medication information from clinical records has many potential applications, and recently published research, systems, and competitions reflect an interest therein. Much of the early extraction work involved rules and lexicons, but more recently machine learning has been applied to the task. Methods We present a hybrid system consisting of two parts. The first part, field detection, uses a cascade of statistical classifiers to identify medication-related named entities. The second part uses simple heuristics to link those entities into medication events. Results The system achieved performance that is comparable to other approaches to the same task. This performance is further improved by adding features that reference external medication name lists. Conclusions This study demonstrates that our hybrid approach outperforms purely statistical or rule-based systems. The study also shows that a cascade of classifiers works better than a single classifier in extracting medication information. The system is available as is upon request from the first author.

  20. Relevant cost information for order acceptance decisions

    NARCIS (Netherlands)

    Wouters, M.J.F.

    1997-01-01

    Some economic considerations for order acceptance decisions are discussed. The relevant economic considerations for order acceptance are widely discussed in the literature: only those costs are relevant which would be avoidable by not accepting the order incremental costs plus opportunity costs .

  1. Divided attention selectively impairs memory for self-relevant information.

    Science.gov (United States)

    Turk, David J; Brady-van den Bos, Mirjam; Collard, Philip; Gillespie-Smith, Karri; Conway, Martin A; Cunningham, Sheila J

    2013-05-01

    Information that is relevant to oneself tends to be remembered more than information that relates to other people, but the role of attention in eliciting this "self-reference effect" is unclear. In the present study, we assessed the importance of attention in self-referential encoding using an ownership paradigm, which required participants to encode items under conditions of imagined ownership by themselves or by another person. Previous work has established that this paradigm elicits a robust self-reference effect, with more "self-owned" items being remembered than "other-owned" items. Access to attentional resources was manipulated using divided-attention tasks at encoding. A significant self-reference effect emerged under full-attention conditions and was related to an increase in episodic recollection for self-owned items, but dividing attention eliminated this memory advantage. These findings are discussed in relation to the nature of self-referential cognition and the importance of attentional resources at encoding in the manifestation of the self-reference effect in memory.

  2. An analysis of contextual information relevant to medical care unexpectedly volunteered to researchers by asthma patients.

    Science.gov (United States)

    Black, Heather L; Priolo, Chantel; Gonzalez, Rodalyn; Geer, Sabrina; Adam, Bariituu; Apter, Andrea J

    2012-09-01

    To describe and categorize contextual information relevant to patients' medical care unexpectedly volunteered to research personnel as part of a patient advocate (PA) intervention to facilitate access health care, communication with medical personnel, and self-management of a chronic disease such as asthma. We adapted a patient navigator intervention, to overcome barriers to access and communication for adults with moderate or severe asthma. Informed by focus groups of patients and providers, our PAs facilitated preparation for a visit with an asthma provider, attended the visit, confirmed understanding, and assisted with post-visit activities. During meetings with researchers, either for PA activities or for data collection, participants frequently volunteered personal and medical information relevant for achieving successful self-management that was not routinely shared with medical personnel. For this project, researchers journaled information not captured by the structured questionnaires and protocol. Using a qualitative analysis, we describe (1) researchers' journals of these unique communications; (2) their relevance for accomplishing self-management; (3) PAs' formal activities including teach-back, advocacy, and facilitating appointment making; and (4) observations of patients' interactions with the clinical practices. In 83 journals, patients' social support (83%), health (68%), and deportment (69%) were described. PA assistance with navigating the medical system (59%), teach-back (46%), and observed interactions with patient and medical staff (76%) were also journaled. Implicit were ways patients and practices could overcome barriers to access and communication. These journals describe the importance of seeking contextual and medically relevant information from all patients and, especially, those with significant morbidities, prompting patients for barriers to access to health care, and confirming understanding of medical information.

  3. Rapid automatic keyword extraction for information retrieval and analysis

    Science.gov (United States)

    Rose, Stuart J [Richland, WA; Cowley,; E, Wendy [Richland, WA; Crow, Vernon L [Richland, WA; Cramer, Nicholas O [Richland, WA

    2012-03-06

    Methods and systems for rapid automatic keyword extraction for information retrieval and analysis. Embodiments can include parsing words in an individual document by delimiters, stop words, or both in order to identify candidate keywords. Word scores for each word within the candidate keywords are then calculated based on a function of co-occurrence degree, co-occurrence frequency, or both. Based on a function of the word scores for words within the candidate keyword, a keyword score is calculated for each of the candidate keywords. A portion of the candidate keywords are then extracted as keywords based, at least in part, on the candidate keywords having the highest keyword scores.

  4. Direct nano ESI time-of-flight mass spectrometric investigations on lanthanide BTP complexes in the extraction-relevant diluent 1-octanol

    International Nuclear Information System (INIS)

    Steppert, M.; Walther, C.; Geist, A.; Fanghanel, Th.

    2009-01-01

    The present work focuses on investigations of a highly selective ligand for Am(III)/Ln(III) separation: bis-triazinyl-pyridine (BTP). By means of nano-electro-spray mass spectrometry, complex formation of BTP with selected elements of the lanthanide series is investigated. We show that the diluent drastically influences complex speciation. Measurements obtained in the extraction-relevant diluent 1-octanol show the occurrence of Ln(BTP) i (i 1-3) species in different relative abundances, depending on the lanthanide used. Here, the relative abundances of the Ln(BTP) 3 complexes correlate with the distribution ratios for extraction to the organic phase of the respective lanthanide. (authors)

  5. Extractive text summarization system to aid data extraction from full text in systematic review development.

    Science.gov (United States)

    Bui, Duy Duc An; Del Fiol, Guilherme; Hurdle, John F; Jonnalagadda, Siddhartha

    2016-12-01

    Extracting data from publication reports is a standard process in systematic review (SR) development. However, the data extraction process still relies too much on manual effort which is slow, costly, and subject to human error. In this study, we developed a text summarization system aimed at enhancing productivity and reducing errors in the traditional data extraction process. We developed a computer system that used machine learning and natural language processing approaches to automatically generate summaries of full-text scientific publications. The summaries at the sentence and fragment levels were evaluated in finding common clinical SR data elements such as sample size, group size, and PICO values. We compared the computer-generated summaries with human written summaries (title and abstract) in terms of the presence of necessary information for the data extraction as presented in the Cochrane review's study characteristics tables. At the sentence level, the computer-generated summaries covered more information than humans do for systematic reviews (recall 91.2% vs. 83.8%, p<0.001). They also had a better density of relevant sentences (precision 59% vs. 39%, p<0.001). At the fragment level, the ensemble approach combining rule-based, concept mapping, and dictionary-based methods performed better than individual methods alone, achieving an 84.7% F-measure. Computer-generated summaries are potential alternative information sources for data extraction in systematic review development. Machine learning and natural language processing are promising approaches to the development of such an extractive summarization system. Copyright © 2016 Elsevier Inc. All rights reserved.

  6. Towards brain-activity-controlled information retrieval: Decoding image relevance from MEG signals.

    Science.gov (United States)

    Kauppi, Jukka-Pekka; Kandemir, Melih; Saarinen, Veli-Matti; Hirvenkari, Lotta; Parkkonen, Lauri; Klami, Arto; Hari, Riitta; Kaski, Samuel

    2015-05-15

    We hypothesize that brain activity can be used to control future information retrieval systems. To this end, we conducted a feasibility study on predicting the relevance of visual objects from brain activity. We analyze both magnetoencephalographic (MEG) and gaze signals from nine subjects who were viewing image collages, a subset of which was relevant to a predetermined task. We report three findings: i) the relevance of an image a subject looks at can be decoded from MEG signals with performance significantly better than chance, ii) fusion of gaze-based and MEG-based classifiers significantly improves the prediction performance compared to using either signal alone, and iii) non-linear classification of the MEG signals using Gaussian process classifiers outperforms linear classification. These findings break new ground for building brain-activity-based interactive image retrieval systems, as well as for systems utilizing feedback both from brain activity and eye movements. Copyright © 2015 Elsevier Inc. All rights reserved.

  7. Information theoretic approach to tactile encoding and discrimination

    OpenAIRE

    Saal, Hannes

    2011-01-01

    The human sense of touch integrates feedback from a multitude of touch receptors, but how this information is represented in the neural responses such that it can be extracted quickly and reliably is still largely an open question. At the same time, dexterous robots equipped with touch sensors are becoming more common, necessitating better methods for representing sequentially updated information and new control strategies that aid in extracting relevant features for object man...

  8. An Evaluation of Applying Knowledge Base to Academic Information Service

    OpenAIRE

    Seok-Hyoung Lee; Hwan-Min Kim; Ho-Seop Choe

    2013-01-01

    Through a series of precise text handling processes, including automatic extraction of information from documents with knowledge from various fields, recognition of entity names, detection of core topics, analysis of the relations between the extracted information and topics, and automatic inference of new knowledge, the most efficient knowledge base of the relevant field is created, and plans to apply these to the information knowledge management and service are the core requirements necessa...

  9. Robust Vehicle and Traffic Information Extraction for Highway Surveillance

    Directory of Open Access Journals (Sweden)

    Yeh Chia-Hung

    2005-01-01

    Full Text Available A robust vision-based traffic monitoring system for vehicle and traffic information extraction is developed in this research. It is challenging to maintain detection robustness at all time for a highway surveillance system. There are three major problems in detecting and tracking a vehicle: (1 the moving cast shadow effect, (2 the occlusion effect, and (3 nighttime detection. For moving cast shadow elimination, a 2D joint vehicle-shadow model is employed. For occlusion detection, a multiple-camera system is used to detect occlusion so as to extract the exact location of each vehicle. For vehicle nighttime detection, a rear-view monitoring technique is proposed to maintain tracking and detection accuracy. Furthermore, we propose a method to improve the accuracy of background extraction, which usually serves as the first step in any vehicle detection processing. Experimental results are given to demonstrate that the proposed techniques are effective and efficient for vision-based highway surveillance.

  10. Using language models to identify relevant new information in inpatient clinical notes.

    Science.gov (United States)

    Zhang, Rui; Pakhomov, Serguei V; Lee, Janet T; Melton, Genevieve B

    2014-01-01

    Redundant information in clinical notes within electronic health record (EHR) systems is ubiquitous and may negatively impact the use of these notes by clinicians, and, potentially, the efficiency of patient care delivery. Automated methods to identify redundant versus relevant new information may provide a valuable tool for clinicians to better synthesize patient information and navigate to clinically important details. In this study, we investigated the use of language models for identification of new information in inpatient notes, and evaluated our methods using expert-derived reference standards. The best method achieved precision of 0.743, recall of 0.832 and F1-measure of 0.784. The average proportion of redundant information was similar between inpatient and outpatient progress notes (76.6% (SD=17.3%) and 76.7% (SD=14.0%), respectively). Advanced practice providers tended to have higher rates of redundancy in their notes compared to physicians. Future investigation includes the addition of semantic components and visualization of new information.

  11. Research of building information extraction and evaluation based on high-resolution remote-sensing imagery

    Science.gov (United States)

    Cao, Qiong; Gu, Lingjia; Ren, Ruizhi; Wang, Lang

    2016-09-01

    Building extraction currently is important in the application of high-resolution remote sensing imagery. At present, quite a few algorithms are available for detecting building information, however, most of them still have some obvious disadvantages, such as the ignorance of spectral information, the contradiction between extraction rate and extraction accuracy. The purpose of this research is to develop an effective method to detect building information for Chinese GF-1 data. Firstly, the image preprocessing technique is used to normalize the image and image enhancement is used to highlight the useful information in the image. Secondly, multi-spectral information is analyzed. Subsequently, an improved morphological building index (IMBI) based on remote sensing imagery is proposed to get the candidate building objects. Furthermore, in order to refine building objects and further remove false objects, the post-processing (e.g., the shape features, the vegetation index and the water index) is employed. To validate the effectiveness of the proposed algorithm, the omission errors (OE), commission errors (CE), the overall accuracy (OA) and Kappa are used at final. The proposed method can not only effectively use spectral information and other basic features, but also avoid extracting excessive interference details from high-resolution remote sensing images. Compared to the original MBI algorithm, the proposed method reduces the OE by 33.14% .At the same time, the Kappa increase by 16.09%. In experiments, IMBI achieved satisfactory results and outperformed other algorithms in terms of both accuracies and visual inspection

  12. Extracting Social Networks and Contact Information From Email and the Web

    National Research Council Canada - National Science Library

    Culotta, Aron; Bekkerman, Ron; McCallum, Andrew

    2005-01-01

    ...-suited for such information extraction tasks. By recursively calling itself on new people discovered on the Web, the system builds a social network with multiple degrees of separation from the user...

  13. Information Extraction of High-Resolution Remotely Sensed Image Based on Multiresolution Segmentation

    Directory of Open Access Journals (Sweden)

    Peng Shao

    2014-08-01

    Full Text Available The principle of multiresolution segmentation was represented in detail in this study, and the canny algorithm was applied for edge-detection of a remotely sensed image based on this principle. The target image was divided into regions based on object-oriented multiresolution segmentation and edge-detection. Furthermore, object hierarchy was created, and a series of features (water bodies, vegetation, roads, residential areas, bare land and other information were extracted by the spectral and geometrical features. The results indicate that the edge-detection has a positive effect on multiresolution segmentation, and overall accuracy of information extraction reaches to 94.6% by the confusion matrix.

  14. Semantic Location Extraction from Crowdsourced Data

    Science.gov (United States)

    Koswatte, S.; Mcdougall, K.; Liu, X.

    2016-06-01

    Crowdsourced Data (CSD) has recently received increased attention in many application areas including disaster management. Convenience of production and use, data currency and abundancy are some of the key reasons for attracting this high interest. Conversely, quality issues like incompleteness, credibility and relevancy prevent the direct use of such data in important applications like disaster management. Moreover, location information availability of CSD is problematic as it remains very low in many crowd sourced platforms such as Twitter. Also, this recorded location is mostly related to the mobile device or user location and often does not represent the event location. In CSD, event location is discussed descriptively in the comments in addition to the recorded location (which is generated by means of mobile device's GPS or mobile communication network). This study attempts to semantically extract the CSD location information with the help of an ontological Gazetteer and other available resources. 2011 Queensland flood tweets and Ushahidi Crowd Map data were semantically analysed to extract the location information with the support of Queensland Gazetteer which is converted to an ontological gazetteer and a global gazetteer. Some preliminary results show that the use of ontologies and semantics can improve the accuracy of place name identification of CSD and the process of location information extraction.

  15. SEMANTIC LOCATION EXTRACTION FROM CROWDSOURCED DATA

    Directory of Open Access Journals (Sweden)

    S. Koswatte

    2016-06-01

    Full Text Available Crowdsourced Data (CSD has recently received increased attention in many application areas including disaster management. Convenience of production and use, data currency and abundancy are some of the key reasons for attracting this high interest. Conversely, quality issues like incompleteness, credibility and relevancy prevent the direct use of such data in important applications like disaster management. Moreover, location information availability of CSD is problematic as it remains very low in many crowd sourced platforms such as Twitter. Also, this recorded location is mostly related to the mobile device or user location and often does not represent the event location. In CSD, event location is discussed descriptively in the comments in addition to the recorded location (which is generated by means of mobile device's GPS or mobile communication network. This study attempts to semantically extract the CSD location information with the help of an ontological Gazetteer and other available resources. 2011 Queensland flood tweets and Ushahidi Crowd Map data were semantically analysed to extract the location information with the support of Queensland Gazetteer which is converted to an ontological gazetteer and a global gazetteer. Some preliminary results show that the use of ontologies and semantics can improve the accuracy of place name identification of CSD and the process of location information extraction.

  16. 76 FR 44337 - Comments and Information Relevant to Mid Decade Review of NORA

    Science.gov (United States)

    2011-07-25

    ... NIOSH-244] Comments and Information Relevant to Mid Decade Review of NORA AGENCY: Department of Health...) is conducting a review of the processes of the National Occupational Research Agenda (NORA). In 2006, NORA entered its second decade with an industry sector-based structure. In 2011, as NORA reaches the...

  17. On the relevance of spectral features for instrument classification

    DEFF Research Database (Denmark)

    Nielsen, Andreas Brinch; Sigurdsson, Sigurdur; Hansen, Lars Kai

    2007-01-01

    Automatic knowledge extraction from music signals is a key component for most music organization and music information retrieval systems. In this paper, we consider the problem of instrument modelling and instrument classification from the rough audio data. Existing systems for automatic instrument...... classification operate normally on a relatively large number of features, from which those related to the spectrum of the audio signal are particularly relevant. In this paper, we confront two different models about the spectral characterization of musical instruments. The first assumes a constant envelope...

  18. A method for automating the extraction of specialized information from the web

    NARCIS (Netherlands)

    Lin, L.; Liotta, A.; Hippisley, A.; Hao, Y.; Liu, J.; Wang, Y.; Cheung, Y-M.; Yin, H.; Jiao, L.; Ma, j.; Jiao, Y-C.

    2005-01-01

    The World Wide Web can be viewed as a gigantic distributed database including millions of interconnected hosts some of which publish information via web servers or peer-to-peer systems. We present here a novel method for the extraction of semantically rich information from the web in a fully

  19. Motivated memory: memory for attitude-relevant information as a function of self-esteem

    NARCIS (Netherlands)

    Wiersema, D.V.; van der Pligt, J.; van Harreveld, F.

    2010-01-01

    In this article we offer a new perspective on the contradictory findings in the literature on memory for attitude-relevant information. We propose that biases in memory are most likely to occur when the attitude involved is connected to personally important values and the self; i.e., if the attitude

  20. Relevance between the degree of industrial competition and fair value information: Study on the listed companies in China

    Directory of Open Access Journals (Sweden)

    Xuemin Zhuang

    2015-05-01

    Full Text Available Purpose: The purpose of this article is to study whether there exists natural relationship between fair value and corporate external market. A series of special phenomenon in the application of fair value arouses our research interests, which present evidences on how competition affects the correlation of fair value information. Design/methodology/approach: this thesis chooses fair value changes gains and losses and calculate the ratio of DFVPSit as the alternative variable of the fair value. In order to effectively inspect the mutual influence between the degree of industry competition and the value relevance of fair value, and reduce the impact of multi-collinearity, we built a regression model on the hypothesis, which supposes that if other conditions are the same, the fair value information has greater value relevance if the degree of the industry competition is greater. To test the hypothesis, we use the comparison of the DFVPSit coefficient absolute value to judge the value relevance of fair value information, and the greater the absolute value is, the higher relevance between the changes in fair value per share profits and losses with the stock prices. Findings: The higher the degree of competition in the industry is, the more fair value information relevance is. Also, there are evidences representing that fair value information often presents negative correlation with the stock price. Originality/value: The main contribution of the article is to show that not only need we make the formulation and implementation of the high quality of fair value accounting standards to suit for both the national conditions and international practice, but also need we further to improve the company's external governance mechanism to promote fair value’s information correlation.

  1. Trend extraction of rail corrugation measured dynamically based on the relevant low-frequency principal components reconstruction

    International Nuclear Information System (INIS)

    Li, Yanfu; Liu, Hongli; Ma, Ziji

    2016-01-01

    Rail corrugation dynamic measurement techniques are critical to guarantee transport security and guide rail maintenance. During the inspection process, low-frequency trends caused by rail fluctuation are usually superimposed on rail corrugation and seriously affect the assessment of rail maintenance quality. In order to extract and remove the nonlinear and non-stationary trends from original mixed signals, a hybrid model based ensemble empirical mode decomposition (EEMD) and modified principal component analysis (MPCA) is proposed in this paper. Compared with the existing de-trending methods based on EMD, this method first considers low-frequency intrinsic mode functions (IMFs) thought to be underlying trend components that maybe contain some unrelated components, such as white noise and low-frequency signal itself, and proposes to use PCA to accurately extract the pure trends from the IMFs containing multiple components. On the other hand, due to the energy contribution ratio between trends and mixed signals is prior unknown, and the principal components (PCs) decomposed by PCA are arranged in order of energy reduction without considering frequency distribution, the proposed method modifies traditional PCA and just selects relevant low-frequency PCs to reconstruct the trends based on the zero-crossing numbers (ZCN) of each PC. Extensive tests are presented to illustrate the effectiveness of the proposed method. The results show the proposed EEMD-PCA-ZCN is an effective tool for trend extraction of rail corrugation measured dynamically. (paper)

  2. IMPACT OF THE CONVERGENCE PROCESS TO INTERNATIONAL FINANCIAL REPORTING STANDARDS ON THE VALUE RELEVANCE OF FINANCIAL INFORMATION

    Directory of Open Access Journals (Sweden)

    Marcelo Alvaro da Silva Macedo

    2012-11-01

    Full Text Available Law 11.638/07 marked the start of a series of changes in the laws that regulate Brazilian accounting practices. The main reason for these changes is the convergence process of local with international accounting standards. As a result of Law 11.638/07, the legal precedent was established to achieve convergence. In that context, the aim of this study is to analyze the impact of the convergence process with international accounting standards on the relevance of financial information, based on data for 2007, without and with the alterations Law 11.638/07 introduced and according to the CPC Pronouncements, applicable as from 2008 onwards. Therefore, a value relevance study is used, applying regression analysis to annual stock price information (dependent variable and net profit per share (NPPS and net equity per share (NEPS as independent variables. The main results show that financial information on NPPS and NEPS for 2007, with and without the legal alterations, are relevant for the capital market. A comparison between both regressions used in the analysis, however, shows an information gain for financial information that includes the changes introduced in the first phase of the accounting convergence process with the international standards.

  3. Extraction of land cover change information from ENVISAT-ASAR data in Chengdu Plain

    Science.gov (United States)

    Xu, Wenbo; Fan, Jinlong; Huang, Jianxi; Tian, Yichen; Zhang, Yong

    2006-10-01

    Land cover data are essential to most global change research objectives, including the assessment of current environmental conditions and the simulation of future environmental scenarios that ultimately lead to public policy development. Chinese Academy of Sciences generated a nationwide land cover database in order to carry out the quantification and spatial characterization of land use/cover changes (LUCC) in 1990s. In order to improve the reliability of the database, we will update the database anytime. But it is difficult to obtain remote sensing data to extract land cover change information in large-scale. It is hard to acquire optical remote sensing data in Chengdu plain, so the objective of this research was to evaluate multitemporal ENVISAT advanced synthetic aperture radar (ASAR) data for extracting land cover change information. Based on the fieldwork and the nationwide 1:100000 land cover database, the paper assesses several land cover changes in Chengdu plain, for example: crop to buildings, forest to buildings, and forest to bare land. The results show that ENVISAT ASAR data have great potential for the applications of extracting land cover change information.

  4. Creation of reliable relevance judgments in information retrieval systems evaluation experimentation through crowdsourcing: a review.

    Science.gov (United States)

    Samimi, Parnia; Ravana, Sri Devi

    2014-01-01

    Test collection is used to evaluate the information retrieval systems in laboratory-based evaluation experimentation. In a classic setting, generating relevance judgments involves human assessors and is a costly and time consuming task. Researchers and practitioners are still being challenged in performing reliable and low-cost evaluation of retrieval systems. Crowdsourcing as a novel method of data acquisition is broadly used in many research fields. It has been proven that crowdsourcing is an inexpensive and quick solution as well as a reliable alternative for creating relevance judgments. One of the crowdsourcing applications in IR is to judge relevancy of query document pair. In order to have a successful crowdsourcing experiment, the relevance judgment tasks should be designed precisely to emphasize quality control. This paper is intended to explore different factors that have an influence on the accuracy of relevance judgments accomplished by workers and how to intensify the reliability of judgments in crowdsourcing experiment.

  5. Evolving spectral transformations for multitemporal information extraction using evolutionary computation

    Science.gov (United States)

    Momm, Henrique; Easson, Greg

    2011-01-01

    Remote sensing plays an important role in assessing temporal changes in land features. The challenge often resides in the conversion of large quantities of raw data into actionable information in a timely and cost-effective fashion. To address this issue, research was undertaken to develop an innovative methodology integrating biologically-inspired algorithms with standard image classification algorithms to improve information extraction from multitemporal imagery. Genetic programming was used as the optimization engine to evolve feature-specific candidate solutions in the form of nonlinear mathematical expressions of the image spectral channels (spectral indices). The temporal generalization capability of the proposed system was evaluated by addressing the task of building rooftop identification from a set of images acquired at different dates in a cross-validation approach. The proposed system generates robust solutions (kappa values > 0.75 for stage 1 and > 0.4 for stage 2) despite the statistical differences between the scenes caused by land use and land cover changes coupled with variable environmental conditions, and the lack of radiometric calibration between images. Based on our results, the use of nonlinear spectral indices enhanced the spectral differences between features improving the clustering capability of standard classifiers and providing an alternative solution for multitemporal information extraction.

  6. Extraction of Graph Information Based on Image Contents and the Use of Ontology

    Science.gov (United States)

    Kanjanawattana, Sarunya; Kimura, Masaomi

    2016-01-01

    A graph is an effective form of data representation used to summarize complex information. Explicit information such as the relationship between the X- and Y-axes can be easily extracted from a graph by applying human intelligence. However, implicit knowledge such as information obtained from other related concepts in an ontology also resides in…

  7. Deep learning relevance

    DEFF Research Database (Denmark)

    Lioma, Christina; Larsen, Birger; Petersen, Casper

    2016-01-01

    train a Recurrent Neural Network (RNN) on existing relevant information to that query. We then use the RNN to "deep learn" a single, synthetic, and we assume, relevant document for that query. We design a crowdsourcing experiment to assess how relevant the "deep learned" document is, compared...... to existing relevant documents. Users are shown a query and four wordclouds (of three existing relevant documents and our deep learned synthetic document). The synthetic document is ranked on average most relevant of all....

  8. Expert vs. novice differences in the detection of relevant information during a chess game: evidence from eye movements.

    Science.gov (United States)

    Sheridan, Heather; Reingold, Eyal M

    2014-01-01

    The present study explored the ability of expert and novice chess players to rapidly distinguish between regions of a chessboard that were relevant to the best move on the board, and regions of the board that were irrelevant. Accordingly, we monitored the eye movements of expert and novice chess players, while they selected white's best move for a variety of chess problems. To manipulate relevancy, we constructed two different versions of each chess problem in the experiment, and we counterbalanced these versions across participants. These two versions of each problem were identical except that a single piece was changed from a bishop to a knight. This subtle change reversed the relevancy map of the board, such that regions that were relevant in one version of the board were now irrelevant (and vice versa). Using this paradigm, we demonstrated that both the experts and novices spent more time fixating the relevant relative to the irrelevant regions of the board. However, the experts were faster at detecting relevant information than the novices, as shown by the finding that experts (but not novices) were able to distinguish between relevant and irrelevant information during the early part of the trial. These findings further demonstrate the domain-related perceptual processing advantage of chess experts, using an experimental paradigm that allowed us to manipulate relevancy under tightly controlled conditions.

  9. Expert versus novice differences in the detection of relevant information during a chess game: Evidence from eye movements

    Directory of Open Access Journals (Sweden)

    Heather eSheridan

    2014-08-01

    Full Text Available The present study explored the ability of expert and novice chess players to rapidly distinguish between regions of a chessboard that were relevant to the best move on the board, and regions of the board that were irrelevant. Accordingly, we monitored the eye movements of expert and novice chess players, while they selected white’s best move for a variety of chess problems. To manipulate relevancy, we constructed two different versions of each chess problem in the experiment, and we counterbalanced these versions across participants. These two versions of each problem were identical except that a single piece was changed from a bishop to a knight. This subtle change reversed the relevancy map of the board, such that regions that were relevant in one version of the board were now irrelevant (and vice versa. Using this paradigm, we demonstrated that both the experts and novices spent more time fixating the relevant relative to the irrelevant regions of the board. However, the experts were faster at detecting relevant information than the novices, as shown by the finding that experts (but not novices were able to distinguish between relevant and irrelevant information during the early part of the trial. These findings further demonstrate the domain-related perceptual processing advantage of chess experts, using an experimental paradigm that allowed us to manipulate relevancy under tightly controlled conditions.

  10. An Accurate Integral Method for Vibration Signal Based on Feature Information Extraction

    Directory of Open Access Journals (Sweden)

    Yong Zhu

    2015-01-01

    Full Text Available After summarizing the advantages and disadvantages of current integral methods, a novel vibration signal integral method based on feature information extraction was proposed. This method took full advantage of the self-adaptive filter characteristic and waveform correction feature of ensemble empirical mode decomposition in dealing with nonlinear and nonstationary signals. This research merged the superiorities of kurtosis, mean square error, energy, and singular value decomposition on signal feature extraction. The values of the four indexes aforementioned were combined into a feature vector. Then, the connotative characteristic components in vibration signal were accurately extracted by Euclidean distance search, and the desired integral signals were precisely reconstructed. With this method, the interference problem of invalid signal such as trend item and noise which plague traditional methods is commendably solved. The great cumulative error from the traditional time-domain integral is effectively overcome. Moreover, the large low-frequency error from the traditional frequency-domain integral is successfully avoided. Comparing with the traditional integral methods, this method is outstanding at removing noise and retaining useful feature information and shows higher accuracy and superiority.

  11. LANGUAGE EXPERIENCE SHAPES PROCESSING OF PITCH RELEVANT INFORMATION IN THE HUMAN BRAINSTEM AND AUDITORY CORTEX: ELECTROPHYSIOLOGICAL EVIDENCE.

    Science.gov (United States)

    Krishnan, Ananthanarayan; Gandour, Jackson T

    2014-12-01

    Pitch is a robust perceptual attribute that plays an important role in speech, language, and music. As such, it provides an analytic window to evaluate how neural activity relevant to pitch undergo transformation from early sensory to later cognitive stages of processing in a well coordinated hierarchical network that is subject to experience-dependent plasticity. We review recent evidence of language experience-dependent effects in pitch processing based on comparisons of native vs. nonnative speakers of a tonal language from electrophysiological recordings in the auditory brainstem and auditory cortex. We present evidence that shows enhanced representation of linguistically-relevant pitch dimensions or features at both the brainstem and cortical levels with a stimulus-dependent preferential activation of the right hemisphere in native speakers of a tone language. We argue that neural representation of pitch-relevant information in the brainstem and early sensory level processing in the auditory cortex is shaped by the perceptual salience of domain-specific features. While both stages of processing are shaped by language experience, neural representations are transformed and fundamentally different at each biological level of abstraction. The representation of pitch relevant information in the brainstem is more fine-grained spectrotemporally as it reflects sustained neural phase-locking to pitch relevant periodicities contained in the stimulus. In contrast, the cortical pitch relevant neural activity reflects primarily a series of transient temporal neural events synchronized to certain temporal attributes of the pitch contour. We argue that experience-dependent enhancement of pitch representation for Chinese listeners most likely reflects an interaction between higher-level cognitive processes and early sensory-level processing to improve representations of behaviorally-relevant features that contribute optimally to perception. It is our view that long

  12. Knowledge discovery: Extracting usable information from large amounts of data

    International Nuclear Information System (INIS)

    Whiteson, R.

    1998-01-01

    The threat of nuclear weapons proliferation is a problem of world wide concern. Safeguards are the key to nuclear nonproliferation and data is the key to safeguards. The safeguards community has access to a huge and steadily growing volume of data. The advantages of this data rich environment are obvious, there is a great deal of information which can be utilized. The challenge is to effectively apply proven and developing technologies to find and extract usable information from that data. That information must then be assessed and evaluated to produce the knowledge needed for crucial decision making. Efficient and effective analysis of safeguards data will depend on utilizing technologies to interpret the large, heterogeneous data sets that are available from diverse sources. With an order-of-magnitude increase in the amount of data from a wide variety of technical, textual, and historical sources there is a vital need to apply advanced computer technologies to support all-source analysis. There are techniques of data warehousing, data mining, and data analysis that can provide analysts with tools that will expedite their extracting useable information from the huge amounts of data to which they have access. Computerized tools can aid analysts by integrating heterogeneous data, evaluating diverse data streams, automating retrieval of database information, prioritizing inputs, reconciling conflicting data, doing preliminary interpretations, discovering patterns or trends in data, and automating some of the simpler prescreening tasks that are time consuming and tedious. Thus knowledge discovery technologies can provide a foundation of support for the analyst. Rather than spending time sifting through often irrelevant information, analysts could use their specialized skills in a focused, productive fashion. This would allow them to make their analytical judgments with more confidence and spend more of their time doing what they do best

  13. Identifying relevant group of miRNAs in cancer using fuzzy mutual information.

    Science.gov (United States)

    Pal, Jayanta Kumar; Ray, Shubhra Sankar; Pal, Sankar K

    2016-04-01

    MicroRNAs (miRNAs) act as a major biomarker of cancer. All miRNAs in human body are not equally important for cancer identification. We propose a methodology, called FMIMS, which automatically selects the most relevant miRNAs for a particular type of cancer. In FMIMS, miRNAs are initially grouped by using a SVM-based algorithm; then the group with highest relevance is determined and the miRNAs in that group are finally ranked for selection according to their redundancy. Fuzzy mutual information is used in computing the relevance of a group and the redundancy of miRNAs within it. Superiority of the most relevant group to all others, in deciding normal or cancer, is demonstrated on breast, renal, colorectal, lung, melanoma and prostate data. The merit of FMIMS as compared to several existing methods is established. While 12 out of 15 selected miRNAs by FMIMS corroborate with those of biological investigations, three of them viz., "hsa-miR-519," "hsa-miR-431" and "hsa-miR-320c" are possible novel predictions for renal cancer, lung cancer and melanoma, respectively. The selected miRNAs are found to be involved in disease-specific pathways by targeting various genes. The method is also able to detect the responsible miRNAs even at the primary stage of cancer. The related code is available at http://www.jayanta.droppages.com/FMIMS.html .

  14. Autism Spectrum Disorder Updates - Relevant Information for Early Interventionists to Consider.

    Science.gov (United States)

    Allen-Meares, Paula; MacDonald, Megan; McGee, Kristin

    2016-01-01

    Autism spectrum disorder (ASD) is a pervasive developmental disorder characterized by deficits in social communication skills as well as repetitive, restricted or stereotyped behaviors (1). Early interventionists are often found at the forefront of assessment, evaluation, and early intervention services for children with ASD. The role of an early intervention specialist may include assessing developmental history, providing group and individual counseling, working in partnership with families on home, school, and community environments, mobilizing school and community resources, and assisting in the development of positive early intervention strategies (2, 3). The commonality among these roles resides in the importance of providing up-to-date, relevant information to families and children. The purpose of this review is to provide pertinent up-to-date knowledge for early interventionists to help inform practice in working with individuals with ASD, including common behavioral models of intervention.

  15. Extraction of temporal information in functional MRI

    Science.gov (United States)

    Singh, M.; Sungkarat, W.; Jeong, Jeong-Won; Zhou, Yongxia

    2002-10-01

    The temporal resolution of functional MRI (fMRI) is limited by the shape of the haemodynamic response function (hrf) and the vascular architecture underlying the activated regions. Typically, the temporal resolution of fMRI is on the order of 1 s. We have developed a new data processing approach to extract temporal information on a pixel-by-pixel basis at the level of 100 ms from fMRI data. Instead of correlating or fitting the time-course of each pixel to a single reference function, which is the common practice in fMRI, we correlate each pixel's time-course to a series of reference functions that are shifted with respect to each other by 100 ms. The reference function yielding the highest correlation coefficient for a pixel is then used as a time marker for that pixel. A Monte Carlo simulation and experimental study of this approach were performed to estimate the temporal resolution as a function of signal-to-noise ratio (SNR) in the time-course of a pixel. Assuming a known and stationary hrf, the simulation and experimental studies suggest a lower limit in the temporal resolution of approximately 100 ms at an SNR of 3. The multireference function approach was also applied to extract timing information from an event-related motor movement study where the subjects flexed a finger on cue. The event was repeated 19 times with the event's presentation staggered to yield an approximately 100-ms temporal sampling of the haemodynamic response over the entire presentation cycle. The timing differences among different regions of the brain activated by the motor task were clearly visualized and quantified by this method. The results suggest that it is possible to achieve a temporal resolution of /spl sim/200 ms in practice with this approach.

  16. Extractive Summarisation of Medical Documents

    Directory of Open Access Journals (Sweden)

    Abeed Sarker

    2012-09-01

    Full Text Available Background Evidence Based Medicine (EBM practice requires practitioners to extract evidence from published medical research when answering clinical queries. Due to the time-consuming nature of this practice, there is a strong motivation for systems that can automatically summarise medical documents and help practitioners find relevant information. Aim The aim of this work is to propose an automatic query-focused, extractive summarisation approach that selects informative sentences from medical documents. MethodWe use a corpus that is specifically designed for summarisation in the EBM domain. We use approximately half the corpus for deriving important statistics associated with the best possible extractive summaries. We take into account factors such as sentence position, length, sentence content, and the type of the query posed. Using the statistics from the first set, we evaluate our approach on a separate set. Evaluation of the qualities of the generated summaries is performed automatically using ROUGE, which is a popular tool for evaluating automatic summaries. Results Our summarisation approach outperforms all baselines (best baseline score: 0.1594; our score 0.1653. Further improvements are achieved when query types are taken into account. Conclusion The quality of extractive summarisation in the medical domain can be significantly improved by incorporating domain knowledge and statistics derived from a specialised corpus. Such techniques can therefore be applied for content selection in end-to-end summarisation systems.

  17. INFORMATION EXTRACTION IN TOMB PIT USING HYPERSPECTRAL DATA

    Directory of Open Access Journals (Sweden)

    X. Yang

    2018-04-01

    Full Text Available Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  18. Information Extraction in Tomb Pit Using Hyperspectral Data

    Science.gov (United States)

    Yang, X.; Hou, M.; Lyu, S.; Ma, S.; Gao, Z.; Bai, S.; Gu, M.; Liu, Y.

    2018-04-01

    Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  19. Human capital information in management reports: An analysis of compliance with the characteristic of the relevance of disclosure

    Directory of Open Access Journals (Sweden)

    Ainhoa Saitua

    2015-06-01

    Full Text Available Purpose: The aim of this paper is to assess the compliance with the characteristic of the relevance of disclosure in Management Reports, particularly dealing with Human Capital (HC information.Design/methodology/approach: We codify all instances where narratives of IBEX-35 stock index companies over a five year period in Spain comply with the recommendations for a “high quality” Management Commentary (MC in terms of the relevance characteristic of the information disclosed (IASB, 2005.Findings: The analysis results show that a greater quantity of information about HC in terms of the number of pages devoted is not always indicative of higher quality in terms of relevance if we look for the application of IASB recommendations.Research limitations/implications: Further research could assess compliance with other qualitative characteristics required by other standards or guidances that were internationaly generaly accepted.Practical implications: Among the areas that require improvement in HC disclosures we highlight forward-looking information.Social implications: We propose that an internationally accepted agreement must be struck to unite all the efforts that are being made to improve narrative information in the MC section, and specifically with reference to HC.Originality/value: This work compiles the HC disclosures identified as best practices that may serve as a reference to other companies.

  20. Assessment of efficiency of functioning the infocommunication systems a special purpose in the conditions of violation quality of relevance information

    Science.gov (United States)

    Parinov, A. V.; Korotkikh, L. P.; Desyatov, D. B.; Stepanov, L. V.

    2018-03-01

    The uniqueness of information processing mechanisms in special-purpose infocommunication systems and the increased interest of intruders lead to an increase in the relevance of the problems associated with their protection. The paper considers the issues of building risk-models for the violation of the relevance and value of information in infocommunication systems for special purposes. Also, special attention is paid to the connection between the qualities of relevance and the value of information obtained as a result of the operation of infocommunication systems for special purposes. Analytical expressions for the risk and damage function in the time range in special-purpose infocommunication systems are obtained, which can serve as a mathematical basis for risk assessment. Further, an analytical expression is obtained to assess the chance of obtaining up-to-date information in the operation of infocommunication systems up to the time the information quality is violated. An analytical expression for estimating the chance can be used to calculate the effectiveness of a special-purpose infocommunication system.

  1. Automatic knowledge extraction in sequencing analysis with multiagent system and grid computing.

    Science.gov (United States)

    González, Roberto; Zato, Carolina; Benito, Rocío; Bajo, Javier; Hernández, Jesús M; De Paz, Juan F; Vera, Vicente; Corchado, Juan M

    2012-12-01

    Advances in bioinformatics have contributed towards a significant increase in available information. Information analysis requires the use of distributed computing systems to best engage the process of data analysis. This study proposes a multiagent system that incorporates grid technology to facilitate distributed data analysis by dynamically incorporating the roles associated to each specific case study. The system was applied to genetic sequencing data to extract relevant information about insertions, deletions or polymorphisms.

  2. Automatic knowledge extraction in sequencing analysis with multiagent system and grid computing

    Directory of Open Access Journals (Sweden)

    González Roberto

    2012-12-01

    Full Text Available Advances in bioinformatics have contributed towards a significant increase in available information. Information analysis requires the use of distributed computing systems to best engage the process of data analysis. This study proposes a multiagent system that incorporates grid technology to facilitate distributed data analysis by dynamically incorporating the roles associated to each specific case study. The system was applied to genetic sequencing data to extract relevant information about insertions, deletions or polymorphisms.

  3. Dual-wavelength phase-shifting digital holography selectively extracting wavelength information from wavelength-multiplexed holograms.

    Science.gov (United States)

    Tahara, Tatsuki; Mori, Ryota; Kikunaga, Shuhei; Arai, Yasuhiko; Takaki, Yasuhiro

    2015-06-15

    Dual-wavelength phase-shifting digital holography that selectively extracts wavelength information from five wavelength-multiplexed holograms is presented. Specific phase shifts for respective wavelengths are introduced to remove the crosstalk components and extract only the object wave at the desired wavelength from the holograms. Object waves in multiple wavelengths are selectively extracted by utilizing 2π ambiguity and the subtraction procedures based on phase-shifting interferometry. Numerical results show the validity of the proposed technique. The proposed technique is also experimentally demonstrated.

  4. Stimulus-response correspondence effect as a function of temporal overlap between relevant and irrelevant information processing.

    Science.gov (United States)

    Wang, Dong-Yuan Debbie; Richard, F Dan; Ray, Brittany

    2016-01-01

    The stimulus-response correspondence (SRC) effect refers to advantages in performance when stimulus and response correspond in dimensions or features, even if the common features are irrelevant to the task. Previous research indicated that the SRC effect depends on the temporal course of stimulus information processing. The current study investigated how the temporal overlap between relevant and irrelevant stimulus processing influences the SRC effect. In this experiment, the irrelevant stimulus (a previously associated tone) preceded the relevant stimulus (a coloured rectangle). The irrelevant and relevant stimuli onset asynchrony was varied to manipulate the temporal overlap between the irrelevant and relevant stimuli processing. Results indicated that the SRC effect size varied as a quadratic function of the temporal overlap between the relevant stimulus and irrelevant stimulus. This finding extends previous experimental observations that the SRC effect size varies in an increasing or decreasing function with reaction time. The current study demonstrated a quadratic function between effect size and the temporal overlap.

  5. Method for extracting relevant electrical parameters from graphene field-effect transistors using a physical model

    International Nuclear Information System (INIS)

    Boscá, A.; Pedrós, J.; Martínez, J.; Calle, F.

    2015-01-01

    Due to its intrinsic high mobility, graphene has proved to be a suitable material for high-speed electronics, where graphene field-effect transistor (GFET) has shown excellent properties. In this work, we present a method for extracting relevant electrical parameters from GFET devices using a simple electrical characterization and a model fitting. With experimental data from the device output characteristics, the method allows to calculate parameters such as the mobility, the contact resistance, and the fixed charge. Differentiated electron and hole mobilities and direct connection with intrinsic material properties are some of the key aspects of this method. Moreover, the method output values can be correlated with several issues during key fabrication steps such as the graphene growth and transfer, the lithographic steps, or the metalization processes, providing a flexible tool for quality control in GFET fabrication, as well as a valuable feedback for improving the material-growth process

  6. Method for extracting relevant electrical parameters from graphene field-effect transistors using a physical model

    Energy Technology Data Exchange (ETDEWEB)

    Boscá, A., E-mail: alberto.bosca@upm.es [Instituto de Sistemas Optoelectrónicos y Microtecnología, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Dpto. de Ingeniería Electrónica, E.T.S.I. de Telecomunicación, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Pedrós, J. [Instituto de Sistemas Optoelectrónicos y Microtecnología, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Campus de Excelencia Internacional, Campus Moncloa UCM-UPM, Madrid 28040 (Spain); Martínez, J. [Instituto de Sistemas Optoelectrónicos y Microtecnología, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Dpto. de Ciencia de Materiales, E.T.S.I de Caminos, Canales y Puertos, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Calle, F. [Instituto de Sistemas Optoelectrónicos y Microtecnología, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Dpto. de Ingeniería Electrónica, E.T.S.I. de Telecomunicación, Universidad Politécnica de Madrid, Madrid 28040 (Spain); Campus de Excelencia Internacional, Campus Moncloa UCM-UPM, Madrid 28040 (Spain)

    2015-01-28

    Due to its intrinsic high mobility, graphene has proved to be a suitable material for high-speed electronics, where graphene field-effect transistor (GFET) has shown excellent properties. In this work, we present a method for extracting relevant electrical parameters from GFET devices using a simple electrical characterization and a model fitting. With experimental data from the device output characteristics, the method allows to calculate parameters such as the mobility, the contact resistance, and the fixed charge. Differentiated electron and hole mobilities and direct connection with intrinsic material properties are some of the key aspects of this method. Moreover, the method output values can be correlated with several issues during key fabrication steps such as the graphene growth and transfer, the lithographic steps, or the metalization processes, providing a flexible tool for quality control in GFET fabrication, as well as a valuable feedback for improving the material-growth process.

  7. Chaotic spectra: How to extract dynamic information

    International Nuclear Information System (INIS)

    Taylor, H.S.; Gomez Llorente, J.M.; Zakrzewski, J.; Kulander, K.C.

    1988-10-01

    Nonlinear dynamics is applied to chaotic unassignable atomic and molecular spectra with the aim of extracting detailed information about regular dynamic motions that exist over short intervals of time. It is shown how this motion can be extracted from high resolution spectra by doing low resolution studies or by Fourier transforming limited regions of the spectrum. These motions mimic those of periodic orbits (PO) and are inserts into the dominant chaotic motion. Considering these inserts and the PO as a dynamically decoupled region of space, resonant scattering theory and stabilization methods enable us to compute ladders of resonant states which interact with the chaotic quasi-continuum computed in principle from basis sets placed off the PO. The interaction of the resonances with the quasicontinuum explains the low resolution spectra seen in such experiments. It also allows one to associate low resolution features with a particular PO. The motion on the PO thereby supplies the molecular movements whose quantization causes the low resolution spectra. Characteristic properties of the periodic orbit based resonances are discussed. The method is illustrated on the photoabsorption spectrum of the hydrogen atom in a strong magnetic field and on the photodissociation spectrum of H 3 + . Other molecular systems which are currently under investigation using this formalism are also mentioned. 53 refs., 10 figs., 2 tabs

  8. Automated Extraction of Substance Use Information from Clinical Texts.

    Science.gov (United States)

    Wang, Yan; Chen, Elizabeth S; Pakhomov, Serguei; Arsoniadis, Elliot; Carter, Elizabeth W; Lindemann, Elizabeth; Sarkar, Indra Neil; Melton, Genevieve B

    2015-01-01

    Within clinical discourse, social history (SH) includes important information about substance use (alcohol, drug, and nicotine use) as key risk factors for disease, disability, and mortality. In this study, we developed and evaluated a natural language processing (NLP) system for automated detection of substance use statements and extraction of substance use attributes (e.g., temporal and status) based on Stanford Typed Dependencies. The developed NLP system leveraged linguistic resources and domain knowledge from a multi-site social history study, Propbank and the MiPACQ corpus. The system attained F-scores of 89.8, 84.6 and 89.4 respectively for alcohol, drug, and nicotine use statement detection, as well as average F-scores of 82.1, 90.3, 80.8, 88.7, 96.6, and 74.5 respectively for extraction of attributes. Our results suggest that NLP systems can achieve good performance when augmented with linguistic resources and domain knowledge when applied to a wide breadth of substance use free text clinical notes.

  9. Associations between presence of relevant information in referrals to radiology and prevalence rates in patients with suspected pulmonary embolism.

    Science.gov (United States)

    Hedner, Charlotta; Sundgren, Pia C; Kelly, Aine Marie

    2013-09-01

    The purpose of this study was to assess if the presence of information including the pretest probability (Wells score), other known risk factors, and symptoms given on referrals for computed tomography (CT) pulmonary angiography correlated with prevalence rates for pulmonary embolism (PE). Also, to evaluate for differences between a university and a regional hospital setting regarding patient characteristics, amount of relevant information provided on referrals, and prevalence rates for pulmonary embolism. Retrospective review of all consecutive referrals (emergency room, inpatient, and outpatient) for CT performed on children and adults for suspected PE from two sites: a tertiary (university) hospital (site 1) and a secondary (regional) hospital (site 2) over a 5-year period. The overall prevalence rate was 510/3641 or 14% of all referrals. Significantly higher number of males had a positive CT compared to women (18% versus 12%, P relevant information on the referral and the probability for positive finding existed, a slight trend was noted (P = .09). In two categories, "hypoxia" and "signs of deep vein thrombosis," the presence of this information conferred a higher probability for pulmonary embolism, P information conferred a higher probability for pulmonary embolism. The amount of relevant clinical information on the request did not correlate with prevalence rates, which may reflect a lack of documentation on the part of emergency physicians who may use a "gestalt" approach. Request forms likely did not capture all relevant patient risks and many factors may interact with each other, both positively and negatively. Pretest probability estimations were rarely performed, despite their inclusion in major society guidelines. Copyright © 2013 AUR. Published by Elsevier Inc. All rights reserved.

  10. Network and Ensemble Enabled Entity Extraction in Informal Text (NEEEEIT) final report

    Energy Technology Data Exchange (ETDEWEB)

    Kegelmeyer, Philip W. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Shead, Timothy M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Dunlavy, Daniel M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2013-09-01

    This SAND report summarizes the activities and outcomes of the Network and Ensemble Enabled Entity Extraction in Information Text (NEEEEIT) LDRD project, which addressed improving the accuracy of conditional random fields for named entity recognition through the use of ensemble methods.

  11. Age and self-relevance effects on information search during decision making.

    Science.gov (United States)

    Hess, Thomas M; Queen, Tara L; Ennis, Gilda E

    2013-09-01

    We investigated how information search strategies used to support decision making were influenced by self-related implications of the task to the individual. Consistent with the notion of selective engagement, we hypothesized that increased self-relevance would result in more adaptive search behaviors and that this effect would be stronger in older adults than in younger adults. We examined search behaviors in 79 younger and 81 older adults using a process-tracing procedure with 2 different decision tasks. The impact of motivation (i.e., self-related task implications) was examined by manipulating social accountability and the age-related relevance of the task. Although age differences in search strategies were not great, older adults were more likely than younger adults to use simpler strategies in contexts with minimal self-implications. Contrary to expectations, young and old alike were more likely to use noncompensatory than compensatory strategies, even when engaged in systematic search, with education being the most important determinant of search behavior. The results support the notion that older adults are adaptive decision makers and that factors other than age may be more important determinants of performance in situations where knowledge can be used to support performance.

  12. EXTRACTING KNOWLEDGE FROM DATA - DATA MINING

    Directory of Open Access Journals (Sweden)

    DIANA ELENA CODREANU

    2011-04-01

    Full Text Available Managers of economic organizations have at their disposal a large volume of information and practically facing an avalanche of information, but they can not operate studying reports containing detailed data volumes without a correlation because of the good an organization may be decided in fractions of time. Thus, to take the best and effective decisions in real time, managers need to have the correct information is presented quickly, in a synthetic way, but relevant to allow for predictions and analysis.This paper wants to highlight the solutions to extract knowledge from data, namely data mining. With this technology not only has to verify some hypotheses, but aims at discovering new knowledge, so that economic organization to cope with fierce competition in the market.

  13. Three-dimensional information extraction from GaoFen-1 satellite images for landslide monitoring

    Science.gov (United States)

    Wang, Shixin; Yang, Baolin; Zhou, Yi; Wang, Futao; Zhang, Rui; Zhao, Qing

    2018-05-01

    To more efficiently use GaoFen-1 (GF-1) satellite images for landslide emergency monitoring, a Digital Surface Model (DSM) can be generated from GF-1 across-track stereo image pairs to build a terrain dataset. This study proposes a landslide 3D information extraction method based on the terrain changes of slope objects. The slope objects are mergences of segmented image objects which have similar aspects; and the terrain changes are calculated from the post-disaster Digital Elevation Model (DEM) from GF-1 and the pre-disaster DEM from GDEM V2. A high mountain landslide that occurred in Wenchuan County, Sichuan Province is used to conduct a 3D information extraction test. The extracted total area of the landslide is 22.58 ha; the displaced earth volume is 652,100 m3; and the average sliding direction is 263.83°. The accuracies of them are 0.89, 0.87 and 0.95, respectively. Thus, the proposed method expands the application of GF-1 satellite images to the field of landslide emergency monitoring.

  14. Lithium NLP: A System for Rich Information Extraction from Noisy User Generated Text on Social Media

    OpenAIRE

    Bhargava, Preeti; Spasojevic, Nemanja; Hu, Guoning

    2017-01-01

    In this paper, we describe the Lithium Natural Language Processing (NLP) system - a resource-constrained, high- throughput and language-agnostic system for information extraction from noisy user generated text on social media. Lithium NLP extracts a rich set of information including entities, topics, hashtags and sentiment from text. We discuss several real world applications of the system currently incorporated in Lithium products. We also compare our system with existing commercial and acad...

  15. System and method for extracting physiological information from remotely detected electromagnetic radiation

    NARCIS (Netherlands)

    2016-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  16. System and method for extracting physiological information from remotely detected electromagnetic radiation

    NARCIS (Netherlands)

    2015-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  17. Effects of self-schema elaboration on affective and cognitive reactions to self-relevant information.

    Science.gov (United States)

    Petersen, L E; Stahlberg, D; Dauenheimer, D

    2000-02-01

    The basic assumption of the integrative self-schema model (ISSM; L.-E. Petersen, 1994; L.-E. Petersen, D. Stahlberg, & D. Dauenheimer, 1996; D. Stahlberg, L.-E. Petersen, & D. Dauenheimer, 1994, 1999) is that self-schema elaboration (schematic vs. aschematic) affects reactions to self-relevant information. This assumption is based on the idea that schematic dimensions occupy a more central position in the cognitive system than aschematic dimensions. In the first study, this basic prediction could be clearly confirmed: The results showed that schematic dimensions possessed stronger cognitive associations with other self-relevant cognitions as well as a higher resistance to change than aschematic dimensions did. In the second study, the main assumptions of the ISSM concerning the affective and cognitive reactions to self-relevant feedback were tested: The ISSM proposes that, on schematic dimensions, reactions to self-relevant feedback will most likely follow principles of self-consistency theory, whereas on aschematic dimensions positive feedback should elicit the most positive reactions that self-enhancement theory would predict. The experimental results clearly confirmed the hypotheses derived from the ISSM for affective reactions. Cognitive reactions, however, were in line with self-consistency principles and were not modified by the elaboration of the self-schema dimension involved.

  18. INFO ANAV, a channel that is consolidated in the communication of information relevant to plant safety

    International Nuclear Information System (INIS)

    Lopera Broto, A. J.; Balbas Gomez, S.

    2012-01-01

    This weekly publication intended to make it to all the people who work at the sites of Asco and Vandellos relevant information for security since we are all responsible for the safe and reliable operation of our plants.

  19. Autism Spectrum Disorder Updates – Relevant Information for Early Interventionists to Consider

    Science.gov (United States)

    Allen-Meares, Paula; MacDonald, Megan; McGee, Kristin

    2016-01-01

    Autism spectrum disorder (ASD) is a pervasive developmental disorder characterized by deficits in social communication skills as well as repetitive, restricted or stereotyped behaviors (1). Early interventionists are often found at the forefront of assessment, evaluation, and early intervention services for children with ASD. The role of an early intervention specialist may include assessing developmental history, providing group and individual counseling, working in partnership with families on home, school, and community environments, mobilizing school and community resources, and assisting in the development of positive early intervention strategies (2, 3). The commonality among these roles resides in the importance of providing up-to-date, relevant information to families and children. The purpose of this review is to provide pertinent up-to-date knowledge for early interventionists to help inform practice in working with individuals with ASD, including common behavioral models of intervention. PMID:27840812

  20. Methods from Information Extraction from LIDAR Intensity Data and Multispectral LIDAR Technology

    Science.gov (United States)

    Scaioni, M.; Höfle, B.; Baungarten Kersting, A. P.; Barazzetti, L.; Previtali, M.; Wujanz, D.

    2018-04-01

    LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on `Information Extraction from LiDAR Intensity Data' has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  1. METHODS FROM INFORMATION EXTRACTION FROM LIDAR INTENSITY DATA AND MULTISPECTRAL LIDAR TECHNOLOGY

    Directory of Open Access Journals (Sweden)

    M. Scaioni

    2018-04-01

    Full Text Available LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on ‘Information Extraction from LiDAR Intensity Data’ has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  2. Electronic processing of informed consents in a global pharmaceutical company environment.

    Science.gov (United States)

    Vishnyakova, Dina; Gobeill, Julien; Oezdemir-Zaech, Fatma; Kreim, Olivier; Vachon, Therese; Clade, Thierry; Haenning, Xavier; Mikhailov, Dmitri; Ruch, Patrick

    2014-01-01

    We present an electronic capture tool to process informed consents, which are mandatory recorded when running a clinical trial. This tool aims at the extraction of information expressing the duration of the consent given by the patient to authorize the exploitation of biomarker-related information collected during clinical trials. The system integrates a language detection module (LDM) to route a document into the appropriate information extraction module (IEM). The IEM is based on language-specific sets of linguistic rules for the identification of relevant textual facts. The achieved accuracy of both the LDM and IEM is 99%. The architecture of the system is described in detail.

  3. Historical maintenance relevant information road-map for a self-learning maintenance prediction procedural approach

    Science.gov (United States)

    Morales, Francisco J.; Reyes, Antonio; Cáceres, Noelia; Romero, Luis M.; Benitez, Francisco G.; Morgado, Joao; Duarte, Emanuel; Martins, Teresa

    2017-09-01

    A large percentage of transport infrastructures are composed of linear assets, such as roads and rail tracks. The large social and economic relevance of these constructions force the stakeholders to ensure a prolonged health/durability. Even though, inevitable malfunctioning, breaking down, and out-of-service periods arise randomly during the life cycle of the infrastructure. Predictive maintenance techniques tend to diminish the appearance of unpredicted failures and the execution of needed corrective interventions, envisaging the adequate interventions to be conducted before failures show up. This communication presents: i) A procedural approach, to be conducted, in order to collect the relevant information regarding the evolving state condition of the assets involved in all maintenance interventions; this reported and stored information constitutes a rich historical data base to train Machine Learning algorithms in order to generate reliable predictions of the interventions to be carried out in further time scenarios. ii) A schematic flow chart of the automatic learning procedure. iii) Self-learning rules from automatic learning from false positive/negatives. The description, testing, automatic learning approach and the outcomes of a pilot case are presented; finally some conclusions are outlined regarding the methodology proposed for improving the self-learning predictive capability.

  4. Automated extraction of chemical structure information from digital raster images

    Directory of Open Access Journals (Sweden)

    Shedden Kerby A

    2009-02-01

    Full Text Available Abstract Background To search for chemical structures in research articles, diagrams or text representing molecules need to be translated to a standard chemical file format compatible with cheminformatic search engines. Nevertheless, chemical information contained in research articles is often referenced as analog diagrams of chemical structures embedded in digital raster images. To automate analog-to-digital conversion of chemical structure diagrams in scientific research articles, several software systems have been developed. But their algorithmic performance and utility in cheminformatic research have not been investigated. Results This paper aims to provide critical reviews for these systems and also report our recent development of ChemReader – a fully automated tool for extracting chemical structure diagrams in research articles and converting them into standard, searchable chemical file formats. Basic algorithms for recognizing lines and letters representing bonds and atoms in chemical structure diagrams can be independently run in sequence from a graphical user interface-and the algorithm parameters can be readily changed-to facilitate additional development specifically tailored to a chemical database annotation scheme. Compared with existing software programs such as OSRA, Kekule, and CLiDE, our results indicate that ChemReader outperforms other software systems on several sets of sample images from diverse sources in terms of the rate of correct outputs and the accuracy on extracting molecular substructure patterns. Conclusion The availability of ChemReader as a cheminformatic tool for extracting chemical structure information from digital raster images allows research and development groups to enrich their chemical structure databases by annotating the entries with published research articles. Based on its stable performance and high accuracy, ChemReader may be sufficiently accurate for annotating the chemical database with links

  5. A neural mechanism of dynamic gating of task-relevant information by top-down influence in primary visual cortex.

    Science.gov (United States)

    Kamiyama, Akikazu; Fujita, Kazuhisa; Kashimori, Yoshiki

    2016-12-01

    Visual recognition involves bidirectional information flow, which consists of bottom-up information coding from retina and top-down information coding from higher visual areas. Recent studies have demonstrated the involvement of early visual areas such as primary visual area (V1) in recognition and memory formation. V1 neurons are not passive transformers of sensory inputs but work as adaptive processor, changing their function according to behavioral context. Top-down signals affect tuning property of V1 neurons and contribute to the gating of sensory information relevant to behavior. However, little is known about the neuronal mechanism underlying the gating of task-relevant information in V1. To address this issue, we focus on task-dependent tuning modulations of V1 neurons in two tasks of perceptual learning. We develop a model of the V1, which receives feedforward input from lateral geniculate nucleus and top-down input from a higher visual area. We show here that the change in a balance between excitation and inhibition in V1 connectivity is necessary for gating task-relevant information in V1. The balance change well accounts for the modulations of tuning characteristic and temporal properties of V1 neuronal responses. We also show that the balance change of V1 connectivity is shaped by top-down signals with temporal correlations reflecting the perceptual strategies of the two tasks. We propose a learning mechanism by which synaptic balance is modulated. To conclude, top-down signal changes the synaptic balance between excitation and inhibition in V1 connectivity, enabling early visual area such as V1 to gate context-dependent information under multiple task performances. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  6. What's New in the Medicine Cabinet?: A Panoramic Review of Clinically Relevant Information for the Busy Dermatologist.

    Science.gov (United States)

    Del Rosso, James Q; Zeichner, Joshua

    2014-01-01

    This article is the first in a periodic series of therapeutic topics with short reviews gleaned from major dermatology meetings, especially Scientific Poster Sessions, and is designed to provide information that may assist the readers in adapting information from the literature to their clinical practice. The topics covered in this issue are discussions of the clinical relevance of newer information about acne pathophysiology, acne in adult women, and topical corticosteroid spray formulations for chronic plaque psoriasis.

  7. Evaluation of relevant information for optimal reflector modeling through data assimilation procedures

    International Nuclear Information System (INIS)

    Argaud, J.P.; Bouriquet, B.; Clerc, T.; Lucet-Sanchez, F.; Poncot, A.

    2015-01-01

    The goal of this study is to look after the amount of information that is mandatory to get a relevant parameters optimisation by data assimilation for physical models in neutronic diffusion calculations, and to determine what is the best information to reach the optimum of accuracy at the cheapest cost. To evaluate the quality of the optimisation, we study the covariance matrix that represents the accuracy of the optimised parameter. This matrix is a classical output of the data assimilation procedure, and it is the main information about accuracy and sensitivity of the parameter optimal determination. We present some results collected in the field of neutronic simulation for PWR type reactor. We seek to optimise the reflector parameters that characterise the neutronic reflector surrounding the whole reactive core. On the basis of the configuration studies, it has been shown that with data assimilation we can determine a global strategy to optimise the quality of the result with respect to the amount of information provided. The consequence of this is a cost reduction in terms of measurement and/or computing time with respect to the basic approach. Another result is that using multi-campaign data rather data from a unique campaign significantly improves the efficiency of parameters optimisation

  8. Age differences in default and reward networks during processing of personally relevant information.

    Science.gov (United States)

    Grady, Cheryl L; Grigg, Omer; Ng, Charisa

    2012-06-01

    We recently found activity in default mode and reward-related regions during self-relevant tasks in young adults. Here we examine the effect of aging on engagement of the default network (DN) and reward network (RN) during these tasks. Previous studies have shown reduced engagement of the DN and reward areas in older adults, but the influence of age on these circuits during self-relevant tasks has not been examined. The tasks involved judging personality traits about one's self or a well known other person. There were no age differences in reaction time on the tasks but older adults had more positive Self and Other judgments, whereas younger adults had more negative judgments. Both groups had increased DN and RN activity during the self-relevant tasks, relative to non-self tasks, but this increase was reduced in older compared to young adults. Functional connectivity of both networks during the tasks was weaker in the older relative to younger adults. Intrinsic functional connectivity, measured at rest, also was weaker in the older adults in the DN, but not in the RN. These results suggest that, in younger adults, the processing of personally relevant information involves robust activation of and functional connectivity within these two networks, in line with current models that emphasize strong links between the self and reward. The finding that older adults had more positive judgments, but weaker engagement and less consistent functional connectivity in these networks, suggests potential brain mechanisms for the "positivity bias" with aging. Copyright © 2012 Elsevier Ltd. All rights reserved.

  9. Autism spectrum disorder updates – relevant information for early interventionists to consider

    Directory of Open Access Journals (Sweden)

    Paula Allen-Meares

    2016-10-01

    Full Text Available Autism spectrum disorder (ASD is a pervasive developmental disorder characterized by deficits in social communication skills as well as repetitive, restricted or stereotyped behaviors (1. Early interventionists are often found at the forefront of assessment, evaluation and early intervention services for children with ASD. The role of an early intervention specialist may include, assessing developmental history, providing group and individual counseling, working in partnership with families on home, school, and community environments, mobilizing school and community resources and assisting in the development of positive early intervention strategies (2, 3. The commonality amongst these roles resides in the importance of providing up-to-date, relevant information to families and children. The purpose of this review is to provide pertinent up-to-date knowledge for early interventionists to help inform practice in working with individuals with ASD, including common behavioral models of intervention.

  10. Earth Science Data Analytics: Preparing for Extracting Knowledge from Information

    Science.gov (United States)

    Kempler, Steven; Barbieri, Lindsay

    2016-01-01

    Data analytics is the process of examining large amounts of data of a variety of types to uncover hidden patterns, unknown correlations and other useful information. Data analytics is a broad term that includes data analysis, as well as an understanding of the cognitive processes an analyst uses to understand problems and explore data in meaningful ways. Analytics also include data extraction, transformation, and reduction, utilizing specific tools, techniques, and methods. Turning to data science, definitions of data science sound very similar to those of data analytics (which leads to a lot of the confusion between the two). But the skills needed for both, co-analyzing large amounts of heterogeneous data, understanding and utilizing relevant tools and techniques, and subject matter expertise, although similar, serve different purposes. Data Analytics takes on a practitioners approach to applying expertise and skills to solve issues and gain subject knowledge. Data Science, is more theoretical (research in itself) in nature, providing strategic actionable insights and new innovative methodologies. Earth Science Data Analytics (ESDA) is the process of examining, preparing, reducing, and analyzing large amounts of spatial (multi-dimensional), temporal, or spectral data using a variety of data types to uncover patterns, correlations and other information, to better understand our Earth. The large variety of datasets (temporal spatial differences, data types, formats, etc.) invite the need for data analytics skills that understand the science domain, and data preparation, reduction, and analysis techniques, from a practitioners point of view. The application of these skills to ESDA is the focus of this presentation. The Earth Science Information Partners (ESIP) Federation Earth Science Data Analytics (ESDA) Cluster was created in recognition of the practical need to facilitate the co-analysis of large amounts of data and information for Earth science. Thus, from a to

  11. Japanese Quality Assurance System Regarding the Provision of Material Accounting Reports and the Safeguards Relevant Information to the IAEA

    International Nuclear Information System (INIS)

    Goto, Y.; Namekawa, M.; Kumekawa, H.; Usui, A.; Sano, K.

    2015-01-01

    The provision of the safeguards relevant reports and information in accordance with the comprehensive safeguards agreement (CSA) and the additional protocol (AP) is the basis for the IAEA safeguards. The government of Japan (Japan Safeguards Office, JSGO) has believed that the correct reports contribute to effective and efficient safeguards therefore the domestic quality assurance system for the reporting to the IAEA was already established at the time of the accession of the CSA in 1977. It consists of Code 10 interpretation (including the seminars for operators in Japan), SSAC's checks for syntax error, code and internal consistency (computer based consistency check between facilities) and the discussion with the IAEA on the facilities' measurement system for bulk-handling facilities, which contributes to the more accurate reports from operators. This spirit has been maintained for the entry into force of the AP. For example, questions and amplification from the IAEA will be taken into account the review of the AP declaration before sending to the IAEA and the open source information such as news article and scientific literature in Japanese is collected and translated into English, and the translated information is provided to the IAEA as the supplementary information, which may contribute to broadening the IAEA information source and to their comprehensive evaluation. The other safeguards relevant information, such as the mail-box information for SNRI at LEU fuel fabrication plants, is also checked by the JSGO's QC software before posting. The software was developed by JSGO and it checks data format, batch IDs, birth/death date, shipper/receiver information and material description code. This paper explains the history of the development of the Japanese quality assurance system regarding the reports and the safeguards relevant information to the IAEA. (author)

  12. EXTRACT

    DEFF Research Database (Denmark)

    Pafilis, Evangelos; Buttigieg, Pier Luigi; Ferrell, Barbra

    2016-01-01

    The microbial and molecular ecology research communities have made substantial progress on developing standards for annotating samples with environment metadata. However, sample manual annotation is a highly labor intensive process and requires familiarity with the terminologies used. We have the...... and text-mining-assisted curation revealed that EXTRACT speeds up annotation by 15-25% and helps curators to detect terms that would otherwise have been missed.Database URL: https://extract.hcmr.gr/......., organism, tissue and disease terms. The evaluators in the BioCreative V Interactive Annotation Task found the system to be intuitive, useful, well documented and sufficiently accurate to be helpful in spotting relevant text passages and extracting organism and environment terms. Comparison of fully manual...

  13. ERP investigation of attentional disengagement from suicide-relevant information in patients with major depressive disorder.

    Science.gov (United States)

    Baik, Seung Yeon; Jeong, Minkyung; Kim, Hyang Sook; Lee, Seung-Hwan

    2018-01-01

    Previous studies suggest the presence of attentional bias towards suicide-relevant information in suicidal individuals. However, the findings are limited by their reliance on behavioral measures. This study investigates the role of difficulty in disengaging attention from suicide-relevant stimuli using the P300 component of event-related potentials (ERPs). Forty-four adults with Major Depressive Disorder (MDD) were administered the spatial cueing task using suicide-relevant and negatively-valenced words as cue stimuli. Disengagement difficulty was measured using reaction time and P300 during invalid trials. P300 amplitudes at Pz were higher in suicide-relevant compared to negatively-valenced word condition on invalid trials for participants with low rates of suicidal behavior. However, no such difference was found among participants with high rates of suicidal behavior. P300 amplitudes for suicide-relevant word condition were negatively correlated with "lifetime suicide ideation and attempt" at Pz. No significant results were found for the reaction time data, indicating that the ERP may be more sensitive in capturing the attentional disengagement effect. The groups were divided according to Suicidal Behaviors Questionnaire-Revised (SBQ-R) total score. Neutral stimulus was not included as cue stimuli. Most participants were under medication during the experiment. Our results indicate that patients with MDD and low rates of suicidal behavior show difficulty in disengaging attention from suicide-relevant stimuli. We suggest that suicide-specific disengagement difficulties may be related to recentness of suicide attempt and that acquired capability for suicide may contribute to reduced disengagement difficulties. Copyright © 2017 Elsevier B.V. All rights reserved.

  14. Differential impact of relevant and irrelevant dimension primes on rule-based and information-integration category learning.

    Science.gov (United States)

    Grimm, Lisa R; Maddox, W Todd

    2013-11-01

    Research has identified multiple category-learning systems with each being "tuned" for learning categories with different task demands and each governed by different neurobiological systems. Rule-based (RB) classification involves testing verbalizable rules for category membership while information-integration (II) classification requires the implicit learning of stimulus-response mappings. In the first study to directly test rule priming with RB and II category learning, we investigated the influence of the availability of information presented at the beginning of the task. Participants viewed lines that varied in length, orientation, and position on the screen, and were primed to focus on stimulus dimensions that were relevant or irrelevant to the correct classification rule. In Experiment 1, we used an RB category structure, and in Experiment 2, we used an II category structure. Accuracy and model-based analyses suggested that a focus on relevant dimensions improves RB task performance later in learning while a focus on an irrelevant dimension improves II task performance early in learning. © 2013.

  15. Automated concept-level information extraction to reduce the need for custom software and rules development.

    Science.gov (United States)

    D'Avolio, Leonard W; Nguyen, Thien M; Goryachev, Sergey; Fiore, Louis D

    2011-01-01

    Despite at least 40 years of promising empirical performance, very few clinical natural language processing (NLP) or information extraction systems currently contribute to medical science or care. The authors address this gap by reducing the need for custom software and rules development with a graphical user interface-driven, highly generalizable approach to concept-level retrieval. A 'learn by example' approach combines features derived from open-source NLP pipelines with open-source machine learning classifiers to automatically and iteratively evaluate top-performing configurations. The Fourth i2b2/VA Shared Task Challenge's concept extraction task provided the data sets and metrics used to evaluate performance. Top F-measure scores for each of the tasks were medical problems (0.83), treatments (0.82), and tests (0.83). Recall lagged precision in all experiments. Precision was near or above 0.90 in all tasks. Discussion With no customization for the tasks and less than 5 min of end-user time to configure and launch each experiment, the average F-measure was 0.83, one point behind the mean F-measure of the 22 entrants in the competition. Strong precision scores indicate the potential of applying the approach for more specific clinical information extraction tasks. There was not one best configuration, supporting an iterative approach to model creation. Acceptable levels of performance can be achieved using fully automated and generalizable approaches to concept-level information extraction. The described implementation and related documentation is available for download.

  16. Extraction of quantifiable information from complex systems

    CERN Document Server

    Dahmen, Wolfgang; Griebel, Michael; Hackbusch, Wolfgang; Ritter, Klaus; Schneider, Reinhold; Schwab, Christoph; Yserentant, Harry

    2014-01-01

    In April 2007, the  Deutsche Forschungsgemeinschaft (DFG) approved the  Priority Program 1324 “Mathematical Methods for Extracting Quantifiable Information from Complex Systems.” This volume presents a comprehensive overview of the most important results obtained over the course of the program.   Mathematical models of complex systems provide the foundation for further technological developments in science, engineering and computational finance.  Motivated by the trend toward steadily increasing computer power, ever more realistic models have been developed in recent years. These models have also become increasingly complex, and their numerical treatment poses serious challenges.   Recent developments in mathematics suggest that, in the long run, much more powerful numerical solution strategies could be derived if the interconnections between the different fields of research were systematically exploited at a conceptual level. Accordingly, a deeper understanding of the mathematical foundations as w...

  17. THE EXTRACTION OF INDOOR BUILDING INFORMATION FROM BIM TO OGC INDOORGML

    Directory of Open Access Journals (Sweden)

    T.-A. Teo

    2017-07-01

    Full Text Available Indoor Spatial Data Infrastructure (indoor-SDI is an important SDI for geosptial analysis and location-based services. Building Information Model (BIM has high degree of details in geometric and semantic information for building. This study proposed direct conversion schemes to extract indoor building information from BIM to OGC IndoorGML. The major steps of the research include (1 topological conversion from building model into indoor network model; and (2 generation of IndoorGML. The topological conversion is a major process of generating and mapping nodes and edges from IFC to indoorGML. Node represents every space (e.g. IfcSpace and objects (e.g. IfcDoor in the building while edge shows the relationships between nodes. According to the definition of IndoorGML, the topological model in the dual space is also represented as a set of nodes and edges. These definitions of IndoorGML are the same as in the indoor network. Therefore, we can extract the necessary data in the indoor network and easily convert them into IndoorGML based on IndoorGML Schema. The experiment utilized a real BIM model to examine the proposed method. The experimental results indicated that the 3D indoor model (i.e. IndoorGML model can be automatically imported from IFC model by the proposed procedure. In addition, the geometric and attribute of building elements are completely and correctly converted from BIM to indoor-SDI.

  18. Geospatial Information Relevant to the Flood Protection Available on The Mainstream Web

    Directory of Open Access Journals (Sweden)

    Kliment Tomáš

    2014-03-01

    Full Text Available Flood protection is one of several disciplines where geospatial data is very important and is a crucial component. Its management, processing and sharing form the foundation for their efficient use; therefore, special attention is required in the development of effective, precise, standardized, and interoperable models for the discovery and publishing of data on the Web. This paper describes the design of a methodology to discover Open Geospatial Consortium (OGC services on the Web and collect descriptive information, i.e., metadata in a geocatalogue. A pilot implementation of the proposed methodology - Geocatalogue of geospatial information provided by OGC services discovered on Google (hereinafter “Geocatalogue” - was used to search for available resources relevant to the area of flood protection. The result is an analysis of the availability of resources discovered through their metadata collected from the OGC services (WMS, WFS, etc. and the resources they provide (WMS layers, WFS objects, etc. within the domain of flood protection.

  19. Adaptive web data extraction policies

    Directory of Open Access Journals (Sweden)

    Provetti, Alessandro

    2008-12-01

    Full Text Available Web data extraction is concerned, among other things, with routine data accessing and downloading from continuously-updated dynamic Web pages. There is a relevant trade-off between the rate at which the external Web sites are accessed and the computational burden on the accessing client. We address the problem by proposing a predictive model, typical of the Operating Systems literature, of the rate-of-update of each Web source. The presented model has been implemented into a new version of the Dynamo project: a middleware that assists in generating informative RSS feeds out of traditional HTML Web sites. To be effective, i.e., make RSS feeds be timely and informative and to be scalable, Dynamo needs a careful tuning and customization of its polling policies, which are described in detail.

  20. ONTOGRABBING: Extracting Information from Texts Using Generative Ontologies

    DEFF Research Database (Denmark)

    Nilsson, Jørgen Fischer; Szymczak, Bartlomiej Antoni; Jensen, P.A.

    2009-01-01

    We describe principles for extracting information from texts using a so-called generative ontology in combination with syntactic analysis. Generative ontologies are introduced as semantic domains for natural language phrases. Generative ontologies extend ordinary finite ontologies with rules...... for producing recursively shaped terms representing the ontological content (ontological semantics) of NL noun phrases and other phrases. We focus here on achieving a robust, often only partial, ontology-driven parsing of and ascription of semantics to a sentence in the text corpus. The aim of the ontological...... analysis is primarily to identify paraphrases, thereby achieving a search functionality beyond mere keyword search with synsets. We further envisage use of the generative ontology as a phrase-based rather than word-based browser into text corpora....

  1. MedEx: a medication information extraction system for clinical narratives

    Science.gov (United States)

    Stenner, Shane P; Doan, Son; Johnson, Kevin B; Waitman, Lemuel R; Denny, Joshua C

    2010-01-01

    Medication information is one of the most important types of clinical data in electronic medical records. It is critical for healthcare safety and quality, as well as for clinical research that uses electronic medical record data. However, medication data are often recorded in clinical notes as free-text. As such, they are not accessible to other computerized applications that rely on coded data. We describe a new natural language processing system (MedEx), which extracts medication information from clinical notes. MedEx was initially developed using discharge summaries. An evaluation using a data set of 50 discharge summaries showed it performed well on identifying not only drug names (F-measure 93.2%), but also signature information, such as strength, route, and frequency, with F-measures of 94.5%, 93.9%, and 96.0% respectively. We then applied MedEx unchanged to outpatient clinic visit notes. It performed similarly with F-measures over 90% on a set of 25 clinic visit notes. PMID:20064797

  2. Methods to extract information on the atomic and molecular states from scientific abstracts

    International Nuclear Information System (INIS)

    Sasaki, Akira; Ueshima, Yutaka; Yamagiwa, Mitsuru; Murata, Masaki; Kanamaru, Toshiyuki; Shirado, Tamotsu; Isahara, Hitoshi

    2005-01-01

    We propose a new application of information technology to recognize and extract expressions of atomic and molecular states from electrical forms of scientific abstracts. Present results will help scientists to understand atomic states as well as the physics discussed in the articles. Combining with the internet search engines, it will make one possible to collect not only atomic and molecular data but broader scientific information over a wide range of research fields. (author)

  3. Aggregation of dialkyl-substituted diphosphonic acids and its effect on metal ion extraction.

    Energy Technology Data Exchange (ETDEWEB)

    Chiarizia, R.; Barrans, R. E., Jr.; Ferraro, J. R. Herlinger, A. W.; McAlister, D. R.

    1999-10-22

    Solvent extraction reagents containing the diphosphonic acid group exhibit an extraordinary affinity for tri-, tetra- and hexavalent actinides. Their use has been considered for actinide separation and pre-concentration procedures. Solvent extraction data obtained with P,P{prime}-di(2-ethylhexyl) methane-, ethane- and butanediphosphonic acids exhibit features that are difficult to explain without Knowledge of the aggregation state of the extractants. Information about the aggregation of the dialkyl-substituted diphosphonic acids in aromatic diluents has been obtained using the complementary techniques of vapor pressure osmometry (VPO), small angle neutron scattering (SANS), infrared spectroscopy and molecular mechanics. The results from these techniques provide an understanding of the aggregation behavior of these extractants that is fully compatible with the solvent extraction data. The most important results and their relevance to solvent extraction are reviewed in this paper.

  4. Post-encoding control of working memory enhances processing of relevant information in rhesus monkeys (Macaca mulatta).

    Science.gov (United States)

    Brady, Ryan J; Hampton, Robert R

    2018-06-01

    Working memory is a system by which a limited amount of information can be kept available for processing after the cessation of sensory input. Because working memory resources are limited, it is adaptive to focus processing on the most relevant information. We used a retro-cue paradigm to determine the extent to which monkey working memory possesses control mechanisms that focus processing on the most relevant representations. Monkeys saw a sample array of images, and shortly after the array disappeared, they were visually cued to a location that had been occupied by one of the sample images. The cue indicated which image should be remembered for the upcoming recognition test. By determining whether the monkeys were more accurate and quicker to respond to cued images compared to un-cued images, we tested the hypothesis that monkey working memory focuses processing on relevant information. We found a memory benefit for the cued image in terms of accuracy and retrieval speed with a memory load of two images. With a memory load of three images, we found a benefit in retrieval speed but only after shortening the onset latency of the retro-cue. Our results demonstrate previously unknown flexibility in the cognitive control of memory in monkeys, suggesting that control mechanisms in working memory likely evolved in a common ancestor of humans and monkeys more than 32 million years ago. Future work should be aimed at understanding the interaction between memory load and the ability to control memory resources, and the role of working memory control in generating differences in cognitive capacity among primates. Copyright © 2018 Elsevier B.V. All rights reserved.

  5. Analysis Methods for Extracting Knowledge from Large-Scale WiFi Monitoring to Inform Building Facility Planning

    DEFF Research Database (Denmark)

    Ruiz-Ruiz, Antonio; Blunck, Henrik; Prentow, Thor Siiger

    2014-01-01

    realistic data to inform facility planning. In this paper, we propose analysis methods to extract knowledge from large sets of network collected WiFi traces to better inform facility management and planning in large building complexes. The analysis methods, which build on a rich set of temporal and spatial......The optimization of logistics in large building com- plexes with many resources, such as hospitals, require realistic facility management and planning. Current planning practices rely foremost on manual observations or coarse unverified as- sumptions and therefore do not properly scale or provide....... Spatio-temporal visualization tools built on top of these methods enable planners to inspect and explore extracted information to inform facility-planning activities. To evaluate the methods, we present results for a large hospital complex covering more than 10 hectares. The evaluation is based on Wi...

  6. Exploring the information and communication technology competence and confidence of nursing students and their perception of its relevance to clinical practice.

    Science.gov (United States)

    Levett-Jones, Tracy; Kenny, Raelene; Van der Riet, Pamela; Hazelton, Michael; Kable, Ashley; Bourgeois, Sharon; Luxford, Yoni

    2009-08-01

    This paper profiles a study that explored nursing students' information and communication technology competence and confidence. It presents selected findings that focus on students' attitudes towards information and communication technology as an educational methodology and their perceptions of its relevance to clinical practice. Information and communication technology is integral to contemporary nursing practice. Development of these skills is important to ensure that graduates are 'work ready' and adequately prepared to practice in increasingly technological healthcare environments. This was a mixed methods study. Students (n=971) from three Australian universities were surveyed using an instrument designed specifically for the study, and 24 students participated in focus groups. The focus group data revealed that a number of students were resistant to the use of information and communication technology as an educational methodology and lacked the requisite skills and confidence to engage successfully with this educational approach. Survey results indicated that 26 per cent of students were unsure about the relevance of information and communication technology to clinical practice and only 50 per cent felt 'very confident' using a computer. While the importance of information and communication technology to student's learning and to their preparedness for practice has been established, it is evident that students' motivation is influenced by their level of confidence and competence, and their understanding of the relevance of information and communication technology to their future careers.

  7. The relevance of the International Classification of Functioning, Disability and Health (ICF) in monitoring and evaluating Community-based Rehabilitation (CBR).

    Science.gov (United States)

    Madden, Rosamond H; Dune, Tinashe; Lukersmith, Sue; Hartley, Sally; Kuipers, Pim; Gargett, Alexandra; Llewellyn, Gwynnyth

    2014-01-01

    To examine the relevance of the International Classification of Functioning, Disability and Health (ICF) to CBR monitoring and evaluation by investigating the relationship between the ICF and information in published CBR monitoring and evaluation reports. A three-stage literature search and analysis method was employed. Studies were identified via online database searches for peer-reviewed journal articles, and hand-searching of CBR network resources, NGO websites and specific journals. From each study "information items" were extracted; extraction consistency among authors was established. Finally, the resulting information items were coded to ICF domains and categories, with consensus on coding being achieved. Thirty-six articles relating to monitoring and evaluating CBR were selected for analysis. Approximately one third of the 2495 information items identified in these articles (788 or 32%) related to concepts of functioning, disability and environment, and could be coded to the ICF. These information items were spread across the entire ICF classification with a concentration on Activities and Participation (49% of the 788 information items) and Environmental Factors (42%). The ICF is a relevant and potentially useful framework and classification, providing building blocks for the systematic recording of information pertaining to functioning and disability, for CBR monitoring and evaluation. Implications for Rehabilitation The application of the ICF, as one of the building blocks for CBR monitoring and evaluation, is a constructive step towards an evidence-base on the efficacy and outcomes of CBR programs. The ICF can be used to provide the infrastructure for functioning and disability information to inform service practitioners and enable national and international comparisons.

  8. Online Capacity Estimation of Lithium-Ion Batteries Based on Novel Feature Extraction and Adaptive Multi-Kernel Relevance Vector Machine

    Directory of Open Access Journals (Sweden)

    Yang Zhang

    2015-11-01

    Full Text Available Prognostics is necessary to ensure the reliability and safety of lithium-ion batteries for hybrid electric vehicles or satellites. This process can be achieved by capacity estimation, which is a direct fading indicator for assessing the state of health of a battery. However, the capacity of a lithium-ion battery onboard is difficult to monitor. This paper presents a data-driven approach for online capacity estimation. First, six novel features are extracted from cyclic charge/discharge cycles and used as indirect health indicators. An adaptive multi-kernel relevance machine (MKRVM based on accelerated particle swarm optimization algorithm is used to determine the optimal parameters of MKRVM and characterize the relationship between extracted features and battery capacity. The overall estimation process comprises offline and online stages. A supervised learning step in the offline stage is established for model verification to ensure the generalizability of MKRVM for online application. Cross-validation is further conducted to validate the performance of the proposed model. Experiment and comparison results show the effectiveness, accuracy, efficiency, and robustness of the proposed approach for online capacity estimation of lithium-ion batteries.

  9. Inducing task-relevant responses to speech in the sleeping brain.

    Science.gov (United States)

    Kouider, Sid; Andrillon, Thomas; Barbosa, Leonardo S; Goupil, Louise; Bekinschtein, Tristan A

    2014-09-22

    Falling asleep leads to a loss of sensory awareness and to the inability to interact with the environment [1]. While this was traditionally thought as a consequence of the brain shutting down to external inputs, it is now acknowledged that incoming stimuli can still be processed, at least to some extent, during sleep [2]. For instance, sleeping participants can create novel sensory associations between tones and odors [3] or reactivate existing semantic associations, as evidenced by event-related potentials [4-7]. Yet, the extent to which the brain continues to process external stimuli remains largely unknown. In particular, it remains unclear whether sensory information can be processed in a flexible and task-dependent manner by the sleeping brain, all the way up to the preparation of relevant actions. Here, using semantic categorization and lexical decision tasks, we studied task-relevant responses triggered by spoken stimuli in the sleeping brain. Awake participants classified words as either animals or objects (experiment 1) or as either words or pseudowords (experiment 2) by pressing a button with their right or left hand, while transitioning toward sleep. The lateralized readiness potential (LRP), an electrophysiological index of response preparation, revealed that task-specific preparatory responses are preserved during sleep. These findings demonstrate that despite the absence of awareness and behavioral responsiveness, sleepers can still extract task-relevant information from external stimuli and covertly prepare for appropriate motor responses. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.

  10. The foundation of the concept of relevance

    DEFF Research Database (Denmark)

    Hjørland, Birger

    2010-01-01

    that what was regarded as the most fundamental view by Saracevic in 1975 has not since been considered (with very few exceptions). Other views, which are based on less fruitful assumptions, have dominated the discourse on relevance in information retrieval and information science. Many authors have...... reexamined the concept of relevance in information science, but have neglected the subject knowledge view, hence basic theoretical assumptions seem not to have been properly addressed. It is as urgent now as it was in 1975 seriously to consider “the subject knowledge view” of relevance (which may also...... be termed “the epistemological view”). The concept of relevance, like other basic concepts, is influenced by overall approaches to information science, such as the cognitive view and the domain-analytic view. There is today a trend toward a social paradigm for information science. This paper offers...

  11. Domain-independent information extraction in unstructured text

    Energy Technology Data Exchange (ETDEWEB)

    Irwin, N.H. [Sandia National Labs., Albuquerque, NM (United States). Software Surety Dept.

    1996-09-01

    Extracting information from unstructured text has become an important research area in recent years due to the large amount of text now electronically available. This status report describes the findings and work done during the second year of a two-year Laboratory Directed Research and Development Project. Building on the first-year`s work of identifying important entities, this report details techniques used to group words into semantic categories and to output templates containing selective document content. Using word profiles and category clustering derived during a training run, the time-consuming knowledge-building task can be avoided. Though the output still lacks in completeness when compared to systems with domain-specific knowledge bases, the results do look promising. The two approaches are compatible and could complement each other within the same system. Domain-independent approaches retain appeal as a system that adapts and learns will soon outpace a system with any amount of a priori knowledge.

  12. Breast cancer and quality of life: medical information extraction from health forums.

    Science.gov (United States)

    Opitz, Thomas; Aze, Jérome; Bringay, Sandra; Joutard, Cyrille; Lavergne, Christian; Mollevi, Caroline

    2014-01-01

    Internet health forums are a rich textual resource with content generated through free exchanges among patients and, in certain cases, health professionals. We tackle the problem of retrieving clinically relevant information from such forums, with relevant topics being defined from clinical auto-questionnaires. Texts in forums are largely unstructured and noisy, calling for adapted preprocessing and query methods. We minimize the number of false negatives in queries by using a synonym tool to achieve query expansion of initial topic keywords. To avoid false positives, we propose a new measure based on a statistical comparison of frequent co-occurrences in a large reference corpus (Web) to keep only relevant expansions. Our work is motivated by a study of breast cancer patients' health-related quality of life (QoL). We consider topics defined from a breast-cancer specific QoL-questionnaire. We quantify and structure occurrences in posts of a specialized French forum and outline important future developments.

  13. Direct toxic effects of aqueous extract of cigarette smoke on cardiac myocytes at clinically relevant concentrations

    International Nuclear Information System (INIS)

    Yamada, Shigeyuki; Zhang Xiuquan; Kadono, Toshie; Matsuoka, Nobuhiro; Rollins, Douglas; Badger, Troy; Rodesch, Christopher K.; Barry, William H.

    2009-01-01

    Aims: Our goal was to determine if clinically relevant concentrations of aqueous extract of cigarette smoke (CSE) have direct deleterious effects on ventricular myocytes during simulated ischemia, and to investigate the mechanisms involved. Methods: CSE was prepared with a smoking chamber. Ischemia was simulated by metabolic inhibition (MI) with cyanide (CN) and 0 glucose. Adult rabbit and mouse ventricular myocyte [Ca 2+ ] i was measured by flow cytometry using fluo-3. Mitochondrial [Ca 2+ ] was measured with confocal microscopy, and Rhod-2 fluorescence. The mitochondrial permeability transition (MPT) was detected by TMRM fluorescence and myocyte contracture. Myocyte oxidative stress was quantified by dichlorofluorescein (DCF) fluorescence with confocal microscopy. Results: CSE 0.1% increased myocyte contracture caused by MI. The nicotine concentration (HPLC) in 0.1% CSE was 15 ng/ml, similar to that in humans after smoking cigarettes. CSE 0.1% increased mitochondrial Ca 2+ uptake, and increased the susceptibility of mitochondria to the MPT. CSE 0.1% increased DCF fluorescence in isolated myocytes, and increased [Ca 2+ ] i in paced myocytes exposed to 2.0 mM CN, 0 glucose (P-MI). These effects were inhibited by the superoxide scavenger Tiron. The effect of CSE on [Ca 2+ ] i during P-MI was also prevented by ranolazine. Conclusions: CSE in clinically relevant concentrations increases myocyte [Ca 2+ ] i during simulated ischemia, and increases myocyte susceptibility to the MPT. These effects appear to be mediated at least in part by oxidative radicals in CSE, and likely contribute to the effects of cigarette smoke to increase myocardial infarct size, and to decrease angina threshold

  14. Relevant Factors in The Post-Merger Systems Integration and Information Technology in Brazilian Banks

    Directory of Open Access Journals (Sweden)

    Marcel Ginotti Pires

    2017-01-01

    Full Text Available This article discusses the factors present in post-merger integration of Systems and Information Technology (SIT that lead to positive and negative results in mergers and acquisitions (M & A. The research comprised three of the largest acquiring banks in Brazil. We adopted two methods of research, qualitative, to operationalize the theoretical concepts and quantitative, to test the hypotheses. We interviewed six executives of banks that held relevant experience in M & A processes. Subsequently, we applied questionnaires to IT professionals who were involved in the SIT integration processes. The results showed that the quality and expertise of the integration teams and managing the integration were the most relevant factors in the processes, with positive results for increased efficiency and the increased capacity of SIT. Negative results were due to failures in exploiting learning opportunities, the loss of employees and the inexpressive record of integration procedures.

  15. Elaboration of a guide including relevant project and logistic information: a case study

    Energy Technology Data Exchange (ETDEWEB)

    Costa, Tchaikowisky M. [Faculdade de Tecnologia e Ciencias (FTC), Itabuna, BA (Brazil); Bresci, Claudio T.; Franca, Carlos M.M. [PETROBRAS, Rio de Janeiro, RJ (Brazil)

    2009-07-01

    For every mobilization of a new enterprise it is necessary to quickly obtain the greatest amount of relative information in regards to location and availability of infra-structure, logistics, and work site amenities. Among this information are reports elaborated for management of the enterprise, (organizational chart, work schedule, objectives, contacts, etc.) as well as geographic anomalies, social-economic and culture of the area to be developed such as territorial extension, land aspects, local population, roads and amenities (fuel stations ,restaurants and hotels), infra-structure of the cities (health, education, entertainment, housing, transport, etc.) and logistically the distance between cities the estimated travel time, ROW access maps and notable points, among other relevant information. With the idea of making this information available for everyone involved in the enterprise, it was elaborated for GASCAC Spread 2A a rapid guide containing all the information mentioned above and made it available for all the vehicles used to transport employees and visitors to the spread. With this, everyone quickly received the majority of information necessary in one place, in a practical, quick, and precise manner, since the information is always used and controlled by the same person. This study includes the model used in the gas pipeline GASCAC Spread 2A project and the methodology used to draft and update the information. Besides the above, a file in the GIS format was prepared containing all necessary planning, execution and tracking information for enterprise activities, from social communication to the execution of the works previously mentioned. Part of the GIS file information was uploaded to Google Earth so as to disclose the information to a greater group of people, bearing in mind that this program is free of charge and easy to use. (author)

  16. Extracting breathing rate information from a wearable reflectance pulse oximeter sensor.

    Science.gov (United States)

    Johnston, W S; Mendelson, Y

    2004-01-01

    The integration of multiple vital physiological measurements could help combat medics and field commanders to better predict a soldier's health condition and enhance their ability to perform remote triage procedures. In this paper we demonstrate the feasibility of extracting accurate breathing rate information from a photoplethysmographic signal that was recorded by a reflectance pulse oximeter sensor mounted on the forehead and subsequently processed by a simple time domain filtering and frequency domain Fourier analysis.

  17. An hybrid method for the Arabic queries disambiguation to improve the relevance calculation in the IRS

    OpenAIRE

    Adil ENAANAI, Aziz SDIGUI DOUKKALI, El habib BENLAHMER

    2012-01-01

    In the information systems, the query’s expansion brings more benefices in the relevant documents extraction. However, the current expansion types are focused on the retrieve of the maximum of documents (reduce the silence). In Arabic, the queries are derived in many morphosemantical variants. Hence the diversity of the semantic interpretations that often creates a problem of ambiguity. Our objective is to prepare the Arabic request before its introduction to the document retrieval system. Th...

  18. Self-referential and anxiety-relevant information processing in subclinical social anxiety: an fMRI study.

    Science.gov (United States)

    Abraham, Anna; Kaufmann, Carolin; Redlich, Ronny; Hermann, Andrea; Stark, Rudolf; Stevens, Stephan; Hermann, Christiane

    2013-03-01

    The fear of negative evaluation is one of the hallmark features of social anxiety. Behavioral evidence thus far largely supports cognitive models which postulate that information processing biases in the face of socially relevant information are a key factor underlying this widespread phobia. So far only one neuroimaging study has explicitly focused on the fear of negative evaluation in social anxiety where the brain responses of social phobics were compared to healthy participants during the processing of self-referential relative to other-referential criticism, praise or neutral information. Only self-referential criticism led to stronger activations in emotion-relevant regions of the brain, such as the amygdala and medial prefrontal cortices (mPFC), in the social phobics. The objective of the current study was to determine whether these findings could be extended to subclinical social anxiety. In doing so, the specificity of this self-referential bias was also examined by including both social and non-social (physical illness-related) threat information as well as a highly health anxious control group in the experimental paradigm. The fMRI findings indicated that the processing of emotional stimuli was accompanied by activations in the amygdala and the ventral mPFC, while self-referential processing was associated with activity in regions such as the mPFC, posterior cingulate and temporal poles. Despite the validation of the paradigm, the results revealed that the previously reported behavioral and brain biases associated with social phobia could not be unequivocally extended to subclinical social anxiety. The divergence between the findings is explored in detail with reference to paradigm differences and conceptual issues.

  19. Significance of wood extractives for wood bonding.

    Science.gov (United States)

    Roffael, Edmone

    2016-02-01

    Wood contains primary extractives, which are present in all woods, and secondary extractives, which are confined in certain wood species. Extractives in wood play a major role in wood-bonding processes, as they can contribute to or determine the bonding relevant properties of wood such as acidity and wettability. Therefore, extractives play an immanent role in bonding of wood chips and wood fibres with common synthetic adhesives such as urea-formaldehyde-resins (UF-resins) and phenol-formaldehyde-resins (PF-resins). Extractives of high acidity accelerate the curing of acid curing UF-resins and decelerate bonding with alkaline hardening PF-resins. Water-soluble extractives like free sugars are detrimental for bonding of wood with cement. Polyphenolic extractives (tannins) can be used as a binder in the wood-based industry. Additionally, extractives in wood can react with formaldehyde and reduce the formaldehyde emission of wood-based panels. Moreover, some wood extractives are volatile organic compounds (VOC) and insofar also relevant to the emission of VOC from wood and wood-based panels.

  20. Study of time-frequency characteristics of single snores: extracting new information for sleep apnea diagnosis

    Energy Technology Data Exchange (ETDEWEB)

    Castillo Escario, Y.; Blanco Almazan, D.; Camara Vazquez, M.A.; Jane Campos, R.

    2016-07-01

    Obstructive sleep apnea (OSA) is a highly prevalent chronic disease, especially in elderly and obese population. Despite constituting a huge health and economic problem, most patients remain undiagnosed due to limitations in current strategies. Therefore, it is essential to find cost-effective diagnostic alternatives. One of these novel approaches is the analysis of acoustic snoring signals. Snoring is an early symptom of OSA which carries pathophysiological information of high diagnostic value. For this reason, the main objective of this work is to study the characteristics of single snores of different types, from healthy and OSA subjects. To do that, we analyzed snoring signals from previous databases and developed an experimental protocol to record simulated OSA-related sounds and characterize the response of two commercial tracheal microphones. Automatic programs for filtering, downsampling, event detection and time-frequency analysis were built in MATLAB. We found that time-frequency maps and spectral parameters (central, mean and peak frequency and energy in the 100-500 Hz band) allow distinguishing regular snores of healthy subjects from non-regular snores and snores of OSA subjects. Regarding the two commercial microphones, we found that one of them was a suitable snoring sensor, while the other had a too restricted frequency response. Future work shall include a higher number of episodes and subjects, but our study has contributed to show how important the differences between regular and non-regular snores can be for OSA diagnosis, and how much clinically relevant information can be extracted from time-frequency maps and spectral parameters of single snores. (Author)

  1. Self-relevant beauty evaluation: Evidence from an event-related potentials study.

    Science.gov (United States)

    Kong, Fanchang; Zhang, Yan; Tian, Yuan; Fan, Cuiying; Zhou, Zongkui

    2015-03-01

    This study examines the electrophysiological correlates of beauty evaluation when participants performed the self-reference task. About 13 (7 men, 6 women) undergraduates participated in the experiment using event-related potentials. Results showed that the response to self-relevant information was faster compared to other-relevant information and no significant differences for self-relevant relative to mother-relevant information were observed. Both physical and interior beauty words for self-relevant information showed an enhanced late positive component as compared to other-relevant information. Physical beauty for self-relevant information yielded a larger late positive component in contrast to mother-relevant information but not for interior beauty. This study indicates that beauty is specific to the person who judges it though an individual and one's mother may hold similar views of interior beauty.

  2. Systematically extracting metal- and solvent-related occupational information from free-text responses to lifetime occupational history questionnaires.

    Science.gov (United States)

    Friesen, Melissa C; Locke, Sarah J; Tornow, Carina; Chen, Yu-Cheng; Koh, Dong-Hee; Stewart, Patricia A; Purdue, Mark; Colt, Joanne S

    2014-06-01

    Lifetime occupational history (OH) questionnaires often use open-ended questions to capture detailed information about study participants' jobs. Exposure assessors use this information, along with responses to job- and industry-specific questionnaires, to assign exposure estimates on a job-by-job basis. An alternative approach is to use information from the OH responses and the job- and industry-specific questionnaires to develop programmable decision rules for assigning exposures. As a first step in this process, we developed a systematic approach to extract the free-text OH responses and convert them into standardized variables that represented exposure scenarios. Our study population comprised 2408 subjects, reporting 11991 jobs, from a case-control study of renal cell carcinoma. Each subject completed a lifetime OH questionnaire that included verbatim responses, for each job, to open-ended questions including job title, main tasks and activities (task), tools and equipment used (tools), and chemicals and materials handled (chemicals). Based on a review of the literature, we identified exposure scenarios (occupations, industries, tasks/tools/chemicals) expected to involve possible exposure to chlorinated solvents, trichloroethylene (TCE) in particular, lead, and cadmium. We then used a SAS macro to review the information reported by study participants to identify jobs associated with each exposure scenario; this was done using previously coded standardized occupation and industry classification codes, and a priori lists of associated key words and phrases related to possibly exposed tasks, tools, and chemicals. Exposure variables representing the occupation, industry, and task/tool/chemicals exposure scenarios were added to the work history records of the study respondents. Our identification of possibly TCE-exposed scenarios in the OH responses was compared to an expert's independently assigned probability ratings to evaluate whether we missed identifying

  3. An analysis of contingent factors for the detection of strategic relevance in business information technologies

    Directory of Open Access Journals (Sweden)

    Antonio Paños Álvarez

    2005-01-01

    Full Text Available Information Technologies are resources able to create competitive advantages for companies. In this analysis, the Resource-based perspective have taken special relevance, because it is argued that this advantages should be identified, reached and maintained. This work is positioned in the analysis of several contingent factors in the process of pointing the possible assesment of these advantages. It is aproaching a portfolio for helping to select what Information Technologies are valuable for what companies and in what activity areas and the study of in what way the sector, the technological innovation profile, the size and the financial capacity of the companies affects this process

  4. The Feature Extraction Based on Texture Image Information for Emotion Sensing in Speech

    Directory of Open Access Journals (Sweden)

    Kun-Ching Wang

    2014-09-01

    Full Text Available In this paper, we present a novel texture image feature for Emotion Sensing in Speech (ESS. This idea is based on the fact that the texture images carry emotion-related information. The feature extraction is derived from time-frequency representation of spectrogram images. First, we transform the spectrogram as a recognizable image. Next, we use a cubic curve to enhance the image contrast. Then, the texture image information (TII derived from the spectrogram image can be extracted by using Laws’ masks to characterize emotional state. In order to evaluate the effectiveness of the proposed emotion recognition in different languages, we use two open emotional databases including the Berlin Emotional Speech Database (EMO-DB and eNTERFACE corpus and one self-recorded database (KHUSC-EmoDB, to evaluate the performance cross-corpora. The results of the proposed ESS system are presented using support vector machine (SVM as a classifier. Experimental results show that the proposed TII-based feature extraction inspired by visual perception can provide significant classification for ESS systems. The two-dimensional (2-D TII feature can provide the discrimination between different emotions in visual expressions except for the conveyance pitch and formant tracks. In addition, the de-noising in 2-D images can be more easily completed than de-noising in 1-D speech.

  5. A multiple relevance feedback strategy with positive and negative models.

    Directory of Open Access Journals (Sweden)

    Yunlong Ma

    Full Text Available A commonly used strategy to improve search accuracy is through feedback techniques. Most existing work on feedback relies on positive information, and has been extensively studied in information retrieval. However, when a query topic is difficult and the results from the first-pass retrieval are very poor, it is impossible to extract enough useful terms from a few positive documents. Therefore, the positive feedback strategy is incapable to improve retrieval in this situation. Contrarily, there is a relatively large number of negative documents in the top of the result list, and it has been confirmed that negative feedback strategy is an important and useful way for adapting this scenario by several recent studies. In this paper, we consider a scenario when the search results are so poor that there are at most three relevant documents in the top twenty documents. Then, we conduct a novel study of multiple strategies for relevance feedback using both positive and negative examples from the first-pass retrieval to improve retrieval accuracy for such difficult queries. Experimental results on these TREC collections show that the proposed language model based multiple model feedback method which is generally more effective than both the baseline method and the methods using only positive or negative model.

  6. Joint Feature Extraction and Classifier Design for ECG-Based Biometric Recognition.

    Science.gov (United States)

    Gutta, Sandeep; Cheng, Qi

    2016-03-01

    Traditional biometric recognition systems often utilize physiological traits such as fingerprint, face, iris, etc. Recent years have seen a growing interest in electrocardiogram (ECG)-based biometric recognition techniques, especially in the field of clinical medicine. In existing ECG-based biometric recognition methods, feature extraction and classifier design are usually performed separately. In this paper, a multitask learning approach is proposed, in which feature extraction and classifier design are carried out simultaneously. Weights are assigned to the features within the kernel of each task. We decompose the matrix consisting of all the feature weights into sparse and low-rank components. The sparse component determines the features that are relevant to identify each individual, and the low-rank component determines the common feature subspace that is relevant to identify all the subjects. A fast optimization algorithm is developed, which requires only the first-order information. The performance of the proposed approach is demonstrated through experiments using the MIT-BIH Normal Sinus Rhythm database.

  7. User perspectives on relevance criteria

    DEFF Research Database (Denmark)

    Maglaughlin, Kelly L.; Sonnenwald, Diane H.

    2002-01-01

    , partially relevant, or not relevant to their information need; and explained their decisions in an interview. Analysis revealed 29 criteria, discussed positively and negatively, that were used by the participants when selecting passages that contributed or detracted from a document's relevance......This study investigates the use of criteria to assess relevant, partially relevant, and not-relevant documents. Study participants identified passages within 20 document representations that they used to make relevance judgments; judged each document representation as a whole to be relevant...... matter, thought catalyst), full text (e.g., audience, novelty, type, possible content, utility), journal/publisher (e.g., novelty, main focus, perceived quality), and personal (e.g., competition, time requirements). Results further indicate that multiple criteria are used when making relevant, partially...

  8. Machine learning classification of surgical pathology reports and chunk recognition for information extraction noise reduction.

    Science.gov (United States)

    Napolitano, Giulio; Marshall, Adele; Hamilton, Peter; Gavin, Anna T

    2016-06-01

    Machine learning techniques for the text mining of cancer-related clinical documents have not been sufficiently explored. Here some techniques are presented for the pre-processing of free-text breast cancer pathology reports, with the aim of facilitating the extraction of information relevant to cancer staging. The first technique was implemented using the freely available software RapidMiner to classify the reports according to their general layout: 'semi-structured' and 'unstructured'. The second technique was developed using the open source language engineering framework GATE and aimed at the prediction of chunks of the report text containing information pertaining to the cancer morphology, the tumour size, its hormone receptor status and the number of positive nodes. The classifiers were trained and tested respectively on sets of 635 and 163 manually classified or annotated reports, from the Northern Ireland Cancer Registry. The best result of 99.4% accuracy - which included only one semi-structured report predicted as unstructured - was produced by the layout classifier with the k nearest algorithm, using the binary term occurrence word vector type with stopword filter and pruning. For chunk recognition, the best results were found using the PAUM algorithm with the same parameters for all cases, except for the prediction of chunks containing cancer morphology. For semi-structured reports the performance ranged from 0.97 to 0.94 and from 0.92 to 0.83 in precision and recall, while for unstructured reports performance ranged from 0.91 to 0.64 and from 0.68 to 0.41 in precision and recall. Poor results were found when the classifier was trained on semi-structured reports but tested on unstructured. These results show that it is possible and beneficial to predict the layout of reports and that the accuracy of prediction of which segments of a report may contain certain information is sensitive to the report layout and the type of information sought. Copyright

  9. YAdumper: extracting and translating large information volumes from relational databases to structured flat files.

    Science.gov (United States)

    Fernández, José M; Valencia, Alfonso

    2004-10-12

    Downloading the information stored in relational databases into XML and other flat formats is a common task in bioinformatics. This periodical dumping of information requires considerable CPU time, disk and memory resources. YAdumper has been developed as a purpose-specific tool to deal with the integral structured information download of relational databases. YAdumper is a Java application that organizes database extraction following an XML template based on an external Document Type Declaration. Compared with other non-native alternatives, YAdumper substantially reduces memory requirements and considerably improves writing performance.

  10. Bengali-English Relevant Cross Lingual Information Access Using Finite Automata

    Science.gov (United States)

    Banerjee, Avishek; Bhattacharyya, Swapan; Hazra, Simanta; Mondal, Shatabdi

    2010-10-01

    CLIR techniques searches unrestricted texts and typically extract term and relationships from bilingual electronic dictionaries or bilingual text collections and use them to translate query and/or document representations into a compatible set of representations with a common feature set. In this paper, we focus on dictionary-based approach by using a bilingual data dictionary with a combination to statistics-based methods to avoid the problem of ambiguity also the development of human computer interface aspects of NLP (Natural Language processing) is the approach of this paper. The intelligent web search with regional language like Bengali is depending upon two major aspect that is CLIA (Cross language information access) and NLP. In our previous work with IIT, KGP we already developed content based CLIA where content based searching in trained on Bengali Corpora with the help of Bengali data dictionary. Here we want to introduce intelligent search because to recognize the sense of meaning of a sentence and it has a better real life approach towards human computer interactions.

  11. From Specific Information Extraction to Inferences: A Hierarchical Framework of Graph Comprehension

    Science.gov (United States)

    2004-09-01

    The skill to interpret the information displayed in graphs is so important to have, the National Council of Teachers of Mathematics has created...guidelines to ensure that students learn these skills ( NCTM : Standards for Mathematics , 2003). These guidelines are based primarily on the extraction of...graphical perception. Human Computer Interaction, 8, 353-388. NCTM : Standards for Mathematics . (2003, 2003). Peebles, D., & Cheng, P. C.-H. (2002

  12. Computer-based information management system for interventional radiology

    International Nuclear Information System (INIS)

    Forman, B.H.; Silverman, S.G.; Mueller, P.R.; Hahn, P.F.; Papanicolaou, N.; Tung, G.A.; Brink, J.A.; Ferrucci, J.T.

    1989-01-01

    The authors authored and implemented a computer-based information management system (CBIMS) for the integrated analysis of data from a variety of abdominal nonvascular interventional procedures. The CBIMS improved on their initial handwritten-card system (which listed only patient name, hospital number, and type of procedure) by capturing relevant patient data in an organized fashion and integrating information for meaningful analysis. Advantages of CBIMS include enhanced compilation of monthly census, easy access to a patient's interventional history, and flexible querying capability that allows easy extraction of subsets of information from the patient database

  13. Does Guiding Toward Task-Relevant Information Help Improve Graph Processing and Graph Comprehension of Individuals with Low or High Numeracy? An Eye-Tracker Experiment.

    Science.gov (United States)

    Keller, Carmen; Junghans, Alex

    2017-11-01

    Individuals with low numeracy have difficulties with understanding complex graphs. Combining the information-processing approach to numeracy with graph comprehension and information-reduction theories, we examined whether high numerates' better comprehension might be explained by their closer attention to task-relevant graphical elements, from which they would expect numerical information to understand the graph. Furthermore, we investigated whether participants could be trained in improving their attention to task-relevant information and graph comprehension. In an eye-tracker experiment ( N = 110) involving a sample from the general population, we presented participants with 2 hypothetical scenarios (stomach cancer, leukemia) showing survival curves for 2 treatments. In the training condition, participants received written instructions on how to read the graph. In the control condition, participants received another text. We tracked participants' eye movements while they answered 9 knowledge questions. The sum constituted graph comprehension. We analyzed visual attention to task-relevant graphical elements by using relative fixation durations and relative fixation counts. The mediation analysis revealed a significant ( P attention to task-relevant information, which did not differ between the 2 conditions. Training had a significant main effect on visual attention ( P attention to task-relevant graphical elements than individuals with low numeracy. With appropriate instructions, both groups can be trained to improve their graph-processing efficiency. Future research should examine (e.g., motivational) mediators between visual attention and graph comprehension to develop appropriate instructions that also result in higher graph comprehension.

  14. Meteorological information in GPS-RO reflected signals

    Directory of Open Access Journals (Sweden)

    K. Boniface

    2011-07-01

    Full Text Available Vertical profiles of the atmosphere can be obtained globally with the radio-occultation technique. However, the lowest layers of the atmosphere are less accurately extracted. A good description of these layers is important for the good performance of Numerical Weather Prediction (NWP systems, and an improvement of the observational data available for the low troposphere would thus be of great interest for data assimilation. We outline here how supplemental meteorological information close to the surface can be extracted whenever reflected signals are available. We separate the reflected signal through a radioholographic filter, and we interpret it with a ray tracing procedure, analyzing the trajectories of the electromagnetic waves over a 3-D field of refractive index. A perturbation approach is then used to perform an inversion, identifying the relevant contribution of the lowest layers of the atmosphere to the properties of the reflected signal, and extracting some supplemental information to the solution of the inversion of the direct propagation signals. It is found that there is a significant amount of useful information in the reflected signal, which is sufficient to extract a stand-alone profile of the low atmosphere, with a precision of approximately 0.1 %. The methodology is applied to one reflection case.

  15. "What is relevant in a text document?": An interpretable machine learning approach.

    Directory of Open Access Journals (Sweden)

    Leila Arras

    Full Text Available Text documents can be described by a number of abstract concepts such as semantic category, writing style, or sentiment. Machine learning (ML models have been trained to automatically map documents to these abstract concepts, allowing to annotate very large text collections, more than could be processed by a human in a lifetime. Besides predicting the text's category very accurately, it is also highly desirable to understand how and why the categorization process takes place. In this paper, we demonstrate that such understanding can be achieved by tracing the classification decision back to individual words using layer-wise relevance propagation (LRP, a recently developed technique for explaining predictions of complex non-linear classifiers. We train two word-based ML models, a convolutional neural network (CNN and a bag-of-words SVM classifier, on a topic categorization task and adapt the LRP method to decompose the predictions of these models onto words. Resulting scores indicate how much individual words contribute to the overall classification decision. This enables one to distill relevant information from text documents without an explicit semantic information extraction step. We further use the word-wise relevance scores for generating novel vector-based document representations which capture semantic information. Based on these document vectors, we introduce a measure of model explanatory power and show that, although the SVM and CNN models perform similarly in terms of classification accuracy, the latter exhibits a higher level of explainability which makes it more comprehensible for humans and potentially more useful for other applications.

  16. Automated selection of relevant information for notification of incident cancer cases within a multisource cancer registry.

    Science.gov (United States)

    Jouhet, V; Defossez, G; Ingrand, P

    2013-01-01

    The aim of this study was to develop and evaluate a selection algorithm of relevant records for the notification of incident cases of cancer on the basis of the individual data available in a multi-source information system. This work was conducted on data for the year 2008 in the general cancer registry of Poitou-Charentes region (France). The selection algorithm hierarchizes information according to its level of relevance for tumoral topography and tumoral morphology independently. The selected data are combined to form composite records. These records are then grouped in respect with the notification rules of the International Agency for Research on Cancer for multiple primary cancers. The evaluation, based on recall, precision and F-measure confronted cases validated manually by the registry's physicians with tumours notified with and without records selection. The analysis involved 12,346 tumours validated among 11,971 individuals. The data used were hospital discharge data (104,474 records), pathology data (21,851 records), healthcare insurance data (7508 records) and cancer care centre's data (686 records). The selection algorithm permitted performances improvement for notification of tumour topography (F-measure 0.926 with vs. 0.857 without selection) and tumour morphology (F-measure 0.805 with vs. 0.750 without selection). These results show that selection of information according to its origin is efficient in reducing noise generated by imprecise coding. Further research is needed for solving the semantic problems relating to the integration of heterogeneous data and the use of non-structured information.

  17. The Criteria People Use in Relevance Decisions on Health Information: An Analysis of User Eye Movements When Browsing a Health Discussion Forum.

    Science.gov (United States)

    Pian, Wenjing; Khoo, Christopher Sg; Chang, Yun-Ke

    2016-06-20

    People are increasingly accessing health-related social media sites, such as health discussion forums, to post and read user-generated health information. It is important to know what criteria people use when deciding the relevance of information found on health social media websites, in different situations. The study attempted to identify the relevance criteria that people use when browsing a health discussion forum, in 3 types of use contexts: when seeking information for their own health issue, when seeking for other people's health issue, and when browsing without a particular health issue in mind. A total of 58 study participants were self-assigned to 1 of the 3 use contexts or information needs and were asked to browse a health discussion forum, HealthBoards.com. In the analysis, browsing a discussion forum was divided into 2 stages: scanning a set of post surrogates (mainly post titles) in the summary result screen and reading a detailed post content (including comments by other users). An eye tracker system was used to capture participants' eye movement behavior and the text they skim over and focus (ie, fixate) on during browsing. By analyzing the text that people's eyes fixated on, the types of health information used in the relevance judgment were determined. Post-experiment interviews elicited participants' comments on the relevance of the information and criteria used. It was found that participants seeking health information for their own health issue focused significantly more on the poster's symptoms, personal history of the disease, and description of the disease (P=.01, .001, and .02). Participants seeking for other people's health issue focused significantly more on cause of disease, disease terminology, and description of treatments and procedures (P=.01, .01, and .02). In contrast, participants browsing with no particular issue in mind focused significantly more on general health topics, hot topics, and rare health issues (P=.01, .01, and .01

  18. The Criteria People Use in Relevance Decisions on Health Information: An Analysis of User Eye Movements When Browsing a Health Discussion Forum

    Science.gov (United States)

    Khoo, Christopher SG; Chang, Yun-Ke

    2016-01-01

    Background People are increasingly accessing health-related social media sites, such as health discussion forums, to post and read user-generated health information. It is important to know what criteria people use when deciding the relevance of information found on health social media websites, in different situations. Objective The study attempted to identify the relevance criteria that people use when browsing a health discussion forum, in 3 types of use contexts: when seeking information for their own health issue, when seeking for other people’s health issue, and when browsing without a particular health issue in mind. Methods A total of 58 study participants were self-assigned to 1 of the 3 use contexts or information needs and were asked to browse a health discussion forum, HealthBoards.com. In the analysis, browsing a discussion forum was divided into 2 stages: scanning a set of post surrogates (mainly post titles) in the summary result screen and reading a detailed post content (including comments by other users). An eye tracker system was used to capture participants’ eye movement behavior and the text they skim over and focus (ie, fixate) on during browsing. By analyzing the text that people’s eyes fixated on, the types of health information used in the relevance judgment were determined. Post-experiment interviews elicited participants’ comments on the relevance of the information and criteria used. Results It was found that participants seeking health information for their own health issue focused significantly more on the poster’s symptoms, personal history of the disease, and description of the disease (P=.01, .001, and .02). Participants seeking for other people’s health issue focused significantly more on cause of disease, disease terminology, and description of treatments and procedures (P=.01, .01, and .02). In contrast, participants browsing with no particular issue in mind focused significantly more on general health topics, hot

  19. On-matrix derivatization extraction of chemical weapons convention relevant alcohols from soil.

    Science.gov (United States)

    Chinthakindi, Sridhar; Purohit, Ajay; Singh, Varoon; Dubey, D K; Pardasani, Deepak

    2013-10-11

    Present study deals with the on-matrix derivatization-extraction of aminoalcohols and thiodiglycols, which are important precursors and/or degradation products of VX analogues and vesicants class of chemical warfare agents (CWAs). The method involved hexamethyldisilazane (HMDS) mediated in situ silylation of analytes on the soil. Subsequent extraction and gas chromatography-mass spectrometry analysis of derivatized analytes offered better recoveries in comparison to the procedure recommended by the Organization for the Prohibition of Chemical Weapons (OPCW). Various experimental conditions such as extraction solvent, reagent and catalyst amount, reaction time and temperature were optimized. Best recoveries of analytes ranging from 45% to 103% were obtained with DCM solvent containing 5%, v/v HMDS and 0.01%, w/v iodine as catalyst. The limits of detection (LOD) and limit of quantification (LOQ) with selected analytes ranged from 8 to 277 and 21 to 665ngmL(-1), respectively, in selected ion monitoring mode. Copyright © 2013 Elsevier B.V. All rights reserved.

  20. Types of Lexicographical Information Needs and their Relevance for Information Science

    OpenAIRE

    Bergenholtz, Henning; Agerbo, Heidi

    2017-01-01

    In some situations, you need information in order to solve a problem that has occurred. In information science, user needs are often described through very specific examples rather than through a classification of situation types in which information needs occur. Furthermore, information science often describes general human needs, typically with a reference to Maslow's classification of needs (1954), instead of actual information needs. Lexicography has also focused on information needs, but...

  1. SPECTRa-T: machine-based data extraction and semantic searching of chemistry e-theses.

    Science.gov (United States)

    Downing, Jim; Harvey, Matt J; Morgan, Peter B; Murray-Rust, Peter; Rzepa, Henry S; Stewart, Diana C; Tonge, Alan P; Townsend, Joe A

    2010-02-22

    The SPECTRa-T project has developed text-mining tools to extract named chemical entities (NCEs), such as chemical names and terms, and chemical objects (COs), e.g., experimental spectral assignments and physical chemistry properties, from electronic theses (e-theses). Although NCEs were readily identified within the two major document formats studied, only the use of structured documents enabled identification of chemical objects and their association with the relevant chemical entity (e.g., systematic chemical name). A corpus of theses was analyzed and it is shown that a high degree of semantic information can be extracted from structured documents. This integrated information has been deposited in a persistent Resource Description Framework (RDF) triple-store that allows users to conduct semantic searches. The strength and weaknesses of several document formats are reviewed.

  2. Acute and chronic aquatic toxicity of aromatic extracts. Summary of relevant test data

    Energy Technology Data Exchange (ETDEWEB)

    Comber, M.I.H.; Den Haan, K.; Djemel, N.; Eadsforth, C.V.; King, D.; Parkerton, T.; Leon Paumen, M.; Dmytrasz, B.; Del Castillo, F.

    2013-09-15

    This report describes the experimental procedures and the results obtained in acute and chronic ecotoxicity tests on several aromatic extracts samples. The samples were tested for toxicity to the rainbow trout (Oncorhynchus mykiss), the crustacean zooplankter, Daphnia magna and the algae, Selenastrum capricornutum using water accommodated fractions. These results assist in determining the environmental hazard posed by aromatic extracts.

  3. Acquisition of data for plasma simulation by automated extraction of terminology from article abstracts

    International Nuclear Information System (INIS)

    Pichl, L.; Suzuki, Manabu; Murata, Masaki; Sasaki, Akira; Kato, Daiji; Murakami, Izumi; Rhee, Yongjoo

    2007-01-01

    Computer simulation of burning plasmas as well as computational plasma modeling in image processing requires a number of accurate data, in addition to a relevant model framework. To this aim, it is very important to recognize, obtain and evaluate data relevant for such a simulation from the literature. This work focuses on the simultaneous search of relevant data across various online databases, extraction of cataloguing and numerical information, and automatic recognition of specific terminology in the text retrieved. The concept is illustrated on the particular terminology of Atomic and Molecular data relevant to edge plasma simulation. The IAEA search engine GENIE and the NIFS search engine Joint Search 2 are compared and discussed. Accurate modeling of the imaged object is considered to be the ultimate challenge in improving the resolution limits of plasma imaging. (author)

  4. 5W1H Information Extraction with CNN-Bidirectional LSTM

    Science.gov (United States)

    Nurdin, A.; Maulidevi, N. U.

    2018-03-01

    In this work, information about who, did what, when, where, why, and how on Indonesian news articles were extracted by combining Convolutional Neural Network and Bidirectional Long Short-Term Memory. Convolutional Neural Network can learn semantically meaningful representations of sentences. Bidirectional LSTM can analyze the relations among words in the sequence. We also use word embedding word2vec for word representation. By combining these algorithms, we obtained F-measure 0.808. Our experiments show that CNN-BLSTM outperforms other shallow methods, namely IBk, C4.5, and Naïve Bayes with the F-measure 0.655, 0.645, and 0.595, respectively.

  5. Comparing deep learning and concept extraction based methods for patient phenotyping from clinical narratives.

    Science.gov (United States)

    Gehrmann, Sebastian; Dernoncourt, Franck; Li, Yeran; Carlson, Eric T; Wu, Joy T; Welt, Jonathan; Foote, John; Moseley, Edward T; Grant, David W; Tyler, Patrick D; Celi, Leo A

    2018-01-01

    In secondary analysis of electronic health records, a crucial task consists in correctly identifying the patient cohort under investigation. In many cases, the most valuable and relevant information for an accurate classification of medical conditions exist only in clinical narratives. Therefore, it is necessary to use natural language processing (NLP) techniques to extract and evaluate these narratives. The most commonly used approach to this problem relies on extracting a number of clinician-defined medical concepts from text and using machine learning techniques to identify whether a particular patient has a certain condition. However, recent advances in deep learning and NLP enable models to learn a rich representation of (medical) language. Convolutional neural networks (CNN) for text classification can augment the existing techniques by leveraging the representation of language to learn which phrases in a text are relevant for a given medical condition. In this work, we compare concept extraction based methods with CNNs and other commonly used models in NLP in ten phenotyping tasks using 1,610 discharge summaries from the MIMIC-III database. We show that CNNs outperform concept extraction based methods in almost all of the tasks, with an improvement in F1-score of up to 26 and up to 7 percentage points in area under the ROC curve (AUC). We additionally assess the interpretability of both approaches by presenting and evaluating methods that calculate and extract the most salient phrases for a prediction. The results indicate that CNNs are a valid alternative to existing approaches in patient phenotyping and cohort identification, and should be further investigated. Moreover, the deep learning approach presented in this paper can be used to assist clinicians during chart review or support the extraction of billing codes from text by identifying and highlighting relevant phrases for various medical conditions.

  6. Respiratory Information Extraction from Electrocardiogram Signals

    KAUST Repository

    Amin, Gamal El Din Fathy

    2010-12-01

    The Electrocardiogram (ECG) is a tool measuring the electrical activity of the heart, and it is extensively used for diagnosis and monitoring of heart diseases. The ECG signal reflects not only the heart activity but also many other physiological processes. The respiratory activity is a prominent process that affects the ECG signal due to the close proximity of the heart and the lungs. In this thesis, several methods for the extraction of respiratory process information from the ECG signal are presented. These methods allow an estimation of the lung volume and the lung pressure from the ECG signal. The potential benefit of this is to eliminate the corresponding sensors used to measure the respiration activity. A reduction of the number of sensors connected to patients will increase patients’ comfort and reduce the costs associated with healthcare. As a further result, the efficiency of diagnosing respirational disorders will increase since the respiration activity can be monitored with a common, widely available method. The developed methods can also improve the detection of respirational disorders that occur while patients are sleeping. Such disorders are commonly diagnosed in sleeping laboratories where the patients are connected to a number of different sensors. Any reduction of these sensors will result in a more natural sleeping environment for the patients and hence a higher sensitivity of the diagnosis.

  7. Protection of safety-relevant information in distributed energy information systems; Schutz sicherheitsrelevanter Informationen in verteilten Energieinformationssystemen

    Energy Technology Data Exchange (ETDEWEB)

    Beenken, Petra

    2010-07-01

    Within the last years there has been an ongoing change in the energy domain. The German renewable energies law EnWG requires a liberalization that leads to a strict separation of domains such as transportation, supply and conversion of energy. Furthermore, climate and environmental protection as well as cost transparency and energy saving in combination with efficiency of resources leads to new challenges for the energy industry. The so called smart grid vision and the concluding design of an ICT-based information structure for the energy domain will help to reach these goals by integrating renewable energy resources, saving fuels and getting a higher energy efficiency. In order to reach these goals, information about current energy generation, energy storage and energy demand is required. Through an efficient network and fast information exchange by means of an energy information network an efficient energy use can be gained. The federated networking of an energy information network like this can tend to a weakness for cyber security within the energy domain. The growing number of people involved and data exchanges will create more potential points of attacks than before. Therefore, a suitable protection of an energy information network is necessary. Through paragraph 9 EnWG the protection goal confidentiality is particularly important. But the implementation of confidentiality must not lead to a violation of availability requirements, which are very important at some point of the energy domain. Additionally to the identification of such crucial side effects, the implementation of confidentiality for distributed, decentral systems is a challenge for the domain. The ENERTRUST security model includes a knowledge base construction, which allows the identification of such side effects or conflicts in the energy domain by applying reasoning techniques. Moreover, it allows the realization of confidentiality from distributed locations through a use and combination of

  8. Information extraction and knowledge graph construction from geoscience literature

    Science.gov (United States)

    Wang, Chengbin; Ma, Xiaogang; Chen, Jianguo; Chen, Jingwen

    2018-03-01

    Geoscience literature published online is an important part of open data, and brings both challenges and opportunities for data analysis. Compared with studies of numerical geoscience data, there are limited works on information extraction and knowledge discovery from textual geoscience data. This paper presents a workflow and a few empirical case studies for that topic, with a focus on documents written in Chinese. First, we set up a hybrid corpus combining the generic and geology terms from geology dictionaries to train Chinese word segmentation rules of the Conditional Random Fields model. Second, we used the word segmentation rules to parse documents into individual words, and removed the stop-words from the segmentation results to get a corpus constituted of content-words. Third, we used a statistical method to analyze the semantic links between content-words, and we selected the chord and bigram graphs to visualize the content-words and their links as nodes and edges in a knowledge graph, respectively. The resulting graph presents a clear overview of key information in an unstructured document. This study proves the usefulness of the designed workflow, and shows the potential of leveraging natural language processing and knowledge graph technologies for geoscience.

  9. Advancing Affect Modeling via Preference Learning and Unsupervised Feature Extraction

    DEFF Research Database (Denmark)

    Martínez, Héctor Pérez

    strategies (error functions and training algorithms) for artificial neural networks are examined across synthetic and psycho-physiological datasets, and compared against support vector machines and Cohen’s method. Results reveal the best training strategies for neural networks and suggest their superiority...... difficulties, ordinal reports such as rankings and ratings can yield more reliable affect annotations than alternative tools. This thesis explores preference learning methods to automatically learn computational models from ordinal annotations of affect. In particular, an extensive collection of training...... over the other examined methods. The second challenge addressed in this thesis refers to the extraction of relevant information from physiological modalities. Deep learning is proposed as an automatic approach to extract input features for models of affect from physiological signals. Experiments...

  10. Use and perceptions of information among family physicians: sources considered accessible, relevant, and reliable.

    Science.gov (United States)

    Kosteniuk, Julie G; Morgan, Debra G; D'Arcy, Carl K

    2013-01-01

    The research determined (1) the information sources that family physicians (FPs) most commonly use to update their general medical knowledge and to make specific clinical decisions, and (2) the information sources FPs found to be most physically accessible, intellectually accessible (easy to understand), reliable (trustworthy), and relevant to their needs. A cross-sectional postal survey of 792 FPs and locum tenens, in full-time or part-time medical practice, currently practicing or on leave of absence in the Canadian province of Saskatchewan was conducted during the period of January to April 2008. Of 666 eligible physicians, 331 completed and returned surveys, resulting in a response rate of 49.7% (331/666). Medical textbooks and colleagues in the main patient care setting were the top 2 sources for the purpose of making specific clinical decisions. Medical textbooks were most frequently considered by FPs to be reliable (trustworthy), and colleagues in the main patient care setting were most physically accessible (easy to access). When making specific clinical decisions, FPs were most likely to use information from sources that they considered to be reliable and generally physically accessible, suggesting that FPs can best be supported by facilitating easy and convenient access to high-quality information.

  11. Extraction of neutron spectral information from Bonner-Sphere data

    CERN Document Server

    Haney, J H; Zaidins, C S

    1999-01-01

    We have extended a least-squares method of extracting neutron spectral information from Bonner-Sphere data which was previously developed by Zaidins et al. (Med. Phys. 5 (1978) 42). A pulse-height analysis with background stripping is employed which provided a more accurate count rate for each sphere. Newer response curves by Mares and Schraube (Nucl. Instr. and Meth. A 366 (1994) 461) were included for the moderating spheres and the bare detector which comprise the Bonner spectrometer system. Finally, the neutron energy spectrum of interest was divided using the philosophy of fuzzy logic into three trapezoidal regimes corresponding to slow, moderate, and fast neutrons. Spectral data was taken using a PuBe source in two different environments and the analyzed data is presented for these cases as slow, moderate, and fast neutron fluences. (author)

  12. 75 FR 20843 - Notice of Workshop To Discuss Policy-Relevant Science to Inform EPA's Integrated Plan for the...

    Science.gov (United States)

    2010-04-21

    ... Policy-Relevant Science to Inform EPA's Integrated Plan for the Review of the Lead National Ambient Air.... Environmental Protection Agency (EPA) is announcing that a workshop entitled, ``Workshop to Discuss Policy... workshop will be open to attendance by interested public observers on a first-come, first-served basis up...

  13. Information Extraction for Clinical Data Mining: A Mammography Case Study.

    Science.gov (United States)

    Nassif, Houssam; Woods, Ryan; Burnside, Elizabeth; Ayvaci, Mehmet; Shavlik, Jude; Page, David

    2009-01-01

    Breast cancer is the leading cause of cancer mortality in women between the ages of 15 and 54. During mammography screening, radiologists use a strict lexicon (BI-RADS) to describe and report their findings. Mammography records are then stored in a well-defined database format (NMD). Lately, researchers have applied data mining and machine learning techniques to these databases. They successfully built breast cancer classifiers that can help in early detection of malignancy. However, the validity of these models depends on the quality of the underlying databases. Unfortunately, most databases suffer from inconsistencies, missing data, inter-observer variability and inappropriate term usage. In addition, many databases are not compliant with the NMD format and/or solely consist of text reports. BI-RADS feature extraction from free text and consistency checks between recorded predictive variables and text reports are crucial to addressing this problem. We describe a general scheme for concept information retrieval from free text given a lexicon, and present a BI-RADS features extraction algorithm for clinical data mining. It consists of a syntax analyzer, a concept finder and a negation detector. The syntax analyzer preprocesses the input into individual sentences. The concept finder uses a semantic grammar based on the BI-RADS lexicon and the experts' input. It parses sentences detecting BI-RADS concepts. Once a concept is located, a lexical scanner checks for negation. Our method can handle multiple latent concepts within the text, filtering out ultrasound concepts. On our dataset, our algorithm achieves 97.7% precision, 95.5% recall and an F 1 -score of 0.97. It outperforms manual feature extraction at the 5% statistical significance level.

  14. Comparing the Influence of Title and URL in Information Retrieval Relevance in Search Engines Results between Human Science and Agriculture Science

    Directory of Open Access Journals (Sweden)

    Parisa Allami

    2012-12-01

    Full Text Available When the World Wide Web provides suitable methods for producing and publishing information to scientists, the Web has become a mediator to publishing information. This environment has been formed billions of web pages that each of them has a special title, special content, special address and special purpose. Search engines provide a variety of facilities limit search results to raise the possibility of relevance in the retrieval results. One of these facilities is the limitation of the keywords and search terms to the title or URL. It can increase the possibility of results relevance significantly. Search engines claim what are limited to title and URL is most relevant. This research tried to compare the results relevant between results limited in title and URL in agricultural and Humanities areas from their users sights also it notice to Comparison of the presence of keywords in the title and URL between two areas and the relationship between search query numbers and matching keywords in title and their URLs. For this purpose, the number of 30 students in each area whom were in MA process and in doing their thesis was chosen. There was a significant relevant of the results that they limited their information needs to title and URL. There was significantly relevance in URL results in agricultural area, but there was not any significant difference between title and URL results in the humanities. For comparing the number of keywords in title and URL in two areas, 30 keywords in each area were chosen. There was not any significantly difference between the number of keywords in the title and URL of websites in two areas. To show relationship between number of search keyword and the matching of title and URL 45 keywords in each area were chosen. They were divided to three parts (one keyword, two keywords and three keywords. It was determined that if search keyword was less, the amount of matching between title and URL was more and if the matching

  15. A Participatory Model for Multi-Document Health Information Summarisation

    Directory of Open Access Journals (Sweden)

    Dinithi Nallaperuma

    2017-03-01

    Full Text Available Increasing availability and access to health information has been a paradigm shift in healthcare provision as it empowers both patients and practitioners alike. Besides awareness, significant time savings and process efficiencies can be achieved through effective summarisation of healthcare information. Relevance and accuracy are key concerns when generating summaries for such documents. Despite advances in automated summarisation approaches, the role of participation has not been explored. In this paper, we propose a new model for multi-document health information summarisation that takes into account the role of participation. The updated IS user participation theory was extended to explicate these roles. The proposed model integrates both extractive and abstractive summarisation processes with continuous participatory inputs to each phase. The model was implemented as a client-server application and evaluated by both domain experts and health information consumers. Results from the evaluation phase indicates the model is successful in generating relevant and accurate summaries for diverse audiences.

  16. Extraction of prospecting information of uranium deposit based on high spatial resolution satellite data. Taking bashibulake region as an example

    International Nuclear Information System (INIS)

    Yang Xu; Liu Dechang; Zhang Jielin

    2008-01-01

    In this study, the signification and content of prospecting information of uranium deposit are expounded. Quickbird high spatial resolution satellite data are used to extract the prospecting information of uranium deposit in Bashibulake area in the north of Tarim Basin. By using the pertinent methods of image processing, the information of ore-bearing bed, ore-control structure and mineralized alteration have been extracted. The results show a high consistency with the field survey. The aim of this study is to explore practicability of high spatial resolution satellite data for prospecting minerals, and to broaden the thinking of prospectation at similar area. (authors)

  17. Extracting Low-Frequency Information from Time Attenuation in Elastic Waveform Inversion

    Science.gov (United States)

    Guo, Xuebao; Liu, Hong; Shi, Ying; Wang, Weihong

    2017-03-01

    Low-frequency information is crucial for recovering background velocity, but the lack of low-frequency information in field data makes inversion impractical without accurate initial models. Laplace-Fourier domain waveform inversion can recover a smooth model from real data without low-frequency information, which can be used for subsequent inversion as an ideal starting model. In general, it also starts with low frequencies and includes higher frequencies at later inversion stages, while the difference is that its ultralow frequency information comes from the Laplace-Fourier domain. Meanwhile, a direct implementation of the Laplace-transformed wavefield using frequency domain inversion is also very convenient. However, because broad frequency bands are often used in the pure time domain waveform inversion, it is difficult to extract the wavefields dominated by low frequencies in this case. In this paper, low-frequency components are constructed by introducing time attenuation into the recorded residuals, and the rest of the method is identical to the traditional time domain inversion. Time windowing and frequency filtering are also applied to mitigate the ambiguity of the inverse problem. Therefore, we can start at low frequencies and to move to higher frequencies. The experiment shows that the proposed method can achieve a good inversion result in the presence of a linear initial model and records without low-frequency information.

  18. [Test Reviews in Child Psychology: Test Users Wish to Obtain Practical Information Relevant to their Respective Field of Work].

    Science.gov (United States)

    Renner, Gerolf; Irblich, Dieter

    2016-11-01

    Test Reviews in Child Psychology: Test Users Wish to Obtain Practical Information Relevant to their Respective Field of Work This study investigated to what extent diagnosticians use reviews of psychometric tests for children and adolescents, how they evaluate their quality, and what they expect concerning content. Test users (n = 323) from different areas of work (notably social pediatrics, early intervention, special education, speech and language therapy) rated test reviews as one of the most important sources of information. Readers of test reviews value practically oriented descriptions and evaluations of tests that are relevant to their respective field of work. They expect independent reviews that critically discuss opportunities and limits of the tests under scrutiny. The results show that authors of test reviews should not only have a background in test theory but should also be familiar with the practical application of tests in various settings.

  19. Bibliography of extraction chromatography, ch. 15

    International Nuclear Information System (INIS)

    Eschrich, H.; Drent, W.

    1975-01-01

    A bibliography of reviews, books and publications dealing with laminar extraction chromatography is given. A tabular survey of experimental data on column extraction chromatographic investigations, stationary phases and support materials is presented. An author index and corporative author index are included. This bibliography refers to various publications relevant to nuclear interest

  20. Comparison of methods of extracting information for meta-analysis of observational studies in nutritional epidemiology

    Directory of Open Access Journals (Sweden)

    Jong-Myon Bae

    2016-01-01

    Full Text Available OBJECTIVES: A common method for conducting a quantitative systematic review (QSR for observational studies related to nutritional epidemiology is the “highest versus lowest intake” method (HLM, in which only the information concerning the effect size (ES of the highest category of a food item is collected on the basis of its lowest category. However, in the interval collapsing method (ICM, a method suggested to enable a maximum utilization of all available information, the ES information is collected by collapsing all categories into a single category. This study aimed to compare the ES and summary effect size (SES between the HLM and ICM. METHODS: A QSR for evaluating the citrus fruit intake and risk of pancreatic cancer and calculating the SES by using the HLM was selected. The ES and SES were estimated by performing a meta-analysis using the fixed-effect model. The directionality and statistical significance of the ES and SES were used as criteria for determining the concordance between the HLM and ICM outcomes. RESULTS: No significant differences were observed in the directionality of SES extracted by using the HLM or ICM. The application of the ICM, which uses a broader information base, yielded more-consistent ES and SES, and narrower confidence intervals than the HLM. CONCLUSIONS: The ICM is advantageous over the HLM owing to its higher statistical accuracy in extracting information for QSR on nutritional epidemiology. The application of the ICM should hence be recommended for future studies.

  1. Culturally-Relevant Online Cancer Education Modules Empower Alaska's Community Health Aides/Practitioners to Disseminate Cancer Information and Reduce Cancer Risk.

    Science.gov (United States)

    Cueva, Katie; Revels, Laura; Cueva, Melany; Lanier, Anne P; Dignan, Mark; Viswanath, K; Fung, Teresa T; Geller, Alan C

    2017-04-12

    To address a desire for timely, medically accurate cancer education in rural Alaska, ten culturally relevant online learning modules were developed with, and for, Alaska's Community Health Aides/Practitioners (CHA/Ps). The project was guided by the framework of Community-Based Participatory Action Research, honored Indigenous Ways of Knowing, and was informed by Empowerment Theory. A total of 428 end-of-module evaluation surveys were completed by 89 unique Alaska CHA/Ps between January and December 2016. CHA/Ps shared that as a result of completing the modules, they were empowered to share cancer information with their patients, families, friends, and communities, as well as engage in cancer risk reduction behaviors such as eating healthier, getting cancer screenings, exercising more, and quitting tobacco. CHA/Ps also reported the modules were informative and respectful of their diverse cultures. These results from end-of-module evaluation surveys suggest that the collaboratively developed, culturally relevant, online cancer education modules have empowered CHA/Ps to reduce cancer risk and disseminate cancer information. "brought me to tears couple of times, and I think it will help in destroying the silence that surrounds cancer".

  2. A COMPARATIVE ANALYSIS OF WEB INFORMATION EXTRACTION TECHNIQUES DEEP LEARNING vs. NAÏVE BAYES vs. BACK PROPAGATION NEURAL NETWORKS IN WEB DOCUMENT EXTRACTION

    Directory of Open Access Journals (Sweden)

    J. Sharmila

    2016-01-01

    Full Text Available Web mining related exploration is getting the chance to be more essential these days in view of the reason that a lot of information is overseen through the web. Web utilization is expanding in an uncontrolled way. A particular framework is required for controlling such extensive measure of information in the web space. Web mining is ordered into three noteworthy divisions: Web content mining, web usage mining and web structure mining. Tak-Lam Wong has proposed a web content mining methodology in the exploration with the aid of Bayesian Networks (BN. In their methodology, they were learning on separating the web data and characteristic revelation in view of the Bayesian approach. Roused from their investigation, we mean to propose a web content mining methodology, in view of a Deep Learning Algorithm. The Deep Learning Algorithm gives the interest over BN on the basis that BN is not considered in any learning architecture planning like to propose system. The main objective of this investigation is web document extraction utilizing different grouping algorithm and investigation. This work extricates the data from the web URL. This work shows three classification algorithms, Deep Learning Algorithm, Bayesian Algorithm and BPNN Algorithm. Deep Learning is a capable arrangement of strategies for learning in neural system which is connected like computer vision, speech recognition, and natural language processing and biometrics framework. Deep Learning is one of the simple classification technique and which is utilized for subset of extensive field furthermore Deep Learning has less time for classification. Naive Bayes classifiers are a group of basic probabilistic classifiers in view of applying Bayes hypothesis with concrete independence assumptions between the features. At that point the BPNN algorithm is utilized for classification. Initially training and testing dataset contains more URL. We extract the content presently from the dataset. The

  3. Zone analysis in biology articles as a basis for information extraction.

    Science.gov (United States)

    Mizuta, Yoko; Korhonen, Anna; Mullen, Tony; Collier, Nigel

    2006-06-01

    In the field of biomedicine, an overwhelming amount of experimental data has become available as a result of the high throughput of research in this domain. The amount of results reported has now grown beyond the limits of what can be managed by manual means. This makes it increasingly difficult for the researchers in this area to keep up with the latest developments. Information extraction (IE) in the biological domain aims to provide an effective automatic means to dynamically manage the information contained in archived journal articles and abstract collections and thus help researchers in their work. However, while considerable advances have been made in certain areas of IE, pinpointing and organizing factual information (such as experimental results) remains a challenge. In this paper we propose tackling this task by incorporating into IE information about rhetorical zones, i.e. classification of spans of text in terms of argumentation and intellectual attribution. As the first step towards this goal, we introduce a scheme for annotating biological texts for rhetorical zones and provide a qualitative and quantitative analysis of the data annotated according to this scheme. We also discuss our preliminary research on automatic zone analysis, and its incorporation into our IE framework.

  4. Types of lexicographical information needs and their relevance for information science

    DEFF Research Database (Denmark)

    Bergenholtz, Henning; Pedersen, Heidi Agerbo

    2017-01-01

    often describes general human needs, typically with a reference to Maslow’s classification of needs (1954), instead of actual information needs. Lexicography has also focused on information needs, but has developed a more abstract classification of types of information needs, though (until more recent...

  5. Api m 10, a genuine A. mellifera venom allergen, is clinically relevant but underrepresented in therapeutic extracts.

    Science.gov (United States)

    Blank, S; Seismann, H; Michel, Y; McIntyre, M; Cifuentes, L; Braren, I; Grunwald, T; Darsow, U; Ring, J; Bredehorst, R; Ollert, M; Spillner, E

    2011-10-01

    Generalized systemic reactions to stinging hymenoptera venom constitute a potentially fatal condition in venom-allergic individuals. Hence, the identification and characterization of all allergens is imperative for improvement of diagnosis and design of effective immunotherapeutic approaches. Our aim was the immunochemical characterization of the carbohydrate-rich protein Api m 10, an Apis mellifera venom component and putative allergen, with focus on the relevance of glycosylation. Furthermore, the presence of Api m 10 in honeybee venom (HBV) and licensed venom immunotherapy preparations was addressed. Api m 10 was produced as soluble, aglycosylated protein in Escherichia coli and as differentially glycosylated protein providing a varying degree of fucosylation in insect cells. IgE reactivity and basophil activation of allergic patients were analyzed. For detection of Api m 10 in different venom preparations, a monoclonal human IgE antibody was generated. Both, the aglycosylated and the glycosylated variant of Api m 10 devoid of cross-reactive carbohydrate determinants (CCD), exhibited IgE reactivity with approximately 50% of HBV-sensitized patients. A corresponding reactivity could be documented for the activation of basophils. Although the detection of the native protein in crude HBV suggested content comparable to other relevant allergens, three therapeutical HBV extracts lacked detectable amounts of this component. Api m 10 is a genuine allergen of A. mellifera venom with IgE sensitizing potential in a significant fraction of allergic patients independent of CCD reactivity. Thus, Api m 10 could become a key element for component-resolved diagnostic tests and improved immunotherapeutic approaches in hymenoptera venom allergy. © 2011 John Wiley & Sons A/S.

  6. Evaluation of needle trap micro-extraction and solid-phase micro-extraction: Obtaining comprehensive information on volatile emissions from in vitro cultures.

    Science.gov (United States)

    Oertel, Peter; Bergmann, Andreas; Fischer, Sina; Trefz, Phillip; Küntzel, Anne; Reinhold, Petra; Köhler, Heike; Schubert, Jochen K; Miekisch, Wolfram

    2018-05-14

    Volatile organic compounds (VOCs) emitted from in vitro cultures may reveal information on species and metabolism. Owing to low nmol L -1 concentration ranges, pre-concentration techniques are required for gas chromatography-mass spectrometry (GC-MS) based analyses. This study was intended to compare the efficiency of established micro-extraction techniques - solid-phase micro-extraction (SPME) and needle-trap micro-extraction (NTME) - for the analysis of complex VOC patterns. For SPME, a 75 μm Carboxen®/polydimethylsiloxane fiber was used. The NTME needle was packed with divinylbenzene, Carbopack X and Carboxen 1000. The headspace was sampled bi-directionally. Seventy-two VOCs were calibrated by reference standard mixtures in the range of 0.041-62.24 nmol L -1 by means of GC-MS. Both pre-concentration methods were applied to profile VOCs from cultures of Mycobacterium avium ssp. paratuberculosis. Limits of detection ranged from 0.004 to 3.93 nmol L -1 (median = 0.030 nmol L -1 ) for NTME and from 0.001 to 5.684 nmol L -1 (median = 0.043 nmol L -1 ) for SPME. NTME showed advantages in assessing polar compounds such as alcohols. SPME showed advantages in reproducibility but disadvantages in sensitivity for N-containing compounds. Micro-extraction techniques such as SPME and NTME are well suited for trace VOC profiling over cultures if the limitations of each technique is taken into account. Copyright © 2018 John Wiley & Sons, Ltd.

  7. Approaching the largest ‘API’: extracting information from the Internet with Python

    Directory of Open Access Journals (Sweden)

    Jonathan E. Germann

    2018-02-01

    Full Text Available This article explores the need for libraries to algorithmically access and manipulate the world’s largest API: the Internet. The billions of pages on the ‘Internet API’ (HTTP, HTML, CSS, XPath, DOM, etc. are easily accessible and manipulable. Libraries can assist in creating meaning through the datafication of information on the world wide web. Because most information is created for human consumption, some programming is required for automated extraction. Python is an easy-to-learn programming language with extensive packages and community support for web page automation. Four packages (Urllib, Selenium, BeautifulSoup, Scrapy in Python can automate almost any web page for all sized projects. An example warrant data project is explained to illustrate how well Python packages can manipulate web pages to create meaning through assembling custom datasets.

  8. Ultrasound pretreatment as an alternative to improve essential oils extraction

    Directory of Open Access Journals (Sweden)

    Flávia Michelon Dalla Nora

    Full Text Available ABSTRACT: Essential oils are substances originated from plants in general. These compounds are well known to have a high biological activity, specially the antioxidant and antimicrobial. Several extraction techniques are employed to obtain these substances. However, the majority of these techniques require a long extraction time. In this sense, innovative and alternative extraction techniques, such as ultrasound, have recently been the target of studies. In view of the small amount of publications using ultrasonic pretreatment, this review aimed to congregate current relevant information on ultrasound-assisted extraction of essential oils. In this sense, theoretical aspects, such as the main factors that influence the performance of this technique as well as the advantages and disadvantages of the use of ultrasound as an environmental friendly alternative technique to improve the extraction of essential oil in comparison to traditional methods, are shown. Considering the available studies in the literature on essential oil extraction using ultrasonic pretreatment, low frequencies ranged from 20 to 50kWz and times ranged from 20 to 40min were used. The use of ultrasonic pretreatment represents a time reduction to near 70% in relation to the conventional hydrodistillation. Also, these conditions enabled a growth in the extraction of bioactive compounds and consequently improving the antioxidant and antimicrobial activities of essential oils.

  9. Extracting and Using Photon Polarization Information in Radiative B Decays

    Energy Technology Data Exchange (ETDEWEB)

    Grossman, Yuval

    2000-05-09

    The authors discuss the uses of conversion electron pairs for extracting photon polarization information in weak radiative B decays. Both cases of leptons produced through a virtual and real photon are considered. Measurements of the angular correlation between the (K-pi) and (e{sup +}e{sup {minus}}) decay planes in B --> K*(--> K-pi)gamma (*)(--> e{sup +}e{sup {minus}}) decays can be used to determine the helicity amplitudes in the radiative B --> K*gamma decays. A large right-handed helicity amplitude in B-bar decays is a signal of new physics. The time-dependent CP asymmetry in the B{sup 0} decay angular correlation is shown to measure sin 2-beta and cos 2-beta with little hadronic uncertainty.

  10. SAR matrices: automated extraction of information-rich SAR tables from large compound data sets.

    Science.gov (United States)

    Wassermann, Anne Mai; Haebel, Peter; Weskamp, Nils; Bajorath, Jürgen

    2012-07-23

    We introduce the SAR matrix data structure that is designed to elucidate SAR patterns produced by groups of structurally related active compounds, which are extracted from large data sets. SAR matrices are systematically generated and sorted on the basis of SAR information content. Matrix generation is computationally efficient and enables processing of large compound sets. The matrix format is reminiscent of SAR tables, and SAR patterns revealed by different categories of matrices are easily interpretable. The structural organization underlying matrix formation is more flexible than standard R-group decomposition schemes. Hence, the resulting matrices capture SAR information in a comprehensive manner.

  11. Extracting chemical information from high-resolution Kβ X-ray emission spectroscopy

    Science.gov (United States)

    Limandri, S.; Robledo, J.; Tirao, G.

    2018-06-01

    High-resolution X-ray emission spectroscopy allows studying the chemical environment of a wide variety of materials. Chemical information can be obtained by fitting the X-ray spectra and observing the behavior of some spectral features. Spectral changes can also be quantified by means of statistical parameters calculated by considering the spectrum as a probability distribution. Another possibility is to perform statistical multivariate analysis, such as principal component analysis. In this work the performance of these procedures for extracting chemical information in X-ray emission spectroscopy spectra for mixtures of Mn2+ and Mn4+ oxides are studied. A detail analysis of the parameters obtained, as well as the associated uncertainties is shown. The methodologies are also applied for Mn oxidation state characterization of double perovskite oxides Ba1+xLa1-xMnSbO6 (with 0 ≤ x ≤ 0.7). The results show that statistical parameters and multivariate analysis are the most suitable for the analysis of this kind of spectra.

  12. Feature extraction and learning using context cue and Rényi entropy based mutual information

    DEFF Research Database (Denmark)

    Pan, Hong; Olsen, Søren Ingvor; Zhu, Yaping

    2015-01-01

    information. In particular, for feature extraction, we develop a new set of kernel descriptors−Context Kernel Descriptors (CKD), which enhance the original KDES by embedding the spatial context into the descriptors. Context cues contained in the context kernel enforce some degree of spatial consistency, thus...... improving the robustness of CKD. For feature learning and reduction, we propose a novel codebook learning method, based on a Rényi quadratic entropy based mutual information measure called Cauchy-Schwarz Quadratic Mutual Information (CSQMI), to learn a compact and discriminative CKD codebook. Projecting...... as the information about the underlying labels of the CKD using CSQMI. Thus the resulting codebook and reduced CKD are discriminative. We verify the effectiveness of our method on several public image benchmark datasets such as YaleB, Caltech-101 and CIFAR-10, as well as a challenging chicken feet dataset of our own...

  13. Task-relevant perceptual features can define categories in visual memory too.

    Science.gov (United States)

    Antonelli, Karla B; Williams, Carrick C

    2017-11-01

    Although Konkle, Brady, Alvarez, and Oliva (2010, Journal of Experimental Psychology: General, 139(3), 558) claim that visual long-term memory (VLTM) is organized on underlying conceptual, not perceptual, information, visual memory results from visual search tasks are not well explained by this theory. We hypothesized that when viewing an object, any task-relevant visual information is critical to the organizational structure of VLTM. In two experiments, we examined the organization of VLTM by measuring the amount of retroactive interference created by objects possessing different combinations of task-relevant features. Based on task instructions, only the conceptual category was task relevant or both the conceptual category and a perceptual object feature were task relevant. Findings indicated that when made task relevant, perceptual object feature information, along with conceptual category information, could affect memory organization for objects in VLTM. However, when perceptual object feature information was task irrelevant, it did not contribute to memory organization; instead, memory defaulted to being organized around conceptual category information. These findings support the theory that a task-defined organizational structure is created in VLTM based on the relevance of particular object features and information.

  14. Coulomb disintegration as an information source for relevant processes in nuclear astrophysics

    International Nuclear Information System (INIS)

    Bertulani, C.A.

    1989-01-01

    The possibility of obtaining the photodisintegration cross section using the equivalent-photon number method first deduced and employed for the Coulomb disintegration processes has been suggested. This is very interesting because there exist radioactive capture processes, related to the photodisintegration through time reversal, that are relevant in astrophysics. In this paper, the recent results of the Karlsruhe and the Texas A and M groups on the Coulomb disintegration of 6 Li and 7 Li and the problems of the method are discussed. The ideas developed in a previous paper (Nucl. Phys. A458 (1986) 188) are confirmed qualitatively. To understand the process quantitatively it is necessary to use a quantum treatment that would imply the introduction of Coulomb excitation effects of higher orders. The Coulomb disintegration of exotic secondary beams is also studied. It is particularly interesting the question about what kind of nuclear structure information, as binding energies of momentum distributions, may be obtained. (Author) [es

  15. 3D numerical simulations of negative hydrogen ion extraction using realistic plasma parameters, geometry of the extraction aperture and full 3D magnetic field map

    Science.gov (United States)

    Mochalskyy, S.; Wünderlich, D.; Ruf, B.; Franzen, P.; Fantz, U.; Minea, T.

    2014-02-01

    Decreasing the co-extracted electron current while simultaneously keeping negative ion (NI) current sufficiently high is a crucial issue on the development plasma source system for ITER Neutral Beam Injector. To support finding the best extraction conditions the 3D Particle-in-Cell Monte Carlo Collision electrostatic code ONIX (Orsay Negative Ion eXtraction) has been developed. Close collaboration with experiments and other numerical models allows performing realistic simulations with relevant input parameters: plasma properties, geometry of the extraction aperture, full 3D magnetic field map, etc. For the first time ONIX has been benchmarked with commercial positive ions tracing code KOBRA3D. A very good agreement in terms of the meniscus position and depth has been found. Simulation of NI extraction with different e/NI ratio in bulk plasma shows high relevance of the direct negative ion extraction from the surface produced NI in order to obtain extracted NI current as in the experimental results from BATMAN testbed.

  16. Information relevant to ensuring that occupational radiation exposures at nuclear power stations will be as low as in reasonably achievable

    International Nuclear Information System (INIS)

    Anon.

    1979-01-01

    Regulations require that all reasonable efforts must be made to maintain exposure to radiation as far below the limits specified in 10 CFR Part 20 as is reasonably achievable. Information is provided relevant to attaining goals and objectives for planning, designing, constructing, operating and decommissioning a light-water-cooled nuclear power station to meet that criterion. Much of the information presented is also applicable to other than light-water-cooled nuclear power stations

  17. Locating relevant patient information in electronic health record data using representations of clinical concepts and database structures.

    Science.gov (United States)

    Pan, Xuequn; Cimino, James J

    2014-01-01

    Clinicians and clinical researchers often seek information in electronic health records (EHRs) that are relevant to some concept of interest, such as a disease or finding. The heterogeneous nature of EHRs can complicate retrieval, risking incomplete results. We frame this problem as the presence of two gaps: 1) a gap between clinical concepts and their representations in EHR data and 2) a gap between data representations and their locations within EHR data structures. We bridge these gaps with a knowledge structure that comprises relationships among clinical concepts (including concepts of interest and concepts that may be instantiated in EHR data) and relationships between clinical concepts and the database structures. We make use of available knowledge resources to develop a reproducible, scalable process for creating a knowledge base that can support automated query expansion from a clinical concept to all relevant EHR data.

  18. Biclustering methods: biological relevance and application in gene expression analysis.

    Directory of Open Access Journals (Sweden)

    Ali Oghabian

    Full Text Available DNA microarray technologies are used extensively to profile the expression levels of thousands of genes under various conditions, yielding extremely large data-matrices. Thus, analyzing this information and extracting biologically relevant knowledge becomes a considerable challenge. A classical approach for tackling this challenge is to use clustering (also known as one-way clustering methods where genes (or respectively samples are grouped together based on the similarity of their expression profiles across the set of all samples (or respectively genes. An alternative approach is to develop biclustering methods to identify local patterns in the data. These methods extract subgroups of genes that are co-expressed across only a subset of samples and may feature important biological or medical implications. In this study we evaluate 13 biclustering and 2 clustering (k-means and hierarchical methods. We use several approaches to compare their performance on two real gene expression data sets. For this purpose we apply four evaluation measures in our analysis: (1 we examine how well the considered (biclustering methods differentiate various sample types; (2 we evaluate how well the groups of genes discovered by the (biclustering methods are annotated with similar Gene Ontology categories; (3 we evaluate the capability of the methods to differentiate genes that are known to be specific to the particular sample types we study and (4 we compare the running time of the algorithms. In the end, we conclude that as long as the samples are well defined and annotated, the contamination of the samples is limited, and the samples are well replicated, biclustering methods such as Plaid and SAMBA are useful for discovering relevant subsets of genes and samples.

  19. Supporting Better Treatments for Meeting Health Consumers' Needs: Extracting Semantics in Social Data for Representing a Consumer Health Ontology

    Science.gov (United States)

    Choi, Yunseon

    2016-01-01

    Introduction: The purpose of this paper is to provide a framework for building a consumer health ontology using social tags. This would assist health users when they are accessing health information and increase the number of documents relevant to their needs. Methods: In order to extract concepts from social tags, this study conducted an…

  20. Scholarly Information Extraction Is Going to Make a Quantum Leap with PubMed Central (PMC).

    Science.gov (United States)

    Matthies, Franz; Hahn, Udo

    2017-01-01

    With the increasing availability of complete full texts (journal articles), rather than their surrogates (titles, abstracts), as resources for text analytics, entirely new opportunities arise for information extraction and text mining from scholarly publications. Yet, we gathered evidence that a range of problems are encountered for full-text processing when biomedical text analytics simply reuse existing NLP pipelines which were developed on the basis of abstracts (rather than full texts). We conducted experiments with four different relation extraction engines all of which were top performers in previous BioNLP Event Extraction Challenges. We found that abstract-trained engines loose up to 6.6% F-score points when run on full-text data. Hence, the reuse of existing abstract-based NLP software in a full-text scenario is considered harmful because of heavy performance losses. Given the current lack of annotated full-text resources to train on, our study quantifies the price paid for this short cut.

  1. Data Assimilation to Extract Soil Moisture Information from SMAP Observations

    Directory of Open Access Journals (Sweden)

    Jana Kolassa

    2017-11-01

    Full Text Available This study compares different methods to extract soil moisture information through the assimilation of Soil Moisture Active Passive (SMAP observations. Neural network (NN and physically-based SMAP soil moisture retrievals were assimilated into the National Aeronautics and Space Administration (NASA Catchment model over the contiguous United States for April 2015 to March 2017. By construction, the NN retrievals are consistent with the global climatology of the Catchment model soil moisture. Assimilating the NN retrievals without further bias correction improved the surface and root zone correlations against in situ measurements from 14 SMAP core validation sites (CVS by 0.12 and 0.16, respectively, over the model-only skill, and reduced the surface and root zone unbiased root-mean-square error (ubRMSE by 0.005 m 3 m − 3 and 0.001 m 3 m − 3 , respectively. The assimilation reduced the average absolute surface bias against the CVS measurements by 0.009 m 3 m − 3 , but increased the root zone bias by 0.014 m 3 m − 3 . Assimilating the NN retrievals after a localized bias correction yielded slightly lower surface correlation and ubRMSE improvements, but generally the skill differences were small. The assimilation of the physically-based SMAP Level-2 passive soil moisture retrievals using a global bias correction yielded similar skill improvements, as did the direct assimilation of locally bias-corrected SMAP brightness temperatures within the SMAP Level-4 soil moisture algorithm. The results show that global bias correction methods may be able to extract more independent information from SMAP observations compared to local bias correction methods, but without accurate quality control and observation error characterization they are also more vulnerable to adverse effects from retrieval errors related to uncertainties in the retrieval inputs and algorithm. Furthermore, the results show that using global bias correction approaches without a

  2. Improving life sciences information retrieval using semantic web technology.

    Science.gov (United States)

    Quan, Dennis

    2007-05-01

    The ability to retrieve relevant information is at the heart of every aspect of research and development in the life sciences industry. Information is often distributed across multiple systems and recorded in a way that makes it difficult to piece together the complete picture. Differences in data formats, naming schemes and network protocols amongst information sources, both public and private, must be overcome, and user interfaces not only need to be able to tap into these diverse information sources but must also assist users in filtering out extraneous information and highlighting the key relationships hidden within an aggregated set of information. The Semantic Web community has made great strides in proposing solutions to these problems, and many efforts are underway to apply Semantic Web techniques to the problem of information retrieval in the life sciences space. This article gives an overview of the principles underlying a Semantic Web-enabled information retrieval system: creating a unified abstraction for knowledge using the RDF semantic network model; designing semantic lenses that extract contextually relevant subsets of information; and assembling semantic lenses into powerful information displays. Furthermore, concrete examples of how these principles can be applied to life science problems including a scenario involving a drug discovery dashboard prototype called BioDash are provided.

  3. Biological network extraction from scientific literature: state of the art and challenges.

    Science.gov (United States)

    Li, Chen; Liakata, Maria; Rebholz-Schuhmann, Dietrich

    2014-09-01

    Networks of molecular interactions explain complex biological processes, and all known information on molecular events is contained in a number of public repositories including the scientific literature. Metabolic and signalling pathways are often viewed separately, even though both types are composed of interactions involving proteins and other chemical entities. It is necessary to be able to combine data from all available resources to judge the functionality, complexity and completeness of any given network overall, but especially the full integration of relevant information from the scientific literature is still an ongoing and complex task. Currently, the text-mining research community is steadily moving towards processing the full body of the scientific literature by making use of rich linguistic features such as full text parsing, to extract biological interactions. The next step will be to combine these with information from scientific databases to support hypothesis generation for the discovery of new knowledge and the extension of biological networks. The generation of comprehensive networks requires technologies such as entity grounding, coordination resolution and co-reference resolution, which are not fully solved and are required to further improve the quality of results. Here, we analyse the state of the art for the extraction of network information from the scientific literature and the evaluation of extraction methods against reference corpora, discuss challenges involved and identify directions for future research. © The Author 2013. Published by Oxford University Press. For Permissions, please email: journals.permissions@oup.com.

  4. The effect of informed consent on stress levels associated with extraction of impacted mandibular third molars.

    Science.gov (United States)

    Casap, Nardy; Alterman, Michael; Sharon, Guy; Samuni, Yuval

    2008-05-01

    To evaluate the effect of informed consent on stress levels associated with removal of impacted mandibular third molars. A total of 60 patients scheduled for extraction of impacted mandibular third molars participated in this study. The patients were unaware of the study's objectives. Data from 20 patients established the baseline levels of electrodermal activity (EDA). The remaining 40 patients were randomly assigned into 2 equal groups receiving either a detailed document of informed consent, disclosing the possible risks involved with the surgery, or a simplified version. Pulse, blood pressure, and EDA were monitored before, during, and after completion of the consent document. Changes in EDA, but not in blood pressure, were measured on completion of either version of the consent document. A greater increase in EDA was associated with the detailed version of the consent document (P = .004). A similar concomitant increase (although nonsignificant) in pulse values was monitored on completion of both versions. Completion of overdisclosed document of informed consent is associated with changes in physiological parameters. The results suggest that overdetailed listing and disclosure before extraction of impacted mandibular third molars can increase patient stress.

  5. The value relevance of direct cash flows under international financial reporting standards

    OpenAIRE

    Clacher, I; de Ricquebourg, AD; Hodgson, A

    2013-01-01

    This study examines if there has been a change in the value relevance of direct cash flow components since the adoption of International Financial Reporting Standards (IFRS) in Australia. Our results show that for both industrial and extractive firms direct cash flow statements are value relevant under Australian Generally Accepted Accounting Principles (AGAAP) and remain so after the adoption of IFRS. In addition, for industrial firms there is a significant increase in the value relevance of...

  6. Method of extracting significant trouble information of nuclear power plants using probabilistic analysis technique

    International Nuclear Information System (INIS)

    Shimada, Yoshio; Miyazaki, Takamasa

    2005-01-01

    In order to analyze and evaluate large amounts of trouble information of overseas nuclear power plants, it is necessary to select information that is significant in terms of both safety and reliability. In this research, a method of efficiently and simply classifying degrees of importance of components in terms of safety and reliability while paying attention to root-cause components appearing in the information was developed. Regarding safety, the reactor core damage frequency (CDF), which is used in the probabilistic analysis of a reactor, was used. Regarding reliability, the automatic plant trip probability (APTP), which is used in the probabilistic analysis of automatic reactor trips, was used. These two aspects were reflected in the development of criteria for classifying degrees of importance of components. By applying these criteria, a simple method of extracting significant trouble information of overseas nuclear power plants was developed. (author)

  7. Clinical evaluation of post-extraction site wound healing.

    Science.gov (United States)

    Adeyemo, Wasiu Lanre; Ladeinde, Akinola Ladipo; Ogunlewe, Mobolanle Olugbemiga

    2006-07-01

    The aim of this prospective study was to evaluate the clinical pattern of post-extraction wound healing with a view to identify the types, incidence, and pattern of healing complications following non-surgical tooth extraction. A total of 311 patients, who were referred for non-surgical (intra-alveolar) extractions, were included in the study. The relevant pre-operative information recorded for each patient included age and gender of the patient, indications for extraction, and tooth/teeth removed. Extractions were performed under local anesthesia with dental forceps, elevators, or both. Patients were evaluated on the third and seventh postoperative days for alveolus healing assessment. Data recorded were: biodata, day of presentation for alveolus healing assessment, day of onset of any symptoms, body temperature (degrees C) in cases of alveolus infection, and presence or absence of pain. Two hundred eighty-two patients (282) with 318 extraction sites were evaluated for alveolus healing. Healing was uneventful in 283 alveoli (89%), while 35 alveoli (11%) developed healing complications. These complications were: localized osteitis 26 (8.2%); acutely infected alveolus 5 (1.6%); and an acutely inflamed alveolus 4 (1.2%). Females developed more complications than males (p=0.003). Most complications were found in molars (60%) and premolars (37.1%). Localized osteitis caused severe pain in all cases, while infected and inflamed alveolus caused mild or no pain. Thirty patients (12%) among those without healing complications experienced mild pain. Most of the post-extraction alveoli healed uneventfully. Apart from alveolar osteitis (AO), post-extraction alveolus healing was also complicated by acutely infected alveoli and acutely inflamed alveoli. This study also demonstrated a painful alveolus is not necessarily a disturbance of post-extraction site wound healing; a thorough clinical examination must, therefore, be made to exclude any of the complications.

  8. Modulation of key biochemical markers relevant to stroke by Antiaris ...

    African Journals Online (AJOL)

    Modulation of key biochemical markers relevant to stroke by Antiaris africana leaf extract following cerebral ischemia/reperfusion injury. Omotayo B. Ilesanmi, Afolabi C. Akinmoladun, Olanrewaju Sam Olayeriju, Ibrahim Olabayode Saliu, M. Tolulope Olaleye, Afolabi A. Akindahunsi ...

  9. A text-mining system for extracting metabolic reactions from full-text articles.

    Science.gov (United States)

    Czarnecki, Jan; Nobeli, Irene; Smith, Adrian M; Shepherd, Adrian J

    2012-07-23

    Increasingly biological text mining research is focusing on the extraction of complex relationships relevant to the construction and curation of biological networks and pathways. However, one important category of pathway - metabolic pathways - has been largely neglected.Here we present a relatively simple method for extracting metabolic reaction information from free text that scores different permutations of assigned entities (enzymes and metabolites) within a given sentence based on the presence and location of stemmed keywords. This method extends an approach that has proved effective in the context of the extraction of protein-protein interactions. When evaluated on a set of manually-curated metabolic pathways using standard performance criteria, our method performs surprisingly well. Precision and recall rates are comparable to those previously achieved for the well-known protein-protein interaction extraction task. We conclude that automated metabolic pathway construction is more tractable than has often been assumed, and that (as in the case of protein-protein interaction extraction) relatively simple text-mining approaches can prove surprisingly effective. It is hoped that these results will provide an impetus to further research and act as a useful benchmark for judging the performance of more sophisticated methods that are yet to be developed.

  10. Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images

    Directory of Open Access Journals (Sweden)

    Koji Iwano

    2007-03-01

    Full Text Available This paper proposes an audio-visual speech recognition method using lip information extracted from side-face images as an attempt to increase noise robustness in mobile environments. Our proposed method assumes that lip images can be captured using a small camera installed in a handset. Two different kinds of lip features, lip-contour geometric features and lip-motion velocity features, are used individually or jointly, in combination with audio features. Phoneme HMMs modeling the audio and visual features are built based on the multistream HMM technique. Experiments conducted using Japanese connected digit speech contaminated with white noise in various SNR conditions show effectiveness of the proposed method. Recognition accuracy is improved by using the visual information in all SNR conditions. These visual features were confirmed to be effective even when the audio HMM was adapted to noise by the MLLR method.

  11. [Evidence-based medicine. 2. Research of clinically relevant biomedical information. Gruppo Italiano per la Medicina Basata sulle Evidenze--GIMBE].

    Science.gov (United States)

    Cartabellotta, A

    1998-05-01

    Evidence-based Medicine is a product of the electronic information age and there are several databases useful for practice it--MEDLINE, EMBASE, specialized compendiums of evidence (Cochrane Library, Best Evidence), practice guidelines--most of them free available through Internet, that offers a growing number of health resources. Because searching best evidence is a basic step to practice Evidence-based Medicine, this second review (the first one has been published in the issue of March 1998) has the aim to provide physicians tools and skills for retrieving relevant biomedical information. Therefore, we discuss about strategies for managing information overload, analyze characteristics, usefulness and limits of medical databases and explain how to use MEDLINE in day-to-day clinical practice.

  12. Information Retrieval and Text Mining Technologies for Chemistry.

    Science.gov (United States)

    Krallinger, Martin; Rabal, Obdulia; Lourenço, Anália; Oyarzabal, Julen; Valencia, Alfonso

    2017-06-28

    Efficient access to chemical information contained in scientific literature, patents, technical reports, or the web is a pressing need shared by researchers and patent attorneys from different chemical disciplines. Retrieval of important chemical information in most cases starts with finding relevant documents for a particular chemical compound or family. Targeted retrieval of chemical documents is closely connected to the automatic recognition of chemical entities in the text, which commonly involves the extraction of the entire list of chemicals mentioned in a document, including any associated information. In this Review, we provide a comprehensive and in-depth description of fundamental concepts, technical implementations, and current technologies for meeting these information demands. A strong focus is placed on community challenges addressing systems performance, more particularly CHEMDNER and CHEMDNER patents tasks of BioCreative IV and V, respectively. Considering the growing interest in the construction of automatically annotated chemical knowledge bases that integrate chemical information and biological data, cheminformatics approaches for mapping the extracted chemical names into chemical structures and their subsequent annotation together with text mining applications for linking chemistry with biological information are also presented. Finally, future trends and current challenges are highlighted as a roadmap proposal for research in this emerging field.

  13. Proceedings of the meeting on uranium exploration, mining and extraction

    International Nuclear Information System (INIS)

    1996-01-01

    Meeting on uranium exploration, mining, and extraction is aimed to expedite information exchange among researchers from the National Atomic Energy Agency (BATAN), their international colleagues, the higher education institutions,and other interested scientific communities on the latest development on Kalan uranium minerals exploration, mining, and extraction. Nuclear Minerals Development Centre (PPBGN) roles in nuclear energy provision, the theme of the meeting, reflect current advancements of the Centre in fulfilling its major tasks and responsibilities. In order to assist PPBGN better to assume its roles and responsibilities, the meeting is expected to bring forth essential solutions for problems and difficulties relevant to PPBGN's activities. Hence, the scope of the meeting will be limited to discussion on the status of nuclear minerals exploration, mining, and extraction technologies in Indonesia as well as the related environmental and workplace safeties in uranium mining and milling. Ten technical papers were presented in meeting, including four topics on exploration status and technology, three subject matter on mining, two presentations on milling, and one paper on environmental and workplace safeties

  14. Relevance as process: judgements in the context of scholarly research

    Directory of Open Access Journals (Sweden)

    Theresa D. Anderson

    2005-01-01

    Full Text Available Introduction. This paper discusses how exploring the research process in-depth and over time contributes to a fuller understanding of interactions with various representations of information. Method. A longitudinal ethnographic study explored decisions made by two informants involved in scholarly research. Relevance assessment and information seeking were observed as part of informants' own ongoing research projects. Fieldwork used methods of discovery that allowed informants to shape the exploration of the practices surrounding the evolving understandings of their topics. Analysis. Inductive analysis was carried out on the qualitative data collected over a two-year period of judgements observed on a document-by-document basis. The paper introduces broad categories that point to the variability and richness of the ways that informants used representations of information resources to make relevance judgements. Results. Relevance judgements appear to be drivers of the search and research processes informants moved through during the observations. Focusing on research goals rather than on retrieval tasks brings us to a fuller understanding of the relationship between ultimate research goals and the articulation of those goals in interactions with information systems. Conclusion. Relevance assessment is a process that unfolds in the doing of a search, the making of judgements and the using of texts and representations of information.

  15. Culicoides obsoletus extract relevant for diagnostics of insect bite hypersensitivity in horses

    NARCIS (Netherlands)

    Meide, van der N.M.A.; Meulenbroeks, C.; Altena, van S.E.C.; Schurink, A.; Ducro, B.J.; Wagner, B.; Leibold, W.; Rohwer, J.; Jacobs, F.; Sloet van Oldruitenborgh-Oosterbaan, M.M.; Savelkoul, H.F.J.; Tijhaar, E.

    2012-01-01

    Insect bite hypersensitivity (IBH) is an allergic dermatitis in horses caused by the bites of Culicoides species. The aim of the present study was to evaluate the applicability of whole body extracts of C. obsoletus (the main species found feeding on horses in the Netherlands), C. nubeculosus

  16. Which Type of Risk Information to Use for Whom? Moderating Role of Outcome-Relevant Involvement in the Effects of Statistical and Exemplified Risk Information on Risk Perceptions.

    Science.gov (United States)

    So, Jiyeon; Jeong, Se-Hoon; Hwang, Yoori

    2017-04-01

    The extant empirical research examining the effectiveness of statistical and exemplar-based health information is largely inconsistent. Under the premise that the inconsistency may be due to an unacknowledged moderator (O'Keefe, 2002), this study examined a moderating role of outcome-relevant involvement (Johnson & Eagly, 1989) in the effects of statistical and exemplified risk information on risk perception. Consistent with predictions based on elaboration likelihood model (Petty & Cacioppo, 1984), findings from an experiment (N = 237) concerning alcohol consumption risks showed that statistical risk information predicted risk perceptions of individuals with high, rather than low, involvement, while exemplified risk information predicted risk perceptions of those with low, rather than high, involvement. Moreover, statistical risk information contributed to negative attitude toward drinking via increased risk perception only for highly involved individuals, while exemplified risk information influenced the attitude through the same mechanism only for individuals with low involvement. Theoretical and practical implications for health risk communication are discussed.

  17. Explaining citizens’ perceptions of international climate-policy relevance

    International Nuclear Information System (INIS)

    Schleich, Joachim; Faure, Corinne

    2017-01-01

    This paper empirically analyses the antecedents of citizens’ perceptions of the relevance of international climate policy. Its use of representative surveys in the USA, China and Germany controls for different environmental attitudes and socio-economic factors between countries. The findings of the micro-econometric analysis suggest that the perceived relevance of international climate policy is positively affected by its perceived effectiveness, approval of the key topics discussed at international climate conferences, and environmental attitudes, but is not affected by perceived procedural justice. A higher level of perceived trust in international climate policy was positively related to perceived relevance in the USA and in China, but not in Germany. Citizens who felt that they were well informed and that their position was represented at climate summits were more likely to perceive international climate policy as relevant in China in particular. Generally, the results show only weak evidence of socio-demographic effects. - Highlights: • Perceptions of climate-policy relevance increase with perceptions of effectiveness. • In China and the USA, trust increases perceptions of climate-policy relevance. • Environmental attitudes are related to perceptions of climate-policy relevance. • In China, well-informed citizens perceive climate policy as more relevant. • Socio-demographics only weakly affect perceptions of climate-policy relevance.

  18. Perceived Relevance of Educative Information on Public (Skin Health: Results of a Representative, Population-Based Telephone Survey

    Directory of Open Access Journals (Sweden)

    Daniela Haluza

    2015-11-01

    Full Text Available Individual skin health attitudes are influenced by various factors, including public education campaigns, mass media, family, and friends. Evidence-based, educative information materials assist communication and decision-making in doctor-patient interactions. The present study aims at assessing the prevailing use of skin health information material and sources and their impact on skin health knowledge, motives to tan, and sun protection. We conducted a questionnaire survey among a representative sample of Austrian residents. Print media and television were perceived as the two most relevant sources for skin health information, whereas the source physician was ranked third. Picking the information source physician increased participants’ skin health knowledge (p = 0.025 and sun-protective behavior (p < 0.001. The study results highlight the demand for targeted health messages to attain lifestyle changes towards photo-protective habits. Providing resources that encourage pro-active counseling in every-day doctor-patient communication could increase skin health knowledge and sun-protective behavior, and thus, curb the rise in skin cancer incidence rates.

  19. EXTRACT: interactive extraction of environment metadata and term suggestion for metagenomic sample annotation.

    Science.gov (United States)

    Pafilis, Evangelos; Buttigieg, Pier Luigi; Ferrell, Barbra; Pereira, Emiliano; Schnetzer, Julia; Arvanitidis, Christos; Jensen, Lars Juhl

    2016-01-01

    The microbial and molecular ecology research communities have made substantial progress on developing standards for annotating samples with environment metadata. However, sample manual annotation is a highly labor intensive process and requires familiarity with the terminologies used. We have therefore developed an interactive annotation tool, EXTRACT, which helps curators identify and extract standard-compliant terms for annotation of metagenomic records and other samples. Behind its web-based user interface, the system combines published methods for named entity recognition of environment, organism, tissue and disease terms. The evaluators in the BioCreative V Interactive Annotation Task found the system to be intuitive, useful, well documented and sufficiently accurate to be helpful in spotting relevant text passages and extracting organism and environment terms. Comparison of fully manual and text-mining-assisted curation revealed that EXTRACT speeds up annotation by 15-25% and helps curators to detect terms that would otherwise have been missed. Database URL: https://extract.hcmr.gr/. © The Author(s) 2016. Published by Oxford University Press.

  20. Extracting business vocabularies from business process models: SBVR and BPMN standards-based approach

    Science.gov (United States)

    Skersys, Tomas; Butleris, Rimantas; Kapocius, Kestutis

    2013-10-01

    Approaches for the analysis and specification of business vocabularies and rules are very relevant topics in both Business Process Management and Information Systems Development disciplines. However, in common practice of Information Systems Development, the Business modeling activities still are of mostly empiric nature. In this paper, basic aspects of the approach for business vocabularies' semi-automated extraction from business process models are presented. The approach is based on novel business modeling-level OMG standards "Business Process Model and Notation" (BPMN) and "Semantics for Business Vocabularies and Business Rules" (SBVR), thus contributing to OMG's vision about Model-Driven Architecture (MDA) and to model-driven development in general.

  1. Resource Extraction in a Political Economy Framework

    OpenAIRE

    Ryszka, Karolina

    2013-01-01

    We analyze resource extraction in a political economy setting by introducing a political leader who optimizes both his own and the society's welfare function. We find that accounting for the private utility of a political elite, its higher discount rate and a different time horizon generally speeds up extraction. The higher than optimal resource extraction is not only relevant in welfare terms, but also regarding possible consequences with respect to climate change. The effect of higher extra...

  2. Architecture and data processing alternatives for the TSE computer. Volume 2: Extraction of topological information from an image by the Tse computer

    Science.gov (United States)

    Jones, J. R.; Bodenheimer, R. E.

    1976-01-01

    A simple programmable Tse processor organization and arithmetic operations necessary for extraction of the desired topological information are described. Hardware additions to this organization are discussed along with trade-offs peculiar to the tse computing concept. An improved organization is presented along with the complementary software for the various arithmetic operations. The performance of the two organizations is compared in terms of speed, power, and cost. Software routines developed to extract the desired information from an image are included.

  3. Information Extraction for Social Media

    NARCIS (Netherlands)

    Habib, M. B.; Keulen, M. van

    2014-01-01

    The rapid growth in IT in the last two decades has led to a growth in the amount of information available online. A new style for sharing information is social media. Social media is a continuously instantly updated source of information. In this position paper, we propose a framework for

  4. A construction scheme of web page comment information extraction system based on frequent subtree mining

    Science.gov (United States)

    Zhang, Xiaowen; Chen, Bingfeng

    2017-08-01

    Based on the frequent sub-tree mining algorithm, this paper proposes a construction scheme of web page comment information extraction system based on frequent subtree mining, referred to as FSM system. The entire system architecture and the various modules to do a brief introduction, and then the core of the system to do a detailed description, and finally give the system prototype.

  5. Evaluation of pattern recognition and feature extraction methods in ADHD prediction.

    Directory of Open Access Journals (Sweden)

    Joao Ricardo Sato

    2012-09-01

    Full Text Available Attention-Deficit/Hyperactivity Disorder is a neurodevelopmental disorder, being one of the most prevalent psychiatric disorders in childhood. The neural substrates associated with this condition, both from structural and functional perspectives, are not yet well established . Recent studies have highlighted the relevance of neuroimaging not only to provide a more solid understanding about the disorder but also for possible clinical support. The ADHD-200 Consortium organized the ADHD-200 global competition making publicly available, hundreds of structural magnetic resonance imaging (MRI and functional MRI (fMRI datasets of both ADHD patients and typically developing controls for research use. In the current study, we evaluate the predictive power of a set of three different feature extraction methods and 10 different pattern recognition methods. The features tested were regional homogeneity (ReHo, amplitude of low frequency fluctuations (ALFF and independent components analysis maps (RSN. Our findings suggest that the combination ALFF+ReHo maps contain relevant information to discriminate ADHD patients from typically developing controls, but with limited accuracy. All classifiers provided almost the same performance in this case. In addition, the combination ALFF+ReHo+RSN was relevant in combined vs inattentive ADHD classification, achieving a score accuracy of 67%. In this latter case, the performances of the classifiers were not equivalent and L2-regularized logistic regression (both in primal and dual space provided the most accurate predictions. The analysis of brain regions containing most discriminative information suggested that in both classifications (ADHD vs typically developing controls and combined vs inattentive, the relevant information is not confined only to a small set of regions but it is spatially distributed across the whole brain.

  6. Using small XML elements to support relevance

    NARCIS (Netherlands)

    G. Ramirez Camps (Georgina); T.H.W. Westerveld (Thijs); A.P. de Vries (Arjen)

    2006-01-01

    htmlabstractSmall XML elements are often estimated relevant by the retrieval model but they are not desirable retrieval units. This paper presents a generic model that exploits the information obtained from small elements. We identify relationships between small and relevant elements and use this

  7. A study on the relevance and influence of the existing regulation and risk informed/performance based regulation

    Energy Technology Data Exchange (ETDEWEB)

    Cheong, B. J.; Koh, Y. J.; Kim, H. S.; Koh, S. H.; Kang, D. H.; Kang, T. W. [Cheju National Univ., Jeju (Korea, Republic of)

    2004-02-15

    The goal of this study is to estimate the Relevance and Influence of the Existing Regulation and the RI-PBR to the institutionalization of the regulatory system. This study reviews the current regulatory system and the status of the RI-PBR implementation of the US NRC and Korea based upon SECY Papers, Risk Informed Regulation Implementation Plan (RIRIP) of the US NRC and other domestic studies. Also the recent trends of the individual technologies regarding the RI-PBR and RIA are summarized.

  8. A Survey: Framework of an Information Retrieval for Malay Translated Hadith Document

    Directory of Open Access Journals (Sweden)

    Zulkefli Nurul Syeilla Syazhween

    2017-01-01

    Full Text Available This paper reviews and analyses the limitation of the existing method used in the IR process in retrieving Malay Translated Hadith documents related to the search request. Traditional Malay Translated Hadith retrieval system has not focused on semantic extraction from text. The bag-of-words representation ignores the conceptual similarity of information in the query text and documents, which produce unsatisfactory retrieval results. Therefore, a more efficient IR framework is needed. This paper claims that the significant information extraction and subject-related information are actually important because the clues from this information can be used to search and find the relevance document to a query. Also, unimportant information can be discarded to represent the document content. So, semantic understanding of query and document is necessary to improve the effectiveness and accuracy of retrieval results for this domain study. Therefore, advance research is needed and it will be experimented in the future work. It is hoped that it will help users to search and find information regarding to the Malay Translated Hadith document.

  9. Medicaid Analytic eXtract (MAX) General Information

    Data.gov (United States)

    U.S. Department of Health & Human Services — The Medicaid Analytic eXtract (MAX) data is a set of person-level data files on Medicaid eligibility, service utilization, and payments. The MAX data are created to...

  10. Unsupervised Symbolization of Signal Time Series for Extraction of the Embedded Information

    Directory of Open Access Journals (Sweden)

    Yue Li

    2017-03-01

    Full Text Available This paper formulates an unsupervised algorithm for symbolization of signal time series to capture the embedded dynamic behavior. The key idea is to convert time series of the digital signal into a string of (spatially discrete symbols from which the embedded dynamic information can be extracted in an unsupervised manner (i.e., no requirement for labeling of time series. The main challenges here are: (1 definition of the symbol assignment for the time series; (2 identification of the partitioning segment locations in the signal space of time series; and (3 construction of probabilistic finite-state automata (PFSA from the symbol strings that contain temporal patterns. The reported work addresses these challenges by maximizing the mutual information measures between symbol strings and PFSA states. The proposed symbolization method has been validated by numerical simulation as well as by experimentation in a laboratory environment. Performance of the proposed algorithm has been compared to that of two commonly used algorithms of time series partitioning.

  11. Value Relevance of Investment Properties: Evidence from the Brazilian Capital Market

    Directory of Open Access Journals (Sweden)

    Ketlyn Alves Gonçalves

    2017-04-01

    Full Text Available This study investigates the relevance to the capital market of the assets recognized as investment properties of companies listed on the BM&F BOVESPA, in the period from 2011 to 2014. The research conducted was based on the Ohlson model (1995 and panel analysis was carried out using linear regression with POLS and Fixed and Random Effects estimators. Two hypothesis were made: (i that Earning and Equity generate accounting information relevant to investors; and (2 that Earning, Equity and Investment Property generate accounting information relevant to investors, assuming that investment properties have incremental effect on the relevance of this information relative only to earning and to equity. Both hypotheses were rejected, so it is concluded that Investment Property assets are not of value relevance in the determination of share price and do not influence the decision making of users of accounting information. The study adds to the limited literature on the value relevance of Investment Property, permitting a better understanding of the impact of accounting disclosures used by companies on their market value.

  12. Developing decision-relevant data and information systems for California water through listening and collaboration

    Science.gov (United States)

    Bales, R. C.; Bernacchi, L.; Conklin, M. H.; Viers, J. H.; Fogg, G. E.; Fisher, A. T.; Kiparsky, M.

    2017-12-01

    California's historic drought of 2011-2015 provided excellent conditions for researchers to listen to water-management challenges from decision makers, particularly with regard to data and information needs for improved decision making. Through the UC Water Security and Sustainability Research Initiative (http://ucwater.org/) we began a multi-year dialog with water-resources decision makers and state agencies that provide data and technical support for water management. Near-term products of that collaboration will be both a vision for a 21st-century water data and information system, and near-term steps to meet immediate legislative deadlines in a way that is consistent with the longer-term vision. While many university-based water researchers engage with state and local agencies on both science and policy challenges, UC Water's focus was on: i) integrated system management, from headwaters through groundwater and agriculture, and on ii) improved decision making through better water information systems. This focus aligned with the recognition by water leaders that fundamental changes in the way the state manages water were overdue. UC Water is focused on three "I"s: improved water information, empowering Institutions to use and to create new information, and enabling decision makers to make smart investments in both green and grey Infrastructure. Effective communication with water decision makers has led to engagement on high-priority programs where large knowledge gaps remain, including more-widespread groundwater recharge of storm flows, restoration of mountain forests in important source-water areas, governance structures for groundwater sustainability, and filling information gaps by bringing new technology to bear on measurement and data programs. Continuing engagement of UC Water researchers in public dialog around water resources, through opinion pieces, feature articles, blogs, white papers, social media, video clips and a feature documentary film have

  13. Does DNA extraction affect the physical and chemical composition of historical cod (Gadus morhua) otoliths?

    DEFF Research Database (Denmark)

    Therkildsen, Nina Overgaard; Eg Nielsen, Einar; Hüssy, Karin

    2010-01-01

    Archived otoliths constitute an important source of historical DNA for use in temporal genetic studies, but such otoliths are also valuable for other research applications, e.g. growth or microchemistry studies, where information about the past is of relevance. Consequently, there are potentially...... conflicting interests regarding how the limited and irreplaceable otolith collections should be used. To resolve this, it is important to find out whether DNA extraction damages otoliths such that they can no longer be used for other research purposes or whether individual otoliths can be used in multiple...... applications. We examined the effects of three different DNA extraction methods on the elemental composition, the morphology, and the clarity of annual growth increments for successful age estimation of Atlantic cod (Gadus morhua) otoliths that had been archived for 0–31 years. The three extraction methods...

  14. Addressing Risk Assessment for Patient Safety in Hospitals through Information Extraction in Medical Reports

    Science.gov (United States)

    Proux, Denys; Segond, Frédérique; Gerbier, Solweig; Metzger, Marie Hélène

    Hospital Acquired Infections (HAI) is a real burden for doctors and risk surveillance experts. The impact on patients' health and related healthcare cost is very significant and a major concern even for rich countries. Furthermore required data to evaluate the threat is generally not available to experts and that prevents from fast reaction. However, recent advances in Computational Intelligence Techniques such as Information Extraction, Risk Patterns Detection in documents and Decision Support Systems allow now to address this problem.

  15. Summarizing Simulation Results using Causally-relevant States

    Science.gov (United States)

    Parikh, Nidhi; Marathe, Madhav; Swarup, Samarth

    2016-01-01

    As increasingly large-scale multiagent simulations are being implemented, new methods are becoming necessary to make sense of the results of these simulations. Even concisely summarizing the results of a given simulation run is a challenge. Here we pose this as the problem of simulation summarization: how to extract the causally-relevant descriptions of the trajectories of the agents in the simulation. We present a simple algorithm to compress agent trajectories through state space by identifying the state transitions which are relevant to determining the distribution of outcomes at the end of the simulation. We present a toy-example to illustrate the working of the algorithm, and then apply it to a complex simulation of a major disaster in an urban area. PMID:28042620

  16. Advancing the Relevance Criteria for Video Search and Visual Summarization

    NARCIS (Netherlands)

    Rudinac, S.

    2013-01-01

    To facilitate finding of relevant information in ever-growing multimedia collections, a number of multimedia information retrieval solutions have been proposed over the past years. The essential element of any such solution is the relevance criterion deployed to select or rank the items from a

  17. Relevant Scatterers Characterization in SAR Images

    Science.gov (United States)

    Chaabouni, Houda; Datcu, Mihai

    2006-11-01

    Recognizing scenes in a single look meter resolution Synthetic Aperture Radar (SAR) images, requires the capability to identify relevant signal signatures in condition of variable image acquisition geometry, arbitrary objects poses and configurations. Among the methods to detect relevant scatterers in SAR images, we can mention the internal coherence. The SAR spectrum splitted in azimuth generates a series of images which preserve high coherence only for particular object scattering. The detection of relevant scatterers can be done by correlation study or Independent Component Analysis (ICA) methods. The present article deals with the state of the art for SAR internal correlation analysis and proposes further extensions using elements of inference based on information theory applied to complex valued signals. The set of azimuth looks images is analyzed using mutual information measures and an equivalent channel capacity is derived. The localization of the "target" requires analysis in a small image window, thus resulting in imprecise estimation of the second order statistics of the signal. For a better precision, a Hausdorff measure is introduced. The method is applied to detect and characterize relevant objects in urban areas.

  18. MememxGATE: Unearthing Latent Content Features for Improved Search and Relevancy Ranking Across Scientific Literature

    Science.gov (United States)

    Wilson, B. D.; McGibbney, L. J.; Mattmann, C. A.; Ramirez, P.; Joyce, M.; Whitehall, K. D.

    2015-12-01

    Quantifying scientific relevancy is of increasing importance to NASA and the research community. Scientific relevancy may be defined by mapping the impacts of a particular NASA mission, instrument, and/or retrieved variables to disciplines such as climate predictions, natural hazards detection and mitigation processes, education, and scientific discoveries. Related to relevancy, is the ability to expose data with similar attributes. This in turn depends upon the ability for us to extract latent, implicit document features from scientific data and resources and make them explicit, accessible and useable for search activities amongst others. This paper presents MemexGATE; a server side application, command line interface and computing environment for running large scale metadata extraction, general architecture text engineering, document classification and indexing tasks over document resources such as social media streams, scientific literature archives, legal documentation, etc. This work builds on existing experiences using MemexGATE (funded, developed and validated through the DARPA Memex Progrjam PI Mattmann) for extracting and leveraging latent content features from document resources within the Materials Research domain. We extend the software functionality capability to the domain of scientific literature with emphasis on the expansion of gazetteer lists, named entity rules, natural language construct labeling (e.g. synonym, antonym, hyponym, etc.) efforts to enable extraction of latent content features from data hosted by wide variety of scientific literature vendors (AGU Meeting Abstract Database, Springer, Wiley Online, Elsevier, etc.) hosting earth science literature. Such literature makes both implicit and explicit references to NASA datasets and relationships between such concepts stored across EOSDIS DAAC's hence we envisage that a significant part of this effort will also include development and understanding of relevancy signals which can ultimately

  19. Strategies for the extraction and analysis of non-extractable polyphenols from plants.

    Science.gov (United States)

    Domínguez-Rodríguez, Gloria; Marina, María Luisa; Plaza, Merichel

    2017-09-08

    The majority of studies based on phenolic compounds from plants are focused on the extractable fraction derived from an aqueous or aqueous-organic extraction. However, an important fraction of polyphenols is ignored due to the fact that they remain retained in the residue of extraction. They are the so-called non-extractable polyphenols (NEPs) which are high molecular weight polymeric polyphenols or individual low molecular weight phenolics associated to macromolecules. The scarce information available about NEPs shows that these compounds possess interesting biological activities. That is why the interest about the study of these compounds has been increasing in the last years. Furthermore, the extraction and characterization of NEPs are considered a challenge because the developed analytical methodologies present some limitations. Thus, the present literature review summarizes current knowledge of NEPs and the different methodologies for the extraction of these compounds, with a particular focus on hydrolysis treatments. Besides, this review provides information on the most recent developments in the purification, separation, identification and quantification of NEPs from plants. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. DEFINING THE RELEVANT OUTCOME MEASURES IN MEDICAL DEVICE ASSESSMENTS: AN ANALYSIS OF THE DEFINITION PROCESS IN HEALTH TECHNOLOGY ASSESSMENT.

    Science.gov (United States)

    Jacobs, Esther; Antoine, Sunya-Lee; Prediger, Barbara; Neugebauer, Edmund; Eikermann, Michaela

    2017-01-01

    Defining relevant outcome measures for clinical trials on medical devices (MD) is complex, as there is a large variety of potentially relevant outcomes. The chosen outcomes vary widely across clinical trials making the assessment in evidence syntheses very challenging. The objective is to provide an overview on the current common procedures of health technology assessment (HTA) institutions in defining outcome measures in MD trials. In 2012-14, the Web pages of 126 institutions involved in HTA were searched for methodological manuals written in English or German that describe methods for the predefinition process of outcome measures. Additionally, the institutions were contacted by email. Relevant information was extracted. All process steps were performed independently by two reviewers. Twenty-four manuals and ten responses from the email request were included in the analysis. Overall, 88.5 percent of the institutions describe the type of outcomes that should be considered in detail and 84.6 percent agree that the main focus should be on patient relevant outcomes. Specifically related to MD, information could be obtained in 26 percent of the included manuals and email responses. Eleven percent of the institutions report a particular consideration of MD related outcomes. This detailed analysis on common procedures of HTA institutions in the context of defining relevant outcome measures for the assessment of MD shows that standardized procedures for MD from the perspective of HTA institutions are not widespread. This leads to the question if a homogenous approach should be implemented in the field of HTA on MD.

  1. The relevance of electrostatics for scanning-gate microscopy

    International Nuclear Information System (INIS)

    Schnez, S; Guettinger, J; Stampfer, C; Ensslin, K; Ihn, T

    2011-01-01

    Scanning-probe techniques have been developed to extract local information from a given physical system. In particular, conductance maps obtained by means of scanning-gate microscopy (SGM), where a conducting tip of an atomic-force microscope is used as a local and movable gate, seem to present an intuitive picture of the underlying physical processes. Here, we argue that the interpretation of such images is complex and not very intuitive under certain circumstances: scanning a graphene quantum dot (QD) in the Coulomb-blockaded regime, we observe an apparent shift of features in scanning-gate images as a function of gate voltages, which cannot be a real shift of the physical system. Furthermore, we demonstrate the appearance of more than one set of Coulomb rings arising from the graphene QD. We attribute these effects to screening between the metallic tip and the gates. Our results are relevant for SGM on any kind of nanostructure, but are of particular importance for nanostructures that are not covered with a dielectric, e.g. graphene or carbon nanotube structures.

  2. A comparison of the value relevance of interim and annual financial statements

    Directory of Open Access Journals (Sweden)

    Mbalenhle Zulu

    2017-03-01

    Aim: It explores whether the value relevance of interim financial statements is higher than the value relevance of annual financial statements. Finally, it investigates whether accounting information published in interim and annual financial statements has incremental value relevance. Setting: Data for the period from 1999 to 2012 were collected from a sample of non-financial companies listed on the Johannesburg Stock Exchange. Method: The Ohlson model to investigate the value relevance of accounting information was used for the study. Results: The results show that interim book value of equity is value relevant while interim earnings are not. Interim financial statements appear to have higher value relevance than annual financial statements. The value relevance of interim and annual accounting information has remained fairly constant over the sample period. Incremental comparisons provide evidence that additional book value of equity and earnings that accrue to a company between interim and annual reporting dates are value relevant. Conclusion: The study was conducted over a long sample period (1999–2012, in an era when a technology-driven economy and more timely reporting media could have had an effect on the value relevance of published accounting information. To the best of our knowledge, this is the first study to evaluate and compare the value relevance of published interim and annual financial statements.

  3. Extracting respiratory information from seismocardiogram signals acquired on the chest using a miniature accelerometer

    International Nuclear Information System (INIS)

    Pandia, Keya; Inan, Omer T; Kovacs, Gregory T A; Giovangrandi, Laurent

    2012-01-01

    Seismocardiography (SCG) is a non-invasive measurement of the vibrations of the chest caused by the heartbeat. SCG signals can be measured using a miniature accelerometer attached to the chest, and are thus well-suited for unobtrusive and long-term patient monitoring. Additionally, SCG contains information relating to both cardiovascular and respiratory systems. In this work, algorithms were developed for extracting three respiration-dependent features of the SCG signal: intensity modulation, timing interval changes within each heartbeat, and timing interval changes between successive heartbeats. Simultaneously with a reference respiration belt, SCG signals were measured from 20 healthy subjects and a respiration rate was estimated using each of the three SCG features and the reference signal. The agreement between each of the three accelerometer-derived respiration rate measurements was computed with respect to the respiration rate derived from the reference respiration belt. The respiration rate obtained from the intensity modulation in the SCG signal was found to be in closest agreement with the respiration rate obtained from the reference respiration belt: the bias was found to be 0.06 breaths per minute with a 95% confidence interval of −0.99 to 1.11 breaths per minute. The limits of agreement between the respiration rates estimated using SCG (intensity modulation) and the reference were within the clinically relevant ranges given in existing literature, demonstrating that SCG could be used for both cardiovascular and respiratory monitoring. Furthermore, phases of each of the three SCG parameters were investigated at four instances of a respiration cycle—start inspiration, peak inspiration, start expiration, and peak expiration—and during breath hold (apnea). The phases of the three SCG parameters observed during the respiration cycle were congruent with existing literature and physiologically expected trends. (paper)

  4. Relevant energy scale of color confinement from lattice QCD

    International Nuclear Information System (INIS)

    Yamamoto, Arata; Suganuma, Hideo

    2009-01-01

    We propose a new lattice framework to extract the relevant gluonic energy scale of QCD phenomena which is based on a 'cut' on link variables in momentum space. This framework is expected to be broadly applicable to all lattice QCD calculations. Using this framework, we quantitatively determine the relevant energy scale of color confinement, through the analyses of the quark-antiquark potential and meson masses. The relevant energy scale of color confinement is found to be below 1.5 GeV in the Landau gauge. In fact, the string tension is almost unchanged even after cutting off the high-momentum gluon component above 1.5 GeV. When the relevant low-energy region is cut, the quark-antiquark potential is approximately reduced to a Coulomb-like potential, and each meson becomes a quasifree quark pair. As an analytical model calculation, we also investigate the dependence of the Richardson potential on the cut, and find the consistent behavior with the lattice result.

  5. Massive open online courses are relevant for postgraduate medical training

    DEFF Research Database (Denmark)

    Subhi, Yousif; Andresen, Kristoffer; Rolskov Bojsen, Signe

    2014-01-01

    INTRODUCTION: The CanMEDS framework describes seven roles in postgraduate training, but training and courses relevant to these roles can be limited. Massive open online courses (MOOCs) - free online courses in which anyone can participate, anywhere - may improve course participation. This study...... investigates the relevance of MOOCs for postgraduate medical training within the CanMEDS framework. MATERIAL AND METHODS: We extracted a list of all courses posted by the two largest MOOC providers, Coursera and EdX, and reviewed all course descriptions and categorised each course into one of three categories...

  6. Factors influencing efficiency of sliding mechanics to close extraction space: a systematic review.

    Science.gov (United States)

    Barlow, M; Kula, K

    2008-05-01

    To review recent literature to determine strength of clinical evidence concerning the influence of various factors on the efficiency (rate of tooth movement) of closing extraction spaces using sliding mechanics. A comprehensive systematic review on prospective clinical trials. An electronic search (1966-2006) of several databases limiting the searches to English and using several keywords was performed. Also a hand search of five key journals specifically searching for prospective clinical trials relevant to orthodontic space closure using sliding mechanics was completed. Outcome Measure - Rate of tooth movement. Ten prospective clinical trials comparing rates of closure under different variables and focusing only on sliding mechanics were selected for review. Of these ten trials on rate of closure, two compared arch wire variables, seven compared material variables used to apply force, and one examined bracket variables. Other articles which were not prospective clinical trials on sliding mechanics, but containing relevant information were examined and included as background information. CONCLUSION - The results of clinical research support laboratory results that nickel-titanium coil springs produce a more consistent force and a faster rate of closure when compared with active ligatures as a method of force delivery to close extraction space along a continuous arch wire; however, elastomeric chain produces similar rates of closure when compared with nickel-titanium springs. Clinical and laboratory research suggest little advantage of 200 g nickel-titanium springs over 150 g springs. More clinical research is needed in this area.

  7. Identification of "pathologs" (disease-related genes from the RIKEN mouse cDNA dataset using human curation plus FACTS, a new biological information extraction system

    Directory of Open Access Journals (Sweden)

    Socha Luis A

    2004-04-01

    Full Text Available Abstract Background A major goal in the post-genomic era is to identify and characterise disease susceptibility genes and to apply this knowledge to disease prevention and treatment. Rodents and humans have remarkably similar genomes and share closely related biochemical, physiological and pathological pathways. In this work we utilised the latest information on the mouse transcriptome as revealed by the RIKEN FANTOM2 project to identify novel human disease-related candidate genes. We define a new term "patholog" to mean a homolog of a human disease-related gene encoding a product (transcript, anti-sense or protein potentially relevant to disease. Rather than just focus on Mendelian inheritance, we applied the analysis to all potential pathologs regardless of their inheritance pattern. Results Bioinformatic analysis and human curation of 60,770 RIKEN full-length mouse cDNA clones produced 2,578 sequences that showed similarity (70–85% identity to known human-disease genes. Using a newly developed biological information extraction and annotation tool (FACTS in parallel with human expert analysis of 17,051 MEDLINE scientific abstracts we identified 182 novel potential pathologs. Of these, 36 were identified by computational tools only, 49 by human expert analysis only and 97 by both methods. These pathologs were related to neoplastic (53%, hereditary (24%, immunological (5%, cardio-vascular (4%, or other (14%, disorders. Conclusions Large scale genome projects continue to produce a vast amount of data with potential application to the study of human disease. For this potential to be realised we need intelligent strategies for data categorisation and the ability to link sequence data with relevant literature. This paper demonstrates the power of combining human expert annotation with FACTS, a newly developed bioinformatics tool, to identify novel pathologs from within large-scale mouse transcript datasets.

  8. Videomicroscopic extraction of specific information on cell proliferation and migration in vitro

    International Nuclear Information System (INIS)

    Debeir, Olivier; Megalizzi, Veronique; Warzee, Nadine; Kiss, Robert; Decaestecker, Christine

    2008-01-01

    In vitro cell imaging is a useful exploratory tool for cell behavior monitoring with a wide range of applications in cell biology and pharmacology. Combined with appropriate image analysis techniques, this approach has been shown to provide useful information on the detection and dynamic analysis of cell events. In this context, numerous efforts have been focused on cell migration analysis. In contrast, the cell division process has been the subject of fewer investigations. The present work focuses on this latter aspect and shows that, in complement to cell migration data, interesting information related to cell division can be extracted from phase-contrast time-lapse image series, in particular cell division duration, which is not provided by standard cell assays using endpoint analyses. We illustrate our approach by analyzing the effects induced by two sigma-1 receptor ligands (haloperidol and 4-IBP) on the behavior of two glioma cell lines using two in vitro cell models, i.e., the low-density individual cell model and the high-density scratch wound model. This illustration also shows that the data provided by our approach are suggestive as to the mechanism of action of compounds, and are thus capable of informing the appropriate selection of further time-consuming and more expensive biological evaluations required to elucidate a mechanism

  9. Passive vapor extraction feasibility study

    International Nuclear Information System (INIS)

    Rohay, V.J.

    1994-01-01

    Demonstration of a passive vapor extraction remediation system is planned for sites in the 200 West Area used in the past for the disposal of waste liquids containing carbon tetrachloride. The passive vapor extraction units will consist of a 4-in.-diameter pipe, a check valve, a canister filled with granular activated carbon, and a wind turbine. The check valve will prevent inflow of air that otherwise would dilute the soil gas and make its subsequent extraction less efficient. The granular activated carbon is used to adsorb the carbon tetrachloride from the air. The wind turbine enhances extraction rates on windy days. Passive vapor extraction units will be designed and operated to meet all applicable or relevant and appropriate requirements. Based on a cost analysis, passive vapor extraction was found to be a cost-effective method for remediation of soils containing lower concentrations of volatile contaminants. Passive vapor extraction used on wells that average 10-stdft 3 /min air flow rates was found to be more cost effective than active vapor extraction for concentrations below 500 parts per million by volume (ppm) of carbon tetrachloride. For wells that average 5-stdft 3 /min air flow rates, passive vapor extraction is more cost effective below 100 ppm

  10. The Relevant Physical Trace in Criminal Investigation

    Directory of Open Access Journals (Sweden)

    Durdica Hazard

    2016-01-01

    Full Text Available A criminal investigation requires the forensic scientist to search and to interpret vestiges of a criminal act that happened in the past. The forensic scientist is one of the many stakeholders who take part in the information quest within the criminal justice system. She reads the investigation scene in search of physical traces that should enable her to tell the story of the offense/crime that allegedly occurred. The challenge for any investigator is to detect and recognize relevant physical traces in order to provide clues for investigation and intelligence purposes, and that will constitute sound and relevant evidence for the court. This article shows how important it is to consider the relevancy of physical traces from the beginning of the investigation and what might influence the evaluation process. The exchange and management of information between the investigation stakeholders are important. Relevancy is a dimension that needs to be understood from the standpoints of law enforcement personnel and forensic scientists with the aim of strengthening investigation and ultimately the overall judicial process.

  11. Improving Nigerian health policymakers' capacity to access and utilize policy relevant evidence: outcome of information and communication technology training workshop.

    Science.gov (United States)

    Uneke, Chigozie Jesse; Ezeoha, Abel Ebeh; Uro-Chukwu, Henry; Ezeonu, Chinonyelum Thecla; Ogbu, Ogbonnaya; Onwe, Friday; Edoga, Chima

    2015-01-01

    Information and communication technology (ICT) tools are known to facilitate communication and processing of information and sharing of knowledge by electronic means. In Nigeria, the lack of adequate capacity on the use of ICT by health sector policymakers constitutes a major impediment to the uptake of research evidence into the policymaking process. The objective of this study was to improve the knowledge and capacity of policymakers to access and utilize policy relevant evidence. A modified "before and after" intervention study design was used in which outcomes were measured on the target participants both before the intervention is implemented and after. A 4-point likert scale according to the degree of adequacy; 1 = grossly inadequate, 4 = very adequate was employed. This study was conducted in Ebonyi State, south-eastern Nigeria and the participants were career health policy makers. A two-day intensive ICT training workshop was organized for policymakers who had 52 participants in attendance. Topics covered included: (i). intersectoral partnership/collaboration; (ii). Engaging ICT in evidence-informed policy making; use of ICT for evidence synthesis; (iv) capacity development on the use of computer, internet and other ICT. The pre-workshop mean of knowledge and capacity for use of ICT ranged from 2.19-3.05, while the post-workshop mean ranged from 2.67-3.67 on 4-point scale. The percentage increase in mean of knowledge and capacity at the end of the workshop ranged from 8.3%-39.1%. Findings of this study suggest that policymakers' ICT competence relevant to evidence-informed policymaking can be enhanced through training workshop.

  12. Using the DOM Tree for Content Extraction

    Directory of Open Access Journals (Sweden)

    David Insa

    2012-10-01

    Full Text Available The main information of a webpage is usually mixed between menus, advertisements, panels, and other not necessarily related information; and it is often difficult to automatically isolate this information. This is precisely the objective of content extraction, a research area of widely interest due to its many applications. Content extraction is useful not only for the final human user, but it is also frequently used as a preprocessing stage of different systems that need to extract the main content in a web document to avoid the treatment and processing of other useless information. Other interesting application where content extraction is particularly used is displaying webpages in small screens such as mobile phones or PDAs. In this work we present a new technique for content extraction that uses the DOM tree of the webpage to analyze the hierarchical relations of the elements in the webpage. Thanks to this information, the technique achieves a considerable recall and precision. Using the DOM structure for content extraction gives us the benefits of other approaches based on the syntax of the webpage (such as characters, words and tags, but it also gives us a very precise information regarding the related components in a block, thus, producing very cohesive blocks.

  13. Benchmarking and Its Relevance to the Library and Information Sector. Interim Findings of "Best Practice Benchmarking in the Library and Information Sector," a British Library Research and Development Department Project.

    Science.gov (United States)

    Kinnell, Margaret; Garrod, Penny

    This British Library Research and Development Department study assesses current activities and attitudes toward quality management in library and information services (LIS) in the academic sector as well as the commercial/industrial sector. Definitions and types of benchmarking are described, and the relevance of benchmarking to LIS is evaluated.…

  14. Android Smartphone Relevance to Military Weather Applications

    Science.gov (United States)

    2011-10-01

    lithium -ion battery that may be replaced by the user (unlike Apple iPod Touch devices), thus spare batteries can be carried. If there is only sporadic...Android Smartphone Relevance to Military Weather Applications by David Sauter ARL-TR-5793 October 2011...Android Smartphone Relevance to Military Weather Applications David Sauter Computational and Information Sciences Directorate, ARL

  15. EOG feature relevance determination for microsleep detection

    OpenAIRE

    Golz Martin; Wollner Sebastian; Sommer David; Schnieder Sebastian

    2017-01-01

    Automatic relevance determination (ARD) was applied to two-channel EOG recordings for microsleep event (MSE) recognition. 10 s immediately before MSE and also before counterexamples of fatigued, but attentive driving were analysed. Two type of signal features were extracted: the maximum cross correlation (MaxCC) and logarithmic power spectral densities (PSD) averaged in spectral bands of 0.5 Hz width ranging between 0 and 8 Hz. Generalised learn-ing vector quantisation (GRLVQ) was used as ARD...

  16. Automated Text Markup for Information Retrieval from an Electronic Textbook of Infectious Disease

    Science.gov (United States)

    Berrios, Daniel C.; Kehler, Andrew; Kim, David K.; Yu, Victor L.; Fagan, Lawrence M.

    1998-01-01

    The information needs of practicing clinicians frequently require textbook or journal searches. Making these sources available in electronic form improves the speed of these searches, but precision (i.e., the fraction of relevant to total documents retrieved) remains low. Improving the traditional keyword search by transforming search terms into canonical concepts does not improve search precision greatly. Kim et al. have designed and built a prototype system (MYCIN II) for computer-based information retrieval from a forthcoming electronic textbook of infectious disease. The system requires manual indexing by experts in the form of complex text markup. However, this mark-up process is time consuming (about 3 person-hours to generate, review, and transcribe the index for each of 218 chapters). We have designed and implemented a system to semiautomate the markup process. The system, information extraction for semiautomated indexing of documents (ISAID), uses query models and existing information-extraction tools to provide support for any user, including the author of the source material, to mark up tertiary information sources quickly and accurately.

  17. Improving information access by relevance and topical feedback

    NARCIS (Netherlands)

    Kaptein, R.; Kamps, J.; Hopfgartner, F.

    2008-01-01

    One of the main bottle-necks in providing more effective information access is the poverty of the query end. With an average query length of about two terms, users provide only a highly ambiguous statement of the, often complex, underlying information need. Implicit and explicit feedback can provide

  18. A Hybrid Approach to Finding Relevant Social Media Content for Complex Domain Specific Information Needs.

    Science.gov (United States)

    Cameron, Delroy; Sheth, Amit P; Jaykumar, Nishita; Thirunarayan, Krishnaprasad; Anand, Gaurish; Smith, Gary A

    2014-12-01

    While contemporary semantic search systems offer to improve classical keyword-based search, they are not always adequate for complex domain specific information needs. The domain of prescription drug abuse, for example, requires knowledge of both ontological concepts and "intelligible constructs" not typically modeled in ontologies. These intelligible constructs convey essential information that include notions of intensity, frequency, interval, dosage and sentiments, which could be important to the holistic needs of the information seeker. In this paper, we present a hybrid approach to domain specific information retrieval that integrates ontology-driven query interpretation with synonym-based query expansion and domain specific rules, to facilitate search in social media on prescription drug abuse. Our framework is based on a context-free grammar (CFG) that defines the query language of constructs interpretable by the search system. The grammar provides two levels of semantic interpretation: 1) a top-level CFG that facilitates retrieval of diverse textual patterns, which belong to broad templates and 2) a low-level CFG that enables interpretation of specific expressions belonging to such textual patterns. These low-level expressions occur as concepts from four different categories of data: 1) ontological concepts, 2) concepts in lexicons (such as emotions and sentiments), 3) concepts in lexicons with only partial ontology representation, called lexico-ontology concepts (such as side effects and routes of administration (ROA)), and 4) domain specific expressions (such as date, time, interval, frequency and dosage) derived solely through rules. Our approach is embodied in a novel Semantic Web platform called PREDOSE, which provides search support for complex domain specific information needs in prescription drug abuse epidemiology. When applied to a corpus of over 1 million drug abuse-related web forum posts, our search framework proved effective in retrieving

  19. Inexperienced clinicians can extract pathoanatomic information from MRI narrative reports with high reproducability for use in research/quality assurance

    DEFF Research Database (Denmark)

    Kent, Peter; Briggs, Andrew M; Albert, Hanne Birgit

    2011-01-01

    Background Although reproducibility in reading MRI images amongst radiologists and clinicians has been studied previously, no studies have examined the reproducibility of inexperienced clinicians in extracting pathoanatomic information from magnetic resonance imaging (MRI) narrative reports and t...

  20. Information Extraction and Interpretation Analysis of Mineral Potential Targets Based on ETM+ Data and GIS technology: A Case Study of Copper and Gold Mineralization in Burma

    International Nuclear Information System (INIS)

    Wenhui, Du; Yongqing, Chen; Nana, Guo; Yinglong, Hao; Pengfei, Zhao; Gongwen, Wang

    2014-01-01

    Mineralization-alteration and structure information extraction plays important roles in mineral resource prospecting and assessment using remote sensing data and the Geographical Information System (GIS) technology. Choosing copper and gold mines in Burma as example, the authors adopt band ratio, threshold segmentation and principal component analysis (PCA) to extract the hydroxyl alteration information using ETM+ remote sensing images. Digital elevation model (DEM) (30m spatial resolution) and ETM+ data was used to extract linear and circular faults that are associated with copper and gold mineralization. Combining geological data and the above information, the weights of evidence method and the C-A fractal model was used to integrate and identify the ore-forming favourable zones in this area. Research results show that the high grade potential targets are located with the known copper and gold deposits, and the integrated information can be used to the next exploration for the mineral resource decision-making

  1. Dramatic lives and relevant becomings

    DEFF Research Database (Denmark)

    Henriksen, Ann-Karina; Miller, Jody

    2012-01-01

    of marginality into positions of relevance. The analysis builds on empirical data from Copenhagen, Denmark, gained through ethnographic fieldwork with the participation of 20 female informants aged 13–22. The theoretical contribution proposes viewing conflicts as multi-linear, multi-causal and non...

  2. The effects of lossy compression on diagnostically relevant seizure information in EEG signals.

    Science.gov (United States)

    Higgins, G; McGinley, B; Faul, S; McEvoy, R P; Glavin, M; Marnane, W P; Jones, E

    2013-01-01

    This paper examines the effects of compression on EEG signals, in the context of automated detection of epileptic seizures. Specifically, it examines the use of lossy compression on EEG signals in order to reduce the amount of data which has to be transmitted or stored, while having as little impact as possible on the information in the signal relevant to diagnosing epileptic seizures. Two popular compression methods, JPEG2000 and SPIHT, were used. A range of compression levels was selected for both algorithms in order to compress the signals with varying degrees of loss. This compression was applied to the database of epileptiform data provided by the University of Freiburg, Germany. The real-time EEG analysis for event detection automated seizure detection system was used in place of a trained clinician for scoring the reconstructed data. Results demonstrate that compression by a factor of up to 120:1 can be achieved, with minimal loss in seizure detection performance as measured by the area under the receiver operating characteristic curve of the seizure detection system.

  3. Using Local Grammar for Entity Extraction from Clinical Reports

    Directory of Open Access Journals (Sweden)

    Aicha Ghoulam

    2015-06-01

    Full Text Available Information Extraction (IE is a natural language processing (NLP task whose aim is to analyze texts written in natural language to extract structured and useful information such as named entities and semantic relations linking these entities. Information extraction is an important task for many applications such as bio-medical literature mining, customer care, community websites, and personal information management. The increasing information available in patient clinical reports is difficult to access. As it is often in an unstructured text form, doctors need tools to enable them access to this information and the ability to search it. Hence, a system for extracting this information in a structured form can benefits healthcare professionals. The work presented in this paper uses a local grammar approach to extract medical named entities from French patient clinical reports. Experimental results show that the proposed approach achieved an F-Measure of 90. 06%.

  4. BioSimplify: an open source sentence simplification engine to improve recall in automatic biomedical information extraction.

    Science.gov (United States)

    Jonnalagadda, Siddhartha; Gonzalez, Graciela

    2010-11-13

    BioSimplify is an open source tool written in Java that introduces and facilitates the use of a novel model for sentence simplification tuned for automatic discourse analysis and information extraction (as opposed to sentence simplification for improving human readability). The model is based on a "shot-gun" approach that produces many different (simpler) versions of the original sentence by combining variants of its constituent elements. This tool is optimized for processing biomedical scientific literature such as the abstracts indexed in PubMed. We tested our tool on its impact to the task of PPI extraction and it improved the f-score of the PPI tool by around 7%, with an improvement in recall of around 20%. The BioSimplify tool and test corpus can be downloaded from https://biosimplify.sourceforge.net.

  5. Information extraction from dynamic PS-InSAR time series using machine learning

    Science.gov (United States)

    van de Kerkhof, B.; Pankratius, V.; Chang, L.; van Swol, R.; Hanssen, R. F.

    2017-12-01

    Due to the increasing number of SAR satellites, with shorter repeat intervals and higher resolutions, SAR data volumes are exploding. Time series analyses of SAR data, i.e. Persistent Scatterer (PS) InSAR, enable the deformation monitoring of the built environment at an unprecedented scale, with hundreds of scatterers per km2, updated weekly. Potential hazards, e.g. due to failure of aging infrastructure, can be detected at an early stage. Yet, this requires the operational data processing of billions of measurement points, over hundreds of epochs, updating this data set dynamically as new data come in, and testing whether points (start to) behave in an anomalous way. Moreover, the quality of PS-InSAR measurements is ambiguous and heterogeneous, which will yield false positives and false negatives. Such analyses are numerically challenging. Here we extract relevant information from PS-InSAR time series using machine learning algorithms. We cluster (group together) time series with similar behaviour, even though they may not be spatially close, such that the results can be used for further analysis. First we reduce the dimensionality of the dataset in order to be able to cluster the data, since applying clustering techniques on high dimensional datasets often result in unsatisfying results. Our approach is to apply t-distributed Stochastic Neighbor Embedding (t-SNE), a machine learning algorithm for dimensionality reduction of high-dimensional data to a 2D or 3D map, and cluster this result using Density-Based Spatial Clustering of Applications with Noise (DBSCAN). The results show that we are able to detect and cluster time series with similar behaviour, which is the starting point for more extensive analysis into the underlying driving mechanisms. The results of the methods are compared to conventional hypothesis testing as well as a Self-Organising Map (SOM) approach. Hypothesis testing is robust and takes the stochastic nature of the observations into account

  6. Behaviour of solvent extraction of niobium in nitric acid

    International Nuclear Information System (INIS)

    Lin Cansheng; Huang Meixin; Zhang Xianzi; Zhang Chonghai

    1988-01-01

    The behaviour of solvent extraction of niobium is discussed. The expractants, includding TBP, HDBP, H 2 MBP, TBP irradiated, HDEHP, TTA and Aliquat-7402, are used. The special influence of molybdenum and zirconium on solvent extraction of niobium and the extraction behaviur of niobium with TBP irradiated are described. The effect of fluorine and uranium in aqueous phase on extraction of niobium is mentioned. It is observed that the interfacial crud has not relevance to D Nb , but niobium-95 can be absorbed on it. The species of extractable niobium, extraction mechanism, and the reason brought niobum into organic phase are discussed. Finally, the idea of increasing decontamination factor for niobium is suggested

  7. A hybrid approach for robust multilingual toponym extraction and disambiguation

    NARCIS (Netherlands)

    Habib, Mena Badieh; van Keulen, Maurice

    Toponym extraction and disambiguation are key topics recently addressed by fields of Information Extraction and Geographical Information Retrieval. Toponym extraction and disambiguation are highly dependent processes. Not only toponym extraction effectiveness affects disambiguation, but also

  8. Automated Trait Extraction using ClearEarth, a Natural Language Processing System for Text Mining in Natural Sciences

    OpenAIRE

    Thessen,Anne; Preciado,Jenette; Jain,Payoj; Martin,James; Palmer,Martha; Bhat,Riyaz

    2018-01-01

    The cTAKES package (using the ClearTK Natural Language Processing toolkit Bethard et al. 2014, http://cleartk.github.io/cleartk/) has been successfully used to automatically read clinical notes in the medical field (Albright et al. 2013, Styler et al. 2014). It is used on a daily basis to automatically process clinical notes and extract relevant information by dozens of medical institutions. ClearEarth is a collaborative project that brings together computational linguistics and domain scient...

  9. The BEL information extraction workflow (BELIEF): evaluation in the BioCreative V BEL and IAT track

    OpenAIRE

    Madan, Sumit; Hodapp, Sven; Senger, Philipp; Ansari, Sam; Szostak, Justyna; Hoeng, Julia; Peitsch, Manuel; Fluck, Juliane

    2016-01-01

    Network-based approaches have become extremely important in systems biology to achieve a better understanding of biological mechanisms. For network representation, the Biological Expression Language (BEL) is well designed to collate findings from the scientific literature into biological network models. To facilitate encoding and biocuration of such findings in BEL, a BEL Information Extraction Workflow (BELIEF) was developed. BELIEF provides a web-based curation interface, the BELIEF Dashboa...

  10. Choice reaching with a LEGO arm robot (CoRLEGO): The motor system guides visual attention to movement-relevant information.

    Science.gov (United States)

    Strauss, Soeren; Woodgate, Philip J W; Sami, Saber A; Heinke, Dietmar

    2015-12-01

    We present an extension of a neurobiologically inspired robotics model, termed CoRLEGO (Choice reaching with a LEGO arm robot). CoRLEGO models experimental evidence from choice reaching tasks (CRT). In a CRT participants are asked to rapidly reach and touch an item presented on the screen. These experiments show that non-target items can divert the reaching movement away from the ideal trajectory to the target item. This is seen as evidence attentional selection of reaching targets can leak into the motor system. Using competitive target selection and topological representations of motor parameters (dynamic neural fields) CoRLEGO is able to mimic this leakage effect. Furthermore if the reaching target is determined by its colour oddity (i.e. a green square among red squares or vice versa), the reaching trajectories become straighter with repetitions of the target colour (colour streaks). This colour priming effect can also be modelled with CoRLEGO. The paper also presents an extension of CoRLEGO. This extension mimics findings that transcranial direct current stimulation (tDCS) over the motor cortex modulates the colour priming effect (Woodgate et al., 2015). The results with the new CoRLEGO suggest that feedback connections from the motor system to the brain's attentional system (parietal cortex) guide visual attention to extract movement-relevant information (i.e. colour) from visual stimuli. This paper adds to growing evidence that there is a close interaction between the motor system and the attention system. This evidence contradicts the traditional conceptualization of the motor system as the endpoint of a serial chain of processing stages. At the end of the paper we discuss CoRLEGO's predictions and also lessons for neurobiologically inspired robotics emerging from this work. Crown Copyright © 2015. Published by Elsevier Ltd. All rights reserved.

  11. Assessing Hospital Physicians' Acceptance of Clinical Information Systems: A Review of the Relevant Literature

    Directory of Open Access Journals (Sweden)

    Bram Pynoo

    2013-06-01

    Full Text Available In view of the tremendous potential benefits of clinical information systems (CIS for the quality of patient care; it is hard to understand why not every CIS is embraced by its targeted users, the physicians. The aim of this study is to propose a framework for assessing hospital physicians' CIS-acceptance that can serve as a guidance for future research into this area. Hereto, a review of the relevant literature was performed in the ISI Web-of-Science database. Eleven studies were withheld from an initial dataset of 797 articles. Results show that just as in business settings, there are four core groups of variables that influence physicians' acceptance of a CIS: its usefulness and ease of use, social norms, and factors in the working environment that facilitate use of the CIS (such as providing computers/workstations, compatibility between the new and existing system.... We also identified some additional variables as predictors of CIS-acceptance.

  12. Geopositioning with a quadcopter: Extracted feature locations and predicted accuracy without a priori sensor attitude information

    Science.gov (United States)

    Dolloff, John; Hottel, Bryant; Edwards, David; Theiss, Henry; Braun, Aaron

    2017-05-01

    This paper presents an overview of the Full Motion Video-Geopositioning Test Bed (FMV-GTB) developed to investigate algorithm performance and issues related to the registration of motion imagery and subsequent extraction of feature locations along with predicted accuracy. A case study is included corresponding to a video taken from a quadcopter. Registration of the corresponding video frames is performed without the benefit of a priori sensor attitude (pointing) information. In particular, tie points are automatically measured between adjacent frames using standard optical flow matching techniques from computer vision, an a priori estimate of sensor attitude is then computed based on supplied GPS sensor positions contained in the video metadata and a photogrammetric/search-based structure from motion algorithm, and then a Weighted Least Squares adjustment of all a priori metadata across the frames is performed. Extraction of absolute 3D feature locations, including their predicted accuracy based on the principles of rigorous error propagation, is then performed using a subset of the registered frames. Results are compared to known locations (check points) over a test site. Throughout this entire process, no external control information (e.g. surveyed points) is used other than for evaluation of solution errors and corresponding accuracy.

  13. Surveying managers to inform a regionally relevant invasive Phragmites australis control research program.

    Science.gov (United States)

    Rohal, C B; Kettenring, K M; Sims, K; Hazelton, E L G; Ma, Z

    2018-01-15

    more pertinent to manager needs and trusted by managers. Such an approach that integrates manager surveys to inform management experiments could be adapted to any developing research program seeking to be relevant to management audiences. Copyright © 2017 Elsevier Ltd. All rights reserved.

  14. An Enhanced Text-Mining Framework for Extracting Disaster Relevant Data through Social Media and Remote Sensing Data Fusion

    Science.gov (United States)

    Scheele, C. J.; Huang, Q.

    2016-12-01

    In the past decade, the rise in social media has led to the development of a vast number of social media services and applications. Disaster management represents one of such applications leveraging massive data generated for event detection, response, and recovery. In order to find disaster relevant social media data, current approaches utilize natural language processing (NLP) methods based on keywords, or machine learning algorithms relying on text only. However, these approaches cannot be perfectly accurate due to the variability and uncertainty in language used on social media. To improve current methods, the enhanced text-mining framework is proposed to incorporate location information from social media and authoritative remote sensing datasets for detecting disaster relevant social media posts, which are determined by assessing the textual content using common text mining methods and how the post relates spatiotemporally to the disaster event. To assess the framework, geo-tagged Tweets were collected for three different spatial and temporal disaster events: hurricane, flood, and tornado. Remote sensing data and products for each event were then collected using RealEarthTM. Both Naive Bayes and Logistic Regression classifiers were used to compare the accuracy within the enhanced text-mining framework. Finally, the accuracies from the enhanced text-mining framework were compared to the current text-only methods for each of the case study disaster events. The results from this study address the need for more authoritative data when using social media in disaster management applications.

  15. Adverse Drug Reaction Identification and Extraction in Social Media: A Scoping Review.

    Science.gov (United States)

    Lardon, Jérémy; Abdellaoui, Redhouane; Bellet, Florelle; Asfari, Hadyl; Souvignet, Julien; Texier, Nathalie; Jaulent, Marie-Christine; Beyens, Marie-Noëlle; Burgun, Anita; Bousquet, Cédric

    2015-07-10

    The underreporting of adverse drug reactions (ADRs) through traditional reporting channels is a limitation in the efficiency of the current pharmacovigilance system. Patients' experiences with drugs that they report on social media represent a new source of data that may have some value in postmarketing safety surveillance. A scoping review was undertaken to explore the breadth of evidence about the use of social media as a new source of knowledge for pharmacovigilance. Daubt et al's recommendations for scoping reviews were followed. The research questions were as follows: How can social media be used as a data source for postmarketing drug surveillance? What are the available methods for extracting data? What are the different ways to use these data? We queried PubMed, Embase, and Google Scholar to extract relevant articles that were published before June 2014 and with no lower date limit. Two pairs of reviewers independently screened the selected studies and proposed two themes of review: manual ADR identification (theme 1) and automated ADR extraction from social media (theme 2). Descriptive characteristics were collected from the publications to create a database for themes 1 and 2. Of the 1032 citations from PubMed and Embase, 11 were relevant to the research question. An additional 13 citations were added after further research on the Internet and in reference lists. Themes 1 and 2 explored 11 and 13 articles, respectively. Ways of approaching the use of social media as a pharmacovigilance data source were identified. This scoping review noted multiple methods for identifying target data, extracting them, and evaluating the quality of medical information from social media. It also showed some remaining gaps in the field. Studies related to the identification theme usually failed to accurately assess the completeness, quality, and reliability of the data that were analyzed from social media. Regarding extraction, no study proposed a generic approach to easily

  16. Pilot information needs survey regarding climate relevant technologies

    International Nuclear Information System (INIS)

    Van Berkel, R.; Van Roekel, A.

    1997-02-01

    The objective of this pilot survey was to arrive at a preliminary understanding of the initial technology and technology information needs in non-Annex II countries in order to support international efforts to facilitate the transfer of technologies and know-how conducive to mitigating and adapting to climate change. The study encompassed two main components, i.e. the development of a survey instrument and the execution of a pilot survey among selected non-Annex II countries. The survey instrument addresses the present status of enabling activities; technology and technology information needs; and issues related to information supply and accessibility. The survey was distributed to national focal points in 20 non-Annex II countries and to at least 35 other stakeholders in five of these non-Annex II countries. A total of 27 completed questionnaires were received, covering 10 non-Annex II countries. 3 refs

  17. Pilot information needs survey regarding climate relevant technologies

    Energy Technology Data Exchange (ETDEWEB)

    Van Berkel, R.; Van Roekel, A.

    1997-02-01

    The objective of this pilot survey was to arrive at a preliminary understanding of the initial technology and technology information needs in non-Annex II countries in order to support international efforts to facilitate the transfer of technologies and know-how conducive to mitigating and adapting to climate change. The study encompassed two main components, i.e. the development of a survey instrument and the execution of a pilot survey among selected non-Annex II countries. The survey instrument addresses the present status of enabling activities; technology and technology information needs; and issues related to information supply and accessibility. The survey was distributed to national focal points in 20 non-Annex II countries and to at least 35 other stakeholders in five of these non-Annex II countries. A total of 27 completed questionnaires were received, covering 10 non-Annex II countries. 3 refs.

  18. Multi-Paradigm and Multi-Lingual Information Extraction as Support for Medical Web Labelling Authorities

    Directory of Open Access Journals (Sweden)

    Martin Labsky

    2010-10-01

    Full Text Available Until recently, quality labelling of medical web content has been a pre-dominantly manual activity. However, the advances in automated text processing opened the way to computerised support of this activity. The core enabling technology is information extraction (IE. However, the heterogeneity of websites offering medical content imposes particular requirements on the IE techniques to be applied. In the paper we discuss these requirements and describe a multi-paradigm approach to IE addressing them. Experiments on multi-lingual data are reported. The research has been carried out within the EU MedIEQ project.

  19. BioSimplify: an open source sentence simplification engine to improve recall in automatic biomedical information extraction

    OpenAIRE

    Jonnalagadda, Siddhartha; Gonzalez, Graciela

    2011-01-01

    BioSimplify is an open source tool written in Java that introduces and facilitates the use of a novel model for sentence simplification tuned for automatic discourse analysis and information extraction (as opposed to sentence simplification for improving human readability). The model is based on a "shot-gun" approach that produces many different (simpler) versions of the original sentence by combining variants of its constituent elements. This tool is optimized for processing biomedical scien...

  20. Photo-generated carriers lose energy during extraction from polymer-fullerene solar cells

    KAUST Repository

    Melianas, Armantas; Etzold, Fabian; Savenije, Tom J.; Laquai, Fré dé ric; Inganä s, Olle; Kemerink, Martijn

    2015-01-01

    motion is boosted by this process, leading to a time-dependent carrier mobility as confirmed by direct experiments. We identify the time and distance scales relevant for carrier extraction and show that the photo-generated carriers are extracted from

  1. The value relevance of environmental emissions

    Directory of Open Access Journals (Sweden)

    Melinda Lydia Nelwan

    2016-07-01

    Full Text Available This study examines whether environmental performance has value relevance by investigating the relations between environmental emissions and stock prices for the U.S. public companies. The previous studies argued that the conjectured relations between accounting performance measures and environmental performance do not have a strong theoretical basis, and the modeling of relations between market per-formance measures and environmental performance do not adequately consider the relevance of accounting performance to market value. Therefore, this study examines whether publicly reported environmental emissions provide incremental information to accounting earnings in pricing companies stocks. It is done among the complete set of industries covered by Toxics Release Inventory (TRI reporting for the period 2007 to 2010. Using Ohlson model but modified to include different types of emis-sions, it is found that ground emissions (underground injection and land emissions are value relevant but other emission types (air and water and transferred-out emis-sions appear to not provide incremental information in the valuation model. The result in this study raise concerns that different types of emissions are assessed differently by the market, confirming that studies should not aggregate such measures.

  2. You had me at "Hello": Rapid extraction of dialect information from spoken words.

    Science.gov (United States)

    Scharinger, Mathias; Monahan, Philip J; Idsardi, William J

    2011-06-15

    Research on the neuronal underpinnings of speaker identity recognition has identified voice-selective areas in the human brain with evolutionary homologues in non-human primates who have comparable areas for processing species-specific calls. Most studies have focused on estimating the extent and location of these areas. In contrast, relatively few experiments have investigated the time-course of speaker identity, and in particular, dialect processing and identification by electro- or neuromagnetic means. We show here that dialect extraction occurs speaker-independently, pre-attentively and categorically. We used Standard American English and African-American English exemplars of 'Hello' in a magnetoencephalographic (MEG) Mismatch Negativity (MMN) experiment. The MMN as an automatic change detection response of the brain reflected dialect differences that were not entirely reducible to acoustic differences between the pronunciations of 'Hello'. Source analyses of the M100, an auditory evoked response to the vowels suggested additional processing in voice-selective areas whenever a dialect change was detected. These findings are not only relevant for the cognitive neuroscience of language, but also for the social sciences concerned with dialect and race perception. Copyright © 2011 Elsevier Inc. All rights reserved.

  3. An image-processing strategy to extract important information suitable for a low-size stimulus pattern in a retinal prosthesis.

    Science.gov (United States)

    Chen, Yili; Fu, Jixiang; Chu, Dawei; Li, Rongmao; Xie, Yaoqin

    2017-11-27

    A retinal prosthesis is designed to help the blind to obtain some sight. It consists of an external part and an internal part. The external part is made up of a camera, an image processor and an RF transmitter. The internal part is made up of an RF receiver, implant chip and microelectrode. Currently, the number of microelectrodes is in the hundreds, and we do not know the mechanism for using an electrode to stimulate the optic nerve. A simple hypothesis is that the pixels in an image correspond to the electrode. The images captured by the camera should be processed by suitable strategies to correspond to stimulation from the electrode. Thus, it is a question of how to obtain the important information from the image captured in the picture. Here, we use the region of interest (ROI), a useful algorithm for extracting the ROI, to retain the important information, and to remove the redundant information. This paper explains the details of the principles and functions of the ROI. Because we are investigating a real-time system, we need a fast processing ROI as a useful algorithm to extract the ROI. Thus, we simplified the ROI algorithm and used it in an outside image-processing digital signal processing (DSP) system of the retinal prosthesis. The results show that our image-processing strategies are suitable for a real-time retinal prosthesis and can eliminate redundant information and provide useful information for expression in a low-size image.

  4. Book value, earnings, dividends, and audit quality on the value relevance of accounting information among Nigerian listed firms

    Directory of Open Access Journals (Sweden)

    Muhammad Yusuf Alkali

    2018-04-01

    Full Text Available The objective of this paper is to determine the effect of International Financial Reporting Standards (IFRS as a new accounting reporting among Nigerian listed firms. This study uses book value, earnings and dividends to fill in the gap using a sample of 126 Nigerian listed firms in the stock market from 2009 to 2013 (pre and Post-IFRS adoption. Data was collected from Thompson Reuters, Bank scope DataStreams and annual reports. The study adopted Ohlson (1995 [Ohlson, J. (1995. Earnings, book-value, and dividends in equity valuation. Contemporary Accounting Research, 11(2, 661–687.] price model that has been frequently used in determining the quality of accounting information studies. The study finds that combined book value, earnings and dividends do not provide statistical significance effects on IFRS after adoption on the quality of accounting information. This could be possible, as dividends do not provide a significant effect in the presence of earnings. Furthermore, the audit big 4 quality provided an effect on the quality of accounting information because of IFRS adoption. Therefore, findings of this study provide additional literature on the decreasing quality of accounting information in an emerging market setting like Nigeria. The study implication is to the policy makers, regulators, and government that accounting information do not provide value relevance among Nigerian listed firms after IFRS adoption.

  5. DNA in ancient bone - where is it located and how should we extract it?

    DEFF Research Database (Denmark)

    Campos, Paula; Craig, Oliver E.; Turner-Walker, Gordon

    2012-01-01

    Despite the widespread use of bones in ancient DNA (aDNA) studies, relatively little concrete information exists in regard to how the DNA in mineralised collagen degrades, or where it survives in the material's architecture. While, at the macrostructural level, physical exclusion of microbes...... and other external contaminants may be an important feature, and, at the ultrastructural level, the adsorption of DNA to hydroxyapatite and/or binding of DNA to Type I collagen may stabilise the DNA, the relative contribution of each, and what other factors may be relevant, are unclear....... The question arises as to whether this may be due to post-collection preservation or just an artefact of the extraction methods used in these different studies? In an attempt to resolve these questions, we examine the efficacy of DNA extraction methods, and the quality and quantity of DNA recovered from both...

  6. Correlation between the extracting solutions, Modified KCl-Olsen and Mehlich 3, used in soil laboratories in Costa Rica

    International Nuclear Information System (INIS)

    Bertsch, Floria; Bejarano, Jose Antonio; Corrales, Marco

    2005-01-01

    The correlation found, between the 2 most commonly used extraction solutions in soil laboratories of Costa Rica, is discussed for Ca, Mg, K, Zn and P determinations in soil analyses. Given the coexistence of extraction methodologies, it is of great relevance to provide users with information allowing an adequate interpretation of the analysis results. Using data exchanged among laboratories, at the national level, relationships between modified KCl-Olsen and Mehlich 3 solutions were established. For all elements determined, except for P, the association between both solutions is very clear and well-defined. Both solutions extract the same amounts of Ca and Mg; Mehlich 3 extracts 1.5 times more K than Modified Olsen. In the case of Zn, in Ca-rich soils (>10 cmol(+) 1 -1 ) Mehlich 3 extracts more Zn, so the critical level must be raised to 3.5 mg 1''- 1 ; whereas, in soils low in Ca ( -1 ), Mehlich 3 extracts less Zn than Modified Olsen, so the critical level must be lowered to 2.5 mg 1 -1 . As for P, the association is not clear at all. (author) [es

  7. Semantics-based information extraction for detecting economic events

    NARCIS (Netherlands)

    A.C. Hogenboom (Alexander); F. Frasincar (Flavius); K. Schouten (Kim); O. van der Meer

    2013-01-01

    textabstractAs today's financial markets are sensitive to breaking news on economic events, accurate and timely automatic identification of events in news items is crucial. Unstructured news items originating from many heterogeneous sources have to be mined in order to extract knowledge useful for

  8. Lung region extraction based on the model information and the inversed MIP method by using chest CT images

    International Nuclear Information System (INIS)

    Tomita, Toshihiro; Miguchi, Ryosuke; Okumura, Toshiaki; Yamamoto, Shinji; Matsumoto, Mitsuomi; Tateno, Yukio; Iinuma, Takeshi; Matsumoto, Toru.

    1997-01-01

    We developed a lung region extraction method based on the model information and the inversed MIP method in the Lung Cancer Screening CT (LSCT). Original model is composed of typical 3-D lung contour lines, a body axis, an apical point, and a convex hull. First, the body axis. the apical point, and the convex hull are automatically extracted from the input image Next, the model is properly transformed to fit to those of input image by the affine transformation. Using the same affine transformation coefficients, typical lung contour lines are also transferred, which correspond to rough contour lines of input image. Experimental results applied for 68 samples showed this method quite promising. (author)

  9. Text-in-context: a method for extracting findings in mixed-methods mixed research synthesis studies.

    Science.gov (United States)

    Sandelowski, Margarete; Leeman, Jennifer; Knafl, Kathleen; Crandell, Jamie L

    2013-06-01

    Our purpose in this paper is to propose a new method for extracting findings from research reports included in mixed-methods mixed research synthesis studies. International initiatives in the domains of systematic review and evidence synthesis have been focused on broadening the conceptualization of evidence, increased methodological inclusiveness and the production of evidence syntheses that will be accessible to and usable by a wider range of consumers. Initiatives in the general mixed-methods research field have been focused on developing truly integrative approaches to data analysis and interpretation. The data extraction challenges described here were encountered, and the method proposed for addressing these challenges was developed, in the first year of the ongoing (2011-2016) study: Mixed-Methods Synthesis of Research on Childhood Chronic Conditions and Family. To preserve the text-in-context of findings in research reports, we describe a method whereby findings are transformed into portable statements that anchor results to relevant information about sample, source of information, time, comparative reference point, magnitude and significance and study-specific conceptions of phenomena. The data extraction method featured here was developed specifically to accommodate mixed-methods mixed research synthesis studies conducted in nursing and other health sciences, but reviewers might find it useful in other kinds of research synthesis studies. This data extraction method itself constitutes a type of integration to preserve the methodological context of findings when statements are read individually and in comparison to each other. © 2012 Blackwell Publishing Ltd.

  10. Multineuronal vectorization is more efficient than time-segmental vectorization for information extraction from neuronal activities in the inferior temporal cortex.

    Science.gov (United States)

    Kaneko, Hidekazu; Tamura, Hiroshi; Tate, Shunta; Kawashima, Takahiro; Suzuki, Shinya S; Fujita, Ichiro

    2010-08-01

    In order for patients with disabilities to control assistive devices with their own neural activity, multineuronal spike trains must be efficiently decoded because only limited computational resources can be used to generate prosthetic control signals in portable real-time applications. In this study, we compare the abilities of two vectorizing procedures (multineuronal and time-segmental) to extract information from spike trains during the same total neuron-seconds. In the multineuronal vectorizing procedure, we defined a response vector whose components represented the spike counts of one to five neurons. In the time-segmental vectorizing procedure, a response vector consisted of components representing a neuron's spike counts for one to five time-segment(s) of a response period of 1 s. Spike trains were recorded from neurons in the inferior temporal cortex of monkeys presented with visual stimuli. We examined whether the amount of information of the visual stimuli carried by these neurons differed between the two vectorizing procedures. The amount of information calculated with the multineuronal vectorizing procedure, but not the time-segmental vectorizing procedure, significantly increased with the dimensions of the response vector. We conclude that the multineuronal vectorizing procedure is superior to the time-segmental vectorizing procedure in efficiently extracting information from neuronal signals. Copyright (c) 2010 Elsevier Ltd. All rights reserved.

  11. Visual content highlighting via automatic extraction of embedded captions on MPEG compressed video

    Science.gov (United States)

    Yeo, Boon-Lock; Liu, Bede

    1996-03-01

    Embedded captions in TV programs such as news broadcasts, documentaries and coverage of sports events provide important information on the underlying events. In digital video libraries, such captions represent a highly condensed form of key information on the contents of the video. In this paper we propose a scheme to automatically detect the presence of captions embedded in video frames. The proposed method operates on reduced image sequences which are efficiently reconstructed from compressed MPEG video and thus does not require full frame decompression. The detection, extraction and analysis of embedded captions help to capture the highlights of visual contents in video documents for better organization of video, to present succinctly the important messages embedded in the images, and to facilitate browsing, searching and retrieval of relevant clips.

  12. The Relevance of Hyperbaric Oxygen to Combat Medicine

    Science.gov (United States)

    2001-06-01

    and Hyperbaric Conditions [les Questions medicales a caractere oprationel liees aux conditions hypobares ou hyperbares ] To order the complete...UNCLASSIFIED Defense Technical Information Center Compilation Part Notice ADPO 11081 TITLE: The Relevance of Hyperbaric Oxygen to Combat Medicine...following component part numbers comprise the compilation report: ADPO11059 thru ADP011100 UNCLASSIFIED 25-1 The Relevance of Hyperbaric Oxygen to

  13. Information Management Processes for Extraction of Student Dropout Indicators in Courses in Distance Mode

    Directory of Open Access Journals (Sweden)

    Renata Maria Abrantes Baracho

    2016-04-01

    Full Text Available This research addresses the use of information management processes in order to extract student dropout indicators in distance mode courses. Distance education in Brazil aims to facilitate access to information. The MEC (Ministry of Education announced, in the second semester of 2013, that the main obstacles faced by institutions offering courses in this mode were students dropping out and the resistance of both educators and students to this mode. The research used a mixed methodology, qualitative and quantitative, to obtain student dropout indicators. The factors found and validated in this research were: the lack of interest from students, insufficient training in the use of the virtual learning environment for students, structural problems in the schools that were chosen to offer the course, students without e-mail, incoherent answers to activities to the course, lack of knowledge on the part of the student when using the computer tool. The scenario considered was a course offered in distance mode called Aluno Integrado (Integrated Student

  14. Feature extraction and sensor selection for NPP initiating event identification

    International Nuclear Information System (INIS)

    Lin, Ting-Han; Wu, Shun-Chi; Chen, Kuang-You; Chou, Hwai-Pwu

    2017-01-01

    Highlights: • A two-stage feature extraction scheme for NPP initiating event identification. • With stBP, interrelations among the sensors can be retained for identification. • With dSFS, sensors that are crucial for identification can be efficiently selected. • Efficacy of the scheme is illustrated with data from the Maanshan NPP simulator. - Abstract: Initiating event identification is essential in managing nuclear power plant (NPP) severe accidents. In this paper, a novel two-stage feature extraction scheme that incorporates the proposed sensor type-wise block projection (stBP) and deflatable sequential forward selection (dSFS) is used to elicit the discriminant information in the data obtained from various NPP sensors to facilitate event identification. With the stBP, the primal features can be extracted without eliminating the interrelations among the sensors of the same type. The extracted features are then subjected to a further dimensionality reduction by selecting the sensors that are most relevant to the events under consideration. This selection is not easy, and a combinatorial optimization technique is normally required. With the dSFS, an optimal sensor set can be found with less computational load. Moreover, its sensor deflation stage allows sensors in the preselected set to be iteratively refined to avoid being trapped into a local optimum. Results from detailed experiments containing data of 12 event categories and a total of 112 events generated with a Taiwan’s Maanshan NPP simulator are presented to illustrate the efficacy of the proposed scheme.

  15. DEVELOPMENT OF AUTOMATIC EXTRACTION METHOD FOR ROAD UPDATE INFORMATION BASED ON PUBLIC WORK ORDER OUTLOOK

    Science.gov (United States)

    Sekimoto, Yoshihide; Nakajo, Satoru; Minami, Yoshitaka; Yamaguchi, Syohei; Yamada, Harutoshi; Fuse, Takashi

    Recently, disclosure of statistic data, representing financial effects or burden for public work, through each web site of national or local government, enables us to discuss macroscopic financial trends. However, it is still difficult to grasp a basic property nationwide how each spot was changed by public work. In this research, our research purpose is to collect road update information reasonably which various road managers provide, in order to realize efficient updating of various maps such as car navigation maps. In particular, we develop the system extracting public work concerned and registering summary including position information to database automatically from public work order outlook, released by each local government, combinating some web mining technologies. Finally, we collect and register several tens of thousands from web site all over Japan, and confirm the feasibility of our method.

  16. Academic Activities Transaction Extraction Based on Deep Belief Network

    Directory of Open Access Journals (Sweden)

    Xiangqian Wang

    2017-01-01

    Full Text Available Extracting information about academic activity transactions from unstructured documents is a key problem in the analysis of academic behaviors of researchers. The academic activities transaction includes five elements: person, activities, objects, attributes, and time phrases. The traditional method of information extraction is to extract shallow text features and then to recognize advanced features from text with supervision. Since the information processing of different levels is completed in steps, the error generated from various steps will be accumulated and affect the accuracy of final results. However, because Deep Belief Network (DBN model has the ability to automatically unsupervise learning of the advanced features from shallow text features, the model is employed to extract the academic activities transaction. In addition, we use character-based feature to describe the raw features of named entities of academic activity, so as to improve the accuracy of named entity recognition. In this paper, the accuracy of the academic activities extraction is compared by using character-based feature vector and word-based feature vector to express the text features, respectively, and with the traditional text information extraction based on Conditional Random Fields. The results show that DBN model is more effective for the extraction of academic activities transaction information.

  17. Extract the Relational Information of Static Features and Motion Features for Human Activities Recognition in Videos

    Directory of Open Access Journals (Sweden)

    Li Yao

    2016-01-01

    Full Text Available Both static features and motion features have shown promising performance in human activities recognition task. However, the information included in these features is insufficient for complex human activities. In this paper, we propose extracting relational information of static features and motion features for human activities recognition. The videos are represented by a classical Bag-of-Word (BoW model which is useful in many works. To get a compact and discriminative codebook with small dimension, we employ the divisive algorithm based on KL-divergence to reconstruct the codebook. After that, to further capture strong relational information, we construct a bipartite graph to model the relationship between words of different feature set. Then we use a k-way partition to create a new codebook in which similar words are getting together. With this new codebook, videos can be represented by a new BoW vector with strong relational information. Moreover, we propose a method to compute new clusters from the divisive algorithm’s projective function. We test our work on the several datasets and obtain very promising results.

  18. Intelligent medical information filtering.

    Science.gov (United States)

    Quintana, Y

    1998-01-01

    This paper describes an intelligent information filtering system to assist users to be notified of updates to new and relevant medical information. Among the major problems users face is the large volume of medical information that is generated each day, and the need to filter and retrieve relevant information. The Internet has dramatically increased the amount of electronically accessible medical information and reduced the cost and time needed to publish. The opportunity of the Internet for the medical profession and consumers is to have more information to make decisions and this could potentially lead to better medical decisions and outcomes. However, without the assistance from professional medical librarians, retrieving new and relevant information from databases and the Internet remains a challenge. Many physicians do not have access to the services of a medical librarian. Most physicians indicate on surveys that they do not prefer to retrieve the literature themselves, or visit libraries because of the lack of recent materials, poor organisation and indexing of materials, lack of appropriate and available material, and lack of time. The information filtering system described in this paper records the online web browsing behaviour of each user and creates a user profile of the index terms found on the web pages visited by the user. A relevance-ranking algorithm then matches the user profiles to the index terms of new health care web pages that are added each day. The system creates customised summaries of new information for each user. A user can then connect to the web site to read the new information. Relevance feedback buttons on each page ask the user to rate the usefulness of the page to their immediate information needs. Errors in relevance ranking are reduced in this system by having both the user profile and medical information represented in the same representation language using a controlled vocabulary. This system also updates the user profiles

  19. A Survey On Various Web Template Detection And Extraction Methods

    Directory of Open Access Journals (Sweden)

    Neethu Mary Varghese

    2015-03-01

    Full Text Available Abstract In todays digital world reliance on the World Wide Web as a source of information is extensive. Users increasingly rely on web based search engines to provide accurate search results on a wide range of topics that interest them. The search engines in turn parse the vast repository of web pages searching for relevant information. However majority of web portals are designed using web templates which are designed to provide consistent look and feel to end users. The presence of these templates however can influence search results leading to inaccurate results being delivered to the users. Therefore to improve the accuracy and reliability of search results identification and removal of web templates from the actual content is essential. A wide range of approaches are commonly employed to achieve this and this paper focuses on the study of the various approaches of template detection and extraction that can be applied across homogenous as well as heterogeneous web pages.

  20. Chemically extracted nanocellulose from sisal fibres by a simple and industrially relevant process

    DEFF Research Database (Denmark)

    Trifol Guzman, Jon; Sillard, Cecile; Plackett, D.

    2017-01-01

    product with a high solids content. An aqueous dispersion of CNF could be obtained directly from this intermediate pulp by simple magnetic stirring. As a proof of concept, the dispersion was used directly for preparing a highly translucent CNF film, illustrating that there are no large aggregates...... in the prepared CNF dispersion. Finally, CNF films with alkali extracts were also prepared, resulting in flatter films with an increased mass yield and improved mechanical strength....

  1. Pressurized Hot Water Extraction of anthocyanins from red onion: A study on extraction and degradation rates

    Energy Technology Data Exchange (ETDEWEB)

    Petersson, Erik V.; Liu Jiayin; Sjoeberg, Per J.R.; Danielsson, Rolf [Uppsala University, Department of Physical and Analytical Chemistry, P.O. Box 599, SE-751 24, Uppsala (Sweden); Turner, Charlotta, E-mail: Charlotta.Turner@kemi.uu.se [Uppsala University, Department of Physical and Analytical Chemistry, P.O. Box 599, SE-751 24, Uppsala (Sweden)

    2010-03-17

    Pressurized Hot Water Extraction (PHWE) is a quick, efficient and environmentally friendly technique for extractions. However, when using PHWE to extract thermally unstable analytes, extraction and degradation effects occur at the same time, and thereby compete. At first, the extraction effect dominates, but degradation effects soon take over. In this paper, extraction and degradation rates of anthocyanins from red onion were studied with experiments in a static batch reactor at 110 deg. C. A total extraction curve was calculated with data from the actual extraction and degradation curves, showing that more anthocyanins, 21-36% depending on the species, could be extracted if no degradation occurred, but then longer extraction times would be required than those needed to reach the peak level in the apparent extraction curves. The results give information about the different kinetic processes competing during an extraction procedure.

  2. Earnings Management, Value Relevance Of Earnings and Book Value of Equity

    OpenAIRE

    Subekti, Imam

    2013-01-01

    Previous studies examining relationship between earnings management and value relevance of accounting information show that earnings management decrease value relevance of accounting information. Generally, the studies apply accruals earnings management. In contrast, the present study applies integrated earnings management proxies i.e. real and accruals earnings manage-ment. Real earnings management proxies are measured by abnormal cash flow of operation, ab-normal production cost, and abnorm...

  3. EOG feature relevance determination for microsleep detection

    Directory of Open Access Journals (Sweden)

    Golz Martin

    2017-09-01

    Full Text Available Automatic relevance determination (ARD was applied to two-channel EOG recordings for microsleep event (MSE recognition. 10 s immediately before MSE and also before counterexamples of fatigued, but attentive driving were analysed. Two type of signal features were extracted: the maximum cross correlation (MaxCC and logarithmic power spectral densities (PSD averaged in spectral bands of 0.5 Hz width ranging between 0 and 8 Hz. Generalised learn-ing vector quantisation (GRLVQ was used as ARD method to show the potential of feature reduction. This is compared to support-vector machines (SVM, in which the feature reduction plays a much smaller role. Cross validation yielded mean normalised relevancies of PSD features in the range of 1.6 – 4.9 % and 1.9 – 10.4 % for horizontal and vertical EOG, respectively. MaxCC relevancies were 0.002 – 0.006 % and 0.002 – 0.06 %, respectively. This shows that PSD features of vertical EOG are indispensable, whereas MaxCC can be neglected. Mean classification accuracies were estimated at 86.6±b 1.3 % and 92.3±b 0.2 % for GRLVQ and SVM, respectively. GRLVQ permits objective feature reduction by inclusion of all processing stages, but is not as accurate as SVM.

  4. EOG feature relevance determination for microsleep detection

    Directory of Open Access Journals (Sweden)

    Golz Martin

    2017-09-01

    Full Text Available Automatic relevance determination (ARD was applied to two-channel EOG recordings for microsleep event (MSE recognition. 10 s immediately before MSE and also before counterexamples of fatigued, but attentive driving were analysed. Two type of signal features were extracted: the maximum cross correlation (MaxCC and logarithmic power spectral densities (PSD averaged in spectral bands of 0.5 Hz width ranging between 0 and 8 Hz. Generalised learn-ing vector quantisation (GRLVQ was used as ARD method to show the potential of feature reduction. This is compared to support-vector machines (SVM, in which the feature reduction plays a much smaller role. Cross validation yielded mean normalised relevancies of PSD features in the range of 1.6 - 4.9 % and 1.9 - 10.4 % for horizontal and vertical EOG, respectively. MaxCC relevancies were 0.002 - 0.006 % and 0.002 - 0.06 %, respectively. This shows that PSD features of vertical EOG are indispensable, whereas MaxCC can be neglected. Mean classification accuracies were estimated at 86.6±b 1.3 % and 92.3±b 0.2 % for GRLVQ and SVM, respec-tively. GRLVQ permits objective feature reduction by inclu-sion of all processing stages, but is not as accurate as SVM.

  5. High-Resolution Remote Sensing Image Building Extraction Based on Markov Model

    Science.gov (United States)

    Zhao, W.; Yan, L.; Chang, Y.; Gong, L.

    2018-04-01

    With the increase of resolution, remote sensing images have the characteristics of increased information load, increased noise, more complex feature geometry and texture information, which makes the extraction of building information more difficult. To solve this problem, this paper designs a high resolution remote sensing image building extraction method based on Markov model. This method introduces Contourlet domain map clustering and Markov model, captures and enhances the contour and texture information of high-resolution remote sensing image features in multiple directions, and further designs the spectral feature index that can characterize "pseudo-buildings" in the building area. Through the multi-scale segmentation and extraction of image features, the fine extraction from the building area to the building is realized. Experiments show that this method can restrain the noise of high-resolution remote sensing images, reduce the interference of non-target ground texture information, and remove the shadow, vegetation and other pseudo-building information, compared with the traditional pixel-level image information extraction, better performance in building extraction precision, accuracy and completeness.

  6. Computations Underlying Social Hierarchy Learning: Distinct Neural Mechanisms for Updating and Representing Self-Relevant Information.

    Science.gov (United States)

    Kumaran, Dharshan; Banino, Andrea; Blundell, Charles; Hassabis, Demis; Dayan, Peter

    2016-12-07

    Knowledge about social hierarchies organizes human behavior, yet we understand little about the underlying computations. Here we show that a Bayesian inference scheme, which tracks the power of individuals, better captures behavioral and neural data compared with a reinforcement learning model inspired by rating systems used in games such as chess. We provide evidence that the medial prefrontal cortex (MPFC) selectively mediates the updating of knowledge about one's own hierarchy, as opposed to that of another individual, a process that underpinned successful performance and involved functional interactions with the amygdala and hippocampus. In contrast, we observed domain-general coding of rank in the amygdala and hippocampus, even when the task did not require it. Our findings reveal the computations underlying a core aspect of social cognition and provide new evidence that self-relevant information may indeed be afforded a unique representational status in the brain. Copyright © 2016 The Authors. Published by Elsevier Inc. All rights reserved.

  7. A METHOD OF EXTRACTING SHORELINE BASED ON SEMANTIC INFORMATION USING DUAL-LENGTH LiDAR DATA

    Directory of Open Access Journals (Sweden)

    C. Yao

    2017-09-01

    Full Text Available Shoreline is a spatial varying separation between water and land. By utilizing dual-wavelength LiDAR point data together with semantic information that shoreline often appears beyond water surface profile and is observable on the beach, the paper generates the shoreline and the details are as follows: (1 Gain the water surface profile: first we obtain water surface by roughly selecting water points based on several features of water body, then apply least square fitting method to get the whole water trend surface. Then we get the ground surface connecting the under -water surface by both TIN progressive filtering method and surface interpolation method. After that, we have two fitting surfaces intersected to get water surface profile of the island. (2 Gain the sandy beach: we grid all points and select the water surface profile grids points as seeds, then extract sandy beach points based on eight-neighborhood method and features, then we get all sandy beaches. (3 Get the island shoreline: first we get the sandy beach shoreline based on intensity information, then we get a threshold value to distinguish wet area and dry area, therefore we get the shoreline of several sandy beaches. In some extent, the shoreline has the same height values within a small area, by using all the sandy shoreline points to fit a plane P, and the intersection line of the ground surface and the shoreline plane P can be regarded as the island shoreline. By comparing with the surveying shoreline, the results show that the proposed method can successfully extract shoreline.

  8. Quantify Water Extraction by TBP/Dodecane via Molecular Dynamics Simulations

    International Nuclear Information System (INIS)

    Khomami, Bamin; Cui, Shengting; De Almeida, Valmor F.

    2013-01-01

    The purpose of this project is to quantify the interfacial transport of water into the most prevalent nuclear reprocessing solvent extractant mixture, namely tri-butyl- phosphate (TBP) and dodecane, via massively parallel molecular dynamics simulations on the most powerful machines available for open research. Specifically, we will accomplish this objective by evolving the water/TBP/dodecane system up to 1 ms elapsed time, and validate the simulation results by direct comparison with experimentally measured water solubility in the organic phase. The significance of this effort is to demonstrate for the first time that the combination of emerging simulation tools and state-of-the-art supercomputers can provide quantitative information on par to experimental measurements for solvent extraction systems of relevance to the nuclear fuel cycle. Results: Initially, the isolated single component, and single phase systems were studied followed by the two-phase, multicomponent counterpart. Specifically, the systems we studied were: pure TBP; pure n-dodecane; TBP/n-dodecane mixture; and the complete extraction system: water-TBP/n-dodecane two phase system to gain deep insight into the water extraction process. We have completely achieved our goal of simulating the molecular extraction of water molecules into the TBP/n-dodecane mixture up to the saturation point, and obtained favorable comparison with experimental data. Many insights into fundamental molecular level processes and physics were obtained from the process. Most importantly, we found that the dipole moment of the extracting agent is crucially important in affecting the interface roughness and the extraction rate of water molecules into the organic phase. In addition, we have identified shortcomings in the existing OPLS-AA force field potential for long-chain alkanes. The significance of this force field is that it is supposed to be optimized for molecular liquid simulations. We found that it failed for dodecane and

  9. Development of an information retrieval tool for biomedical patents.

    Science.gov (United States)

    Alves, Tiago; Rodrigues, Rúben; Costa, Hugo; Rocha, Miguel

    2018-06-01

    The volume of biomedical literature has been increasing in the last years. Patent documents have also followed this trend, being important sources of biomedical knowledge, technical details and curated data, which are put together along the granting process. The field of Biomedical text mining (BioTM) has been creating solutions for the problems posed by the unstructured nature of natural language, which makes the search of information a challenging task. Several BioTM techniques can be applied to patents. From those, Information Retrieval (IR) includes processes where relevant data are obtained from collections of documents. In this work, the main goal was to build a patent pipeline addressing IR tasks over patent repositories to make these documents amenable to BioTM tasks. The pipeline was developed within @Note2, an open-source computational framework for BioTM, adding a number of modules to the core libraries, including patent metadata and full text retrieval, PDF to text conversion and optical character recognition. Also, user interfaces were developed for the main operations materialized in a new @Note2 plug-in. The integration of these tools in @Note2 opens opportunities to run BioTM tools over patent texts, including tasks from Information Extraction, such as Named Entity Recognition or Relation Extraction. We demonstrated the pipeline's main functions with a case study, using an available benchmark dataset from BioCreative challenges. Also, we show the use of the plug-in with a user query related to the production of vanillin. This work makes available all the relevant content from patents to the scientific community, decreasing drastically the time required for this task, and provides graphical interfaces to ease the use of these tools. Copyright © 2018 Elsevier B.V. All rights reserved.

  10. Information Extraction From Chemical Patents

    Directory of Open Access Journals (Sweden)

    Sandra Bergmann

    2012-01-01

    Full Text Available The development of new chemicals or pharmaceuticals is preceded by an indepth analysis of published patents in this field. This information retrieval is a costly and time inefficient step when done by a human reader, yet it is mandatory for potential success of an investment. The goal of the research project UIMA-HPC is to automate and hence speed-up the process of knowledge mining about patents. Multi-threaded analysis engines, developed according to UIMA (Unstructured Information Management Architecture standards, process texts and images in thousands of documents in parallel. UNICORE (UNiform Interface to COmputing Resources workflow control structures make it possible to dynamically allocate resources for every given task to gain best cpu-time/realtime ratios in an HPC environment.

  11. Phenolic Extracts from Clerodendrum volubile Leaves Inhibit Cholinergic and Monoaminergic Enzymes Relevant to the Management of Some Neurodegenerative Diseases.

    Science.gov (United States)

    Oboh, Ganiyu; Ogunruku, Omodesola O; Oyeleye, Sunday I; Olasehinde, Tosin A; Ademosun, Ayokunle O; Boligon, Aline Augusti

    2017-05-04

    This study investigated the inhibitory effects of phenolic-rich extracts from Clerodendrum volubile leaves on cholinergic [acetylcholinesterase (AChE) and butyrylcholinesterase (BChE)] and monoaminergic [monoamine oxidase (MAO)] enzymes' activities and pro-oxidants [Fe 2+ and quinolinic acid-(QA)] induced lipid peroxidation in rats brain homogenates in vitro. Free phenolic extracts (FPE) and bound phenolic extracts (BPE) were obtained via solvent extraction, and the total phenol and flavonoid contents were evaluated. The phenolic constituents of the extracts were also determined using high performance liquid chromatography coupled with diode array detector (HPLC-DAD). Our findings revealed that FPE had higher AChE (2.06 μg/mL), BChE (2.79 μg/mL), and MAO (2.81 μg/mL) inhibitory effects than BPE [AChE, 2.80 μg/mL; BChE, 3.40 μg/mL; MAO, 3.39 μg/mL]. Furthermore, FPE also had significantly (P rich extracts from C. volubile could be part of the mechanism of actions behind its use for memory/cognitive function as obtained in folklore. However, FPE exhibited significantly higher enzymes, inhibitory and antioxidant potentials than BPE.

  12. Extraction Methods for the Isolation of Isoflavonoids from Plant Material

    Directory of Open Access Journals (Sweden)

    Blicharski Tomasz

    2017-03-01

    Full Text Available The purpose of this review is to describe and compare selected traditional and modern extraction methods employed in the isolation of isoflavonoids from plants. Conventional methods such as maceration, percolation, or Soxhlet extraction are still frequently used in phytochemical analysis. Despite their flexibility, traditional extraction techniques have significant drawbacks, including the need for a significant investment of time, energy, and starting material, and a requirement for large amounts of potentially toxic solvents. Moreover, these techniques are difficult to automate, produce considerable amount of waste and pose a risk of degradation of thermolabile compounds. Modern extraction methods, such as: ultrasound-assisted extraction, microwave-assisted extraction, accelerated solvent extraction, supercritical fluid extraction, and negative pressure cavitation extraction, can be regarded as remedies for the aforementioned problems. This manuscript discusses the use of the most relevant extraction techniques in the process of isolation of isoflavonoids, secondary metabolites that have been found to have a plethora of biological and pharmacological activities.

  13. NetNorM: Capturing cancer-relevant information in somatic exome mutation data with gene networks for cancer stratification and prognosis.

    Science.gov (United States)

    Le Morvan, Marine; Zinovyev, Andrei; Vert, Jean-Philippe

    2017-06-01

    Genome-wide somatic mutation profiles of tumours can now be assessed efficiently and promise to move precision medicine forward. Statistical analysis of mutation profiles is however challenging due to the low frequency of most mutations, the varying mutation rates across tumours, and the presence of a majority of passenger events that hide the contribution of driver events. Here we propose a method, NetNorM, to represent whole-exome somatic mutation data in a form that enhances cancer-relevant information using a gene network as background knowledge. We evaluate its relevance for two tasks: survival prediction and unsupervised patient stratification. Using data from 8 cancer types from The Cancer Genome Atlas (TCGA), we show that it improves over the raw binary mutation data and network diffusion for these two tasks. In doing so, we also provide a thorough assessment of somatic mutations prognostic power which has been overlooked by previous studies because of the sparse and binary nature of mutations.

  14. An innovative method for extracting isotopic information from low-resolution gamma spectra

    International Nuclear Information System (INIS)

    Miko, D.; Estep, R.J.; Rawool-Sullivan, M.W.

    1998-01-01

    A method is described for the extraction of isotopic information from attenuated gamma ray spectra using the gross-count material basis set (GC-MBS) model. This method solves for the isotopic composition of an unknown mixture of isotopes attenuated through an absorber of unknown material. For binary isotopic combinations the problem is nonlinear in only one variable and is easily solved using standard line optimization techniques. Results are presented for NaI spectrum analyses of various binary combinations of enriched uranium, depleted uranium, low burnup Pu, 137 Cs, and 133 Ba attenuated through a suite of absorbers ranging in Z from polyethylene through lead. The GC-MBS method results are compared to those computed using ordinary response function fitting and with a simple net peak area method. The GC-MBS method was found to be significantly more accurate than the other methods over the range of absorbers and isotopic blends studied

  15. Comparison of extraction techniques for isolation of steroid oestrogens in environmentally relevant concentrations from sediment

    Czech Academy of Sciences Publication Activity Database

    Sadílek, Jan; Spálovská, P.; Vrana, B.; Vávrová, M.; Maršálek, Blahoslav; Šimek, Z.

    2016-01-01

    Roč. 96, č. 11 (2016), s. 1022-1037 ISSN 0306-7319 Institutional support: RVO:67985939 Keywords : microwave - assisted extraction * estrogens * sediment Subject RIV: CB - Analytical Chemistry , Separation Impact factor: 1.208, year: 2016

  16. [Mood-congruent effect in self-relevant information processing: a study using an autobiographical memory recall task].

    Science.gov (United States)

    Itoh, M

    2000-10-01

    The pattern of the mood-congruent effect in an autobiographical memory recall task was investigated. Each subject was randomly assigned to one of three experimental conditions: positive mood, negative mood (induced with music), and control groups (no specific mood). Subjects were then presented with a word at a time from a list of trait words, which were pleasant or unpleasant. They decided whether they could recall any of their autobiographical memories related to the word, and responded with "yes" or "no" buttons as rapidly and accurately as possible. After the task, they were given five minutes for an incidental free recall test. Results indicated that the mood-congruent effect was found regardless of whether there was an autobiographical memory related to the word or not in both positive and negative mood states. The effect of moods on self-relevant information processing was discussed.

  17. Passage relevance models for genomics search

    Directory of Open Access Journals (Sweden)

    Frieder Ophir

    2009-03-01

    Full Text Available Abstract We present a passage relevance model for integrating syntactic and semantic evidence of biomedical concepts and topics using a probabilistic graphical model. Component models of topics, concepts, terms, and document are represented as potential functions within a Markov Random Field. The probability of a passage being relevant to a biologist's information need is represented as the joint distribution across all potential functions. Relevance model feedback of top ranked passages is used to improve distributional estimates of query concepts and topics in context, and a dimensional indexing strategy is used for efficient aggregation of concept and term statistics. By integrating multiple sources of evidence including dependencies between topics, concepts, and terms, we seek to improve genomics literature passage retrieval precision. Using this model, we are able to demonstrate statistically significant improvements in retrieval precision using a large genomics literature corpus.

  18. Health Information Infrastructure for People with Intellectual and Developmental Disabilities (I/DD) Living in Supported Accommodation: Communication, Co-Ordination and Integration of Health Information.

    Science.gov (United States)

    Dahm, Maria R; Georgiou, Andrew; Balandin, Susan; Hill, Sophie; Hemsley, Bronwyn

    2017-10-25

    People with intellectual and/or developmental disability (I/DD) commonly have complex health care needs, but little is known about how their health information is managed in supported accommodation, and across health services providers. This study aimed to describe the current health information infrastructure (i.e., how data and information are collected, stored, communicated, and used) for people with I/DD living in supported accommodation in Australia. It involved a scoping review and synthesis of research, policies, and health documents relevant in this setting. Iterative database and hand searches were conducted across peer-reviewed articles internationally in English and grey literature in Australia (New South Wales) up to September 2015. Data were extracted from the selected relevant literature and analyzed for content themes. Expert stakeholders were consulted to verify the authors' interpretations of the information and content categories. The included 286 sources (peer-reviewed n = 27; grey literature n = 259) reflect that the health information for people with I/DD in supported accommodation is poorly communicated, coordinated and integrated across isolated systems. 'Work-as-imagined' as outlined in policies, does not align with 'work-as-done' in reality. This gap threatens the quality of care and safety of people with I/DD in these settings. The effectiveness of the health information infrastructure and services for people with I/DD can be improved by integrating the information sources and placing people with I/DD and their supporters at the centre of the information exchange process.

  19. Real-time hypothesis driven feature extraction on parallel processing architectures

    DEFF Research Database (Denmark)

    Granmo, O.-C.; Jensen, Finn Verner

    2002-01-01

    the problem of higher-order feature-content/feature-feature correlation, causally complexly interacting features are identified through Bayesian network d-separation analysis and combined into joint features. When used on a moderately complex object-tracking case, the technique is able to select...... extraction, which selectively extract relevant features one-by-one, have in some cases achieved real-time performance on single processing element architectures. In this paperwe propose a novel technique which combines the above two approaches. Features are selectively extracted in parallelizable sets...

  20. EnvMine: A text-mining system for the automatic extraction of contextual information

    Directory of Open Access Journals (Sweden)

    de Lorenzo Victor

    2010-06-01

    Full Text Available Abstract Background For ecological studies, it is crucial to count on adequate descriptions of the environments and samples being studied. Such a description must be done in terms of their physicochemical characteristics, allowing a direct comparison between different environments that would be difficult to do otherwise. Also the characterization must include the precise geographical location, to make possible the study of geographical distributions and biogeographical patterns. Currently, there is no schema for annotating these environmental features, and these data have to be extracted from textual sources (published articles. So far, this had to be performed by manual inspection of the corresponding documents. To facilitate this task, we have developed EnvMine, a set of text-mining tools devoted to retrieve contextual information (physicochemical variables and geographical locations from textual sources of any kind. Results EnvMine is capable of retrieving the physicochemical variables cited in the text, by means of the accurate identification of their associated units of measurement. In this task, the system achieves a recall (percentage of items retrieved of 92% with less than 1% error. Also a Bayesian classifier was tested for distinguishing parts of the text describing environmental characteristics from others dealing with, for instance, experimental settings. Regarding the identification of geographical locations, the system takes advantage of existing databases such as GeoNames to achieve 86% recall with 92% precision. The identification of a location includes also the determination of its exact coordinates (latitude and longitude, thus allowing the calculation of distance between the individual locations. Conclusion EnvMine is a very efficient method for extracting contextual information from different text sources, like published articles or web pages. This tool can help in determining the precise location and physicochemical

  1. the relevance of libraries and information communicaton technology

    African Journals Online (AJOL)

    GRACE

    Information and Communications Technology (ICT) could be used to improve .... accountable, efficient and effective interaction between the public, business and ... agencies, research institutions and private organizations, such as print and ...

  2. Mining of the social network extraction

    Science.gov (United States)

    Nasution, M. K. M.; Hardi, M.; Syah, R.

    2017-01-01

    The use of Web as social media is steadily gaining ground in the study of social actor behaviour. However, information in Web can be interpreted in accordance with the ability of the method such as superficial methods for extracting social networks. Each method however has features and drawbacks: it cannot reveal the behaviour of social actors, but it has the hidden information about them. Therefore, this paper aims to reveal such information in the social networks mining. Social behaviour could be expressed through a set of words extracted from the list of snippets.

  3. Developing a Process Model for the Forensic Extraction of Information from Desktop Search Applications

    Directory of Open Access Journals (Sweden)

    Timothy Pavlic

    2008-03-01

    Full Text Available Desktop search applications can contain cached copies of files that were deleted from the file system. Forensic investigators see this as a potential source of evidence, as documents deleted by suspects may still exist in the cache. Whilst there have been attempts at recovering data collected by desktop search applications, there is no methodology governing the process, nor discussion on the most appropriate means to do so. This article seeks to address this issue by developing a process model that can be applied when developing an information extraction application for desktop search applications, discussing preferred methods and the limitations of each. This work represents a more structured approach than other forms of current research.

  4. An evidence perspective on topical relevance types and its implications for exploratory and task-based retrieval

    Directory of Open Access Journals (Sweden)

    Xiaoli Huang

    2006-01-01

    Full Text Available Introduction. The concept of relevance lies at the heart of intellectual access and information retrieval, indeed of reasoning and communication in general; in turn, topical relevance lies at the heart of relevance. The common view of topical relevance is limited to topic matching, resulting in information retrieval systems' failure to detect more complex topical connections which are needed to respond to diversified user situations and tasks. Method. Based on the role a piece of information plays in the overall structure of an argument, we have identified four topical relevance types: Direct, Indirect (circumstantial, Context, and Comparison. In the process of creating a speech retrieval test collection, graduate history students made 27,000 topical relevance assessments between Holocaust survivor interview segments and real user topics, using the four relevance types, each on a scale of 0 to 4. They recorded justifications for their assessments and kept detailed Topic Notes. Analysis. We analysed these relevance assessments using a grounded theory approach to arrive at a finer classification of topical relevance types. Results. For example, indirect relevance(a piece of information is connected to the topic indirectly through inference, circumstantial evidence was refined to Generic Indirect Relevance, Backward Inference (abduction, Forward Inference (deduction, and Inference from Cases (induction, with each subtype being further illustrated and explicated by examples. Conclusion. Each of these refined types of topical relevance plays a special role in reasoning, making a conclusive argument, or performing a task. Incorporating them into information retrieval systems allows users more flexibility and a better focus on their tasks. They can also be used in teaching reasoning skills.

  5. Smart Extraction and Analysis System for Clinical Research.

    Science.gov (United States)

    Afzal, Muhammad; Hussain, Maqbool; Khan, Wajahat Ali; Ali, Taqdir; Jamshed, Arif; Lee, Sungyoung

    2017-05-01

    With the increasing use of electronic health records (EHRs), there is a growing need to expand the utilization of EHR data to support clinical research. The key challenge in achieving this goal is the unavailability of smart systems and methods to overcome the issue of data preparation, structuring, and sharing for smooth clinical research. We developed a robust analysis system called the smart extraction and analysis system (SEAS) that consists of two subsystems: (1) the information extraction system (IES), for extracting information from clinical documents, and (2) the survival analysis system (SAS), for a descriptive and predictive analysis to compile the survival statistics and predict the future chance of survivability. The IES subsystem is based on a novel permutation-based pattern recognition method that extracts information from unstructured clinical documents. Similarly, the SAS subsystem is based on a classification and regression tree (CART)-based prediction model for survival analysis. SEAS is evaluated and validated on a real-world case study of head and neck cancer. The overall information extraction accuracy of the system for semistructured text is recorded at 99%, while that for unstructured text is 97%. Furthermore, the automated, unstructured information extraction has reduced the average time spent on manual data entry by 75%, without compromising the accuracy of the system. Moreover, around 88% of patients are found in a terminal or dead state for the highest clinical stage of disease (level IV). Similarly, there is an ∼36% probability of a patient being alive if at least one of the lifestyle risk factors was positive. We presented our work on the development of SEAS to replace costly and time-consuming manual methods with smart automatic extraction of information and survival prediction methods. SEAS has reduced the time and energy of human resources spent unnecessarily on manual tasks.

  6. Sequence complexity and work extraction

    International Nuclear Information System (INIS)

    Merhav, Neri

    2015-01-01

    We consider a simplified version of a solvable model by Mandal and Jarzynski, which constructively demonstrates the interplay between work extraction and the increase of the Shannon entropy of an information reservoir which is in contact with a physical system. We extend Mandal and Jarzynski’s main findings in several directions: first, we allow sequences of correlated bits rather than just independent bits. Secondly, at least for the case of binary information, we show that, in fact, the Shannon entropy is only one measure of complexity of the information that must increase in order for work to be extracted. The extracted work can also be upper bounded in terms of the increase in other quantities that measure complexity, like the predictability of future bits from past ones. Third, we provide an extension to the case of non-binary information (i.e. a larger alphabet), and finally, we extend the scope to the case where the incoming bits (before the interaction) form an individual sequence, rather than a random one. In this case, the entropy before the interaction can be replaced by the Lempel–Ziv (LZ) complexity of the incoming sequence, a fact that gives rise to an entropic meaning of the LZ complexity, not only in information theory, but also in physics. (paper)

  7. Fixed kernel regression for voltammogram feature extraction

    International Nuclear Information System (INIS)

    Acevedo Rodriguez, F J; López-Sastre, R J; Gil-Jiménez, P; Maldonado Bascón, S; Ruiz-Reyes, N

    2009-01-01

    Cyclic voltammetry is an electroanalytical technique for obtaining information about substances under analysis without the need for complex flow systems. However, classifying the information in voltammograms obtained using this technique is difficult. In this paper, we propose the use of fixed kernel regression as a method for extracting features from these voltammograms, reducing the information to a few coefficients. The proposed approach has been applied to a wine classification problem with accuracy rates of over 98%. Although the method is described here for extracting voltammogram information, it can be used for other types of signals

  8. Extraction of indirectly captured information for use in a comparison of offline pH measurement technologies.

    Science.gov (United States)

    Ritchie, Elspeth K; Martin, Elaine B; Racher, Andy; Jaques, Colin

    2017-06-10

    Understanding the causes of discrepancies in pH readings of a sample can allow more robust pH control strategies to be implemented. It was found that 59.4% of differences between two offline pH measurement technologies for an historical dataset lay outside an expected instrument error range of ±0.02pH. A new variable, Osmo Res , was created using multiple linear regression (MLR) to extract information indirectly captured in the recorded measurements for osmolality. Principal component analysis and time series analysis were used to validate the expansion of the historical dataset with the new variable Osmo Res . MLR was used to identify variables strongly correlated (p<0.05) with differences in pH readings by the two offline pH measurement technologies. These included concentrations of specific chemicals (e.g. glucose) and Osmo Res, indicating culture medium and bolus feed additions as possible causes of discrepancies between the offline pH measurement technologies. Temperature was also identified as statistically significant. It is suggested that this was a result of differences in pH-temperature compensations employed by the pH measurement technologies. In summary, a method for extracting indirectly captured information has been demonstrated, and it has been shown that competing pH measurement technologies were not necessarily interchangeable at the desired level of control (±0.02pH). Copyright © 2017 Elsevier B.V. All rights reserved.

  9. Tag-Based Social Image Search: Toward Relevant and Diverse Results

    Science.gov (United States)

    Yang, Kuiyuan; Wang, Meng; Hua, Xian-Sheng; Zhang, Hong-Jiang

    Recent years have witnessed a great success of social media websites. Tag-based image search is an important approach to access the image content of interest on these websites. However, the existing ranking methods for tag-based image search frequently return results that are irrelevant or lack of diversity. This chapter presents a diverse relevance ranking scheme which simultaneously takes relevance and diversity into account by exploring the content of images and their associated tags. First, it estimates the relevance scores of images with respect to the query term based on both visual information of images and semantic information of associated tags. Then semantic similarities of social images are estimated based on their tags. Based on the relevance scores and the similarities, the ranking list is generated by a greedy ordering algorithm which optimizes Average Diverse Precision (ADP), a novel measure that is extended from the conventional Average Precision (AP). Comprehensive experiments and user studies demonstrate the effectiveness of the approach.

  10. Enhancing Public Access to Relevant and Valued Medical Information: Fresh Directions for RadiologyInfo.org.

    Science.gov (United States)

    Rubin, Geoffrey D; Krishnaraj, Arun; Mahesh, Mahadevappa; Rajendran, Ramji R; Fishman, Elliot K

    2017-05-01

    RadiologyInfo.org is a public information portal designed to support patient care and broaden public awareness of the essential role radiology plays in overall patient health care. Over the past 14 years, RadiologyInfo.org has evolved considerably to provide access to more than 220 mixed-media descriptions of tests, treatments, and diseases through a spectrum of mobile and desktop platforms, social media, and downloadable documents in both English and Spanish. In 2014, the RSNA-ACR Public Information Website Committee, which stewards RadiologyInfo.org, developed 3- to 5-year strategic and implementation plans for the website. The process was informed by RadiologyInfo.org user surveys, formal stakeholder interviews, focus groups, and usability testing. Metrics were established as key performance indicators to assess progress toward the stated goals of (1) optimizing content to enhance patient-centeredness, (2) enhancing reach and engagement, and (3) maintaining sustainability. Major changes resulting from this process include a complete redesign of the website, the replacement of text-rich PowerPoint presentations with conversational videos, and the development of an affiliate network. Over the past year, visits to RadiologyInfo.org have increased by 60.27% to 1,424,523 in August 2016 from 235 countries and territories. Twenty-two organizations have affiliated with RadiologyInfo.org with new organizations being added on a monthly basis. RadiologyInfo provides a tangible demonstration of how radiologists can engage directly with the global public to educate them on the value of radiology in their health care and to allay concerns and dispel misconceptions. Regular self-assessment and responsive planning will ensure its continued growth and relevance. Copyright © 2016 American College of Radiology. Published by Elsevier Inc. All rights reserved.

  11. Extraction and analysis of reducing alteration information of oil-gas in Bashibulake uranium ore district based on ASTER remote sensing data

    International Nuclear Information System (INIS)

    Ye Fawang; Liu Dechang; Zhao Yingjun; Yang Xu

    2008-01-01

    Beginning with the analysis of the spectral characteristics of sandstone with reducing alteration of oil-gas in Bashibulake ore district, the extract technology of reducing alteration information based on ASTER data is presented. Several remote sensing anomaly zones of reducing alteration information similar with that in uranium deposit are interpreted in study area. On the basis of above study, these alteration anomaly information are further classified by using the advantage of ASTER data with multi-band in SWIR, the geological significance for alteration anomaly information is respectively discussed. As a result, alteration anomalies good for uranium prospecting are really selected, which provides some important information for uranium exploration in outland of Bashibulake uranium ore area. (authors)

  12. The Implementation of Cosine Similarity to Calculate Text Relevance between Two Documents

    Science.gov (United States)

    Gunawan, D.; Sembiring, C. A.; Budiman, M. A.

    2018-03-01

    Rapidly increasing number of web pages or documents leads to topic specific filtering in order to find web pages or documents efficiently. This is a preliminary research that uses cosine similarity to implement text relevance in order to find topic specific document. This research is divided into three parts. The first part is text-preprocessing. In this part, the punctuation in a document will be removed, then convert the document to lower case, implement stop word removal and then extracting the root word by using Porter Stemming algorithm. The second part is keywords weighting. Keyword weighting will be used by the next part, the text relevance calculation. Text relevance calculation will result the value between 0 and 1. The closer value to 1, then both documents are more related, vice versa.

  13. Information Extraction of Tourist Geological Resources Based on 3d Visualization Remote Sensing Image

    Science.gov (United States)

    Wang, X.

    2018-04-01

    Tourism geological resources are of high value in admiration, scientific research and universal education, which need to be protected and rationally utilized. In the past, most of the remote sensing investigations of tourism geological resources used two-dimensional remote sensing interpretation method, which made it difficult for some geological heritages to be interpreted and led to the omission of some information. This aim of this paper is to assess the value of a method using the three-dimensional visual remote sensing image to extract information of geological heritages. skyline software system is applied to fuse the 0.36 m aerial images and 5m interval DEM to establish the digital earth model. Based on the three-dimensional shape, color tone, shadow, texture and other image features, the distribution of tourism geological resources in Shandong Province and the location of geological heritage sites were obtained, such as geological structure, DaiGu landform, granite landform, Volcanic landform, sandy landform, Waterscapes, etc. The results show that using this method for remote sensing interpretation is highly recognizable, making the interpretation more accurate and comprehensive.

  14. Information technology portfolio in supply chain management using factor analysis

    Directory of Open Access Journals (Sweden)

    Ahmad Jaafarnejad

    2013-11-01

    Full Text Available The adoption of information technology (IT along with supply chain management (SCM has become increasingly a necessity among most businesses. This enhances supply chain (SC performance and helps companies achieve the organizational competitiveness. IT systems capture and analyze information and enable management to make decisions by considering a global scope across the entire SC. This paper reviews the existing literature on IT in SCM and considers pertinent criteria. Using principal component analysis (PCA of factor analysis (FA, a number of related criteria are divided into smaller groups. Finally, SC managers can develop an IT portfolio in SCM using mean values of few extracted components on the relevance –emergency matrix. A numerical example is provided to explain details of the proposed method.

  15. Addressing Information Proliferation: Applications of Information Extraction and Text Mining

    Science.gov (United States)

    Li, Jingjing

    2013-01-01

    The advent of the Internet and the ever-increasing capacity of storage media have made it easy to store, deliver, and share enormous volumes of data, leading to a proliferation of information on the Web, in online libraries, on news wires, and almost everywhere in our daily lives. Since our ability to process and absorb this information remains…

  16. Roles of libraries in information society | Awojobi | Information ...

    African Journals Online (AJOL)

    Through the revi ew of relevant literature, the paper examined the roles of libraries in information society. The information society is characterised by proliferation of information through the use of Information and Communication Technology (ICT). One major principle of an information society is accessibility to information ...

  17. Mutual information, neural networks and the renormalization group

    Science.gov (United States)

    Koch-Janusz, Maciej; Ringel, Zohar

    2018-06-01

    Physical systems differing in their microscopic details often display strikingly similar behaviour when probed at macroscopic scales. Those universal properties, largely determining their physical characteristics, are revealed by the powerful renormalization group (RG) procedure, which systematically retains `slow' degrees of freedom and integrates out the rest. However, the important degrees of freedom may be difficult to identify. Here we demonstrate a machine-learning algorithm capable of identifying the relevant degrees of freedom and executing RG steps iteratively without any prior knowledge about the system. We introduce an artificial neural network based on a model-independent, information-theoretic characterization of a real-space RG procedure, which performs this task. We apply the algorithm to classical statistical physics problems in one and two dimensions. We demonstrate RG flow and extract the Ising critical exponent. Our results demonstrate that machine-learning techniques can extract abstract physical concepts and consequently become an integral part of theory- and model-building.

  18. Quantum measurement information as a key to energy extraction from local vacuums

    International Nuclear Information System (INIS)

    Hotta, Masahiro

    2008-01-01

    In this paper, a protocol is proposed in which energy extraction from local vacuum states is possible by using quantum measurement information for the vacuum state of quantum fields. In the protocol, Alice, who stays at a spatial point, excites the ground state of the fields by a local measurement. Consequently, wave packets generated by Alice's measurement propagate the vacuum to spatial infinity. Let us assume that Bob stays away from Alice and fails to catch the excitation energy when the wave packets pass in front of him. Next Alice announces her local measurement result to Bob by classical communication. Bob performs a local unitary operation depending on the measurement result. In this process, positive energy is released from the fields to Bob's apparatus of the unitary operation. In the field systems, wave packets are generated with negative energy around Bob's location. Soon afterwards, the negative-energy wave packets begin to chase after the positive-energy wave packets generated by Alice and form loosely bound states.

  19. Prominent feature extraction for review analysis: an empirical study

    Science.gov (United States)

    Agarwal, Basant; Mittal, Namita

    2016-05-01

    Sentiment analysis (SA) research has increased tremendously in recent times. SA aims to determine the sentiment orientation of a given text into positive or negative polarity. Motivation for SA research is the need for the industry to know the opinion of the users about their product from online portals, blogs, discussion boards and reviews and so on. Efficient features need to be extracted for machine-learning algorithm for better sentiment classification. In this paper, initially various features are extracted such as unigrams, bi-grams and dependency features from the text. In addition, new bi-tagged features are also extracted that conform to predefined part-of-speech patterns. Furthermore, various composite features are created using these features. Information gain (IG) and minimum redundancy maximum relevancy (mRMR) feature selection methods are used to eliminate the noisy and irrelevant features from the feature vector. Finally, machine-learning algorithms are used for classifying the review document into positive or negative class. Effects of different categories of features are investigated on four standard data-sets, namely, movie review and product (book, DVD and electronics) review data-sets. Experimental results show that composite features created from prominent features of unigram and bi-tagged features perform better than other features for sentiment classification. mRMR is a better feature selection method as compared with IG for sentiment classification. Boolean Multinomial Naïve Bayes) algorithm performs better than support vector machine classifier for SA in terms of accuracy and execution time.

  20. Social relevance enhances memory for impressions in older adults.

    Science.gov (United States)

    Cassidy, Brittany S; Gutchess, Angela H

    2012-01-01

    Previous research has demonstrated that older adults have difficulty retrieving contextual material over items alone. Recent research suggests this deficit can be reduced by adding emotional context, allowing for the possibility that memory for social impressions may show less age-related decline than memory for other types of contextual information. Two studies investigated how orienting to social or self-relevant aspects of information contributed to the learning and retrieval of impressions in young and older adults. Participants encoded impressions of others in conditions varying in the use of self-reference (Experiment 1) and interpersonal meaningfulness (Experiment 2), and completed memory tasks requiring the retrieval of specific traits. For both experiments, age groups remembered similar numbers of impressions. In Experiment 1 using more self-relevant encoding contexts increased memory for impressions over orienting to stimuli in a non-social way, regardless of age. In Experiment 2 older adults had enhanced memory for impressions presented in an interpersonally meaningful relative to a personally irrelevant way, whereas young adults were unaffected by this manipulation. The results provide evidence that increasing social relevance ameliorates age differences in memory for impressions, and enhances older adults' ability to successfully retrieve contextual information.

  1. Workplace Learning - How We Keep Track of Relevant Information

    OpenAIRE

    Bischoff, Kerstin; Herder, Eelco; Nejdl, Wolfgang

    2007-01-01

    At the workplace, learning is often a by-product of working on complex projects, requiring self-steered, need-driven and goal-oriented retrieval of information just in time from documents or peers. The personal desktop provides one rich source for learning material and for adaptation of learning resources. Data within that personal information space enables learning from previous experience, sharing tacit and explicit knowledge, and allows for establishing context and context-aware delivery o...

  2. Relevant test set using feature selection algorithm for early detection ...

    African Journals Online (AJOL)

    The objective of feature selection is to find the most relevant features for classification. Thus, the dimensionality of the information will be reduced and may improve classification's accuracy. This paper proposed a minimum set of relevant questions that can be used for early detection of dyslexia. In this research, we ...

  3. DEXTER: Disease-Expression Relation Extraction from Text.

    Science.gov (United States)

    Gupta, Samir; Dingerdissen, Hayley; Ross, Karen E; Hu, Yu; Wu, Cathy H; Mazumder, Raja; Vijay-Shanker, K

    2018-01-01

    Gene expression levels affect biological processes and play a key role in many diseases. Characterizing expression profiles is useful for clinical research, and diagnostics and prognostics of diseases. There are currently several high-quality databases that capture gene expression information, obtained mostly from large-scale studies, such as microarray and next-generation sequencing technologies, in the context of disease. The scientific literature is another rich source of information on gene expression-disease relationships that not only have been captured from large-scale studies but have also been observed in thousands of small-scale studies. Expression information obtained from literature through manual curation can extend expression databases. While many of the existing databases include information from literature, they are limited by the time-consuming nature of manual curation and have difficulty keeping up with the explosion of publications in the biomedical field. In this work, we describe an automated text-mining tool, Disease-Expression Relation Extraction from Text (DEXTER) to extract information from literature on gene and microRNA expression in the context of disease. One of the motivations in developing DEXTER was to extend the BioXpress database, a cancer-focused gene expression database that includes data derived from large-scale experiments and manual curation of publications. The literature-based portion of BioXpress lags behind significantly compared to expression information obtained from large-scale studies and can benefit from our text-mined results. We have conducted two different evaluations to measure the accuracy of our text-mining tool and achieved average F-scores of 88.51 and 81.81% for the two evaluations, respectively. Also, to demonstrate the ability to extract rich expression information in different disease-related scenarios, we used DEXTER to extract information on differential expression information for 2024 genes in lung

  4. About increasing informativity of diagnostic system of asynchronous electric motor by extracting additional information from values of consumed current parameter

    Science.gov (United States)

    Zhukovskiy, Y.; Korolev, N.; Koteleva, N.

    2018-05-01

    This article is devoted to expanding the possibilities of assessing the technical state of the current consumption of asynchronous electric drives, as well as increasing the information capacity of diagnostic methods, in conditions of limited access to equipment and incompleteness of information. The method of spectral analysis of the electric drive current can be supplemented by an analysis of the components of the current of the Park's vector. The research of the hodograph evolution in the moment of appearance and development of defects was carried out using the example of current asymmetry in the phases of an induction motor. The result of the study is the new diagnostic parameters of the asynchronous electric drive. During the research, it was proved that the proposed diagnostic parameters allow determining the type and level of the defect. At the same time, there is no need to stop the equipment and taky it out of service for repair. Modern digital control and monitoring systems can use the proposed parameters based on the stator current of an electrical machine to improve the accuracy and reliability of obtaining diagnostic patterns and predicting their changes in order to improve the equipment maintenance systems. This approach can also be used in systems and objects where there are significant parasitic vibrations and unsteady loads. The extraction of useful information can be carried out in electric drive systems in the structure of which there is a power electric converter.

  5. Efficient Extraction of Content from Enriched Geospatial and Networked Data

    DEFF Research Database (Denmark)

    Qu, Qiang

    forth, which makes it possible to extract relevant and interesting information that can then be utilized in different applications. However, web content is often semantically rich, structurally complex, and highly dynamic. This dissertation addresses some of the challenges posed by the use of such data......Social network services such as Google Places and Twitter have led to a proliferation of user-generated web content that is constantly shared among users. These services enable access to various types of content, covering geospatial locations, textual descriptions, social relationships, and so...... by merging edges and nodes in the original graph. Generalized, compressed graphs provide a way to interpret large networks. The dissertation reports on studies that compare the proposed solutions with respect to their tradeoffs between result complexity and quality. The findings suggest that the solutions...

  6. GeoDeepDive: Towards a Machine Reading-Ready Digital Library and Information Integration Resource

    Science.gov (United States)

    Husson, J. M.; Peters, S. E.; Livny, M.; Ross, I.

    2015-12-01

    Recent developments in machine reading and learning approaches to text and data mining hold considerable promise for accelerating the pace and quality of literature-based data synthesis, but these advances have outpaced even basic levels of access to the published literature. For many geoscience domains, particularly those based on physical samples and field-based descriptions, this limitation is significant. Here we describe a general infrastructure to support published literature-based machine reading and learning approaches to information integration and knowledge base creation. This infrastructure supports rate-controlled automated fetching of original documents, along with full bibliographic citation metadata, from remote servers, the secure storage of original documents, and the utilization of considerable high-throughput computing resources for the pre-processing of these documents by optical character recognition, natural language parsing, and other document annotation and parsing software tools. New tools and versions of existing tools can be automatically deployed against original documents when they are made available. The products of these tools (text/XML files) are managed by MongoDB and are available for use in data extraction applications. Basic search and discovery functionality is provided by ElasticSearch, which is used to identify documents of potential relevance to a given data extraction task. Relevant files derived from the original documents are then combined into basic starting points for application building; these starting points are kept up-to-date as new relevant documents are incorporated into the digital library. Currently, our digital library stores contains more than 360K documents supplied by Elsevier and the USGS and we are actively seeking additional content providers. By focusing on building a dependable infrastructure to support the retrieval, storage, and pre-processing of published content, we are establishing a foundation for

  7. Interventions to assist health consumers to find reliable online health information: a comprehensive review.

    Directory of Open Access Journals (Sweden)

    Kenneth Lee

    Full Text Available BACKGROUND: Health information on the Internet is ubiquitous, and its use by health consumers prevalent. Finding and understanding relevant online health information, and determining content reliability, pose real challenges for many health consumers. PURPOSE: To identify the types of interventions that have been implemented to assist health consumers to find reliable online health information, and where possible, describe and compare the types of outcomes studied. DATA SOURCES: PubMed, PsycINFO, CINAHL Plus and Cochrane Library databases; WorldCat and Scirus 'gray literature' search engines; and manual review of reference lists of selected publications. STUDY SELECTION: Publications were selected by firstly screening title, abstract, and then full text. DATA EXTRACTION: Seven publications met the inclusion criteria, and were summarized in a data extraction form. The form incorporated the PICOS (Population Intervention Comparators Outcomes and Study Design Model. Two eligible gray literature papers were also reported. DATA SYNTHESIS: Relevant data from included studies were tabulated to enable descriptive comparison. A brief critique of each study was included in the tables. This review was unable to follow systematic review methods due to the paucity of research and humanistic interventions reported. LIMITATIONS: While extensive, the gray literature search may have had limited reach in some countries. The paucity of research on this topic limits conclusions that may be drawn. CONCLUSIONS: The few eligible studies predominantly adopted a didactic approach to assisting health consumers, whereby consumers were either taught how to find credible websites, or how to use the Internet. Common types of outcomes studied include knowledge and skills pertaining to Internet use and searching for reliable health information. These outcomes were predominantly self-assessed by participants. There is potential for further research to explore other avenues for

  8. Antidiabetic Evaluation of Momordica charantia L Fruit Extracts

    Science.gov (United States)

    Tahira, S; Hussain, F

    2014-01-01

    To investigate hypoglycaemic, hypolipidaemic and pancreatic beta cell regeneration activities of Momordica charantia L fruits (MC). Alloxan-induced diabetic rabbits were treated with methanolic and ethanolic MC extract. Effects of plant extracts and the drug glibenclamide on serum glucose, lipid profile and pancreatic beta cell were determined after two weeks of treatment. Serum glucose and lipid profiles were assayed by kit methods. Pancreatic tissue histopathology was performed to study pancreatic beta cell regeneration. Momordica charantia extracts produced significant hypoglycaemic effects (p Momordica charantia supplementations were unable to normalize glucose and lipid profiles. Glibenclamide, a standard drug, not only lowered hyperglycaemia and hyperlipidaemia but also restored the normal levels. Regeneration of pancreatic beta cells by MC extracts was minimal, with fractional improvement produced by glibenclamide. The most significant finding of the present study was a 28% reduction in hyperglycaemia by MC ethanol extracts. To determine reliable antidiabetic potentials of MC, identification of the relevant antidiabetic components and underlying mechanisms is warranted. PMID:25429471

  9. relevance of information warfare models to critical infrastructure

    African Journals Online (AJOL)

    ismith

    Critical infrastructure models, strategies and policies should take information ... gain an advantage over a competitor or adversary through the use of one's own .... digital communications system, where the vehicles are analogous to bits or packets, ..... performance degraded, causing an increase in traffic finding a new route.

  10. Automatically extracting functionally equivalent proteins from SwissProt

    Directory of Open Access Journals (Sweden)

    Martin Andrew CR

    2008-10-01

    Full Text Available Abstract Background There is a frequent need to obtain sets of functionally equivalent homologous proteins (FEPs from different species. While it is usually the case that orthology implies functional equivalence, this is not always true; therefore datasets of orthologous proteins are not appropriate. The information relevant to extracting FEPs is contained in databanks such as UniProtKB/Swiss-Prot and a manual analysis of these data allow FEPs to be extracted on a one-off basis. However there has been no resource allowing the easy, automatic extraction of groups of FEPs – for example, all instances of protein C. We have developed FOSTA, an automatically generated database of FEPs annotated as having the same function in UniProtKB/Swiss-Prot which can be used for large-scale analysis. The method builds a candidate list of homologues and filters out functionally diverged proteins on the basis of functional annotations using a simple text mining approach. Results Large scale evaluation of our FEP extraction method is difficult as there is no gold-standard dataset against which the method can be benchmarked. However, a manual analysis of five protein families confirmed a high level of performance. A more extensive comparison with two manually verified functional equivalence datasets also demonstrated very good performance. Conclusion In summary, FOSTA provides an automated analysis of annotations in UniProtKB/Swiss-Prot to enable groups of proteins already annotated as functionally equivalent, to be extracted. Our results demonstrate that the vast majority of UniProtKB/Swiss-Prot functional annotations are of high quality, and that FOSTA can interpret annotations successfully. Where FOSTA is not successful, we are able to highlight inconsistencies in UniProtKB/Swiss-Prot annotation. Most of these would have presented equal difficulties for manual interpretation of annotations. We discuss limitations and possible future extensions to FOSTA, and

  11. Multidimensional Models of Information Need

    OpenAIRE

    Yun-jie (Calvin) Xu; Kai Huang (Joseph) Tan

    2009-01-01

    User studies in information science have recognised relevance as a multidimensional construct. An implication of multidimensional relevance is that a user's information need should be modeled by multiple data structures to represent different relevance dimensions. While the extant literature has attempted to model multiple dimensions of a user's information need, the fundamental assumption that a multidimensional model is better than a uni-dimensional model has not been addressed. This study ...

  12. Morally-Relevant Similarities and Differences Between Assisted Dying Practices in Paradigm and Non-Paradigm Circumstances: Could They Inform Regulatory Decisions?

    Science.gov (United States)

    Kirby, Jeffrey

    2017-12-01

    There has been contentious debate over the years about whether there are morally relevant similarities and differences between the three practices of continuous deep sedation until death, physician-assisted suicide, and voluntary euthanasia. Surprisingly little academic attention has been paid to a comparison of the uses of these practices in the two types of circumstances in which they are typically performed. A comparative domains of ethics analysis methodological approach is used in the paper to compare 1) the use of the three practices in paradigm circumstances, and 2) the use of the practices in paradigm circumstances to their use in non-paradigm circumstances. The analytical outcomes suggest that a bright moral line cannot be demonstrated between any two of the practices in paradigm circumstances, and that there are significant, morally-relevant distinctions between their use in paradigm and non-paradigm circumstances. A thought experiment is employed to illustrate how these outcomes could possibly inform the decisions of hypothetical deliberators who are engaged in the collaborative development of assisted dying regulatory frameworks.

  13. Enhanced Attenuation Technologies: Passive Soil Vapor Extraction

    Energy Technology Data Exchange (ETDEWEB)

    Vangelas, K.; Looney, B.; Kamath, R.; Adamson, D.; Newell, C.

    2010-03-15

    Passive soil vapor extraction (PSVE) is an enhanced attenuation (EA) approach that removes volatile contaminants from soil. The extraction is driven by natural pressure gradients between the subsurface and atmosphere (Barometric Pumping), or by renewable sources of energy such as wind or solar power (Assisted PSVE). The technology is applicable for remediating sites with low levels of contamination and for transitioning sites from active source technologies such as active soil vapor extraction (ASVE) to natural attenuation. PSVE systems are simple to design and operate and are more cost effective than active systems in many scenarios. Thus, PSVE is often appropriate as an interim-remedial or polishing strategy. Over the past decade, PSVE has been demonstrated in the U.S. and in Europe. These demonstrations provide practical information to assist in selecting, designing and implementing the technology. These demonstrations indicate that the technology can be effective in achieving remedial objectives in a timely fashion. The keys to success include: (1) Application at sites where the residual source quantities, and associated fluxes to groundwater, are relatively low; (2) Selection of the appropriate passive energy source - barometric pumping in cases with a deep vadose zone and barrier (e.g., clay) layers that separate the subsurface from the atmosphere and renewable energy assisted PSVE in other settings and where higher flow rates are required. (3) Provision of sufficient access to the contaminated vadose zones through the spacing and number of extraction wells. This PSVE technology report provides a summary of the relevant technical background, real-world case study performance, key design and cost considerations, and a scenario-based cost evaluation. The key design and cost considerations are organized into a flowchart that dovetails with the Enhanced Attenuation: Chlorinated Organics Guidance of the Interstate Technology and Regulatory Council (ITRC). The PSVE

  14. Testing the reliability of information extracted from ancient zircon

    Science.gov (United States)

    Kielman, Ross; Whitehouse, Martin; Nemchin, Alexander

    2015-04-01

    Studies combining zircon U-Pb chronology, trace element distribution as well as O and Hf isotope systematics are a powerful way to gain understanding of the processes shaping Earth's evolution, especially in detrital populations where constraints from the original host are missing. Such studies of the Hadean detrital zircon population abundant in sedimentary rocks in Western Australia have involved analysis of an unusually large number of individual grains, but also highlighted potential problems with the approach, only apparent when multiple analyses are obtained from individual grains. A common feature of the Hadean as well as many early Archaean zircon populations is their apparent inhomogeneity, which reduces confidence in conclusions based on studies combining chemistry and isotopic characteristics of zircon. In order to test the reliability of information extracted from early Earth zircon, we report results from one of the first in-depth multi-method study of zircon from a relatively simple early Archean magmatic rock, used as an analogue to ancient detrital zircon. The approach involves making multiple SIMS analyses in individual grains in order to be comparable to the most advanced studies of detrital zircon populations. The investigated sample is a relatively undeformed, non-migmatitic ca. 3.8 Ga tonalite collected a few kms south of the Isua Greenstone Belt, southwest Greenland. Extracted zircon grains can be combined into three different groups based on the behavior of their U-Pb systems: (i) grains that show internally consistent and concordant ages and define an average age of 3805±15 Ma, taken to be the age of the rock, (ii) grains that are distributed close to the concordia line, but with significant variability between multiple analyses, suggesting an ancient Pb loss and (iii) grains that have multiple analyses distributed along a discordia pointing towards a zero intercept, indicating geologically recent Pb-loss. This overall behavior has

  15. Developing an Approach to Prioritize River Restoration using Data Extracted from Flood Risk Information System Databases.

    Science.gov (United States)

    Vimal, S.; Tarboton, D. G.; Band, L. E.; Duncan, J. M.; Lovette, J. P.; Corzo, G.; Miles, B.

    2015-12-01

    Prioritizing river restoration requires information on river geometry. In many states in the US detailed river geometry has been collected for floodplain mapping and is available in Flood Risk Information Systems (FRIS). In particular, North Carolina has, for its 100 Counties, developed a database of numerous HEC-RAS models which are available through its Flood Risk Information System (FRIS). These models that include over 260 variables were developed and updated by numerous contractors. They contain detailed surveyed or LiDAR derived cross-sections and modeled flood extents for different extreme event return periods. In this work, over 4700 HEC-RAS models' data was integrated and upscaled to utilize detailed cross-section information and 100-year modelled flood extent information to enable river restoration prioritization for the entire state of North Carolina. We developed procedures to extract geomorphic properties such as entrenchment ratio, incision ratio, etc. from these models. Entrenchment ratio quantifies the vertical containment of rivers and thereby their vulnerability to flooding and incision ratio quantifies the depth per unit width. A map of entrenchment ratio for the whole state was derived by linking these model results to a geodatabase. A ranking of highly entrenched counties enabling prioritization for flood allowance and mitigation was obtained. The results were shared through HydroShare and web maps developed for their visualization using Google Maps Engine API.

  16. Spray-dried extracts from Syzygium cumini seeds: physicochemical and biological evaluation

    Directory of Open Access Journals (Sweden)

    Maria Paula G. Peixoto

    2013-02-01

    Full Text Available Aqueous extract from seeds of Syzygium cumini (L. Skeels, Myrtaceae, obtained by dynamic maceration was spray-dried and characterized by its physico-chemical and antihyperglycaemic action. The extract showed to possess high amount of polyphenols, significant in vitro free radical scavenger activity using the DPPH method and an antihyperglycaemic effect in alloxan-induced experimental diabetes. S. cumini spray-dried extracts were obtained using silicon dioxide and cassava starch as adjuvants. The powders showed acceptable flowability, compactability, and low hygroscopicity at 43% relative humidity. Besides, the spray-dried extracts showed in vivo antihyperglycaemic and in vitro scavenger activity comparable to the lyophilized extract. Thus, experimental data indicates that the extract from S. cumini has a relevant activity and that spray-drying could be adequately used to perform the technological processing of S. cumini fluid extracts.

  17. Balancing relevance criteria through multi-objective optimization

    NARCIS (Netherlands)

    van Doorn, J.; Odijk, D.; Roijers, D.M.; de Rijke, M.

    2016-01-01

    Offline evaluation of information retrieval systems typically focuses on a single effectiveness measure that models the utility for a typical user. Such a measure usually combines a behavior-based rank discount with a notion of document utility that captures the single relevance criterion of

  18. New bioinformatic tool for quick identification of functionally relevant endogenous retroviral inserts in human genome.

    Science.gov (United States)

    Garazha, Andrew; Ivanova, Alena; Suntsova, Maria; Malakhova, Galina; Roumiantsev, Sergey; Zhavoronkov, Alex; Buzdin, Anton

    2015-01-01

    Endogenous retroviruses (ERVs) and LTR retrotransposons (LRs) occupy ∼8% of human genome. Deep sequencing technologies provide clues to understanding of functional relevance of individual ERVs/LRs by enabling direct identification of transcription factor binding sites (TFBS) and other landmarks of functional genomic elements. Here, we performed the genome-wide identification of human ERVs/LRs containing TFBS according to the ENCODE project. We created the first interactive ERV/LRs database that groups the individual inserts according to their familial nomenclature, number of mapped TFBS and divergence from their consensus sequence. Information on any particular element can be easily extracted by the user. We also created a genome browser tool, which enables quick mapping of any ERV/LR insert according to genomic coordinates, known human genes and TFBS. These tools can be used to easily explore functionally relevant individual ERV/LRs, and for studying their impact on the regulation of human genes. Overall, we identified ∼110,000 ERV/LR genomic elements having TFBS. We propose a hypothesis of "domestication" of ERV/LR TFBS by the genome milieu including subsequent stages of initial epigenetic repression, partial functional release, and further mutation-driven reshaping of TFBS in tight coevolution with the enclosing genomic loci.

  19. Analysis of Technique to Extract Data from the Web for Improved Performance

    Science.gov (United States)

    Gupta, Neena; Singh, Manish

    2010-11-01

    The World Wide Web rapidly guides the world into a newly amazing electronic world, where everyone can publish anything in electronic form and extract almost all the information. Extraction of information from semi structured or unstructured documents, such as web pages, is a useful yet complex task. Data extraction, which is important for many applications, extracts the records from the HTML files automatically. Ontologies can achieve a high degree of accuracy in data extraction. We analyze method for data extraction OBDE (Ontology-Based Data Extraction), which automatically extracts the query result records from the web with the help of agents. OBDE first constructs an ontology for a domain according to information matching between the query interfaces and query result pages from different web sites within the same domain. Then, the constructed domain ontology is used during data extraction to identify the query result section in a query result page and to align and label the data values in the extracted records. The ontology-assisted data extraction method is fully automatic and overcomes many of the deficiencies of current automatic data extraction methods.

  20. Social Web Data Analytics : Relevance, Redundancy, Diversity

    NARCIS (Netherlands)

    Tao, K.

    2014-01-01

    In the past decade, the Social Web has evolved into both an essential channel for people to exchange information and a new type of mass media. The immense amount of data produced presents new possibilities and challenges: algorithms and technologies need to be developed to extract and infer useful