WorldWideScience

Sample records for domain-independent information extraction

  1. Domain-independent information extraction in unstructured text

    Energy Technology Data Exchange (ETDEWEB)

    Irwin, N.H. [Sandia National Labs., Albuquerque, NM (United States). Software Surety Dept.

    1996-09-01

    Extracting information from unstructured text has become an important research area in recent years due to the large amount of text now electronically available. This status report describes the findings and work done during the second year of a two-year Laboratory Directed Research and Development Project. Building on the first-year`s work of identifying important entities, this report details techniques used to group words into semantic categories and to output templates containing selective document content. Using word profiles and category clustering derived during a training run, the time-consuming knowledge-building task can be avoided. Though the output still lacks in completeness when compared to systems with domain-specific knowledge bases, the results do look promising. The two approaches are compatible and could complement each other within the same system. Domain-independent approaches retain appeal as a system that adapts and learns will soon outpace a system with any amount of a priori knowledge.

  2. Post-processing of Deep Web Information Extraction Based on Domain Ontology

    Directory of Open Access Journals (Sweden)

    PENG, T.

    2013-11-01

    Full Text Available Many methods are utilized to extract and process query results in deep Web, which rely on the different structures of Web pages and various designing modes of databases. However, some semantic meanings and relations are ignored. So, in this paper, we present an approach for post-processing deep Web query results based on domain ontology which can utilize the semantic meanings and relations. A block identification model (BIM based on node similarity is defined to extract data blocks that are relevant to specific domain after reducing noisy nodes. Feature vector of domain books is obtained by result set extraction model (RSEM based on vector space model (VSM. RSEM, in combination with BIM, builds the domain ontology on books which can not only remove the limit of Web page structures when extracting data information, but also make use of semantic meanings of domain ontology. After extracting basic information of Web pages, a ranking algorithm is adopted to offer an ordered list of data records to users. Experimental results show that BIM and RSEM extract data blocks and build domain ontology accurately. In addition, relevant data records and basic information are extracted and ranked. The performances precision and recall show that our proposed method is feasible and efficient.

  3. Understanding disciplinary vocabularies using a full-text enabled domain-independent term extraction approach.

    Science.gov (United States)

    Yan, Erjia; Williams, Jake; Chen, Zheng

    2017-01-01

    Publication metadata help deliver rich analyses of scholarly communication. However, research concepts and ideas are more effectively expressed through unstructured fields such as full texts. Thus, the goals of this paper are to employ a full-text enabled method to extract terms relevant to disciplinary vocabularies, and through them, to understand the relationships between disciplines. This paper uses an efficient, domain-independent term extraction method to extract disciplinary vocabularies from a large multidisciplinary corpus of PLoS ONE publications. It finds a power-law pattern in the frequency distributions of terms present in each discipline, indicating a semantic richness potentially sufficient for further study and advanced analysis. The salient relationships amongst these vocabularies become apparent in application of a principal component analysis. For example, Mathematics and Computer and Information Sciences were found to have similar vocabulary use patterns along with Engineering and Physics; while Chemistry and the Social Sciences were found to exhibit contrasting vocabulary use patterns along with the Earth Sciences and Chemistry. These results have implications to studies of scholarly communication as scholars attempt to identify the epistemological cultures of disciplines, and as a full text-based methodology could lead to machine learning applications in the automated classification of scholarly work according to disciplinary vocabularies.

  4. Balance and Gait Represent Independent Domains of Mobility in Parkinson Disease

    Science.gov (United States)

    Horak, Fay B.; Carlson-Kuhta, Patricia; Nutt, John G.; Salarian, Arash

    2016-01-01

    Background The Instrumented Stand and Walk (ISAW) test, which includes 30 seconds of stance, step initiation, gait, and turning, results in many objective balance and gait metrics from body-worn inertial sensors. However, it is not clear which metrics provide independent information about mobility. Objective It was hypothesized that balance and gait represent several independent domains of mobility and that not all domains would be abnormal in individuals with Parkinson disease (PD) or would change with levodopa therapy. Design This was a cross-sectional study. Methods A factor analysis approach was used to identify independent measures of mobility extracted from the ISAW in 100 participants with PD and 21 control participants. First, a covariance analysis showed that postural sway measures were independent of gait measures. Then, the factor analysis revealed 6 independent factors (mobility domains: sway area, sway frequency, arm swing asymmetry, trunk motion during gait, gait speed, and cadence) that accounted for 87% of the variance of performance across participants. Results Sway area, gait speed, and trunk motion differed between the PD group in the off-levodopa state and the control group, but sway frequency (but not sway area) differed between the PD group in the off-levodopa state and the control group. Four of the 6 factors changed significantly with levodopa (off to on): sway area, sway frequency, trunk motion during gait, and cadence. When participants were on levodopa, the sway area increased compared with off levodopa, becoming more abnormal, whereas the other 3 significant metrics moved toward, but did not reach, the healthy control values. Limitations Exploratory factor analysis was limited to the PD population. Conclusions The different sensitivity various balance and gait domains to PD and to levodopa also support neural control of at least 6 independent mobility domains, each of which warrants clinical assessment for impairments in mobility. PMID

  5. Independent component analysis of edge information for face recognition

    CERN Document Server

    Karande, Kailash Jagannath

    2013-01-01

    The book presents research work on face recognition using edge information as features for face recognition with ICA algorithms. The independent components are extracted from edge information. These independent components are used with classifiers to match the facial images for recognition purpose. In their study, authors have explored Canny and LOG edge detectors as standard edge detection methods. Oriented Laplacian of Gaussian (OLOG) method is explored to extract the edge information with different orientations of Laplacian pyramid. Multiscale wavelet model for edge detection is also propos

  6. Fine-grained information extraction from German transthoracic echocardiography reports.

    Science.gov (United States)

    Toepfer, Martin; Corovic, Hamo; Fette, Georg; Klügl, Peter; Störk, Stefan; Puppe, Frank

    2015-11-12

    Information extraction techniques that get structured representations out of unstructured data make a large amount of clinically relevant information about patients accessible for semantic applications. These methods typically rely on standardized terminologies that guide this process. Many languages and clinical domains, however, lack appropriate resources and tools, as well as evaluations of their applications, especially if detailed conceptualizations of the domain are required. For instance, German transthoracic echocardiography reports have not been targeted sufficiently before, despite of their importance for clinical trials. This work therefore aimed at development and evaluation of an information extraction component with a fine-grained terminology that enables to recognize almost all relevant information stated in German transthoracic echocardiography reports at the University Hospital of Würzburg. A domain expert validated and iteratively refined an automatically inferred base terminology. The terminology was used by an ontology-driven information extraction system that outputs attribute value pairs. The final component has been mapped to the central elements of a standardized terminology, and it has been evaluated according to documents with different layouts. The final system achieved state-of-the-art precision (micro average.996) and recall (micro average.961) on 100 test documents that represent more than 90 % of all reports. In particular, principal aspects as defined in a standardized external terminology were recognized with f 1=.989 (micro average) and f 1=.963 (macro average). As a result of keyword matching and restraint concept extraction, the system obtained high precision also on unstructured or exceptionally short documents, and documents with uncommon layout. The developed terminology and the proposed information extraction system allow to extract fine-grained information from German semi-structured transthoracic echocardiography reports

  7. Arrhythmia Classification Based on Multi-Domain Feature Extraction for an ECG Recognition System

    Directory of Open Access Journals (Sweden)

    Hongqiang Li

    2016-10-01

    Full Text Available Automatic recognition of arrhythmias is particularly important in the diagnosis of heart diseases. This study presents an electrocardiogram (ECG recognition system based on multi-domain feature extraction to classify ECG beats. An improved wavelet threshold method for ECG signal pre-processing is applied to remove noise interference. A novel multi-domain feature extraction method is proposed; this method employs kernel-independent component analysis in nonlinear feature extraction and uses discrete wavelet transform to extract frequency domain features. The proposed system utilises a support vector machine classifier optimized with a genetic algorithm to recognize different types of heartbeats. An ECG acquisition experimental platform, in which ECG beats are collected as ECG data for classification, is constructed to demonstrate the effectiveness of the system in ECG beat classification. The presented system, when applied to the MIT-BIH arrhythmia database, achieves a high classification accuracy of 98.8%. Experimental results based on the ECG acquisition experimental platform show that the system obtains a satisfactory classification accuracy of 97.3% and is able to classify ECG beats efficiently for the automatic identification of cardiac arrhythmias.

  8. Cross domains Arabic named entity recognition system

    Science.gov (United States)

    Al-Ahmari, S. Saad; Abdullatif Al-Johar, B.

    2016-07-01

    Named Entity Recognition (NER) plays an important role in many Natural Language Processing (NLP) applications such as; Information Extraction (IE), Question Answering (QA), Text Clustering, Text Summarization and Word Sense Disambiguation. This paper presents the development and implementation of domain independent system to recognize three types of Arabic named entities. The system works based on a set of domain independent grammar-rules along with Arabic part of speech tagger in addition to gazetteers and lists of trigger words. The experimental results shown, that the system performed as good as other systems with better results in some cases of cross-domains corpora.

  9. Independent component analysis for automatic note extraction from musical trills

    Science.gov (United States)

    Brown, Judith C.; Smaragdis, Paris

    2004-05-01

    The method of principal component analysis, which is based on second-order statistics (or linear independence), has long been used for redundancy reduction of audio data. The more recent technique of independent component analysis, enforcing much stricter statistical criteria based on higher-order statistical independence, is introduced and shown to be far superior in separating independent musical sources. This theory has been applied to piano trills and a database of trill rates was assembled from experiments with a computer-driven piano, recordings of a professional pianist, and commercially available compact disks. The method of independent component analysis has thus been shown to be an outstanding, effective means of automatically extracting interesting musical information from a sea of redundant data.

  10. A domain-independent descriptive design model and its application to structured reflection on design processes

    NARCIS (Netherlands)

    Reymen, Isabelle; Hammer, D.K.; Kroes, P.A.; van Aken, Joan Ernst; van Aken, J.E.; Dorst, C.H.; Bax, M.F.T.; Basten, T

    2006-01-01

    Domain-independent models of the design process are an important means for facilitating interdisciplinary communication and for supporting multidisciplinary design. Many so-called domain-independent models are, however, not really domain independent. We state that to be domain independent, the

  11. Is there any need for domain-dependent control information? A reply

    Energy Technology Data Exchange (ETDEWEB)

    Minton, S. [USC Information Sciences Inst., Marina del Rey, CA (United States)

    1996-12-31

    In this paper, we consider the role that domain-dependent control knowledge plays in problem solving systems. Ginsberg and Geddis have claimed that domain-dependent control information has no place in declarative systems; instead, they say, such information should be derived from declarative facts about the domain plus domain-independent principles. We dispute their conclusion, arguing that it is impractical to generate control knowledge solely on the basis of logical derivations. We propose that simplifying abstractions are crucial for deriving control knowledge, and, as a result, empirical utility evaluation of the resulting rules will frequently be necessary to validate the utility of derived control knowledge. We illustrate our arguments with examples from two implemented systems.

  12. Cross domains Arabic named entity recognition system

    KAUST Repository

    Al-Ahmari, S. Saad

    2016-07-11

    Named Entity Recognition (NER) plays an important role in many Natural Language Processing (NLP) applications such as; Information Extraction (IE), Question Answering (QA), Text Clustering, Text Summarization and Word Sense Disambiguation. This paper presents the development and implementation of domain independent system to recognize three types of Arabic named entities. The system works based on a set of domain independent grammar-rules along with Arabic part of speech tagger in addition to gazetteers and lists of trigger words. The experimental results shown, that the system performed as good as other systems with better results in some cases of cross-domains corpora. © (2016) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.

  13. Cross domains Arabic named entity recognition system

    KAUST Repository

    Al-Ahmari, S. Saad; Abdullatif Al-Johar, B.

    2016-01-01

    Named Entity Recognition (NER) plays an important role in many Natural Language Processing (NLP) applications such as; Information Extraction (IE), Question Answering (QA), Text Clustering, Text Summarization and Word Sense Disambiguation. This paper presents the development and implementation of domain independent system to recognize three types of Arabic named entities. The system works based on a set of domain independent grammar-rules along with Arabic part of speech tagger in addition to gazetteers and lists of trigger words. The experimental results shown, that the system performed as good as other systems with better results in some cases of cross-domains corpora. © (2016) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.

  14. Information Extraction with Character-level Neural Networks and Free Noisy Supervision

    OpenAIRE

    Meerkamp, Philipp; Zhou, Zhengyi

    2016-01-01

    We present an architecture for information extraction from text that augments an existing parser with a character-level neural network. The network is trained using a measure of consistency of extracted data with existing databases as a form of noisy supervision. Our architecture combines the ability of constraint-based information extraction systems to easily incorporate domain knowledge and constraints with the ability of deep neural networks to leverage large amounts of data to learn compl...

  15. Extracting meronomy relations from domain-specific, textual corporate databases

    NARCIS (Netherlands)

    Ittoo, R.A.; Bouma, G.; Maruster, L.; Wortmann, J.C.; Hopfe, C.J.; Rezgui, Y.; Métais, E.; Preece, A.; Li, H.

    2010-01-01

    Various techniques for learning meronymy relationships from open-domain corpora exist. However, extracting meronymy relationships from domain-specific, textual corporate databases has been overlooked, despite numerous application opportunities particularly in domains like product development and/or

  16. Unsupervised information extraction by text segmentation

    CERN Document Server

    Cortez, Eli

    2013-01-01

    A new unsupervised approach to the problem of Information Extraction by Text Segmentation (IETS) is proposed, implemented and evaluated herein. The authors' approach relies on information available on pre-existing data to learn how to associate segments in the input string with attributes of a given domain relying on a very effective set of content-based features. The effectiveness of the content-based features is also exploited to directly learn from test data structure-based features, with no previous human-driven training, a feature unique to the presented approach. Based on the approach, a

  17. Using text mining techniques to extract phenotypic information from the PhenoCHF corpus.

    Science.gov (United States)

    Alnazzawi, Noha; Thompson, Paul; Batista-Navarro, Riza; Ananiadou, Sophia

    2015-01-01

    Phenotypic information locked away in unstructured narrative text presents significant barriers to information accessibility, both for clinical practitioners and for computerised applications used for clinical research purposes. Text mining (TM) techniques have previously been applied successfully to extract different types of information from text in the biomedical domain. They have the potential to be extended to allow the extraction of information relating to phenotypes from free text. To stimulate the development of TM systems that are able to extract phenotypic information from text, we have created a new corpus (PhenoCHF) that is annotated by domain experts with several types of phenotypic information relating to congestive heart failure. To ensure that systems developed using the corpus are robust to multiple text types, it integrates text from heterogeneous sources, i.e., electronic health records (EHRs) and scientific articles from the literature. We have developed several different phenotype extraction methods to demonstrate the utility of the corpus, and tested these methods on a further corpus, i.e., ShARe/CLEF 2013. Evaluation of our automated methods showed that PhenoCHF can facilitate the training of reliable phenotype extraction systems, which are robust to variations in text type. These results have been reinforced by evaluating our trained systems on the ShARe/CLEF corpus, which contains clinical records of various types. Like other studies within the biomedical domain, we found that solutions based on conditional random fields produced the best results, when coupled with a rich feature set. PhenoCHF is the first annotated corpus aimed at encoding detailed phenotypic information. The unique heterogeneous composition of the corpus has been shown to be advantageous in the training of systems that can accurately extract phenotypic information from a range of different text types. Although the scope of our annotation is currently limited to a single

  18. Equivalent water height extracted from GRACE gravity field model with robust independent component analysis

    Science.gov (United States)

    Guo, Jinyun; Mu, Dapeng; Liu, Xin; Yan, Haoming; Dai, Honglei

    2014-08-01

    The Level-2 monthly GRACE gravity field models issued by Center for Space Research (CSR), GeoForschungs Zentrum (GFZ), and Jet Propulsion Laboratory (JPL) are treated as observations used to extract the equivalent water height (EWH) with the robust independent component analysis (RICA). The smoothing radii of 300, 400, and 500 km are tested, respectively, in the Gaussian smoothing kernel function to reduce the observation Gaussianity. Three independent components are obtained by RICA in the spatial domain; the first component matches the geophysical signal, and the other two match the north-south strip and the other noises. The first mode is used to estimate EWHs of CSR, JPL, and GFZ, and compared with the classical empirical decorrelation method (EDM). The EWH STDs for 12 months in 2010 extracted by RICA and EDM show the obvious fluctuation. The results indicate that the sharp EWH changes in some areas have an important global effect, like in Amazon, Mekong, and Zambezi basins.

  19. Towards an information extraction and knowledge formation framework based on Shannon entropy

    Directory of Open Access Journals (Sweden)

    Iliescu Dragoș

    2017-01-01

    Full Text Available Information quantity subject is approached in this paperwork, considering the specific domain of nonconforming product management as information source. This work represents a case study. Raw data were gathered from a heavy industrial works company, information extraction and knowledge formation being considered herein. Involved method for information quantity estimation is based on Shannon entropy formula. Information and entropy spectrum are decomposed and analysed for extraction of specific information and knowledge-that formation. The result of the entropy analysis point out the information needed to be acquired by the involved organisation, this being presented as a specific knowledge type.

  20. Information Pre-Processing using Domain Meta-Ontology and Rule Learning System

    Science.gov (United States)

    Ranganathan, Girish R.; Biletskiy, Yevgen

    Around the globe, extraordinary amounts of documents are being created by Enterprises and by users outside these Enterprises. The documents created in the Enterprises constitute the main focus of the present chapter. These documents are used to perform numerous amounts of machine processing. While using thesedocuments for machine processing, lack of semantics of the information in these documents may cause misinterpretation of the information, thereby inhibiting the productiveness of computer assisted analytical work. Hence, it would be profitable to the Enterprises if they use well defined domain ontologies which will serve as rich source(s) of semantics for the information in the documents. These domain ontologies can be created manually, semi-automatically or fully automatically. The focus of this chapter is to propose an intermediate solution which will enable relatively easy creation of these domain ontologies. The process of extracting and capturing domain ontologies from these voluminous documents requires extensive involvement of domain experts and application of methods of ontology learning that are substantially labor intensive; therefore, some intermediate solutions which would assist in capturing domain ontologies must be developed. This chapter proposes a solution in this direction which involves building a meta-ontology that will serve as an intermediate information source for the main domain ontology. This chapter proposes a solution in this direction which involves building a meta-ontology as a rapid approach in conceptualizing a domain of interest from huge amount of source documents. This meta-ontology can be populated by ontological concepts, attributes and relations from documents, and then refined in order to form better domain ontology either through automatic ontology learning methods or some other relevant ontology building approach.

  1. Quantitative grain-scale ferroic domain volume fractions and domain switching strains from three-dimensional X-ray diffraction data

    DEFF Research Database (Denmark)

    Oddershede, Jette; Majkut, Marta; Caosyd, Qinghua

    2015-01-01

    A method for the extension of the three-dimensional X-ray diffraction technique to allow the extraction of domain volume fractions in polycrystalline ferroic materials is presented. This method gives access to quantitative domain volume fractions of hundreds of independent embedded grains within...... are applied to tetragonal coarse-grained Ba0.88Ca0.12Zr0.06Ti0.94O3 and rhombohedral fine-grained (0.82)Bi0.5Na0.5TiO3–(0.18)Bi0.5K0.5TiO3 electroceramic materials. The fitted volume fraction information is used to calculate grain-scale non-180° ferroelectric domain switching strains. The absolute errors...

  2. Information extraction system

    Science.gov (United States)

    Lemmond, Tracy D; Hanley, William G; Guensche, Joseph Wendell; Perry, Nathan C; Nitao, John J; Kidwell, Paul Brandon; Boakye, Kofi Agyeman; Glaser, Ron E; Prenger, Ryan James

    2014-05-13

    An information extraction system and methods of operating the system are provided. In particular, an information extraction system for performing meta-extraction of named entities of people, organizations, and locations as well as relationships and events from text documents are described herein.

  3. Extracting breathing rate information from a wearable reflectance pulse oximeter sensor.

    Science.gov (United States)

    Johnston, W S; Mendelson, Y

    2004-01-01

    The integration of multiple vital physiological measurements could help combat medics and field commanders to better predict a soldier's health condition and enhance their ability to perform remote triage procedures. In this paper we demonstrate the feasibility of extracting accurate breathing rate information from a photoplethysmographic signal that was recorded by a reflectance pulse oximeter sensor mounted on the forehead and subsequently processed by a simple time domain filtering and frequency domain Fourier analysis.

  4. An Accurate Integral Method for Vibration Signal Based on Feature Information Extraction

    Directory of Open Access Journals (Sweden)

    Yong Zhu

    2015-01-01

    Full Text Available After summarizing the advantages and disadvantages of current integral methods, a novel vibration signal integral method based on feature information extraction was proposed. This method took full advantage of the self-adaptive filter characteristic and waveform correction feature of ensemble empirical mode decomposition in dealing with nonlinear and nonstationary signals. This research merged the superiorities of kurtosis, mean square error, energy, and singular value decomposition on signal feature extraction. The values of the four indexes aforementioned were combined into a feature vector. Then, the connotative characteristic components in vibration signal were accurately extracted by Euclidean distance search, and the desired integral signals were precisely reconstructed. With this method, the interference problem of invalid signal such as trend item and noise which plague traditional methods is commendably solved. The great cumulative error from the traditional time-domain integral is effectively overcome. Moreover, the large low-frequency error from the traditional frequency-domain integral is successfully avoided. Comparing with the traditional integral methods, this method is outstanding at removing noise and retaining useful feature information and shows higher accuracy and superiority.

  5. Extracting Low-Frequency Information from Time Attenuation in Elastic Waveform Inversion

    Science.gov (United States)

    Guo, Xuebao; Liu, Hong; Shi, Ying; Wang, Weihong

    2017-03-01

    Low-frequency information is crucial for recovering background velocity, but the lack of low-frequency information in field data makes inversion impractical without accurate initial models. Laplace-Fourier domain waveform inversion can recover a smooth model from real data without low-frequency information, which can be used for subsequent inversion as an ideal starting model. In general, it also starts with low frequencies and includes higher frequencies at later inversion stages, while the difference is that its ultralow frequency information comes from the Laplace-Fourier domain. Meanwhile, a direct implementation of the Laplace-transformed wavefield using frequency domain inversion is also very convenient. However, because broad frequency bands are often used in the pure time domain waveform inversion, it is difficult to extract the wavefields dominated by low frequencies in this case. In this paper, low-frequency components are constructed by introducing time attenuation into the recorded residuals, and the rest of the method is identical to the traditional time domain inversion. Time windowing and frequency filtering are also applied to mitigate the ambiguity of the inverse problem. Therefore, we can start at low frequencies and to move to higher frequencies. The experiment shows that the proposed method can achieve a good inversion result in the presence of a linear initial model and records without low-frequency information.

  6. Robustness Design for CNN Templates with Performance of Extracting Closed Domain

    International Nuclear Information System (INIS)

    Li Weidong; Min Lequan

    2006-01-01

    The cellular neural/nonlinear network (CNN) is a powerful tool for image and video signal processing, robotic and biological visions. This paper introduces a kind of CNNs with performance of extracting closed domains in binary images, and gives a general method for designing templates of such a kind of CNNs. One theorem provides parameter inequalities for determining parameter intervals for implementing prescribed image processing functions, respectively. Examples for extracting closed domains in binary scale images are given.

  7. Information analysis of iris biometrics for the needs of cryptology key extraction

    Directory of Open Access Journals (Sweden)

    Adamović Saša

    2013-01-01

    Full Text Available The paper presents a rigorous analysis of iris biometric information for the synthesis of an optimized system for the extraction of a high quality cryptology key. Estimations of local entropy and mutual information were identified as segments of the iris most suitable for this purpose. In order to optimize parameters, corresponding wavelets were transformed, in order to obtain the highest possible entropy and mutual information lower in the transformation domain, which set frameworks for the synthesis of systems for the extraction of truly random sequences of iris biometrics, without compromising authentication properties. [Projekat Ministarstva nauke Republike Srbije, br. TR32054 i br. III44006

  8. Social cognition and neurocognition as independent domains in psychosis.

    Science.gov (United States)

    van Hooren, S; Versmissen, D; Janssen, I; Myin-Germeys, I; à Campo, J; Mengelers, R; van Os, J; Krabbendam, L

    2008-08-01

    Patients with psychosis display alterations in social cognition as well as in the realm of neurocognition. It is unclear, however, to what degree these cognitive domains represent two separate dimensions of liability or the pleiotropic expression of a single deficit. The purpose of the present study was to investigate (i) to what extent alterations in social cognition represent an independent area of vulnerability to psychosis, separate from neurocognitive deficits and (ii) whether social cognition is one construct or can be divided into several subcomponents. Five social cognition and three neurocognitive tasks were completed by 186 participants with different levels of vulnerability for psychosis: 44 patients with psychotic disorder; 47 subjects at familial risk; 41 subjects at psychometric risk and 54 control subjects. The social cognition tasks covered important basic subcomponents of social cognition, i.e. mentalisation (or theory of mind), data gathering bias (jumping to conclusions), source monitoring and attribution style. Neurocognitive tasks assessed speed of information processing, inhibition, cognitive shifting and strategy-driven retrieval from semantic memory. The results of factor analysis suggested that neurocognition and social cognition are two separate areas of vulnerability in psychosis. Furthermore, the social cognition measures lacked significant overlap, suggesting a multidimensional construct. Cognitive liabilities to psychosis are manifold, and include key processes underlying basic person-environment interactions in daily life, independent of cognition quantified by neuropsychological tests.

  9. Automated Extraction of Substance Use Information from Clinical Texts.

    Science.gov (United States)

    Wang, Yan; Chen, Elizabeth S; Pakhomov, Serguei; Arsoniadis, Elliot; Carter, Elizabeth W; Lindemann, Elizabeth; Sarkar, Indra Neil; Melton, Genevieve B

    2015-01-01

    Within clinical discourse, social history (SH) includes important information about substance use (alcohol, drug, and nicotine use) as key risk factors for disease, disability, and mortality. In this study, we developed and evaluated a natural language processing (NLP) system for automated detection of substance use statements and extraction of substance use attributes (e.g., temporal and status) based on Stanford Typed Dependencies. The developed NLP system leveraged linguistic resources and domain knowledge from a multi-site social history study, Propbank and the MiPACQ corpus. The system attained F-scores of 89.8, 84.6 and 89.4 respectively for alcohol, drug, and nicotine use statement detection, as well as average F-scores of 82.1, 90.3, 80.8, 88.7, 96.6, and 74.5 respectively for extraction of attributes. Our results suggest that NLP systems can achieve good performance when augmented with linguistic resources and domain knowledge when applied to a wide breadth of substance use free text clinical notes.

  10. An information theory framework for dynamic functional domain connectivity.

    Science.gov (United States)

    Vergara, Victor M; Miller, Robyn; Calhoun, Vince

    2017-06-01

    Dynamic functional network connectivity (dFNC) analyzes time evolution of coherent activity in the brain. In this technique dynamic changes are considered for the whole brain. This paper proposes an information theory framework to measure information flowing among subsets of functional networks call functional domains. Our method aims at estimating bits of information contained and shared among domains. The succession of dynamic functional states is estimated at the domain level. Information quantity is based on the probabilities of observing each dynamic state. Mutual information measurement is then obtained from probabilities across domains. Thus, we named this value the cross domain mutual information (CDMI). Strong CDMIs were observed in relation to the subcortical domain. Domains related to sensorial input, motor control and cerebellum form another CDMI cluster. Information flow among other domains was seldom found. Other methods of dynamic connectivity focus on whole brain dFNC matrices. In the current framework, information theory is applied to states estimated from pairs of multi-network functional domains. In this context, we apply information theory to measure information flow across functional domains. Identified CDMI clusters point to known information pathways in the basal ganglia and also among areas of sensorial input, patterns found in static functional connectivity. In contrast, CDMI across brain areas of higher level cognitive processing follow a different pattern that indicates scarce information sharing. These findings show that employing information theory to formally measured information flow through brain domains reveals additional features of functional connectivity. Copyright © 2017 Elsevier B.V. All rights reserved.

  11. Multimedia Information Extraction

    CERN Document Server

    Maybury, Mark T

    2012-01-01

    The advent of increasingly large consumer collections of audio (e.g., iTunes), imagery (e.g., Flickr), and video (e.g., YouTube) is driving a need not only for multimedia retrieval but also information extraction from and across media. Furthermore, industrial and government collections fuel requirements for stock media access, media preservation, broadcast news retrieval, identity management, and video surveillance.  While significant advances have been made in language processing for information extraction from unstructured multilingual text and extraction of objects from imagery and vid

  12. Challenges in Managing Information Extraction

    Science.gov (United States)

    Shen, Warren H.

    2009-01-01

    This dissertation studies information extraction (IE), the problem of extracting structured information from unstructured data. Example IE tasks include extracting person names from news articles, product information from e-commerce Web pages, street addresses from emails, and names of emerging music bands from blogs. IE is all increasingly…

  13. Information Architecture Used to Manage Multi-Domain Data Analysis in Intensively Managed Landscape - Critical Zone Observatory

    Science.gov (United States)

    Kooper, R.; Angelo, B.; Marini, L.; Kumar, P.; Muste, M.

    2016-12-01

    The Intensively Managed Landscapes-Critical Zone Observatory (IML-CZO) is a multi-agency partnership that aims to understand the coevoluationary dynamics of change in the context of the landscape, soil, and biota. The Data Management aspect of IML-CZO provides data preservation and analysis for each of the scientific domains as they pursue environmental monitoring throughout the midwestern United States. Data Management is facilitated via data ingestion and storage through Clowder, an open-source, scalable data repository for organizing and analyzing data; and Geodashboard, a web application that provides exploring, querying, visualizing and downloading the data ingested into Clowder. The data collected covers many domains including geology, hydrology, and bioengineering. The data across these domains varies greatly; from real-time streams of environmental measurements to individual soil samples that are sent through a series of laboratories for analysis. All data can be uploaded to Clowder where metadata can be extracted or dynamically calculated based on the nature of the information. Geodashboard was created to provide scientists with a tool to explore data across these varying domains, and to visualize the extracted data from Clowder. Once Clowder has extracted the data, it is available for querying from a REST API for standardized and streamlined access. Users are able to explore the data on multiple axis, and are able to download data across multiple domains in a standardized format for further analysis and research. IML-CZO's Clowder has over 60 users and over 180 datasets. There are over 1.1 million extracted data points that date back to 1992, and it is continually growing.

  14. Domain Independent Vocabulary Generation and Its Use in Category-based Small Footprint Language Model

    Directory of Open Access Journals (Sweden)

    KIM, K.-H.

    2011-02-01

    Full Text Available The work in this paper pertains to domain independent vocabulary generation and its use in category-based small footprint Language Model (LM. Two major constraints of the conventional LMs in the embedded environment are memory capacity limitation and data sparsity for the domain-specific application. This data sparsity adversely affects vocabulary coverage and LM performance. To overcome these constraints, we define a set of domain independent categories using a Part-Of-Speech (POS tagged corpus. Also, we generate a domain independent vocabulary based on this set using the corpus and knowledge base. Then, we propose a mathematical framework for a category-based LM using this set. In this LM, one word can be assigned assign multiple categories. In order to reduce its memory requirements, we propose a tree-based data structure. In addition, we determine the history length of a category n-gram, and the independent assumption applying to a category history generation. The proposed vocabulary generation method illustrates at least 13.68% relative improvement in coverage for a SMS text corpus, where data are sparse due to the difficulties in data collection. The proposed category-based LM requires only 215KB which is 55% and 13% compared to the conventional category-based LM and the word-based LM, respectively. It successively improves the performance, achieving 54.9% and 60.6% perplexity reduction compared to the conventional category-based LM and the word-based LM in terms of normalized perplexity.

  15. Quantifying information transfer by protein domains: Analysis of the Fyn SH2 domain structure

    Directory of Open Access Journals (Sweden)

    Serrano Luis

    2008-10-01

    Full Text Available Abstract Background Efficient communication between distant sites within a protein is essential for cooperative biological response. Although often associated with large allosteric movements, more subtle changes in protein dynamics can also induce long-range correlations. However, an appropriate formalism that directly relates protein structural dynamics to information exchange between functional sites is still lacking. Results Here we introduce a method to analyze protein dynamics within the framework of information theory and show that signal transduction within proteins can be considered as a particular instance of communication over a noisy channel. In particular, we analyze the conformational correlations between protein residues and apply the concept of mutual information to quantify information exchange. Mapping out changes of mutual information on the protein structure then allows visualizing how distal communication is achieved. We illustrate the approach by analyzing information transfer by the SH2 domain of Fyn tyrosine kinase, obtained from Monte Carlo dynamics simulations. Our analysis reveals that the Fyn SH2 domain forms a noisy communication channel that couples residues located in the phosphopeptide and specificity binding sites and a number of residues at the other side of the domain near the linkers that connect the SH2 domain to the SH3 and kinase domains. We find that for this particular domain, communication is affected by a series of contiguous residues that connect distal sites by crossing the core of the SH2 domain. Conclusion As a result, our method provides a means to directly map the exchange of biological information on the structure of protein domains, making it clear how binding triggers conformational changes in the protein structure. As such it provides a structural road, next to the existing attempts at sequence level, to predict long-range interactions within protein structures.

  16. KneeTex: an ontology-driven system for information extraction from MRI reports.

    Science.gov (United States)

    Spasić, Irena; Zhao, Bo; Jones, Christopher B; Button, Kate

    2015-01-01

    In the realm of knee pathology, magnetic resonance imaging (MRI) has the advantage of visualising all structures within the knee joint, which makes it a valuable tool for increasing diagnostic accuracy and planning surgical treatments. Therefore, clinical narratives found in MRI reports convey valuable diagnostic information. A range of studies have proven the feasibility of natural language processing for information extraction from clinical narratives. However, no study focused specifically on MRI reports in relation to knee pathology, possibly due to the complexity of knee anatomy and a wide range of conditions that may be associated with different anatomical entities. In this paper we describe KneeTex, an information extraction system that operates in this domain. As an ontology-driven information extraction system, KneeTex makes active use of an ontology to strongly guide and constrain text analysis. We used automatic term recognition to facilitate the development of a domain-specific ontology with sufficient detail and coverage for text mining applications. In combination with the ontology, high regularity of the sublanguage used in knee MRI reports allowed us to model its processing by a set of sophisticated lexico-semantic rules with minimal syntactic analysis. The main processing steps involve named entity recognition combined with coordination, enumeration, ambiguity and co-reference resolution, followed by text segmentation. Ontology-based semantic typing is then used to drive the template filling process. We adopted an existing ontology, TRAK (Taxonomy for RehAbilitation of Knee conditions), for use within KneeTex. The original TRAK ontology expanded from 1,292 concepts, 1,720 synonyms and 518 relationship instances to 1,621 concepts, 2,550 synonyms and 560 relationship instances. This provided KneeTex with a very fine-grained lexico-semantic knowledge base, which is highly attuned to the given sublanguage. Information extraction results were evaluated

  17. Techniques to extract physical modes in model-independent analysis of rings

    International Nuclear Information System (INIS)

    Wang, C.-X.

    2004-01-01

    A basic goal of Model-Independent Analysis is to extract the physical modes underlying the beam histories collected at a large number of beam position monitors so that beam dynamics and machine properties can be deduced independent of specific machine models. Here we discuss techniques to achieve this goal, especially the Principal Component Analysis and the Independent Component Analysis.

  18. Information Warfare in the Cyber Domain

    National Research Council Canada - National Science Library

    Takemoto, Glenn

    2001-01-01

    ...). This paper lays a foundation by defining the terminology associated with Information Warfare in the Cyber Domain, reviews the threat and illustrates the vulnerabilities of our information systems...

  19. Image encryption using random sequence generated from generalized information domain

    International Nuclear Information System (INIS)

    Zhang Xia-Yan; Wu Jie-Hua; Zhang Guo-Ji; Li Xuan; Ren Ya-Zhou

    2016-01-01

    A novel image encryption method based on the random sequence generated from the generalized information domain and permutation–diffusion architecture is proposed. The random sequence is generated by reconstruction from the generalized information file and discrete trajectory extraction from the data stream. The trajectory address sequence is used to generate a P-box to shuffle the plain image while random sequences are treated as keystreams. A new factor called drift factor is employed to accelerate and enhance the performance of the random sequence generator. An initial value is introduced to make the encryption method an approximately one-time pad. Experimental results show that the random sequences pass the NIST statistical test with a high ratio and extensive analysis demonstrates that the new encryption scheme has superior security. (paper)

  20. Developing a complex independent component analysis technique to extract non-stationary patterns from geophysical time-series

    Science.gov (United States)

    Forootan, Ehsan; Kusche, Jürgen

    2016-04-01

    Geodetic/geophysical observations, such as the time series of global terrestrial water storage change or sea level and temperature change, represent samples of physical processes and therefore contain information about complex physical interactionswith many inherent time scales. Extracting relevant information from these samples, for example quantifying the seasonality of a physical process or its variability due to large-scale ocean-atmosphere interactions, is not possible by rendering simple time series approaches. In the last decades, decomposition techniques have found increasing interest for extracting patterns from geophysical observations. Traditionally, principal component analysis (PCA) and more recently independent component analysis (ICA) are common techniques to extract statistical orthogonal (uncorrelated) and independent modes that represent the maximum variance of observations, respectively. PCA and ICA can be classified as stationary signal decomposition techniques since they are based on decomposing the auto-covariance matrix or diagonalizing higher (than two)-order statistical tensors from centered time series. However, the stationary assumption is obviously not justifiable for many geophysical and climate variables even after removing cyclic components e.g., the seasonal cycles. In this paper, we present a new decomposition method, the complex independent component analysis (CICA, Forootan, PhD-2014), which can be applied to extract to non-stationary (changing in space and time) patterns from geophysical time series. Here, CICA is derived as an extension of real-valued ICA (Forootan and Kusche, JoG-2012), where we (i) define a new complex data set using a Hilbert transformation. The complex time series contain the observed values in their real part, and the temporal rate of variability in their imaginary part. (ii) An ICA algorithm based on diagonalization of fourth-order cumulants is then applied to decompose the new complex data set in (i

  1. Integrated Phoneme Subspace Method for Speech Feature Extraction

    Directory of Open Access Journals (Sweden)

    Park Hyunsin

    2009-01-01

    Full Text Available Speech feature extraction has been a key focus in robust speech recognition research. In this work, we discuss data-driven linear feature transformations applied to feature vectors in the logarithmic mel-frequency filter bank domain. Transformations are based on principal component analysis (PCA, independent component analysis (ICA, and linear discriminant analysis (LDA. Furthermore, this paper introduces a new feature extraction technique that collects the correlation information among phoneme subspaces and reconstructs feature space for representing phonemic information efficiently. The proposed speech feature vector is generated by projecting an observed vector onto an integrated phoneme subspace (IPS based on PCA or ICA. The performance of the new feature was evaluated for isolated word speech recognition. The proposed method provided higher recognition accuracy than conventional methods in clean and reverberant environments.

  2. Text Processing of Domain-Related Information for Individuals with High and Low Domain Knowledge.

    Science.gov (United States)

    Spilich, George J.; And Others

    1979-01-01

    The way in which previously acquired knowledge affects the processing on new domain-related information was investigated. Text processing was studied in two groups differing in knowledge of the domain of baseball. A knowledge structure for the domain was constructed, and text propositions were classified. (SW)

  3. Decoding rule search domain in the left inferior frontal gyrus

    Science.gov (United States)

    Babcock, Laura; Vallesi, Antonino

    2018-01-01

    Traditionally, the left hemisphere has been thought to extract mainly verbal patterns of information, but recent evidence has shown that the left Inferior Frontal Gyrus (IFG) is active during inductive reasoning in both the verbal and spatial domains. We aimed to understand whether the left IFG supports inductive reasoning in a domain-specific or domain-general fashion. To do this we used Multi-Voxel Pattern Analysis to decode the representation of domain during a rule search task. Thirteen participants were asked to extract the rule underlying streams of letters presented in different spatial locations. Each rule was either verbal (letters forming words) or spatial (positions forming geometric figures). Our results show that domain was decodable in the left prefrontal cortex, suggesting that this region represents domain-specific information, rather than processes common to the two domains. A replication study with the same participants tested two years later confirmed these findings, though the individual representations changed, providing evidence for the flexible nature of representations. This study extends our knowledge on the neural basis of goal-directed behaviors and on how information relevant for rule extraction is flexibly mapped in the prefrontal cortex. PMID:29547623

  4. Information extraction from multi-institutional radiology reports.

    Science.gov (United States)

    Hassanpour, Saeed; Langlotz, Curtis P

    2016-01-01

    The radiology report is the most important source of clinical imaging information. It documents critical information about the patient's health and the radiologist's interpretation of medical findings. It also communicates information to the referring physicians and records that information for future clinical and research use. Although efforts to structure some radiology report information through predefined templates are beginning to bear fruit, a large portion of radiology report information is entered in free text. The free text format is a major obstacle for rapid extraction and subsequent use of information by clinicians, researchers, and healthcare information systems. This difficulty is due to the ambiguity and subtlety of natural language, complexity of described images, and variations among different radiologists and healthcare organizations. As a result, radiology reports are used only once by the clinician who ordered the study and rarely are used again for research and data mining. In this work, machine learning techniques and a large multi-institutional radiology report repository are used to extract the semantics of the radiology report and overcome the barriers to the re-use of radiology report information in clinical research and other healthcare applications. We describe a machine learning system to annotate radiology reports and extract report contents according to an information model. This information model covers the majority of clinically significant contents in radiology reports and is applicable to a wide variety of radiology study types. Our automated approach uses discriminative sequence classifiers for named-entity recognition to extract and organize clinically significant terms and phrases consistent with the information model. We evaluated our information extraction system on 150 radiology reports from three major healthcare organizations and compared its results to a commonly used non-machine learning information extraction method. We

  5. Secretory Vesicle Priming by CAPS Is Independent of Its SNARE-Binding MUN Domain

    Directory of Open Access Journals (Sweden)

    Cuc Quynh Nguyen Truong

    2014-11-01

    Full Text Available Priming of secretory vesicles is a prerequisite for their Ca2+-dependent fusion with the plasma membrane. The key vesicle priming proteins, Munc13s and CAPSs, are thought to mediate vesicle priming by regulating the conformation of the t-SNARE syntaxin, thereby facilitating SNARE complex assembly. Munc13s execute their priming function through their MUN domain. Given that the MUN domain of Ca2+-dependent activator protein for secretion (CAPS also binds syntaxin, it was assumed that CAPSs prime vesicles through the same mechanism as Munc13s. We studied naturally occurring splice variants of CAPS2 in CAPS1/CAPS2-deficient cells and found that CAPS2 primes vesicles independently of its MUN domain. Instead, the pleckstrin homology domain of CAPS2 seemingly is essential for its priming function. Our findings indicate a priming mode for secretory vesicles. This process apparently requires membrane phospholipids, does not involve the binding or direct conformational regulation of syntaxin by MUN domains of CAPSs, and is therefore not redundant with Munc13 action.

  6. Zone analysis in biology articles as a basis for information extraction.

    Science.gov (United States)

    Mizuta, Yoko; Korhonen, Anna; Mullen, Tony; Collier, Nigel

    2006-06-01

    In the field of biomedicine, an overwhelming amount of experimental data has become available as a result of the high throughput of research in this domain. The amount of results reported has now grown beyond the limits of what can be managed by manual means. This makes it increasingly difficult for the researchers in this area to keep up with the latest developments. Information extraction (IE) in the biological domain aims to provide an effective automatic means to dynamically manage the information contained in archived journal articles and abstract collections and thus help researchers in their work. However, while considerable advances have been made in certain areas of IE, pinpointing and organizing factual information (such as experimental results) remains a challenge. In this paper we propose tackling this task by incorporating into IE information about rhetorical zones, i.e. classification of spans of text in terms of argumentation and intellectual attribution. As the first step towards this goal, we introduce a scheme for annotating biological texts for rhetorical zones and provide a qualitative and quantitative analysis of the data annotated according to this scheme. We also discuss our preliminary research on automatic zone analysis, and its incorporation into our IE framework.

  7. Performance comparison of six independent components analysis algorithms for fetal signal extraction from real fMCG data

    International Nuclear Information System (INIS)

    Hild, Kenneth E; Alleva, Giovanna; Nagarajan, Srikantan; Comani, Silvia

    2007-01-01

    In this study we compare the performance of six independent components analysis (ICA) algorithms on 16 real fetal magnetocardiographic (fMCG) datasets for the application of extracting the fetal cardiac signal. We also compare the extraction results for real data with the results previously obtained for synthetic data. The six ICA algorithms are FastICA, CubICA, JADE, Infomax, MRMI-SIG and TDSEP. The results obtained using real fMCG data indicate that the FastICA method consistently outperforms the others in regard to separation quality and that the performance of an ICA method that uses temporal information suffers in the presence of noise. These two results confirm the previous results obtained using synthetic fMCG data. There were also two notable differences between the studies based on real and synthetic data. The differences are that all six ICA algorithms are independent of gestational age and sensor dimensionality for synthetic data, but depend on gestational age and sensor dimensionality for real data. It is possible to explain these differences by assuming that the number of point sources needed to completely explain the data is larger than the dimensionality used in the ICA extraction

  8. ONTOGRABBING: Extracting Information from Texts Using Generative Ontologies

    DEFF Research Database (Denmark)

    Nilsson, Jørgen Fischer; Szymczak, Bartlomiej Antoni; Jensen, P.A.

    2009-01-01

    We describe principles for extracting information from texts using a so-called generative ontology in combination with syntactic analysis. Generative ontologies are introduced as semantic domains for natural language phrases. Generative ontologies extend ordinary finite ontologies with rules...... for producing recursively shaped terms representing the ontological content (ontological semantics) of NL noun phrases and other phrases. We focus here on achieving a robust, often only partial, ontology-driven parsing of and ascription of semantics to a sentence in the text corpus. The aim of the ontological...... analysis is primarily to identify paraphrases, thereby achieving a search functionality beyond mere keyword search with synsets. We further envisage use of the generative ontology as a phrase-based rather than word-based browser into text corpora....

  9. Supporting the information domains of fall-risk management in home care via health information technology.

    Science.gov (United States)

    Alhuwail, Dari; Koru, Güneş; Mills, Mary Etta

    2016-01-01

    In the United States, home care clinicians often start the episode of care devoid of relevant fall-risk information. By collecting and analyzing qualitative data from 30 clinicians in one home health agency, this case study aimed to understand how the currently adopted information technology solutions supported the clinicians' fall-risk management (FRM) information domains, and explored opportunities to adopt other solutions to better support FRM. The currently adopted electronic health record system and fall-reporting application served only some information domains with a limited capacity. Substantial improvement in addressing the FRM information domains is possible by effectively modifying the existing solutions and purposefully adopting new solutions.

  10. Solution structure of tensin2 SH2 domain and its phosphotyrosine-independent interaction with DLC-1.

    Directory of Open Access Journals (Sweden)

    Kun Dai

    Full Text Available Src homology 2 (SH2 domain is a conserved module involved in various biological processes. Tensin family member was reported to be involved in tumor suppression by interacting with DLC-1 (deleted-in-liver-cancer-1 via its SH2 domain. We explore here the important questions that what the structure of tensin2 SH2 domain is, and how it binds to DLC-1, which might reveal a novel binding mode.Tensin2 SH2 domain adopts a conserved SH2 fold that mainly consists of five β-strands flanked by two α-helices. Most SH2 domains recognize phosphorylated ligands specifically. However, tensin2 SH2 domain was identified to interact with nonphosphorylated ligand (DLC-1 as well as phosphorylated ligand.We determined the solution structure of tensin2 SH2 domain using NMR spectroscopy, and revealed the interactions between tensin2 SH2 domain and its ligands in a phosphotyrosine-independent manner.

  11. Quantifying information transfer by protein domains: Analysis of the Fyn SH2 domain structure

    DEFF Research Database (Denmark)

    Lenaerts, Tom; Ferkinghoff-Borg, Jesper; Stricher, Francois

    2008-01-01

    instance of communication over a noisy channel. In particular, we analyze the conformational correlations between protein residues and apply the concept of mutual information to quantify information exchange. Mapping out changes of mutual information on the protein structure then allows visualizing how...... distal communication is achieved. We illustrate the approach by analyzing information transfer by the SH2 domain of Fyn tyrosine kinase, obtained from Monte Carlo dynamics simulations. Our analysis reveals that the Fyn SH2 domain forms a noisy communication channel that couples residues located......Background: Efficient communication between distant sites within a protein is essential for cooperative biological response. Although often associated with large allosteric movements, more subtle changes in protein dynamics can also induce long-range correlations. However, an appropriate formalism...

  12. Kernel-Based Learning for Domain-Specific Relation Extraction

    Science.gov (United States)

    Basili, Roberto; Giannone, Cristina; Del Vescovo, Chiara; Moschitti, Alessandro; Naggar, Paolo

    In a specific process of business intelligence, i.e. investigation on organized crime, empirical language processing technologies can play a crucial role. The analysis of transcriptions on investigative activities, such as police interrogatories, for the recognition and storage of complex relations among people and locations is a very difficult and time consuming task, ultimately based on pools of experts. We discuss here an inductive relation extraction platform that opens the way to much cheaper and consistent workflows. The presented empirical investigation shows that accurate results, comparable to the expert teams, can be achieved, and parametrization allows to fine tune the system behavior for fitting domain-specific requirements.

  13. Extracting useful information from images

    DEFF Research Database (Denmark)

    Kucheryavskiy, Sergey

    2011-01-01

    The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic and heter......The paper presents an overview of methods for extracting useful information from digital images. It covers various approaches that utilized different properties of images, like intensity distribution, spatial frequencies content and several others. A few case studies including isotropic...

  14. MiDas: automatic extraction of a common domain of discourse in sleep medicine for multi-center data integration.

    Science.gov (United States)

    Sahoo, Satya S; Ogbuji, Chimezie; Luo, Lingyun; Dong, Xiao; Cui, Licong; Redline, Susan S; Zhang, Guo-Qiang

    2011-01-01

    Clinical studies often use data dictionaries with controlled sets of terms to facilitate data collection, limited interoperability and sharing at a local site. Multi-center retrospective clinical studies require that these data dictionaries, originating from individual participating centers, be harmonized in preparation for the integration of the corresponding clinical research data. Domain ontologies are often used to facilitate multi-center data integration by modeling terms from data dictionaries in a logic-based language, but interoperability among domain ontologies (using automated techniques) is an unresolved issue. Although many upper-level reference ontologies have been proposed to address this challenge, our experience in integrating multi-center sleep medicine data highlights the need for an upper level ontology that models a common set of terms at multiple-levels of abstraction, which is not covered by the existing upper-level ontologies. We introduce a methodology underpinned by a Minimal Domain of Discourse (MiDas) algorithm to automatically extract a minimal common domain of discourse (upper-domain ontology) from an existing domain ontology. Using the Multi-Modality, Multi-Resource Environment for Physiological and Clinical Research (Physio-MIMI) multi-center project in sleep medicine as a use case, we demonstrate the use of MiDas in extracting a minimal domain of discourse for sleep medicine, from Physio-MIMI's Sleep Domain Ontology (SDO). We then extend the resulting domain of discourse with terms from the data dictionary of the Sleep Heart and Health Study (SHHS) to validate MiDas. To illustrate the wider applicability of MiDas, we automatically extract the respective domains of discourse from 6 sample domain ontologies from the National Center for Biomedical Ontologies (NCBO) and the OBO Foundry.

  15. How to Program a Domain Independent Tracer for Explanations

    Science.gov (United States)

    Ishizaka, Alessio; Lusti, Markus

    2006-01-01

    Explanations are essential in the teaching process. Tracers are one possibility to provide students with explanations in an intelligent tutoring system. Their development can be divided into four steps: (a) the definition of the trace model; (b) the extraction of the information from this model; (c) the analysis and abstraction of the extracted…

  16. Analysis of Technique to Extract Data from the Web for Improved Performance

    Science.gov (United States)

    Gupta, Neena; Singh, Manish

    2010-11-01

    The World Wide Web rapidly guides the world into a newly amazing electronic world, where everyone can publish anything in electronic form and extract almost all the information. Extraction of information from semi structured or unstructured documents, such as web pages, is a useful yet complex task. Data extraction, which is important for many applications, extracts the records from the HTML files automatically. Ontologies can achieve a high degree of accuracy in data extraction. We analyze method for data extraction OBDE (Ontology-Based Data Extraction), which automatically extracts the query result records from the web with the help of agents. OBDE first constructs an ontology for a domain according to information matching between the query interfaces and query result pages from different web sites within the same domain. Then, the constructed domain ontology is used during data extraction to identify the query result section in a query result page and to align and label the data values in the extracted records. The ontology-assisted data extraction method is fully automatic and overcomes many of the deficiencies of current automatic data extraction methods.

  17. Feature extraction for magnetic domain images of magneto-optical recording films using gradient feature segmentation

    International Nuclear Information System (INIS)

    Quanqing, Zhu.; Xinsai, Wang; Xuecheng, Zou; Haihua, Li; Xiaofei, Yang

    2002-01-01

    In this paper, we present a method to realize feature extraction on low contrast magnetic domain images of magneto-optical recording films. The method is based on the following three steps: first, Lee-filtering method is adopted to realize pre-filtering and noise reduction; this is followed by gradient feature segmentation, which separates the object area from the background area; finally the common linking method is adopted and the characteristic parameters of magnetic domain are calculated. We describe these steps with particular emphasis on the gradient feature segmentation. The results show that this method has advantages over other traditional ones for feature extraction of low contrast images

  18. Extraction of Information of Audio-Visual Contents

    Directory of Open Access Journals (Sweden)

    Carlos Aguilar

    2011-10-01

    Full Text Available In this article we show how it is possible to use Channel Theory (Barwise and Seligman, 1997 for modeling the process of information extraction realized by audiences of audio-visual contents. To do this, we rely on the concepts pro- posed by Channel Theory and, especially, its treatment of representational systems. We then show how the information that an agent is capable of extracting from the content depends on the number of channels he is able to establish between the content and the set of classifications he is able to discriminate. The agent can endeavor the extraction of information through these channels from the totality of content; however, we discuss the advantages of extracting from its constituents in order to obtain a greater number of informational items that represent it. After showing how the extraction process is endeavored for each channel, we propose a method of representation of all the informative values an agent can obtain from a content using a matrix constituted by the channels the agent is able to establish on the content (source classifications, and the ones he can understand as individual (destination classifications. We finally show how this representation allows reflecting the evolution of the informative items through the evolution of audio-visual content.

  19. Integrating angle-frequency domain synchronous averaging technique with feature extraction for gear fault diagnosis

    Science.gov (United States)

    Zhang, Shengli; Tang, J.

    2018-01-01

    Gear fault diagnosis relies heavily on the scrutiny of vibration responses measured. In reality, gear vibration signals are noisy and dominated by meshing frequencies as well as their harmonics, which oftentimes overlay the fault related components. Moreover, many gear transmission systems, e.g., those in wind turbines, constantly operate under non-stationary conditions. To reduce the influences of non-synchronous components and noise, a fault signature enhancement method that is built upon angle-frequency domain synchronous averaging is developed in this paper. Instead of being averaged in the time domain, the signals are processed in the angle-frequency domain to solve the issue of phase shifts between signal segments due to uncertainties caused by clearances, input disturbances, and sampling errors, etc. The enhanced results are then analyzed through feature extraction algorithms to identify the most distinct features for fault classification and identification. Specifically, Kernel Principal Component Analysis (KPCA) targeting at nonlinearity, Multilinear Principal Component Analysis (MPCA) targeting at high dimensionality, and Locally Linear Embedding (LLE) targeting at local similarity among the enhanced data are employed and compared to yield insights. Numerical and experimental investigations are performed, and the results reveal the effectiveness of angle-frequency domain synchronous averaging in enabling feature extraction and classification.

  20. Bibliographic information interchange with data independence

    International Nuclear Information System (INIS)

    Macedo, L.F.P. de

    1980-01-01

    A technique for bibliographic information processing, at the standardized interchange format ISO 2709 or ANSI Z39.2, which allows to obtain programs with data independence is developed. A format definition language and manipulation commands which enable data storage and retrieval at the logical level are specified. With this technique it is possible to implement programs which can accept data from any information system, reducing, therefore, the requirements of a common bibliographic information interchange format. (Author) [pt

  1. Scenario Customization for Information Extraction

    National Research Council Canada - National Science Library

    Yangarber, Roman

    2001-01-01

    Information Extraction (IE) is an emerging NLP technology, whose function is to process unstructured, natural language text, to locate specific pieces of information, or facts, in the text, and to use these facts to fill a database...

  2. Can we replace curation with information extraction software?

    Science.gov (United States)

    Karp, Peter D

    2016-01-01

    Can we use programs for automated or semi-automated information extraction from scientific texts as practical alternatives to professional curation? I show that error rates of current information extraction programs are too high to replace professional curation today. Furthermore, current IEP programs extract single narrow slivers of information, such as individual protein interactions; they cannot extract the large breadth of information extracted by professional curators for databases such as EcoCyc. They also cannot arbitrate among conflicting statements in the literature as curators can. Therefore, funding agencies should not hobble the curation efforts of existing databases on the assumption that a problem that has stymied Artificial Intelligence researchers for more than 60 years will be solved tomorrow. Semi-automated extraction techniques appear to have significantly more potential based on a review of recent tools that enhance curator productivity. But a full cost-benefit analysis for these tools is lacking. Without such analysis it is possible to expend significant effort developing information-extraction tools that automate small parts of the overall curation workflow without achieving a significant decrease in curation costs.Database URL. © The Author(s) 2016. Published by Oxford University Press.

  3. Transductive Pattern Learning for Information Extraction

    National Research Council Canada - National Science Library

    McLernon, Brian; Kushmerick, Nicholas

    2006-01-01

    .... We present TPLEX, a semi-supervised learning algorithm for information extraction that can acquire extraction patterns from a small amount of labelled text in conjunction with a large amount of unlabelled text...

  4. Transactions in domain-specific information systems

    Science.gov (United States)

    Zacek, Jaroslav

    2017-07-01

    Substantial number of the current information system (IS) implementations is based on transaction approach. In addition, most of the implementations are domain-specific (e.g. accounting IS, resource planning IS). Therefore, we have to have a generic transaction model to build and verify domain-specific IS. The paper proposes a new transaction model for domain-specific ontologies. This model is based on value oriented business process modelling technique. The transaction model is formalized by the Petri Net theory. First part of the paper presents common business processes and analyses related to business process modeling. Second part defines the transactional model delimited by REA enterprise ontology paradigm and introduces states of the generic transaction model. The generic model proposal is defined and visualized by the Petri Net modelling tool. Third part shows application of the generic transaction model. Last part of the paper concludes results and discusses a practical usability of the generic transaction model.

  5. Information Technology (IT) and applied domain education in West ...

    African Journals Online (AJOL)

    Information Technology (IT) is one of the important name in today's world. Information Technology (IT) is important factor for buzzword, education, societal application and development, mass communication, politics, hospitality, health. Information Technology (IT) is an interdisciplinary domain which is mainly combination of ...

  6. Optical Aperture Synthesis Object's Information Extracting Based on Wavelet Denoising

    International Nuclear Information System (INIS)

    Fan, W J; Lu, Y

    2006-01-01

    Wavelet denoising is studied to improve OAS(optical aperture synthesis) object's Fourier information extracting. Translation invariance wavelet denoising based on Donoho wavelet soft threshold denoising is researched to remove Pseudo-Gibbs in wavelet soft threshold image. OAS object's information extracting based on translation invariance wavelet denoising is studied. The study shows that wavelet threshold denoising can improve the precision and the repetition of object's information extracting from interferogram, and the translation invariance wavelet denoising information extracting is better than soft threshold wavelet denoising information extracting

  7. AERIS: An Integrated Domain Information System for Aerospace Science and Technology

    Science.gov (United States)

    Hatua, Sudip Ranjan; Madalli, Devika P.

    2011-01-01

    Purpose: The purpose of this paper is to discuss the methodology in building an integrated domain information system with illustrations that provide proof of concept. Design/methodology/approach: The present work studies the usual search engine approach to information and its pitfalls. A methodology was adopted for construction of a domain-based…

  8. Bell nonlocality: a resource for device-independent quantum information protocols

    Science.gov (United States)

    Acin, Antonio

    2015-05-01

    Bell nonlocality is not only one of the most fundamental properties of quantum physics, but has also recently acquired the status of an information resource for device-independent quantum information protocols. In the device-independent approach, protocols are designed so that their performance is independent of the internal working of the devices used in the implementation. We discuss all these ideas and argue that device-independent protocols are especially relevant or cryptographic applications, as they are insensitive to hacking attacks exploiting imperfections on the modelling of the devices.

  9. Blind Extraction of Chaotic Signals by Using the Fast Independent Component Analysis Algorithm

    International Nuclear Information System (INIS)

    Hong-Bin, Chen; Jiu-Chao, Feng; Yong, Fang

    2008-01-01

    We report the results of using the fast independent component analysis (FastICA) algorithm to realize blind extraction of chaotic signals. Two cases are taken into consideration: namely, the mixture is noiseless or contaminated by noise. Pre-whitening is employed to reduce the effect of noise before using the FastICA algorithm. The correlation coefficient criterion is adopted to evaluate the performance, and the success rate is defined as a new criterion to indicate the performance with respect to noise or different mixing matrices. Simulation results show that the FastICA algorithm can extract the chaotic signals effectively. The impact of noise, the length of a signal frame, the number of sources and the number of observed mixtures on the performance is investigated in detail. It is also shown that regarding a noise as an independent source is not always correct

  10. Cholesterol Bilayer Domains in the Eye Lens Health: A Review.

    Science.gov (United States)

    Widomska, Justyna; Subczynski, Witold K; Mainali, Laxman; Raguz, Marija

    2017-12-01

    The most unique biochemical characteristic of the eye lens fiber cell plasma membrane is its extremely high cholesterol content, the need for which is still unclear. It is evident, however, that the disturbance of Chol homeostasis may result in damages associated with cataracts. Electron paramagnetic resonance methods allow discrimination of two types of lipid domains in model membranes overloaded with Chol, namely, phospholipid-cholesterol domains and pure Chol bilayer domains. These domains are also detected in human lens lipid membranes prepared from the total lipids extracted from lens cortices and nuclei of donors from different age groups. Independent of the age-related changes in phospholipid composition, the physical properties of phospholipid-Chol domains remain the same for all age groups and are practically identical for cortical and nuclear membranes. The presence of Chol bilayer domains in these membranes provides a buffering capacity for cholesterol concentration in the surrounding phospholipid-Chol domains, keeping it at a constant saturating level and thus keeping the physical properties of the membrane consistent with and independent of changes in phospholipid composition. It seems that the presence of Chol bilayer domains plays an integral role in the regulation of cholesterol-dependent processes in fiber cell plasm membranes and in the maintenance of fiber cell membrane homeostasis.

  11. Dimension-independent likelihood-informed MCMC

    KAUST Repository

    Cui, Tiangang

    2015-10-08

    Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters that represent the discretization of an underlying function. This work introduces a family of Markov chain Monte Carlo (MCMC) samplers that can adapt to the particular structure of a posterior distribution over functions. Two distinct lines of research intersect in the methods developed here. First, we introduce a general class of operator-weighted proposal distributions that are well defined on function space, such that the performance of the resulting MCMC samplers is independent of the discretization of the function. Second, by exploiting local Hessian information and any associated low-dimensional structure in the change from prior to posterior distributions, we develop an inhomogeneous discretization scheme for the Langevin stochastic differential equation that yields operator-weighted proposals adapted to the non-Gaussian structure of the posterior. The resulting dimension-independent and likelihood-informed (DILI) MCMC samplers may be useful for a large class of high-dimensional problems where the target probability measure has a density with respect to a Gaussian reference measure. Two nonlinear inverse problems are used to demonstrate the efficiency of these DILI samplers: an elliptic PDE coefficient inverse problem and path reconstruction in a conditioned diffusion.

  12. Dimension-independent likelihood-informed MCMC

    KAUST Repository

    Cui, Tiangang; Law, Kody; Marzouk, Youssef M.

    2015-01-01

    Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters that represent the discretization of an underlying function. This work introduces a family of Markov chain Monte Carlo (MCMC) samplers that can adapt to the particular structure of a posterior distribution over functions. Two distinct lines of research intersect in the methods developed here. First, we introduce a general class of operator-weighted proposal distributions that are well defined on function space, such that the performance of the resulting MCMC samplers is independent of the discretization of the function. Second, by exploiting local Hessian information and any associated low-dimensional structure in the change from prior to posterior distributions, we develop an inhomogeneous discretization scheme for the Langevin stochastic differential equation that yields operator-weighted proposals adapted to the non-Gaussian structure of the posterior. The resulting dimension-independent and likelihood-informed (DILI) MCMC samplers may be useful for a large class of high-dimensional problems where the target probability measure has a density with respect to a Gaussian reference measure. Two nonlinear inverse problems are used to demonstrate the efficiency of these DILI samplers: an elliptic PDE coefficient inverse problem and path reconstruction in a conditioned diffusion.

  13. Cause Information Extraction from Financial Articles Concerning Business Performance

    Science.gov (United States)

    Sakai, Hiroyuki; Masuyama, Shigeru

    We propose a method of extracting cause information from Japanese financial articles concerning business performance. Our method acquires cause informtion, e. g. “_??__??__??__??__??__??__??__??__??__??_ (zidousya no uriage ga koutyou: Sales of cars were good)”. Cause information is useful for investors in selecting companies to invest. Our method extracts cause information as a form of causal expression by using statistical information and initial clue expressions automatically. Our method can extract causal expressions without predetermined patterns or complex rules given by hand, and is expected to be applied to other tasks for acquiring phrases that have a particular meaning not limited to cause information. We compared our method with our previous one originally proposed for extracting phrases concerning traffic accident causes and experimental results showed that our new method outperforms our previous one.

  14. Semantics-driven modelling of user preferences for information retrieval in the biomedical domain.

    Science.gov (United States)

    Gladun, Anatoly; Rogushina, Julia; Valencia-García, Rafael; Béjar, Rodrigo Martínez

    2013-03-01

    A large amount of biomedical and genomic data are currently available on the Internet. However, data are distributed into heterogeneous biological information sources, with little or even no organization. Semantic technologies provide a consistent and reliable basis with which to confront the challenges involved in the organization, manipulation and visualization of data and knowledge. One of the knowledge representation techniques used in semantic processing is the ontology, which is commonly defined as a formal and explicit specification of a shared conceptualization of a domain of interest. The work presented here introduces a set of interoperable algorithms that can use domain and ontological information to improve information-retrieval processes. This work presents an ontology-based information-retrieval system for the biomedical domain. This system, with which some experiments have been carried out that are described in this paper, is based on the use of domain ontologies for the creation and normalization of lightweight ontologies that represent user preferences in a determined domain in order to improve information-retrieval processes.

  15. Gearbox fault diagnosis based on time-frequency domain synchronous averaging and feature extraction technique

    Science.gov (United States)

    Zhang, Shengli; Tang, Jiong

    2016-04-01

    Gearbox is one of the most vulnerable subsystems in wind turbines. Its healthy status significantly affects the efficiency and function of the entire system. Vibration based fault diagnosis methods are prevalently applied nowadays. However, vibration signals are always contaminated by noise that comes from data acquisition errors, structure geometric errors, operation errors, etc. As a result, it is difficult to identify potential gear failures directly from vibration signals, especially for the early stage faults. This paper utilizes synchronous averaging technique in time-frequency domain to remove the non-synchronous noise and enhance the fault related time-frequency features. The enhanced time-frequency information is further employed in gear fault classification and identification through feature extraction algorithms including Kernel Principal Component Analysis (KPCA), Multilinear Principal Component Analysis (MPCA), and Locally Linear Embedding (LLE). Results show that the LLE approach is the most effective to classify and identify different gear faults.

  16. Entropy based classifier for cross-domain opinion mining

    Directory of Open Access Journals (Sweden)

    Jyoti S. Deshmukh

    2018-01-01

    Full Text Available In recent years, the growth of social network has increased the interest of people in analyzing reviews and opinions for products before they buy them. Consequently, this has given rise to the domain adaptation as a prominent area of research in sentiment analysis. A classifier trained from one domain often gives poor results on data from another domain. Expression of sentiment is different in every domain. The labeling cost of each domain separately is very high as well as time consuming. Therefore, this study has proposed an approach that extracts and classifies opinion words from one domain called source domain and predicts opinion words of another domain called target domain using a semi-supervised approach, which combines modified maximum entropy and bipartite graph clustering. A comparison of opinion classification on reviews on four different product domains is presented. The results demonstrate that the proposed method performs relatively well in comparison to the other methods. Comparison of SentiWordNet of domain-specific and domain-independent words reveals that on an average 72.6% and 88.4% words, respectively, are correctly classified.

  17. Sample-based XPath Ranking for Web Information Extraction

    NARCIS (Netherlands)

    Jundt, Oliver; van Keulen, Maurice

    Web information extraction typically relies on a wrapper, i.e., program code or a configuration that specifies how to extract some information from web pages at a specific website. Manually creating and maintaining wrappers is a cumbersome and error-prone task. It may even be prohibitive as some

  18. Ontology-Based Information Extraction for Business Intelligence

    Science.gov (United States)

    Saggion, Horacio; Funk, Adam; Maynard, Diana; Bontcheva, Kalina

    Business Intelligence (BI) requires the acquisition and aggregation of key pieces of knowledge from multiple sources in order to provide valuable information to customers or feed statistical BI models and tools. The massive amount of information available to business analysts makes information extraction and other natural language processing tools key enablers for the acquisition and use of that semantic information. We describe the application of ontology-based extraction and merging in the context of a practical e-business application for the EU MUSING Project where the goal is to gather international company intelligence and country/region information. The results of our experiments so far are very promising and we are now in the process of building a complete end-to-end solution.

  19. 77 FR 17098 - Proposed Extension of Existing Information Collection; Independent Contractor Registration and...

    Science.gov (United States)

    2012-03-23

    ... Information Collection; Independent Contractor Registration and Identification AGENCY: Mine Safety and Health...-00040, Independent Contractor Register. OMB last approved this information collection request (ICR) on...); or 202-693-9441 (facsimile). SUPPLEMENTARY INFORMATION: I. Background Independent contractors...

  20. A Two-Step Resume Information Extraction Algorithm

    Directory of Open Access Journals (Sweden)

    Jie Chen

    2018-01-01

    Full Text Available With the rapid growth of Internet-based recruiting, there are a great number of personal resumes among recruiting systems. To gain more attention from the recruiters, most resumes are written in diverse formats, including varying font size, font colour, and table cells. However, the diversity of format is harmful to data mining, such as resume information extraction, automatic job matching, and candidates ranking. Supervised methods and rule-based methods have been proposed to extract facts from resumes, but they strongly rely on hierarchical structure information and large amounts of labelled data, which are hard to collect in reality. In this paper, we propose a two-step resume information extraction approach. In the first step, raw text of resume is identified as different resume blocks. To achieve the goal, we design a novel feature, Writing Style, to model sentence syntax information. Besides word index and punctuation index, word lexical attribute and prediction results of classifiers are included in Writing Style. In the second step, multiple classifiers are employed to identify different attributes of fact information in resumes. Experimental results on a real-world dataset show that the algorithm is feasible and effective.

  1. Extractive Summarisation of Medical Documents

    Directory of Open Access Journals (Sweden)

    Abeed Sarker

    2012-09-01

    Full Text Available Background Evidence Based Medicine (EBM practice requires practitioners to extract evidence from published medical research when answering clinical queries. Due to the time-consuming nature of this practice, there is a strong motivation for systems that can automatically summarise medical documents and help practitioners find relevant information. Aim The aim of this work is to propose an automatic query-focused, extractive summarisation approach that selects informative sentences from medical documents. MethodWe use a corpus that is specifically designed for summarisation in the EBM domain. We use approximately half the corpus for deriving important statistics associated with the best possible extractive summaries. We take into account factors such as sentence position, length, sentence content, and the type of the query posed. Using the statistics from the first set, we evaluate our approach on a separate set. Evaluation of the qualities of the generated summaries is performed automatically using ROUGE, which is a popular tool for evaluating automatic summaries. Results Our summarisation approach outperforms all baselines (best baseline score: 0.1594; our score 0.1653. Further improvements are achieved when query types are taken into account. Conclusion The quality of extractive summarisation in the medical domain can be significantly improved by incorporating domain knowledge and statistics derived from a specialised corpus. Such techniques can therefore be applied for content selection in end-to-end summarisation systems.

  2. The Agent of extracting Internet Information with Lead Order

    Science.gov (United States)

    Mo, Zan; Huang, Chuliang; Liu, Aijun

    In order to carry out e-commerce better, advanced technologies to access business information are in need urgently. An agent is described to deal with the problems of extracting internet information that caused by the non-standard and skimble-scamble structure of Chinese websites. The agent designed includes three modules which respond to the process of extracting information separately. A method of HTTP tree and a kind of Lead algorithm is proposed to generate a lead order, with which the required web can be retrieved easily. How to transform the extracted information structuralized with natural language is also discussed.

  3. Cyber indicators of compromise: a domain ontology for security information and event management

    Science.gov (United States)

    2017-03-01

    heuristics, mapping, and detection. CybOX is aimed at supporting a broad range of important cyber security domains to include [31]: • Digital...REPORT TYPE AND DATES COVERED Master’s thesis 4. TITLE AND SUBTITLE CYBER INDICATORS OF COMPROMISE: A DOMAIN ONTOLOGY FOR SECURITY INFORMATION AND...Distribution is unlimited. CYBER INDICATORS OF COMPROMISE: A DOMAIN ONTOLOGY FOR SECURITY INFORMATION AND EVENT MANAGEMENT Marsha D. Rowell

  4. A method of directly extracting multiwave angle-domain common-image gathers

    Science.gov (United States)

    Han, Jianguang; Wang, Yun

    2017-10-01

    Angle-domain common-image gathers (ADCIGs) can provide an effective way for migration velocity analysis and amplitude versus angle analysis in oil-gas seismic exploration. On the basis of multi-component Gaussian beam prestack depth migration (GB-PSDM), an alternative method of directly extracting multiwave ADCIGs is presented in this paper. We first introduce multi-component GB-PSDM, where a wavefield separation is proceeded to obtain the separated PP- and PS-wave seismic records before migration imaging for multiwave seismic data. Then, the principle of extracting PP- and PS-ADCIGs using GB-PSDM is presented. The propagation angle can be obtained using the real-value travel time of Gaussian beam in the course of GB-PSDM, which can be used to calculate the incidence and reflection angles. Two kinds of ADCIGs can be extracted for the PS-wave, one of which is P-wave incidence ADCIGs and the other one is S-wave reflection ADCIGs. In this paper, we use the incident angle to plot the ADCIGs for both PP- and PS-waves. Finally, tests of synthetic examples show that the method introduced here is accurate and effective.

  5. Independent regulation of reovirus membrane penetration and apoptosis by the mu1 phi domain.

    Science.gov (United States)

    Danthi, Pranav; Coffey, Caroline M; Parker, John S L; Abel, Ty W; Dermody, Terence S

    2008-12-01

    Apoptosis plays an important role in the pathogenesis of reovirus encephalitis. Reovirus outer-capsid protein mu1, which functions to penetrate host cell membranes during viral entry, is the primary regulator of apoptosis following reovirus infection. Ectopic expression of full-length and truncated forms of mu1 indicates that the mu1 phi domain is sufficient to elicit a cell death response. To evaluate the contribution of the mu1 phi domain to the induction of apoptosis following reovirus infection, phi mutant viruses were generated by reverse genetics and analyzed for the capacity to penetrate cell membranes and elicit apoptosis. We found that mutations in phi diminish reovirus membrane penetration efficiency by preventing conformational changes that lead to generation of key reovirus entry intermediates. Independent of effects on membrane penetration, amino acid substitutions in phi affect the apoptotic potential of reovirus, suggesting that phi initiates apoptosis subsequent to cytosolic delivery. In comparison to wild-type virus, apoptosis-defective phi mutant viruses display diminished neurovirulence following intracranial inoculation of newborn mice. These results indicate that the phi domain of mu1 plays an important regulatory role in reovirus-induced apoptosis and disease.

  6. Independent regulation of reovirus membrane penetration and apoptosis by the mu1 phi domain.

    Directory of Open Access Journals (Sweden)

    Pranav Danthi

    2008-12-01

    Full Text Available Apoptosis plays an important role in the pathogenesis of reovirus encephalitis. Reovirus outer-capsid protein mu1, which functions to penetrate host cell membranes during viral entry, is the primary regulator of apoptosis following reovirus infection. Ectopic expression of full-length and truncated forms of mu1 indicates that the mu1 phi domain is sufficient to elicit a cell death response. To evaluate the contribution of the mu1 phi domain to the induction of apoptosis following reovirus infection, phi mutant viruses were generated by reverse genetics and analyzed for the capacity to penetrate cell membranes and elicit apoptosis. We found that mutations in phi diminish reovirus membrane penetration efficiency by preventing conformational changes that lead to generation of key reovirus entry intermediates. Independent of effects on membrane penetration, amino acid substitutions in phi affect the apoptotic potential of reovirus, suggesting that phi initiates apoptosis subsequent to cytosolic delivery. In comparison to wild-type virus, apoptosis-defective phi mutant viruses display diminished neurovirulence following intracranial inoculation of newborn mice. These results indicate that the phi domain of mu1 plays an important regulatory role in reovirus-induced apoptosis and disease.

  7. On domain modelling of the service system with its application to enterprise information systems

    Science.gov (United States)

    Wang, J. W.; Wang, H. F.; Ding, J. L.; Furuta, K.; Kanno, T.; Ip, W. H.; Zhang, W. J.

    2016-01-01

    Information systems are a kind of service systems and they are throughout every element of a modern industrial and business system, much like blood in our body. Types of information systems are heterogeneous because of extreme uncertainty in changes in modern industrial and business systems. To effectively manage information systems, modelling of the work domain (or domain) of information systems is necessary. In this paper, a domain modelling framework for the service system is proposed and its application to the enterprise information system is outlined. The framework is defined based on application of a general domain modelling tool called function-context-behaviour-principle-state-structure (FCBPSS). The FCBPSS is based on a set of core concepts, namely: function, context, behaviour, principle, state and structure and system decomposition. Different from many other applications of FCBPSS in systems engineering, the FCBPSS is applied to both infrastructure and substance systems, which is novel and effective to modelling of service systems including enterprise information systems. It is to be noted that domain modelling of systems (e.g. enterprise information systems) is a key to integration of heterogeneous systems and to coping with unanticipated situations facing to systems.

  8. Data Leakage Prevention for Secure Cross-Domain Information Exchange

    OpenAIRE

    Nordbotten, Nils Agne; Engelstad, Paal E.; Kongsgård, Kyrre Wahl; Haakseth, Raymond; Mancini, Federico

    2017-01-01

    Cross-domain information exchange is an increasingly important capability for conducting efficient and secure operations, both within coalitions and within single nations. A data guard is a common cross-domain sharing solution that inspects the security labels of exported data objects and validates that they are such that they can be released according to policy. While we see that guard solutions can be implemented with high assurance, we find that obtaining an equivalent level of assurance i...

  9. Dimension-Independent Likelihood-Informed MCMC

    KAUST Repository

    Cui, Tiangang; Law, Kody; Marzouk, Youssef

    2015-01-01

    Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters, which in principle can be described as functions. By exploiting low-dimensional structure in the change from prior to posterior [distributions], we introduce a suite of MCMC samplers that can adapt to the complex structure of the posterior distribution, yet are well-defined on function space. Posterior sampling in nonlinear inverse problems arising from various partial di erential equations and also a stochastic differential equation are used to demonstrate the e ciency of these dimension-independent likelihood-informed samplers.

  10. Dimension-Independent Likelihood-Informed MCMC

    KAUST Repository

    Cui, Tiangang

    2015-01-07

    Many Bayesian inference problems require exploring the posterior distribution of high-dimensional parameters, which in principle can be described as functions. By exploiting low-dimensional structure in the change from prior to posterior [distributions], we introduce a suite of MCMC samplers that can adapt to the complex structure of the posterior distribution, yet are well-defined on function space. Posterior sampling in nonlinear inverse problems arising from various partial di erential equations and also a stochastic differential equation are used to demonstrate the e ciency of these dimension-independent likelihood-informed samplers.

  11. Domain XML semantic integration based on extraction rules and ontology mapping

    Directory of Open Access Journals (Sweden)

    Huayu LI

    2016-08-01

    Full Text Available A plenty of XML documents exist in petroleum engineering field, but traditional XML integration solution can’t provide semantic query, which leads to low data use efficiency. In light of WeXML(oil&gas well XML data semantic integration and query requirement, this paper proposes a semantic integration method based on extraction rules and ontology mapping. The method firstly defines a series of extraction rules with which elements and properties of WeXML Schema are mapped to classes and properties in WeOWL ontology, respectively; secondly, an algorithm is used to transform WeXML documents into WeOWL instances. Because WeOWL provides limited semantics, ontology mappings between two ontologies are then built to explain class and property of global ontology with terms of WeOWL, and semantic query based on global domain concepts model is provided. By constructing a WeXML data semantic integration prototype system, the proposed transformational rule, the transfer algorithm and the mapping rule are tested.

  12. Image feature extraction in encrypted domain with privacy-preserving SIFT.

    Science.gov (United States)

    Hsu, Chao-Yung; Lu, Chun-Shien; Pei, Soo-Chang

    2012-11-01

    Privacy has received considerable attention but is still largely ignored in the multimedia community. Consider a cloud computing scenario where the server is resource-abundant, and is capable of finishing the designated tasks. It is envisioned that secure media applications with privacy preservation will be treated seriously. In view of the fact that scale-invariant feature transform (SIFT) has been widely adopted in various fields, this paper is the first to target the importance of privacy-preserving SIFT (PPSIFT) and to address the problem of secure SIFT feature extraction and representation in the encrypted domain. As all of the operations in SIFT must be moved to the encrypted domain, we propose a privacy-preserving realization of the SIFT method based on homomorphic encryption. We show through the security analysis based on the discrete logarithm problem and RSA that PPSIFT is secure against ciphertext only attack and known plaintext attack. Experimental results obtained from different case studies demonstrate that the proposed homomorphic encryption-based privacy-preserving SIFT performs comparably to the original SIFT and that our method is useful in SIFT-based privacy-preserving applications.

  13. Fetal ECG extraction using independent component analysis by Jade approach

    Science.gov (United States)

    Giraldo-Guzmán, Jader; Contreras-Ortiz, Sonia H.; Lasprilla, Gloria Isabel Bautista; Kotas, Marian

    2017-11-01

    Fetal ECG monitoring is a useful method to assess the fetus health and detect abnormal conditions. In this paper we propose an approach to extract fetal ECG from abdomen and chest signals using independent component analysis based on the joint approximate diagonalization of eigenmatrices approach. The JADE approach avoids redundancy, what reduces matrix dimension and computational costs. Signals were filtered with a high pass filter to eliminate low frequency noise. Several levels of decomposition were tested until the fetal ECG was recognized in one of the separated sources output. The proposed method shows fast and good performance.

  14. Music video shot segmentation using independent component analysis and keyframe extraction based on image complexity

    Science.gov (United States)

    Li, Wei; Chen, Ting; Zhang, Wenjun; Shi, Yunyu; Li, Jun

    2012-04-01

    In recent years, Music video data is increasing at an astonishing speed. Shot segmentation and keyframe extraction constitute a fundamental unit in organizing, indexing, retrieving video content. In this paper a unified framework is proposed to detect the shot boundaries and extract the keyframe of a shot. Music video is first segmented to shots by illumination-invariant chromaticity histogram in independent component (IC) analysis feature space .Then we presents a new metric, image complexity, to extract keyframe in a shot which is computed by ICs. Experimental results show the framework is effective and has a good performance.

  15. Time Domain Feature Extraction Technique for earth's electric field signal prior to the Earthquake

    International Nuclear Information System (INIS)

    Astuti, W; Sediono, W; Akmeliawati, R; Salami, M J E

    2013-01-01

    Earthquake is one of the most destructive of natural disasters that killed many people and destroyed a lot of properties. By considering these catastrophic effects, it is highly important of knowing ahead of earthquakes in order to reduce the number of victims and material losses. Earth's electric field is one of the features that can be used to predict earthquakes (EQs), since it has significant changes in the amplitude of the signal prior to the earthquake. This paper presents a detailed analysis of the earth's electric field due to earthquakes which occurred in Greece, between January 1, 2008 and June 30, 2008. In that period of time, 13 earthquakes had occurred. 6 of them were recorded with magnitudes greater than Ms=5R (5R), while 7 of them were recorded with magnitudes greater than Ms=6R (6R). Time domain feature extraction technique is applied to analyze the 1st significant changes in the earth's electric field prior to the earthquake. Two different time domain feature extraction techniques are applied in this work, namely Simple Square Integral (SSI) and Root Mean Square (RMS). The 1st significant change of the earth's electric field signal in each of monitoring sites is extracted using those two techniques. The feature extraction result can be used as input parameter for an earthquake prediction system

  16. Context and Domain Knowledge Enhanced Entity Spotting in Informal Text

    Science.gov (United States)

    Gruhl, Daniel; Nagarajan, Meena; Pieper, Jan; Robson, Christine; Sheth, Amit

    This paper explores the application of restricted relationship graphs (RDF) and statistical NLP techniques to improve named entity annotation in challenging Informal English domains. We validate our approach using on-line forums discussing popular music. Named entity annotation is particularly difficult in this domain because it is characterized by a large number of ambiguous entities, such as the Madonna album "Music" or Lilly Allen's pop hit "Smile".

  17. Bridging domains : a comparison between information processing in Archaea and Eukarya

    NARCIS (Netherlands)

    Koning, de B.

    2015-01-01

    Bridging Domains

    A Comparison between Information Processing in Archaea and Eukarya

    Studying Information Processing

    Living cells evolved complex systems to handle the flow of information both

  18. Automated extraction of chemical structure information from digital raster images

    Directory of Open Access Journals (Sweden)

    Shedden Kerby A

    2009-02-01

    Full Text Available Abstract Background To search for chemical structures in research articles, diagrams or text representing molecules need to be translated to a standard chemical file format compatible with cheminformatic search engines. Nevertheless, chemical information contained in research articles is often referenced as analog diagrams of chemical structures embedded in digital raster images. To automate analog-to-digital conversion of chemical structure diagrams in scientific research articles, several software systems have been developed. But their algorithmic performance and utility in cheminformatic research have not been investigated. Results This paper aims to provide critical reviews for these systems and also report our recent development of ChemReader – a fully automated tool for extracting chemical structure diagrams in research articles and converting them into standard, searchable chemical file formats. Basic algorithms for recognizing lines and letters representing bonds and atoms in chemical structure diagrams can be independently run in sequence from a graphical user interface-and the algorithm parameters can be readily changed-to facilitate additional development specifically tailored to a chemical database annotation scheme. Compared with existing software programs such as OSRA, Kekule, and CLiDE, our results indicate that ChemReader outperforms other software systems on several sets of sample images from diverse sources in terms of the rate of correct outputs and the accuracy on extracting molecular substructure patterns. Conclusion The availability of ChemReader as a cheminformatic tool for extracting chemical structure information from digital raster images allows research and development groups to enrich their chemical structure databases by annotating the entries with published research articles. Based on its stable performance and high accuracy, ChemReader may be sufficiently accurate for annotating the chemical database with links

  19. Advances in spectral inversion of time-domain induced polarization

    DEFF Research Database (Denmark)

    Fiandaca, Gianluca; Auken, Esben; Christiansen, Anders Vest

    The extraction of spectral information in the inversion process of time-domain (TD) induced polarization (IP) data is changing the use of the TDIP method. Data interpretation is evolving from a qualitative description of the subsurface, able only to discriminate the presence of contrasts in charg......The extraction of spectral information in the inversion process of time-domain (TD) induced polarization (IP) data is changing the use of the TDIP method. Data interpretation is evolving from a qualitative description of the subsurface, able only to discriminate the presence of contrasts...... in chargeability parameters, towards a quantitative analysis of the investigated media, which allows for detailed soil- and rock-type characterization. In this work a review of the recent advances in spectral inversion of TDIP data is presented, in terms of: supported IP parameterizations; modelling of transmitter...

  20. REVIEW PAPER ON THE DEEP WEB DATA EXTRACTION

    OpenAIRE

    Prof. V. S. Patil*1, Miss Sneha Sitafale2, Miss Priyanka Kale3, Miss Poonam Bhujbal 4 , Miss Mohini Dandge 5 .

    2018-01-01

    Deep web data extraction is the process of extracting a set of data records and the items that they contain from a query result page. Such structured data can be later integrated into results from other data sources and given to the user in a single, cohesive view. Domain identification is used to identify the query interfaces related to the domain from the forms obtained in the search process. The surface web contains a large amount of unfiltered information, whereas the deep web includes hi...

  1. Finding Commonalities: Social Information Processing and Domain Theory in the Study of Aggression

    Science.gov (United States)

    Nucci, Larry

    2004-01-01

    The Arsenio and Lemerise (this issue) proposal integrating social information processing (SIP) and domain theory to study children's aggression is evaluated from a domain theory perspective. Basic tenets of domain theory rendering it compatible with SIP are discussed as well as points of divergence. Focus is directed to the proposition that…

  2. Bilinear modeling of EMG signals to extract user-independent features for multiuser myoelectric interface.

    Science.gov (United States)

    Matsubara, Takamitsu; Morimoto, Jun

    2013-08-01

    In this study, we propose a multiuser myoelectric interface that can easily adapt to novel users. When a user performs different motions (e.g., grasping and pinching), different electromyography (EMG) signals are measured. When different users perform the same motion (e.g., grasping), different EMG signals are also measured. Therefore, designing a myoelectric interface that can be used by multiple users to perform multiple motions is difficult. To cope with this problem, we propose for EMG signals a bilinear model that is composed of two linear factors: 1) user dependent and 2) motion dependent. By decomposing the EMG signals into these two factors, the extracted motion-dependent factors can be used as user-independent features. We can construct a motion classifier on the extracted feature space to develop the multiuser interface. For novel users, the proposed adaptation method estimates the user-dependent factor through only a few interactions. The bilinear EMG model with the estimated user-dependent factor can extract the user-independent features from the novel user data. We applied our proposed method to a recognition task of five hand gestures for robotic hand control using four-channel EMG signals measured from subject forearms. Our method resulted in 73% accuracy, which was statistically significantly different from the accuracy of standard nonmultiuser interfaces, as the result of a two-sample t -test at a significance level of 1%.

  3. Extracting Information from Multimedia Meeting Collections

    OpenAIRE

    Gatica-Perez, Daniel; Zhang, Dong; Bengio, Samy

    2005-01-01

    Multimedia meeting collections, composed of unedited audio and video streams, handwritten notes, slides, and electronic documents that jointly constitute a raw record of complex human interaction processes in the workplace, have attracted interest due to the increasing feasibility of recording them in large quantities, by the opportunities for information access and retrieval applications derived from the automatic extraction of relevant meeting information, and by the challenges that the ext...

  4. Multi-level learning: improving the prediction of protein, domain and residue interactions by allowing information flow between levels

    Directory of Open Access Journals (Sweden)

    McDermott Drew

    2009-08-01

    Full Text Available Abstract Background Proteins interact through specific binding interfaces that contain many residues in domains. Protein interactions thus occur on three different levels of a concept hierarchy: whole-proteins, domains, and residues. Each level offers a distinct and complementary set of features for computationally predicting interactions, including functional genomic features of whole proteins, evolutionary features of domain families and physical-chemical features of individual residues. The predictions at each level could benefit from using the features at all three levels. However, it is not trivial as the features are provided at different granularity. Results To link up the predictions at the three levels, we propose a multi-level machine-learning framework that allows for explicit information flow between the levels. We demonstrate, using representative yeast interaction networks, that our algorithm is able to utilize complementary feature sets to make more accurate predictions at the three levels than when the three problems are approached independently. To facilitate application of our multi-level learning framework, we discuss three key aspects of multi-level learning and the corresponding design choices that we have made in the implementation of a concrete learning algorithm. 1 Architecture of information flow: we show the greater flexibility of bidirectional flow over independent levels and unidirectional flow; 2 Coupling mechanism of the different levels: We show how this can be accomplished via augmenting the training sets at each level, and discuss the prevention of error propagation between different levels by means of soft coupling; 3 Sparseness of data: We show that the multi-level framework compounds data sparsity issues, and discuss how this can be dealt with by building local models in information-rich parts of the data. Our proof-of-concept learning algorithm demonstrates the advantage of combining levels, and opens up

  5. Signal extraction and wave field separation in tunnel seismic prediction by independent component analysis

    Science.gov (United States)

    Yue, Y.; Jiang, T.; Zhou, Q.

    2017-12-01

    In order to ensure the rationality and the safety of tunnel excavation, the advanced geological prediction has been become an indispensable step in tunneling. However, the extraction of signal and the separation of P and S waves directly influence the accuracy of geological prediction. Generally, the raw data collected in TSP system is low quality because of the numerous disturb factors in tunnel projects, such as the power interference and machine vibration interference. It's difficult for traditional method (band-pass filtering) to remove interference effectively as well as bring little loss to signal. The power interference, machine vibration interference and the signal are original variables and x, y, z component as observation signals, each component of the representation is a linear combination of the original variables, which satisfy applicable conditions of independent component analysis (ICA). We perform finite-difference simulations of elastic wave propagation to synthetic a tunnel seismic reflection record. The method of ICA was adopted to process the three-component data, and the results show that extract the estimates of signal and the signals are highly correlated (the coefficient correlation is up to more than 0.93). In addition, the estimates of interference that separated from ICA and the interference signals are also highly correlated, and the coefficient correlation is up to more than 0.99. Thus, simulation results showed that the ICA is an ideal method for extracting high quality data from mixed signals. For the separation of P and S waves, the conventional separation techniques are based on physical characteristics of wave propagation, which require knowledge of the near-surface P and S waves velocities and density. Whereas the ICA approach is entirely based on statistical differences between P and S waves, and the statistical technique does not require a priori information. The concrete results of the wave field separation will be presented in

  6. 76 FR 37353 - Federal Acquisition Regulation; Information Collection; Certification of Independent Price...

    Science.gov (United States)

    2011-06-27

    ...; Information Collection; Certification of Independent Price Determination and Parent Company and Identifying... requirement concerning certification of independent price determination and parent company and identifying... laws, offerors on Government contracts must complete the certificate of independent price determination...

  7. 76 FR 16735 - Federal Acquisition Regulation; Information Collection; Certification of Independent Price...

    Science.gov (United States)

    2011-03-25

    ...; Information Collection; Certification of Independent Price Determination and Parent Company and Identifying... requirement concerning certification of independent price determination and parent company and identifying... violating such laws, offerors on Government contracts must complete the certificate of independent price...

  8. NAMED ENTITY RECOGNITION FROM BIOMEDICAL TEXT -AN INFORMATION EXTRACTION TASK

    Directory of Open Access Journals (Sweden)

    N. Kanya

    2016-07-01

    Full Text Available Biomedical Text Mining targets the Extraction of significant information from biomedical archives. Bio TM encompasses Information Retrieval (IR and Information Extraction (IE. The Information Retrieval will retrieve the relevant Biomedical Literature documents from the various Repositories like PubMed, MedLine etc., based on a search query. The IR Process ends up with the generation of corpus with the relevant document retrieved from the Publication databases based on the query. The IE task includes the process of Preprocessing of the document, Named Entity Recognition (NER from the documents and Relationship Extraction. This process includes Natural Language Processing, Data Mining techniques and machine Language algorithm. The preprocessing task includes tokenization, stop word Removal, shallow parsing, and Parts-Of-Speech tagging. NER phase involves recognition of well-defined objects such as genes, proteins or cell-lines etc. This process leads to the next phase that is extraction of relationships (IE. The work was based on machine learning algorithm Conditional Random Field (CRF.

  9. Towards Device-Independent Information Processing on General Quantum Networks

    Science.gov (United States)

    Lee, Ciarán M.; Hoban, Matty J.

    2018-01-01

    The violation of certain Bell inequalities allows for device-independent information processing secure against nonsignaling eavesdroppers. However, this only holds for the Bell network, in which two or more agents perform local measurements on a single shared source of entanglement. To overcome the practical constraints that entangled systems can only be transmitted over relatively short distances, large-scale multisource networks have been employed. Do there exist analogs of Bell inequalities for such networks, whose violation is a resource for device independence? In this Letter, the violation of recently derived polynomial Bell inequalities will be shown to allow for device independence on multisource networks, secure against nonsignaling eavesdroppers.

  10. Semantic Information Extraction of Lanes Based on Onboard Camera Videos

    Science.gov (United States)

    Tang, L.; Deng, T.; Ren, C.

    2018-04-01

    In the field of autonomous driving, semantic information of lanes is very important. This paper proposes a method of automatic detection of lanes and extraction of semantic information from onboard camera videos. The proposed method firstly detects the edges of lanes by the grayscale gradient direction, and improves the Probabilistic Hough transform to fit them; then, it uses the vanishing point principle to calculate the lane geometrical position, and uses lane characteristics to extract lane semantic information by the classification of decision trees. In the experiment, 216 road video images captured by a camera mounted onboard a moving vehicle were used to detect lanes and extract lane semantic information. The results show that the proposed method can accurately identify lane semantics from video images.

  11. Information geometry near randomness and near independence

    CERN Document Server

    Arwini, Khadiga A

    2008-01-01

    This volume will be useful to practising scientists and students working in the application of statistical models to real materials or to processes with perturbations of a Poisson process, a uniform process, or a state of independence for a bivariate process. We use information geometry to provide a common differential geometric framework for a wide range of illustrative applications including amino acid sequence spacings in protein chains, cryptology studies, clustering of communications and galaxies, cosmological voids, coupled spatial statistics in stochastic fibre networks and stochastic porous media, quantum chaology. Introduction sections are provided to mathematical statistics, differential geometry and the information geometry of spaces of probability density functions.

  12. Integrating Information Extraction Agents into a Tourism Recommender System

    Science.gov (United States)

    Esparcia, Sergio; Sánchez-Anguix, Víctor; Argente, Estefanía; García-Fornes, Ana; Julián, Vicente

    Recommender systems face some problems. On the one hand information needs to be maintained updated, which can result in a costly task if it is not performed automatically. On the other hand, it may be interesting to include third party services in the recommendation since they improve its quality. In this paper, we present an add-on for the Social-Net Tourism Recommender System that uses information extraction and natural language processing techniques in order to automatically extract and classify information from the Web. Its goal is to maintain the system updated and obtain information about third party services that are not offered by service providers inside the system.

  13. Phosphorylation of both nucleoplasmin domains is required for activation of its chromatin decondensation activity

    DEFF Research Database (Denmark)

    Bañuelos, Sonia; Omaetxebarria, Miren J; Ramos, Isbaal

    2007-01-01

    Nucleoplasmin (NP) is a histone chaperone involved in nucleosome assembly, chromatin decondensation at fertilization, and apoptosis. To carry out these activities NP has to interact with different types of histones, an interaction that is regulated by phosphorylation. Here we have identified...... are found at the tail domain, flanking the nuclear localization signal. Phosphorylation-mimicking mutations render a recombinant protein as active in chromatin decondensation as hyperphosphorylated NP isolated from Xenopus laevis eggs. Comparison of mutants in which the core and tail domains of the protein...... were independently or simultaneously "activated" indicates that activation or phosphorylation of both protein domains is required for NP to efficiently extract linker-type histones from chromatin....

  14. The Logic Process Formalism of the Informational Domain

    Directory of Open Access Journals (Sweden)

    2007-01-01

    Full Text Available The performance of present-day informational technologies has two main properties: the universality of the structures used and the flexibility of the final user's interfaces. The first determines the potential cover area of the informational domain. The second determines the diversity and efficiency of processing methods of the proceedings being automated. The mentioned aspects are of great importance in agriculture and ecology because there are complex processes and considerable volumes of used information. For example, the meteoro-logical processes are a part of the ecological one like habitats' existential conditions and are known as a complex prognostic problem. The latter needs considerable computational resources to solve the appropriate equations. Likewise, agriculture as a controlled activity under strong impact from natural conditions has the same high requirements for diverse structures and flexibility of information processing.

  15. Optimal Information Extraction of Laser Scanning Dataset by Scale-Adaptive Reduction

    Science.gov (United States)

    Zang, Y.; Yang, B.

    2018-04-01

    3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  16. OPTIMAL INFORMATION EXTRACTION OF LASER SCANNING DATASET BY SCALE-ADAPTIVE REDUCTION

    Directory of Open Access Journals (Sweden)

    Y. Zang

    2018-04-01

    Full Text Available 3D laser technology is widely used to collocate the surface information of object. For various applications, we need to extract a good perceptual quality point cloud from the scanned points. To solve the problem, most of existing methods extract important points based on a fixed scale. However, geometric features of 3D object come from various geometric scales. We propose a multi-scale construction method based on radial basis function. For each scale, important points are extracted from the point cloud based on their importance. We apply a perception metric Just-Noticeable-Difference to measure degradation of each geometric scale. Finally, scale-adaptive optimal information extraction is realized. Experiments are undertaken to evaluate the effective of the proposed method, suggesting a reliable solution for optimal information extraction of object.

  17. Information handbook on independent spent fuel storage installations

    International Nuclear Information System (INIS)

    Raddatz, M.G.; Waters, M.D.

    1996-12-01

    In this information handbook, the staff of the U.S. Nuclear Regulatory Commission describes (1) background information regarding the licensing and history of independent spent fuel storage installations (ISFSIs), (2) a discussion of the licensing process, (3) a description of all currently approved or certified models of dry cask storage systems (DCSSs), and (4) a description of sites currently storing spent fuel in an ISFSI. Storage of spent fuel at ISFSIs must be in accordance with the provisions of 10 CFR Part 72. The staff has provided this handbook for information purposes only. The accuracy of any information herein is not guaranteed. For verification or for more details, the reader should refer to the respective docket files for each DCSS and ISFSI site. The information in this handbook is current as of September 1, 1996

  18. Drosophila Pumilio protein contains multiple autonomous repression domains that regulate mRNAs independently of Nanos and brain tumor.

    Science.gov (United States)

    Weidmann, Chase A; Goldstrohm, Aaron C

    2012-01-01

    Drosophila melanogaster Pumilio is an RNA-binding protein that potently represses specific mRNAs. In developing embryos, Pumilio regulates a key morphogen, Hunchback, in collaboration with the cofactor Nanos. To investigate repression by Pumilio and Nanos, we created cell-based assays and found that Pumilio inhibits translation and enhances mRNA decay independent of Nanos. Nanos robustly stimulates repression through interactions with the Pumilio RNA-binding domain. We programmed Pumilio to recognize a new binding site, which garners repression of new target mRNAs. We show that cofactors Brain Tumor and eIF4E Homologous Protein are not obligatory for Pumilio and Nanos activity. The conserved RNA-binding domain of Pumilio was thought to be sufficient for its function. Instead, we demonstrate that three unique domains in the N terminus of Pumilio possess the major repressive activity and can function autonomously. The N termini of insect and vertebrate Pumilio and Fem-3 binding factors (PUFs) are related, and we show that corresponding regions of human PUM1 and PUM2 have repressive activity. Other PUF proteins lack these repression domains. Our findings suggest that PUF proteins have evolved new regulatory functions through protein sequences appended to their conserved PUF repeat RNA-binding domains.

  19. Extraction of repetitive transients with frequency domain multipoint kurtosis for bearing fault diagnosis

    Science.gov (United States)

    Liao, Yuhe; Sun, Peng; Wang, Baoxiang; Qu, Lei

    2018-05-01

    The appearance of repetitive transients in a vibration signal is one typical feature of faulty rolling element bearings. However, accurate extraction of these fault-related characteristic components has always been a challenging task, especially when there is interference from large amplitude impulsive noises. A frequency domain multipoint kurtosis (FDMK)-based fault diagnosis method is proposed in this paper. The multipoint kurtosis is redefined in the frequency domain and the computational accuracy is improved. An envelope autocorrelation function is also presented to estimate the fault characteristic frequency, which is used to set the frequency hunting zone of the FDMK. Then, the FDMK, instead of kurtosis, is utilized to generate a fast kurtogram and only the optimal band with maximum FDMK value is selected for envelope analysis. Negative interference from both large amplitude impulsive noise and shaft rotational speed related harmonic components are therefore greatly reduced. The analysis results of simulation and experimental data verify the capability and feasibility of this FDMK-based method

  20. Applying a text mining framework to the extraction of numerical parameters from scientific literature in the biotechnology domain

    Directory of Open Access Journals (Sweden)

    André SANTOS

    2012-07-01

    Full Text Available Scientific publications are the main vehicle to disseminate information in the field of biotechnology for wastewater treatment. Indeed, the new research paradigms and the application of high-throughput technologies have increased the rate of publication considerably. The problem is that manual curation becomes harder, prone-to-errors and time-consuming, leading to a probable loss of information and inefficient knowledge acquisition. As a result, research outputs are hardly reaching engineers, hampering the calibration of mathematical models used to optimize the stability and performance of biotechnological systems. In this context, we have developed a data curation workflow, based on text mining techniques, to extract numerical parameters from scientific literature, and applied it to the biotechnology domain. A workflow was built to process wastewater-related articles with the main goal of identifying physico-chemical parameters mentioned in the text. This work describes the implementation of the workflow, identifies achievements and current limitations in the overall process, and presents the results obtained for a corpus of 50 full-text documents.

  1. Applying a text mining framework to the extraction of numerical parameters from scientific literature in the biotechnology domain

    Directory of Open Access Journals (Sweden)

    Anália LOURENÇO

    2013-07-01

    Full Text Available Scientific publications are the main vehicle to disseminate information in the field of biotechnology for wastewater treatment. Indeed, the new research paradigms and the application of high-throughput technologies have increased the rate of publication considerably. The problem is that manual curation becomes harder, prone-to-errors and time-consuming, leading to a probable loss of information and inefficient knowledge acquisition. As a result, research outputs are hardly reaching engineers, hampering the calibration of mathematical models used to optimize the stability and performance of biotechnological systems. In this context, we have developed a data curation workflow, based on text mining techniques, to extract numerical parameters from scientific literature, and applied it to the biotechnology domain. A workflow was built to process wastewater-related articles with the main goal of identifying physico-chemical parameters mentioned in the text. This work describes the implementation of the workflow, identifies achievements and current limitations in the overall process, and presents the results obtained for a corpus of 50 full-text documents.

  2. Social Information-Processing Patterns of Maltreated Children in Two Social Domains

    Science.gov (United States)

    Keil, Vivien; Price, Joseph M.

    2009-01-01

    This study examined relations among social information-processing (SIP) variables in the domains of peer provocation and peer group entry. Using Crick and Dodge's [Crick, N. R., & Dodge, K. A. (1994). "A review and reformulation of social information-processing mechanisms in children's social adjustment." "Psychological Bulletin," 115, 74-101]…

  3. Knowledge Dictionary for Information Extraction on the Arabic Text Data

    Directory of Open Access Journals (Sweden)

    Wahyu Jauharis Saputra

    2013-04-01

    Full Text Available Information extraction is an early stage of a process of textual data analysis. Information extraction is required to get information from textual data that can be used for process analysis, such as classification and categorization. A textual data is strongly influenced by the language. Arabic is gaining a significant attention in many studies because Arabic language is very different from others, and in contrast to other languages, tools and research on the Arabic language is still lacking. The information extracted using the knowledge dictionary is a concept of expression. A knowledge dictionary is usually constructed manually by an expert and this would take a long time and is specific to a problem only. This paper proposed a method for automatically building a knowledge dictionary. Dictionary knowledge is formed by classifying sentences having the same concept, assuming that they will have a high similarity value. The concept that has been extracted can be used as features for subsequent computational process such as classification or categorization. Dataset used in this paper was the Arabic text dataset. Extraction result was tested by using a decision tree classification engine and the highest precision value obtained was 71.0% while the highest recall value was 75.0%. 

  4. Towards automatic music transcription: note extraction based on independent subspace analysis

    Science.gov (United States)

    Wellhausen, Jens; Hoynck, Michael

    2005-01-01

    Due to the increasing amount of music available electronically the need of automatic search, retrieval and classification systems for music becomes more and more important. In this paper an algorithm for automatic transcription of polyphonic piano music into MIDI data is presented, which is a very interesting basis for database applications, music analysis and music classification. The first part of the algorithm performs a note accurate temporal audio segmentation. In the second part, the resulting segments are examined using Independent Subspace Analysis to extract sounding notes. Finally, the results are used to build a MIDI file as a new representation of the piece of music which is examined.

  5. Collaborative Yet Independent: Information Practices in the Physical Sciences

    CERN Document Server

    Meyer, Eric T; Kyriakidou-Zacharoudiou, Avgousta; Power, Lucy; Williams, Peter; Venters, Will; Terras, Melissa; Wyatt, Sally

    2011-12-31

    In many ways, the physical sciences are at the forefront of using digital tools and methods to work with information and data. However, the fields and disciplines that make up the physical sciences are by no means uniform, and physical scientists find, use, and disseminate information in a variety of ways. This report examines information practices in the physical sciences across seven cases, and demonstrates the richly varied ways in which physical scientists work, collaborate, and share information and data. This report details seven case studies in the physical sciences. For each case, qualitative interviews and focus groups were used to understand the domain. Quantitative data gathered from a survey of participants highlights different information strategies employed across the cases, and identifies important software used for research. Finally, conclusions from across the cases are drawn, and recommendations are made. This report is the third in a series commissioned by the Research Information Network...

  6. The intervening domain from MeCP2 enhances the DNA affinity of the methyl binding domain and provides an independent DNA interaction site.

    Science.gov (United States)

    Claveria-Gimeno, Rafael; Lanuza, Pilar M; Morales-Chueca, Ignacio; Jorge-Torres, Olga C; Vega, Sonia; Abian, Olga; Esteller, Manel; Velazquez-Campoy, Adrian

    2017-01-31

    Methyl-CpG binding protein 2 (MeCP2) preferentially interacts with methylated DNA and it is involved in epigenetic regulation and chromatin remodelling. Mutations in MeCP2 are linked to Rett syndrome, the leading cause of intellectual retardation in girls and causing mental, motor and growth impairment. Unstructured regions in MeCP2 provide the plasticity for establishing interactions with multiple binding partners. We present a biophysical characterization of the methyl binding domain (MBD) from MeCP2 reporting the contribution of flanking domains to its structural stability and dsDNA interaction. The flanking disordered intervening domain (ID) increased the structural stability of MBD, modified its dsDNA binding profile from an entropically-driven moderate-affinity binding to an overwhelmingly enthalpically-driven high-affinity binding. Additionally, ID provided an additional site for simultaneously and autonomously binding an independent dsDNA molecule, which is a key feature linked to the chromatin remodelling and looping activity of MeCP2, as well as its ability to interact with nucleosomes replacing histone H1. The dsDNA interaction is characterized by an unusually large heat capacity linked to a cluster of water molecules trapped within the binding interface. The dynamics of disordered regions together with extrinsic factors are key determinants of MeCP2 global structural properties and functional capabilities.

  7. EEMD Independent Extraction for Mixing Features of Rotating Machinery Reconstructed in Phase Space

    Directory of Open Access Journals (Sweden)

    Zaichao Ma

    2015-04-01

    Full Text Available Empirical Mode Decomposition (EMD, due to its adaptive decomposition property for the non-linear and non-stationary signals, has been widely used in vibration analyses for rotating machinery. However, EMD suffers from mode mixing, which is difficult to extract features independently. Although the improved EMD, well known as the ensemble EMD (EEMD, has been proposed, mode mixing is alleviated only to a certain degree. Moreover, EEMD needs to determine the amplitude of added noise. In this paper, we propose Phase Space Ensemble Empirical Mode Decomposition (PSEEMD integrating Phase Space Reconstruction (PSR and Manifold Learning (ML for modifying EEMD. We also provide the principle and detailed procedure of PSEEMD, and the analyses on a simulation signal and an actual vibration signal derived from a rubbing rotor are performed. The results show that PSEEMD is more efficient and convenient than EEMD in extracting the mixing features from the investigated signal and in optimizing the amplitude of the necessary added noise. Additionally PSEEMD can extract the weak features interfered with a certain amount of noise.

  8. Multi-Filter String Matching and Human-Centric Entity Matching for Information Extraction

    Science.gov (United States)

    Sun, Chong

    2012-01-01

    More and more information is being generated in text documents, such as Web pages, emails and blogs. To effectively manage this unstructured information, one broadly used approach includes locating relevant content in documents, extracting structured information and integrating the extracted information for querying, mining or further analysis. In…

  9. METHOD OF RARE TERM CONTRASTIVE EXTRACTION FROM NATURAL LANGUAGE TEXTS

    Directory of Open Access Journals (Sweden)

    I. A. Bessmertny

    2017-01-01

    Full Text Available The paper considers a problem of automatic domain term extraction from documents corpus by means of a contrast collection. Existing contrastive methods successfully extract often used terms but mishandle rare terms. This could yield poorness of the resulting thesaurus. Assessment of point-wise mutual information is one of the known statistical methods of term extraction and it finds rare terms successfully. Although, it extracts many false terms at that. The proposed approach consists of point-wise mutual information application for rare terms extraction and filtering of candidates by criterion of joint occurrence with the other candidates. We build “documents-by-terms” matrix that is subjected to singular value decomposition to eliminate noise and reveal strong interconnections. Then we pass on to the resulting matrix “terms-by-terms” that reproduces strength of interconnections between words. This approach was approved on a documents collection from “Geology” domain with the use of contrast documents from such topics as “Politics”, “Culture”, “Economics” and “Accidents” on some Internet resources. The experimental results demonstrate operability of this method for rare terms extraction.

  10. Characterizing Health-Related Information Needs of Domain Experts

    OpenAIRE

    Znaidi , Eya; Tamine , Lynda; Chouquet , Cécile; Latiri , Chira

    2013-01-01

    International audience; In information retrieval literature, understanding the users’ intents behind the queries is critically important to gain a better insight of how to select relevant results. While many studies investigated how users in general carry out exploratory health searches in digital environments, a few focused on how are the queries formulated, specifically by domain expert users. This study intends to fill this gap by studying 173 health expert queries issued from 3 medical in...

  11. Research on Crowdsourcing Emergency Information Extraction of Based on Events' Frame

    Science.gov (United States)

    Yang, Bo; Wang, Jizhou; Ma, Weijun; Mao, Xi

    2018-01-01

    At present, the common information extraction method cannot extract the structured emergency event information accurately; the general information retrieval tool cannot completely identify the emergency geographic information; these ways also do not have an accurate assessment of these results of distilling. So, this paper proposes an emergency information collection technology based on event framework. This technique is to solve the problem of emergency information picking. It mainly includes emergency information extraction model (EIEM), complete address recognition method (CARM) and the accuracy evaluation model of emergency information (AEMEI). EIEM can be structured to extract emergency information and complements the lack of network data acquisition in emergency mapping. CARM uses a hierarchical model and the shortest path algorithm and allows the toponomy pieces to be joined as a full address. AEMEI analyzes the results of the emergency event and summarizes the advantages and disadvantages of the event framework. Experiments show that event frame technology can solve the problem of emergency information drawing and provides reference cases for other applications. When the emergency disaster is about to occur, the relevant departments query emergency's data that has occurred in the past. They can make arrangements ahead of schedule which defense and reducing disaster. The technology decreases the number of casualties and property damage in the country and world. This is of great significance to the state and society.

  12. From remote sensing data about information extraction for 3D geovisualization - Development of a workflow

    International Nuclear Information System (INIS)

    Tiede, D.

    2010-01-01

    ) object based methods for the extraction of information are a key element to link individual workflow elements - improvements, especially concerning the level of automation in this area are developed and demonstrated; (2) in addition to established geovisualisation techniques, the application of recent developments in 3D geovisualisation using freely available virtual globes can be meaningful to communicate results - analytical 3D views, a new method to effectively provide relevant information in virtual globes is presented; (3) transferability of the workflow to different fields of application is shown to be successful - up to a certain degree even independently from underlying data sources. (author) [de

  13. A rapid extraction of landslide disaster information research based on GF-1 image

    Science.gov (United States)

    Wang, Sai; Xu, Suning; Peng, Ling; Wang, Zhiyi; Wang, Na

    2015-08-01

    In recent years, the landslide disasters occurred frequently because of the seismic activity. It brings great harm to people's life. It has caused high attention of the state and the extensive concern of society. In the field of geological disaster, landslide information extraction based on remote sensing has been controversial, but high resolution remote sensing image can improve the accuracy of information extraction effectively with its rich texture and geometry information. Therefore, it is feasible to extract the information of earthquake- triggered landslides with serious surface damage and large scale. Taking the Wenchuan county as the study area, this paper uses multi-scale segmentation method to extract the landslide image object through domestic GF-1 images and DEM data, which uses the estimation of scale parameter tool to determine the optimal segmentation scale; After analyzing the characteristics of landslide high-resolution image comprehensively and selecting spectrum feature, texture feature, geometric features and landform characteristics of the image, we can establish the extracting rules to extract landslide disaster information. The extraction results show that there are 20 landslide whose total area is 521279.31 .Compared with visual interpretation results, the extraction accuracy is 72.22%. This study indicates its efficient and feasible to extract earthquake landslide disaster information based on high resolution remote sensing and it provides important technical support for post-disaster emergency investigation and disaster assessment.

  14. The Road to Independently Understandable Information

    Science.gov (United States)

    Habermann, T.; Robinson, E.

    2017-12-01

    The turn of the 21st century was a pivotal time in the Earth and Space Science information ecosystem. The Content Standard for Digital Geospatial Metadata (CSDGM) had existed for nearly a decade and ambitious new standards were just emerging. The U.S. Federal Geospatial Data Committee (FGDC) had extended many of the concepts from CSDGM into the International community with ISO 19115:2003 and the Consultative Committee for Space Data Systems (CCSDS) had migrated their Open Archival Information System (OAIS) Reference Model into an international standard (ISO 14721:2003). The OAIS model outlined the roles and responsibilities of archives with the principle role being preserving information and making it available to users, a "designated community", as a service to the data producer. It was mandatory for the archive to ensure that information is "independently understandable" to the designated community and to maintain that understanding through on-going partnerships between archives and designated communities. Standards can play a role in supporting these partnerships as designated communities expand across disciplinary and geographic boundaries. The ISO metadata standards include many capabilities that might make critical contributions to this goal. These include connections to resources outside of the metadata record (i.e. documentation) and mechanisms for ongoing incorporation of user feedback into the metadata stream. We will demonstrate these capabilities with examples of how they can increase understanding.

  15. Independent consultants : fiduciary duties, trade secrets and confidential information

    International Nuclear Information System (INIS)

    Fraser, B.; Wilhelmson, M.

    1999-01-01

    Because of cutbacks and downsizing within the petroleum industry, it has become one of the largest users of contract personnel to provide services that were previously provided by senior levels of management. This paper explored the application of common law fiduciary duties and the duty of confidence to the independent workforce. It examined to what extent fiduciary duties apply to independent consultants and the best way for the employer to protect itself from the potential misuse of confidential information. Part 1 of the paper described fiduciary duties. A fiduciary relationship was described as one that possesses the following three characteristics: (1) the fiduciary has scope for the exercise of some discretion of power, (2) the fiduciary can unilaterally exercise that power to affect the beneficiary's practical interests, and (3) the beneficiary is peculiarly vulnerable to the fiduciary holding the discretion of power. Three examples of how the courts treated some arguments regarding fiduciaries were presented. Part 2 of this paper discussed how trade secrets and confidential information should be handled. It was explained, that regardless of fiduciary duties, the unauthorized use of confidential information by an independent contractor can give rise to liability and an award of damages by the courts. Some examples where the Supreme Court of Canada found breach of confidence by a party were presented. Information communicated from an employer to an employee can be divided into the following three categories: (1) publicly accessible and not confidential, (2) confidential and must be treated as such while an employee is employed, but if learned becomes part of the employees skill and knowledge, and (3) specific confidential trade secrets that cannot lawfully be used for anyone's benefit but the employer's. Issues regarding defining rights and obligations by contract were also discussed

  16. Extraction of CT dose information from DICOM metadata: automated Matlab-based approach.

    Science.gov (United States)

    Dave, Jaydev K; Gingold, Eric L

    2013-01-01

    The purpose of this study was to extract exposure parameters and dose-relevant indexes of CT examinations from information embedded in DICOM metadata. DICOM dose report files were identified and retrieved from a PACS. An automated software program was used to extract from these files information from the structured elements in the DICOM metadata relevant to exposure. Extracting information from DICOM metadata eliminated potential errors inherent in techniques based on optical character recognition, yielding 100% accuracy.

  17. GPR Detection of Buried Symmetrically Shaped Mine-like Objects using Selective Independent Component Analysis

    DEFF Research Database (Denmark)

    Karlsen, Brian; Sørensen, Helge Bjarup Dissing; Larsen, Jan

    2003-01-01

    from small-scale anti-personal (AP) mines to large-scale anti-tank (AT) mines were designed. Large-scale SF-GPR measurements on this series of mine-like objects buried in soil were performed. The SF-GPR data was acquired using a wideband monostatic bow-tie antenna operating in the frequency range 750......This paper addresses the detection of mine-like objects in stepped-frequency ground penetrating radar (SF-GPR) data as a function of object size, object content, and burial depth. The detection approach is based on a Selective Independent Component Analysis (SICA). SICA provides an automatic...... ranking of components, which enables the suppression of clutter, hence extraction of components carrying mine information. The goal of the investigation is to evaluate various time and frequency domain ICA approaches based on SICA. Performance comparison is based on a series of mine-like objects ranging...

  18. Federated management of information for TeleCARE

    NARCIS (Netherlands)

    Afsarmanesh, H.; Guevara-Masis, V.; Hertzberger, L.O.; Camarinha-Matos, L.M.

    2004-01-01

    Distributed information management plays a fundamental role within the base infrastructure supporting the elderly care domain. Specificities of this domain include the autonomy and independence of its involved actors, the critical data that is handled about individuals, and the variety of

  19. Mechanisms for integration of information models across related domains

    Science.gov (United States)

    Atkinson, Rob

    2010-05-01

    It is well recognised that there are opportunities and challenges in cross-disciplinary data integration. A significant barrier, however, is creating a conceptual model of the combined domains and the area of integration. For example, a groundwater domain application may require information from several related domains: geology, hydrology, water policy, etc. Each domain may have its own data holdings and conceptual models, but these will share various common concepts (eg. The concept of an aquifer). These areas of semantic overlap present significant challenges, firstly to choose a single representation (model) of a concept that appears in multiple disparate models,, then to harmonise these other models with the single representation. In addition, models may exist at different levels of abstraction depending on how closely aligned they are with a particular implementation. This makes it hard for modellers in one domain to introduce elements from another domain without either introducing a specific style of implementation, or conversely dealing with a set of abstract patterns that are hard to integrate with existing implementations. Models are easier to integrate if they are broken down into small units, with common concepts implemented using common models from well-known, and predictably managed shared libraries. This vision however requires development of a set of mechanisms (tools and procedures) for implementing and exploiting libraries of model components. These mechanisms need to handle publication, discovery, subscription, versioning and implementation of models in different forms. In this presentation a coherent suite of such mechanisms is proposed, using a scenario based on re-use of geosciences models. This approach forms the basis of a comprehensive strategy to empower domain modellers to create more interoperable systems. The strategy address a range of concerns and practice, and includes methodologies, an accessible toolkit, improvements to available

  20. Slicing-independent RISC activation requires the argonaute PAZ domain.

    Science.gov (United States)

    Gu, Shuo; Jin, Lan; Huang, Yong; Zhang, Feijie; Kay, Mark A

    2012-08-21

    Small RNAs regulate genetic networks through a ribonucleoprotein complex called the RNA-induced silencing complex (RISC), which, in mammals, contains at its center one of four Argonaute proteins (Ago1-Ago4). A key regulatory event in the RNA interference (RNAi) and microRNA (miRNA) pathways is Ago loading, wherein double-stranded small-RNA duplexes are incorporated into RISC (pre-RISC) and then become single-stranded (mature RISC), a process that is not well understood. The Agos contain an evolutionarily conserved PAZ (Piwi/Argonaute/Zwille) domain whose primary function is to bind the 3' end of small RNAs. We created multiple PAZ-domain-disrupted mutant Ago proteins and studied their biochemical properties and biological functionality in cells. We found that the PAZ domain is dispensable for Ago loading of slicing-competent RISC. In contrast, in the absence of slicer activity or slicer-substrate duplex RNAs, PAZ-disrupted Agos bound duplex small interfering RNAs, but were unable to unwind or eject the passenger strand and form functional RISC complexes. We have discovered that the highly conserved PAZ domain plays an important role in RISC activation, providing new mechanistic insights into how miRNAs regulate genes, as well as new insights for future design of miRNA- and RNAi-based therapeutics. Copyright © 2012 Elsevier Ltd. All rights reserved.

  1. Advanced applications of natural language processing for performing information extraction

    CERN Document Server

    Rodrigues, Mário

    2015-01-01

    This book explains how can be created information extraction (IE) applications that are able to tap the vast amount of relevant information available in natural language sources: Internet pages, official documents such as laws and regulations, books and newspapers, and social web. Readers are introduced to the problem of IE and its current challenges and limitations, supported with examples. The book discusses the need to fill the gap between documents, data, and people, and provides a broad overview of the technology supporting IE. The authors present a generic architecture for developing systems that are able to learn how to extract relevant information from natural language documents, and illustrate how to implement working systems using state-of-the-art and freely available software tools. The book also discusses concrete applications illustrating IE uses.   ·         Provides an overview of state-of-the-art technology in information extraction (IE), discussing achievements and limitations for t...

  2. A Review on the Needs of Independent Information Channel for Regulatory Body

    International Nuclear Information System (INIS)

    Raza, Hamid Saeed; Choi, Kwang Sik

    2010-01-01

    For a regulatory body it is very important to be independent not only in its actions but also in its information gathering channels. It is thus required for an organization to get requisite information for their decision making, as well as to prevail over information asymmetries. For corporate culture, information channels sometimes prove to be extensive enough to rune valuable resources and sometime base delays. The paper will reveal nature of an independent information channel. This review will additionally confer a scenario; a regulatory body may face in case of solely depending upon licensee for its information gathering. Working in the wake of this paper comprises review of allied published materials. It might be more helpful for interested ones to persist this effort by ensuring stakeholder involvement i.e. by making this sensitive topic alive through seminars and working groups

  3. An Effective Approach to Biomedical Information Extraction with Limited Training Data

    Science.gov (United States)

    Jonnalagadda, Siddhartha

    2011-01-01

    In the current millennium, extensive use of computers and the internet caused an exponential increase in information. Few research areas are as important as information extraction, which primarily involves extracting concepts and the relations between them from free text. Limitations in the size of training data, lack of lexicons and lack of…

  4. Conserved quantities in background independent theories

    Energy Technology Data Exchange (ETDEWEB)

    Markopoulou, Fotini [Perimeter Institute for Theoretical Physics, 35 King Street North, Waterloo, Ontario N2J 2W9 (Canada); Department of Physics, University of Waterloo, Waterloo, Ontario N2L 3G1 (Canada)

    2007-05-15

    We discuss the difficulties that background independent theories based on quantum geometry encounter in deriving general relativity as the low energy limit. We follow a geometrogenesis scenario of a phase transition from a pre-geometric theory to a geometric phase which suggests that a first step towards the low energy limit is searching for the effective collective excitations that will characterize it. Using the correspondence between the pre-geometric background independent theory and a quantum information processor, we are able to use the method of noiseless subsystems to extract such coherent collective excitations. We illustrate this in the case of locally evolving graphs.

  5. High-Resolution Remote Sensing Image Building Extraction Based on Markov Model

    Science.gov (United States)

    Zhao, W.; Yan, L.; Chang, Y.; Gong, L.

    2018-04-01

    With the increase of resolution, remote sensing images have the characteristics of increased information load, increased noise, more complex feature geometry and texture information, which makes the extraction of building information more difficult. To solve this problem, this paper designs a high resolution remote sensing image building extraction method based on Markov model. This method introduces Contourlet domain map clustering and Markov model, captures and enhances the contour and texture information of high-resolution remote sensing image features in multiple directions, and further designs the spectral feature index that can characterize "pseudo-buildings" in the building area. Through the multi-scale segmentation and extraction of image features, the fine extraction from the building area to the building is realized. Experiments show that this method can restrain the noise of high-resolution remote sensing images, reduce the interference of non-target ground texture information, and remove the shadow, vegetation and other pseudo-building information, compared with the traditional pixel-level image information extraction, better performance in building extraction precision, accuracy and completeness.

  6. SH2 Domain Histochemistry.

    Science.gov (United States)

    Buhs, Sophia; Nollau, Peter

    2017-01-01

    Among posttranslational modifications, the phosphorylation of tyrosine residues is a key modification in cell signaling. Because of its biological importance, characterization of the cellular state of tyrosine phosphorylation is of great interest. Based on the unique properties of endogenously expressed SH2 domains recognizing tyrosine phosphorylated signaling proteins with high specificity we have developed an alternative approach, coined SH2 profiling, enabling us to decipher complex patterns of tyrosine phosphorylation in various normal and cancerous tissues. So far, SH2 profiling has largely been applied for the analysis of protein extracts with the limitation that information on spatial distribution and intensity of tyrosine phosphorylation within a tissue is lost. Here, we describe a novel SH2 domain based strategy for differential characterization of the state of tyrosine phosphorylation in formaldehyde-fixed and paraffin-embedded tissues. This approach demonstrates that SH2 domains may serve as very valuable tools for the analysis of the differential state of tyrosine phosphorylation in primary tissues fixed and processed under conditions frequently applied by routine pathology laboratories.

  7. Interoperable domain models : The ISO land administration domain model LADM and its external classes

    NARCIS (Netherlands)

    Lemmen, C.H.J.; Van Oosterom, P.J.M.; Uitermark, H.T.; Zevenbergen, J.A.; Cooper, A.K.

    2011-01-01

    This paper provides a brief overview of one of the first spatial domain standards: a standard for the domain of Land Administration (LA). This standard is in the draft stage of development now (May 2011). The development of domain standards is a logical follow up after domain-independent standards,

  8. Enhancing public access to legal information : A proposal for a new official legal information generic top-level domain

    NARCIS (Netherlands)

    Mitee, Leesi Ebenezer

    2017-01-01

    Abstract: This article examines the use of a new legal information generic Top-Level Domain (gTLD) as a viable tool for easy identification of official legal information websites (OLIWs) and enhancing global public access to their resources. This intervention is necessary because of the existence of

  9. MedTime: a temporal information extraction system for clinical narratives.

    Science.gov (United States)

    Lin, Yu-Kai; Chen, Hsinchun; Brown, Randall A

    2013-12-01

    Temporal information extraction from clinical narratives is of critical importance to many clinical applications. We participated in the EVENT/TIMEX3 track of the 2012 i2b2 clinical temporal relations challenge, and presented our temporal information extraction system, MedTime. MedTime comprises a cascade of rule-based and machine-learning pattern recognition procedures. It achieved a micro-averaged f-measure of 0.88 in both the recognitions of clinical events and temporal expressions. We proposed and evaluated three time normalization strategies to normalize relative time expressions in clinical texts. The accuracy was 0.68 in normalizing temporal expressions of dates, times, durations, and frequencies. This study demonstrates and evaluates the integration of rule-based and machine-learning-based approaches for high performance temporal information extraction from clinical narratives. Copyright © 2013 Elsevier Inc. All rights reserved.

  10. Improving the performance of DomainDiscovery of protein domain boundary assignment using inter-domain linker index

    Directory of Open Access Journals (Sweden)

    Zomaya Albert Y

    2006-12-01

    Full Text Available Abstract Background Knowledge of protein domain boundaries is critical for the characterisation and understanding of protein function. The ability to identify domains without the knowledge of the structure – by using sequence information only – is an essential step in many types of protein analyses. In this present study, we demonstrate that the performance of DomainDiscovery is improved significantly by including the inter-domain linker index value for domain identification from sequence-based information. Improved DomainDiscovery uses a Support Vector Machine (SVM approach and a unique training dataset built on the principle of consensus among experts in defining domains in protein structure. The SVM was trained using a PSSM (Position Specific Scoring Matrix, secondary structure, solvent accessibility information and inter-domain linker index to detect possible domain boundaries for a target sequence. Results Improved DomainDiscovery is compared with other methods by benchmarking against a structurally non-redundant dataset and also CASP5 targets. Improved DomainDiscovery achieves 70% accuracy for domain boundary identification in multi-domains proteins. Conclusion Improved DomainDiscovery compares favourably to the performance of other methods and excels in the identification of domain boundaries for multi-domain proteins as a result of introducing support vector machine with benchmark_2 dataset.

  11. a Statistical Texture Feature for Building Collapse Information Extraction of SAR Image

    Science.gov (United States)

    Li, L.; Yang, H.; Chen, Q.; Liu, X.

    2018-04-01

    Synthetic Aperture Radar (SAR) has become one of the most important ways to extract post-disaster collapsed building information, due to its extreme versatility and almost all-weather, day-and-night working capability, etc. In view of the fact that the inherent statistical distribution of speckle in SAR images is not used to extract collapsed building information, this paper proposed a novel texture feature of statistical models of SAR images to extract the collapsed buildings. In the proposed feature, the texture parameter of G0 distribution from SAR images is used to reflect the uniformity of the target to extract the collapsed building. This feature not only considers the statistical distribution of SAR images, providing more accurate description of the object texture, but also is applied to extract collapsed building information of single-, dual- or full-polarization SAR data. The RADARSAT-2 data of Yushu earthquake which acquired on April 21, 2010 is used to present and analyze the performance of the proposed method. In addition, the applicability of this feature to SAR data with different polarizations is also analysed, which provides decision support for the data selection of collapsed building information extraction.

  12. A Hybrid Approach to Finding Relevant Social Media Content for Complex Domain Specific Information Needs.

    Science.gov (United States)

    Cameron, Delroy; Sheth, Amit P; Jaykumar, Nishita; Thirunarayan, Krishnaprasad; Anand, Gaurish; Smith, Gary A

    2014-12-01

    While contemporary semantic search systems offer to improve classical keyword-based search, they are not always adequate for complex domain specific information needs. The domain of prescription drug abuse, for example, requires knowledge of both ontological concepts and "intelligible constructs" not typically modeled in ontologies. These intelligible constructs convey essential information that include notions of intensity, frequency, interval, dosage and sentiments, which could be important to the holistic needs of the information seeker. In this paper, we present a hybrid approach to domain specific information retrieval that integrates ontology-driven query interpretation with synonym-based query expansion and domain specific rules, to facilitate search in social media on prescription drug abuse. Our framework is based on a context-free grammar (CFG) that defines the query language of constructs interpretable by the search system. The grammar provides two levels of semantic interpretation: 1) a top-level CFG that facilitates retrieval of diverse textual patterns, which belong to broad templates and 2) a low-level CFG that enables interpretation of specific expressions belonging to such textual patterns. These low-level expressions occur as concepts from four different categories of data: 1) ontological concepts, 2) concepts in lexicons (such as emotions and sentiments), 3) concepts in lexicons with only partial ontology representation, called lexico-ontology concepts (such as side effects and routes of administration (ROA)), and 4) domain specific expressions (such as date, time, interval, frequency and dosage) derived solely through rules. Our approach is embodied in a novel Semantic Web platform called PREDOSE, which provides search support for complex domain specific information needs in prescription drug abuse epidemiology. When applied to a corpus of over 1 million drug abuse-related web forum posts, our search framework proved effective in retrieving

  13. DKIE: Open Source Information Extraction for Danish

    DEFF Research Database (Denmark)

    Derczynski, Leon; Field, Camilla Vilhelmsen; Bøgh, Kenneth Sejdenfaden

    2014-01-01

    Danish is a major Scandinavian language spoken daily by around six million people. However, it lacks a unified, open set of NLP tools. This demonstration will introduce DKIE, an extensible open-source toolkit for processing Danish text. We implement an information extraction architecture for Danish...

  14. Digital Image Watermarking in Transform Domains

    International Nuclear Information System (INIS)

    EL-Shazly, E.H.M.

    2012-01-01

    Fast development of internet and availability of huge digital content make it easy to create, modify and copy digital media such as audio, video and images. This causes a problem for owners of that content and hence a need to copy right protection tool was essential. First, encryption was proposed but it ensures protection during transmission only and once decryption occurred any one can modify the data. at that point watermarking was introduced as a solution to such problem. Watermarking is a process of inserting a low energy signal in to a high energy one so that it doesn't affect the main signal features. A good digital image watermarking technique should satisfy four requirements: 1) Embedding of a watermark should not degrade the host image visual quality (imperceptibility). 2) The embedded watermark should stick to the host image so that it couldn’t be removed by common image processing operation and could be extracted from the attacked watermarked image (robustness). 3) Knowing the embedding and extraction procedures is sufficient but not enough to extract the watermark; extra keys should be needed (security). 4) The watermarking technique should allow embedding and extraction of more than one watermark each independent of the other (capacity). This thesis presents a watermarking scheme that full fill the mentioned four requirements by jointing transform domains with Fractional Fourier Transform Domain (FracFT). More work on cascaded Discrete Wavelet Transform DWT with FracFT was done to develop a joint transform simply called Fractional Wavelet Transform (FWT). The proposed schemes were tested with different image processing attacks to verify its robustness. Finally, the watermarked image is transmitted over simulated MC CDMA channel to prove robustness in real transmission conditions case.

  15. Device and machine independent recording of graphic information in the list 'Bilderbuch'

    International Nuclear Information System (INIS)

    Egloff, P.; Foest, G.; Maiss, G.; Rocholl, M.; Thulke, A.; Tschammer, V.

    1976-07-01

    'Bilderbuch' (picture book) is a data file in which any kind of graphic information can be stored in an optimum fashion with respect to space. It represents the core of the GRAFIX system, an interconnected computer system. Graphic information can be transferred between computers; it is independent of the graphic terminals connected to the computers. The representation is independent of the operating systems of the computers and the different versions of programming languages. The structure of 'Bilderbuch' is described. (WB) [de

  16. Transliteration normalization for Information Extraction and Machine Translation

    Directory of Open Access Journals (Sweden)

    Yuval Marton

    2014-12-01

    Full Text Available Foreign name transliterations typically include multiple spelling variants. These variants cause data sparseness and inconsistency problems, increase the Out-of-Vocabulary (OOV rate, and present challenges for Machine Translation, Information Extraction and other natural language processing (NLP tasks. This work aims to identify and cluster name spelling variants using a Statistical Machine Translation method: word alignment. The variants are identified by being aligned to the same “pivot” name in another language (the source-language in Machine Translation settings. Based on word-to-word translation and transliteration probabilities, as well as the string edit distance metric, names with similar spellings in the target language are clustered and then normalized to a canonical form. With this approach, tens of thousands of high-precision name transliteration spelling variants are extracted from sentence-aligned bilingual corpora in Arabic and English (in both languages. When these normalized name spelling variants are applied to Information Extraction tasks, improvements over strong baseline systems are observed. When applied to Machine Translation tasks, a large improvement potential is shown.

  17. Assessing the influence of health literacy on health information behaviors: A multi-domain skills-based approach.

    Science.gov (United States)

    Suri, Venkata Ratnadeep; Majid, Shaheen; Chang, Yun-Ke; Foo, Schubert

    2016-06-01

    The aim of this study is to investigate the relationship between five domain-specific skills of health literacy: Find Health Information (FHI), Appraise Health Information (AHI), Understand Health Information to act (UHI), Actively Manage One's Health (AMH), and E-health literacy (e-Heals), and health information seeking behaviors and three categories of health outcomes. A survey was implemented and data was collected from 1062 college going adults and analyzed using bivariate tests and multiple regression analysis. Among the five domain-specific Health Literacy skills, AHI and e-Heals were significantly associated with the use of traditional sources and the Internet for healthcare information respectively. Similarly and AMH and e-Heals were significantly associated with the use of traditional sources and the Internet for health lifestyle information respectively. Lastly AHI, AMH and e-Heals were significantly associated with the three categories of outcomes, and AFH was significantly associated with cognitive and instrumental outcomes, but not doctor-patient communication outcomes. Consumers' ability to use different health sources for both healthcare and health lifestyle information, and the three categories of health outcomes are associated with different domain-specific health literacy skills. Health literacy initiatives may be improved by focusing on clients to develop domain-specific skills that increase the likelihood of using health information sources and accrue benefits. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  18. End-to-end information extraction without token-level supervision

    DEFF Research Database (Denmark)

    Palm, Rasmus Berg; Hovy, Dirk; Laws, Florian

    2017-01-01

    Most state-of-the-art information extraction approaches rely on token-level labels to find the areas of interest in text. Unfortunately, these labels are time-consuming and costly to create, and consequently, not available for many real-life IE tasks. To make matters worse, token-level labels...... and output text. We evaluate our model on the ATIS data set, MIT restaurant corpus and the MIT movie corpus and compare to neural baselines that do use token-level labels. We achieve competitive results, within a few percentage points of the baselines, showing the feasibility of E2E information extraction...

  19. Overview of image processing tools to extract physical information from JET videos

    Science.gov (United States)

    Craciunescu, T.; Murari, A.; Gelfusa, M.; Tiseanu, I.; Zoita, V.; EFDA Contributors, JET

    2014-11-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  20. Overview of image processing tools to extract physical information from JET videos

    International Nuclear Information System (INIS)

    Craciunescu, T; Tiseanu, I; Zoita, V; Murari, A; Gelfusa, M

    2014-01-01

    In magnetic confinement nuclear fusion devices such as JET, the last few years have witnessed a significant increase in the use of digital imagery, not only for the surveying and control of experiments, but also for the physical interpretation of results. More than 25 cameras are routinely used for imaging on JET in the infrared (IR) and visible spectral regions. These cameras can produce up to tens of Gbytes per shot and their information content can be very different, depending on the experimental conditions. However, the relevant information about the underlying physical processes is generally of much reduced dimensionality compared to the recorded data. The extraction of this information, which allows full exploitation of these diagnostics, is a challenging task. The image analysis consists, in most cases, of inverse problems which are typically ill-posed mathematically. The typology of objects to be analysed is very wide, and usually the images are affected by noise, low levels of contrast, low grey-level in-depth resolution, reshaping of moving objects, etc. Moreover, the plasma events have time constants of ms or tens of ms, which imposes tough conditions for real-time applications. On JET, in the last few years new tools and methods have been developed for physical information retrieval. The methodology of optical flow has allowed, under certain assumptions, the derivation of information about the dynamics of video objects associated with different physical phenomena, such as instabilities, pellets and filaments. The approach has been extended in order to approximate the optical flow within the MPEG compressed domain, allowing the manipulation of the large JET video databases and, in specific cases, even real-time data processing. The fast visible camera may provide new information that is potentially useful for disruption prediction. A set of methods, based on the extraction of structural information from the visual scene, have been developed for the

  1. Relative information entropy in cosmology: The problem of information entanglement

    Energy Technology Data Exchange (ETDEWEB)

    Czinner, Viktor G., E-mail: czinner.viktor@wigner.mta.hu [Centro de Matemática, Universidade do Minho, Campus de Gualtar, 4710-057 Braga (Portugal); HAS Wigner Research Centre for Physics, H-1525 Budapest, P.O. Box 49 (Hungary); Mena, Filipe C., E-mail: fmena@math.uminho.pt [Centro de Matemática, Universidade do Minho, Campus de Gualtar, 4710-057 Braga (Portugal)

    2016-07-10

    The necessary information to distinguish a local inhomogeneous mass density field from its spatial average on a compact domain of the universe can be measured by relative information entropy. The Kullback–Leibler (KL) formula arises very naturally in this context, however, it provides a very complicated way to compute the mutual information between spatially separated but causally connected regions of the universe in a realistic, inhomogeneous model. To circumvent this issue, by considering a parametric extension of the KL measure, we develop a simple model to describe the mutual information which is entangled via the gravitational field equations. We show that the Tsallis relative entropy can be a good approximation in the case of small inhomogeneities, and for measuring the independent relative information inside the domain, we propose the Rényi relative entropy formula.

  2. An Estimator of Mutual Information and its Application to Independence Testing

    Directory of Open Access Journals (Sweden)

    Joe Suzuki

    2016-03-01

    Full Text Available This paper proposes a novel estimator of mutual information for discrete and continuous variables. The main feature of this estimator is that it is zero for a large sample size n if and only if the two variables are independent. The estimator can be used to construct several histograms, compute estimations of mutual information, and choose the maximum value. We prove that the number of histograms constructed has an upper bound of O(log n and apply this fact to the search. We compare the performance of the proposed estimator with an estimator of the Hilbert-Schmidt independence criterion (HSIC, though the proposed method is based on the minimum description length (MDL principle and the HSIC provides a statistical test. The proposed method completes the estimation in O(n log n time, whereas the HSIC kernel computation requires O(n3 time. We also present examples in which the HSIC fails to detect independence but the proposed method successfully detects it.

  3. A semantic-based method for extracting concept definitions from scientific publications: evaluation in the autism phenotype domain.

    Science.gov (United States)

    Hassanpour, Saeed; O'Connor, Martin J; Das, Amar K

    2013-08-12

    A variety of informatics approaches have been developed that use information retrieval, NLP and text-mining techniques to identify biomedical concepts and relations within scientific publications or their sentences. These approaches have not typically addressed the challenge of extracting more complex knowledge such as biomedical definitions. In our efforts to facilitate knowledge acquisition of rule-based definitions of autism phenotypes, we have developed a novel semantic-based text-mining approach that can automatically identify such definitions within text. Using an existing knowledge base of 156 autism phenotype definitions and an annotated corpus of 26 source articles containing such definitions, we evaluated and compared the average rank of correctly identified rule definition or corresponding rule template using both our semantic-based approach and a standard term-based approach. We examined three separate scenarios: (1) the snippet of text contained a definition already in the knowledge base; (2) the snippet contained an alternative definition for a concept in the knowledge base; and (3) the snippet contained a definition not in the knowledge base. Our semantic-based approach had a higher average rank than the term-based approach for each of the three scenarios (scenario 1: 3.8 vs. 5.0; scenario 2: 2.8 vs. 4.9; and scenario 3: 4.5 vs. 6.2), with each comparison significant at the p-value of 0.05 using the Wilcoxon signed-rank test. Our work shows that leveraging existing domain knowledge in the information extraction of biomedical definitions significantly improves the correct identification of such knowledge within sentences. Our method can thus help researchers rapidly acquire knowledge about biomedical definitions that are specified and evolving within an ever-growing corpus of scientific publications.

  4. Mining knowledge from text repositories using information extraction ...

    Indian Academy of Sciences (India)

    Information extraction (IE); text mining; text repositories; knowledge discovery from .... general purpose English words. However ... of precision and recall, as extensive experimentation is required due to lack of public tagged corpora. 4. Mining ...

  5. SAM: Secure Access of Media Independent Information Service with User Anonymity

    Directory of Open Access Journals (Sweden)

    Li Guangsong

    2010-01-01

    Full Text Available Seamless handover across different access technologies is very important in the future wireless networks. To optimize vertical handover in heterogeneous networks, IEEE 802.21 standard defines Media Independent Handover (MIH services. The MIH services can be a new target to attackers, which will be the main concern for equipment vendors and service providers. In this paper, we focus specifically on security of Media Independent Information Service (MIIS and present a new access authentication scheme with user anonymity for MIIS. The protocol can be used to establish a secure channel between the mobile node and the information server. Security and performance of the protocol are also analyzed in this paper.

  6. Tagline: Information Extraction for Semi-Structured Text Elements in Medical Progress Notes

    Science.gov (United States)

    Finch, Dezon Kile

    2012-01-01

    Text analysis has become an important research activity in the Department of Veterans Affairs (VA). Statistical text mining and natural language processing have been shown to be very effective for extracting useful information from medical documents. However, neither of these techniques is effective at extracting the information stored in…

  7. A Biologically Inspired Approach to Frequency Domain Feature Extraction for EEG Classification

    Directory of Open Access Journals (Sweden)

    Nurhan Gursel Ozmen

    2018-01-01

    Full Text Available Classification of electroencephalogram (EEG signal is important in mental decoding for brain-computer interfaces (BCI. We introduced a feature extraction approach based on frequency domain analysis to improve the classification performance on different mental tasks using single-channel EEG. This biologically inspired method extracts the most discriminative spectral features from power spectral densities (PSDs of the EEG signals. We applied our method on a dataset of six subjects who performed five different imagination tasks: (i resting state, (ii mental arithmetic, (iii imagination of left hand movement, (iv imagination of right hand movement, and (v imagination of letter “A.” Pairwise and multiclass classifications were performed in single EEG channel using Linear Discriminant Analysis and Support Vector Machines. Our method produced results (mean classification accuracy of 83.06% for binary classification and 91.85% for multiclassification that are on par with the state-of-the-art methods, using single-channel EEG with low computational cost. Among all task pairs, mental arithmetic versus letter imagination yielded the best result (mean classification accuracy of 90.29%, indicating that this task pair could be the most suitable pair for a binary class BCI. This study contributes to the development of single-channel BCI, as well as finding the best task pair for user defined applications.

  8. Design and development of semantic web-based system for computer science domain-specific information retrieval

    Directory of Open Access Journals (Sweden)

    Ritika Bansal

    2016-09-01

    Full Text Available In semantic web-based system, the concept of ontology is used to search results by contextual meaning of input query instead of keyword matching. From the research literature, there seems to be a need for a tool which can provide an easy interface for complex queries in natural language that can retrieve the domain-specific information from the ontology. This research paper proposes an IRSCSD system (Information retrieval system for computer science domain as a solution. This system offers advanced querying and browsing of structured data with search results automatically aggregated and rendered directly in a consistent user-interface, thus reducing the manual effort of users. So, the main objective of this research is design and development of semantic web-based system for integrating ontology towards domain-specific retrieval support. Methodology followed is a piecemeal research which involves the following stages. First Stage involves the designing of framework for semantic web-based system. Second stage builds the prototype for the framework using Protégé tool. Third Stage deals with the natural language query conversion into SPARQL query language using Python-based QUEPY framework. Fourth Stage involves firing of converted SPARQL queries to the ontology through Apache's Jena API to fetch the results. Lastly, evaluation of the prototype has been done in order to ensure its efficiency and usability. Thus, this research paper throws light on framework development for semantic web-based system that assists in efficient retrieval of domain-specific information, natural language query interpretation into semantic web language, creation of domain-specific ontology and its mapping with related ontology. This research paper also provides approaches and metrics for ontology evaluation on prototype ontology developed to study the performance based on accessibility of required domain-related information.

  9. Mars Target Encyclopedia: Information Extraction for Planetary Science

    Science.gov (United States)

    Wagstaff, K. L.; Francis, R.; Gowda, T.; Lu, Y.; Riloff, E.; Singh, K.

    2017-06-01

    Mars surface targets / and published compositions / Seek and ye will find. We used text mining methods to extract information from LPSC abstracts about the composition of Mars surface targets. Users can search by element, mineral, or target.

  10. The Role of Domain Knowledge in Cognitive Modeling of Information Search

    NARCIS (Netherlands)

    Karanam, S.; Jorge-Botana, Guillermo; Olmos, Ricardo; van Oostendorp, H.

    2017-01-01

    Computational cognitive models developed so far do not incorporate individual differences in domain knowledge in predicting user clicks on search result pages. We address this problem using a cognitive model of information search which enables us to use two semantic spaces having a low (non-expert

  11. Optimum detection for extracting maximum information from symmetric qubit sets

    International Nuclear Information System (INIS)

    Mizuno, Jun; Fujiwara, Mikio; Sasaki, Masahide; Akiba, Makoto; Kawanishi, Tetsuya; Barnett, Stephen M.

    2002-01-01

    We demonstrate a class of optimum detection strategies for extracting the maximum information from sets of equiprobable real symmetric qubit states of a single photon. These optimum strategies have been predicted by Sasaki et al. [Phys. Rev. A 59, 3325 (1999)]. The peculiar aspect is that the detections with at least three outputs suffice for optimum extraction of information regardless of the number of signal elements. The cases of ternary (or trine), quinary, and septenary polarization signals are studied where a standard von Neumann detection (a projection onto a binary orthogonal basis) fails to access the maximum information. Our experiments demonstrate that it is possible with present technologies to attain about 96% of the theoretical limit

  12. Study on methods and techniques of aeroradiometric weak information extraction for sandstone-hosted uranium deposits based on GIS

    International Nuclear Information System (INIS)

    Han Shaoyang; Ke Dan; Hou Huiqun

    2005-01-01

    The weak information extraction is one of the important research contents in the current sandstone-type uranium prospecting in China. This paper introduces the connotation of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information extraction, and discusses the formation theories of aeroradiometric weak information and establishes some effective mathematic models for weak information extraction. Models for weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are realized based on GIS software platform. Application tests of weak information extraction are completed in known uranium mineralized areas. Research results prove that the prospective areas of sandstone-type uranium deposits can be rapidly delineated by extracting aeroradiometric weak information. (authors)

  13. Domain duplication, divergence, and loss events in vertebrate Msx paralogs reveal phylogenomically informed disease markers.

    Science.gov (United States)

    Finnerty, John R; Mazza, Maureen E; Jezewski, Peter A

    2009-01-20

    Msx originated early in animal evolution and is implicated in human genetic disorders. To reconstruct the functional evolution of Msx and inform the study of human mutations, we analyzed the phylogeny and synteny of 46 metazoan Msx proteins and tracked the duplication, diversification and loss of conserved motifs. Vertebrate Msx sequences sort into distinct Msx1, Msx2 and Msx3 clades. The sister-group relationship between MSX1 and MSX2 reflects their derivation from the 4p/5q chromosomal paralogon, a derivative of the original "MetaHox" cluster. We demonstrate physical linkage between Msx and other MetaHox genes (Hmx, NK1, Emx) in a cnidarian. Seven conserved domains, including two Groucho repression domains (N- and C-terminal), were present in the ancestral Msx. In cnidarians, the Groucho domains are highly similar. In vertebrate Msx1, the N-terminal Groucho domain is conserved, while the C-terminal domain diverged substantially, implying a novel function. In vertebrate Msx2 and Msx3, the C-terminal domain was lost. MSX1 mutations associated with ectodermal dysplasia or orofacial clefting disorders map to conserved domains in a non-random fashion. Msx originated from a MetaHox ancestor that also gave rise to Tlx, Demox, NK, and possibly EHGbox, Hox and ParaHox genes. Duplication, divergence or loss of domains played a central role in the functional evolution of Msx. Duplicated domains allow pleiotropically expressed proteins to evolve new functions without disrupting existing interaction networks. Human missense sequence variants reside within evolutionarily conserved domains, likely disrupting protein function. This phylogenomic evaluation of candidate disease markers will inform clinical and functional studies.

  14. Domain duplication, divergence, and loss events in vertebrate Msx paralogs reveal phylogenomically informed disease markers

    Directory of Open Access Journals (Sweden)

    Finnerty John R

    2009-01-01

    Full Text Available Abstract Background Msx originated early in animal evolution and is implicated in human genetic disorders. To reconstruct the functional evolution of Msx and inform the study of human mutations, we analyzed the phylogeny and synteny of 46 metazoan Msx proteins and tracked the duplication, diversification and loss of conserved motifs. Results Vertebrate Msx sequences sort into distinct Msx1, Msx2 and Msx3 clades. The sister-group relationship between MSX1 and MSX2 reflects their derivation from the 4p/5q chromosomal paralogon, a derivative of the original "MetaHox" cluster. We demonstrate physical linkage between Msx and other MetaHox genes (Hmx, NK1, Emx in a cnidarian. Seven conserved domains, including two Groucho repression domains (N- and C-terminal, were present in the ancestral Msx. In cnidarians, the Groucho domains are highly similar. In vertebrate Msx1, the N-terminal Groucho domain is conserved, while the C-terminal domain diverged substantially, implying a novel function. In vertebrate Msx2 and Msx3, the C-terminal domain was lost. MSX1 mutations associated with ectodermal dysplasia or orofacial clefting disorders map to conserved domains in a non-random fashion. Conclusion Msx originated from a MetaHox ancestor that also gave rise to Tlx, Demox, NK, and possibly EHGbox, Hox and ParaHox genes. Duplication, divergence or loss of domains played a central role in the functional evolution of Msx. Duplicated domains allow pleiotropically expressed proteins to evolve new functions without disrupting existing interaction networks. Human missense sequence variants reside within evolutionarily conserved domains, likely disrupting protein function. This phylogenomic evaluation of candidate disease markers will inform clinical and functional studies.

  15. Systematically extracting metal- and solvent-related occupational information from free-text responses to lifetime occupational history questionnaires.

    Science.gov (United States)

    Friesen, Melissa C; Locke, Sarah J; Tornow, Carina; Chen, Yu-Cheng; Koh, Dong-Hee; Stewart, Patricia A; Purdue, Mark; Colt, Joanne S

    2014-06-01

    Lifetime occupational history (OH) questionnaires often use open-ended questions to capture detailed information about study participants' jobs. Exposure assessors use this information, along with responses to job- and industry-specific questionnaires, to assign exposure estimates on a job-by-job basis. An alternative approach is to use information from the OH responses and the job- and industry-specific questionnaires to develop programmable decision rules for assigning exposures. As a first step in this process, we developed a systematic approach to extract the free-text OH responses and convert them into standardized variables that represented exposure scenarios. Our study population comprised 2408 subjects, reporting 11991 jobs, from a case-control study of renal cell carcinoma. Each subject completed a lifetime OH questionnaire that included verbatim responses, for each job, to open-ended questions including job title, main tasks and activities (task), tools and equipment used (tools), and chemicals and materials handled (chemicals). Based on a review of the literature, we identified exposure scenarios (occupations, industries, tasks/tools/chemicals) expected to involve possible exposure to chlorinated solvents, trichloroethylene (TCE) in particular, lead, and cadmium. We then used a SAS macro to review the information reported by study participants to identify jobs associated with each exposure scenario; this was done using previously coded standardized occupation and industry classification codes, and a priori lists of associated key words and phrases related to possibly exposed tasks, tools, and chemicals. Exposure variables representing the occupation, industry, and task/tool/chemicals exposure scenarios were added to the work history records of the study respondents. Our identification of possibly TCE-exposed scenarios in the OH responses was compared to an expert's independently assigned probability ratings to evaluate whether we missed identifying

  16. Mixed Domains Enhance Charge Generation and Extraction in Bulk-Heterojunction Solar Cells with Small-Molecule Donors

    KAUST Repository

    Alqahtani, Obaid; Babics, Maxime; Gorenflot, Julien; Savikhin, Victoria; Ferron, Thomas; Balawi, Ahmed H.; Paulke, Andreas; Kan, Zhipeng; Pope, Michael; Clulow, Andrew J.; Wolf, Jannic Sebastian; Burn, Paul L.; Gentle, Ian R.; Neher, Dieter; Toney, Michael F.; Laquai, Fré dé ric; Beaujuge, Pierre; Collins, Brian A.

    2018-01-01

    The interplay between nanomorphology and efficiency of polymer-fullerene bulk-heterojunction (BHJ) solar cells has been the subject of intense research, but the generality of these concepts for small-molecule (SM) BHJs remains unclear. Here, the relation between performance; charge generation, recombination, and extraction dynamics; and nanomorphology achievable with two SM donors benzo[1,2-b:4,5-b]dithiophene-pyrido[3,4-b]-pyrazine BDT(PPTh), namely SM1 and SM2, differing by their side-chains, are examined as a function of solution additive composition. The results show that the additive 1,8-diiodooctane acts as a plasticizer in the blends, increases domain size, and promotes ordering/crystallinity. Surprisingly, the system with high domain purity (SM1) exhibits both poor exciton harvesting and severe charge trapping, alleviated only slightly with increased crystallinity. In contrast, the system consisting of mixed domains and lower crystallinity (SM2) shows both excellent exciton harvesting and low charge recombination losses. Importantly, the onset of large, pure crystallites in the latter (SM2) system reduces efficiency, pointing to possible differences in the ideal morphologies for SM-based BHJ solar cells compared with polymer-fullerene devices. In polymer-based systems, tie chains between pure polymer crystals establish a continuous charge transport network, whereas SM-based active layers may in some cases require mixed domains that enable both aggregation and charge percolation to the electrodes.

  17. Mixed Domains Enhance Charge Generation and Extraction in Bulk-Heterojunction Solar Cells with Small-Molecule Donors

    KAUST Repository

    Alqahtani, Obaid

    2018-03-25

    The interplay between nanomorphology and efficiency of polymer-fullerene bulk-heterojunction (BHJ) solar cells has been the subject of intense research, but the generality of these concepts for small-molecule (SM) BHJs remains unclear. Here, the relation between performance; charge generation, recombination, and extraction dynamics; and nanomorphology achievable with two SM donors benzo[1,2-b:4,5-b]dithiophene-pyrido[3,4-b]-pyrazine BDT(PPTh), namely SM1 and SM2, differing by their side-chains, are examined as a function of solution additive composition. The results show that the additive 1,8-diiodooctane acts as a plasticizer in the blends, increases domain size, and promotes ordering/crystallinity. Surprisingly, the system with high domain purity (SM1) exhibits both poor exciton harvesting and severe charge trapping, alleviated only slightly with increased crystallinity. In contrast, the system consisting of mixed domains and lower crystallinity (SM2) shows both excellent exciton harvesting and low charge recombination losses. Importantly, the onset of large, pure crystallites in the latter (SM2) system reduces efficiency, pointing to possible differences in the ideal morphologies for SM-based BHJ solar cells compared with polymer-fullerene devices. In polymer-based systems, tie chains between pure polymer crystals establish a continuous charge transport network, whereas SM-based active layers may in some cases require mixed domains that enable both aggregation and charge percolation to the electrodes.

  18. Text Mining in Biomedical Domain with Emphasis on Document Clustering.

    Science.gov (United States)

    Renganathan, Vinaitheerthan

    2017-07-01

    With the exponential increase in the number of articles published every year in the biomedical domain, there is a need to build automated systems to extract unknown information from the articles published. Text mining techniques enable the extraction of unknown knowledge from unstructured documents. This paper reviews text mining processes in detail and the software tools available to carry out text mining. It also reviews the roles and applications of text mining in the biomedical domain. Text mining processes, such as search and retrieval of documents, pre-processing of documents, natural language processing, methods for text clustering, and methods for text classification are described in detail. Text mining techniques can facilitate the mining of vast amounts of knowledge on a given topic from published biomedical research articles and draw meaningful conclusions that are not possible otherwise.

  19. Joining RDC data from flexible protein domains

    International Nuclear Information System (INIS)

    Sgheri, Luca

    2010-01-01

    We study the inverse problem of determining the conformational freedom of two protein domains from residual dipolar coupling (RDC) measurements. For each paramagnetic ion attached to one of the domains we obtain a magnetic susceptibility tensor χ from the RDC of couples of atoms of that domain, and a mean paramagnetic susceptibility tensor χ-bar from the RDC of couples of atoms of the other domain. The latter is an integral average of rotations of χ which depends on the conformational freedom of the two domains. In this paper we consider the case when we have data from paramagnetic ions attached separately to each of the domains. We prove that in this case not all the elements of χ and χ-bar are independent. We derive the mathematical equations for the compatibility of the measurements and show how these relations can be used in the presence of noisy data to determine a compatible set of χ and χ-bar with an unconstrained minimization. If available, information about the shape of the noise can be included in the target function. We show that in this case the compatible set obtained has a reduced error with respect to the noisy data

  20. Towards a Conceptual Design of a Cross-Domain Integrative Information System for the Geosciences

    Science.gov (United States)

    Zaslavsky, I.; Richard, S. M.; Valentine, D. W.; Malik, T.; Gupta, A.

    2013-12-01

    As geoscientists increasingly focus on studying processes that span multiple research domains, there is an increased need for cross-domain interoperability solutions that can scale to the entire geosciences, bridging information and knowledge systems, models, software tools, as well as connecting researchers and organization. Creating a community-driven cyberinfrastructure (CI) to address the grand challenges of integrative Earth science research and education is the focus of EarthCube, a new research initiative of the U.S. National Science Foundation. We are approaching EarthCube design as a complex socio-technical system of systems, in which communication between various domain subsystems, people and organizations enables more comprehensive, data-intensive research designs and knowledge sharing. In particular, we focus on integrating 'traditional' layered CI components - including information sources, catalogs, vocabularies, services, analysis and modeling tools - with CI components supporting scholarly communication, self-organization and social networking (e.g. research profiles, Q&A systems, annotations), in a manner that follows and enhances existing patterns of data, information and knowledge exchange within and across geoscience domains. We describe an initial architecture design focused on enabling the CI to (a) provide an environment for scientifically sound information and software discovery and reuse; (b) evolve by factoring in the impact of maturing movements like linked data, 'big data', and social collaborations, as well as experience from work on large information systems in other domains; (c) handle the ever increasing volume, complexity and diversity of geoscience information; (d) incorporate new information and analytical requirements, tools, and techniques, and emerging types of earth observations and models; (e) accommodate different ideas and approaches to research and data stewardship; (f) be responsive to the existing and anticipated needs

  1. Cross document ontology based information for multimedia retrieval

    NARCIS (Netherlands)

    Reidsma, Dennis; Kuper, Jan; Declerck, T.; Saggion, H.; Cunningham, H.; Ganter, B.; de Moor, A.

    2003-01-01

    This paper describes the MUMIS project, which applies ontology based Information Extraction to improve the results of Information Retrieval in multimedia archives. It makes use of a domain specific ontology, multilingual lexicons and reasoning algorithms to automatically create a semantic annotation

  2. A COMPARATIVE ANALYSIS OF WEB INFORMATION EXTRACTION TECHNIQUES DEEP LEARNING vs. NAÏVE BAYES vs. BACK PROPAGATION NEURAL NETWORKS IN WEB DOCUMENT EXTRACTION

    Directory of Open Access Journals (Sweden)

    J. Sharmila

    2016-01-01

    Full Text Available Web mining related exploration is getting the chance to be more essential these days in view of the reason that a lot of information is overseen through the web. Web utilization is expanding in an uncontrolled way. A particular framework is required for controlling such extensive measure of information in the web space. Web mining is ordered into three noteworthy divisions: Web content mining, web usage mining and web structure mining. Tak-Lam Wong has proposed a web content mining methodology in the exploration with the aid of Bayesian Networks (BN. In their methodology, they were learning on separating the web data and characteristic revelation in view of the Bayesian approach. Roused from their investigation, we mean to propose a web content mining methodology, in view of a Deep Learning Algorithm. The Deep Learning Algorithm gives the interest over BN on the basis that BN is not considered in any learning architecture planning like to propose system. The main objective of this investigation is web document extraction utilizing different grouping algorithm and investigation. This work extricates the data from the web URL. This work shows three classification algorithms, Deep Learning Algorithm, Bayesian Algorithm and BPNN Algorithm. Deep Learning is a capable arrangement of strategies for learning in neural system which is connected like computer vision, speech recognition, and natural language processing and biometrics framework. Deep Learning is one of the simple classification technique and which is utilized for subset of extensive field furthermore Deep Learning has less time for classification. Naive Bayes classifiers are a group of basic probabilistic classifiers in view of applying Bayes hypothesis with concrete independence assumptions between the features. At that point the BPNN algorithm is utilized for classification. Initially training and testing dataset contains more URL. We extract the content presently from the dataset. The

  3. Information Extraction, Data Integration, and Uncertain Data Management: The State of The Art

    NARCIS (Netherlands)

    Habib, Mena Badieh; van Keulen, Maurice

    2011-01-01

    Information Extraction, data Integration, and uncertain data management are different areas of research that got vast focus in the last two decades. Many researches tackled those areas of research individually. However, information extraction systems should have integrated with data integration

  4. Information extraction from muon radiography data

    International Nuclear Information System (INIS)

    Borozdin, K.N.; Asaki, T.J.; Chartrand, R.; Hengartner, N.W.; Hogan, G.E.; Morris, C.L.; Priedhorsky, W.C.; Schirato, R.C.; Schultz, L.J.; Sottile, M.J.; Vixie, K.R.; Wohlberg, B.E.; Blanpied, G.

    2004-01-01

    Scattering muon radiography was proposed recently as a technique of detection and 3-d imaging for dense high-Z objects. High-energy cosmic ray muons are deflected in matter in the process of multiple Coulomb scattering. By measuring the deflection angles we are able to reconstruct the configuration of high-Z material in the object. We discuss the methods for information extraction from muon radiography data. Tomographic methods widely used in medical images have been applied to a specific muon radiography information source. Alternative simple technique based on the counting of high-scattered muons in the voxels seems to be efficient in many simulated scenes. SVM-based classifiers and clustering algorithms may allow detection of compact high-Z object without full image reconstruction. The efficiency of muon radiography can be increased using additional informational sources, such as momentum estimation, stopping power measurement, and detection of muonic atom emission.

  5. Off the Trodden Path: Thinking Through the Military Exploration of the Information Domain

    National Research Council Canada - National Science Library

    O'Connell, Ed

    1997-01-01

    Trends in today's security environment point to a changed information domain on the horizon--a cyberspace of increased density, interconnectivity and collaboration, where links and nodes have disappeared...

  6. Managing the information systems in the industrial domain

    Directory of Open Access Journals (Sweden)

    Jaime Campos

    2016-12-01

    Full Text Available The purpose of the paper is to understand the developments of the Information Systems (IS and Information technology, i.e. Information and Communication Technologies (ICTs in the area of Condition Monitoring and Maintenance, especially e-maintenance. Therefore, the paper goes through and categorises the literature findings with the support of two models, i.e. the three era and the IS capability models to analyse and understand the maturity of the ICTs in the domain. The result shows that many researchers are working with software applications that are in the data processing era, which is highly important for the technologies and then the management of IS era. Conversely, both in the Management Information Systems and the Strategic Thinking era fewer findings have been made so far. Further on, in the fourth era, named the Information System (IS capability, the research findings are mostly at the rudimentary stage. The implementation of Web technologies, such as the Web 2.0, i.e. Social media technologies, can facilitate the learning aspects in the maintenance department as well as collaboration between employees, which are important aspects of the last era, i.e. in the IS capability era. The paper highlights interesting organisational as well as ICT aspects that should be considered when planning, developing and implementing the e-maintenance IS.

  7. Akt1 binds focal adhesion kinase via the Akt1 kinase domain independently of the pleckstrin homology domain.

    Science.gov (United States)

    Basson, M D; Zeng, B; Wang, S

    2015-10-01

    Akt1 and focal adhesion kinase (FAK) are protein kinases that play key roles in normal cell signaling. Individually, aberrant expression of these kinases has been linked to a variety of cancers. Together, Akt1/FAK interactions facilitate cancer metastasis by increasing cell adhesion under conditions of increased extracellular pressure. Pathological and iatrogenic sources of pressure arise from tumor growth against constraining stroma or direct perioperative manipulation. We previously reported that 15 mmHg increased extracellular pressure causes Akt1 to both directly interact with FAK and to phosphorylate and activate it. We investigated the nature of the Akt1/FAK binding by creating truncations of recombinant FAK, conjugated to glutathione S-transferase (GST), to pull down full-length Akt1. Western blots probing for Akt1 showed that FAK/Akt1 binding persisted in FAK truncations consisting of only amino acids 1-126, FAK(NT1), which contains the F1 subdomain of its band 4.1, ezrin, radixin, and moesin (FERM) domain. Using FAK(NT1) as bait, we then pulled down truncated versions of recombinant Akt1 conjugated to HA (human influenza hemagglutinin). Probes for GST-FAK(NT1) showed Akt1-FAK binding to occur in the absence of the both the Akt1 (N)-terminal pleckstrin homology (PH) domain and its adjacent hinge region. The Akt1 (C)-terminal regulatory domain was equally unnecessary for Akt1/FAK co-immunoprecipitation. Truncations involving the Akt1 catalytic domain showed that the domain by itself was enough to pull down FAK. Additionally, a fragment spanning from the PH domain to half way through the catalytic domain demonstrated increased FAK binding compared to full length Akt1. These results begin to delineate the Akt1/FAK interaction and can be used to manipulate their force-activated signal interactions. Furthermore, the finding that the N-terminal half of the Akt1 catalytic domain binds so strongly to FAK when cleaved from the rest of the protein may suggest a means

  8. Information access in the art history domain. Evaluating a federated search engine for Rembrandt research

    NARCIS (Netherlands)

    Verberne, S.; Boves, L.W.J.; Bosch, A.P.J. van den

    2016-01-01

    The art history domain is an interesting case for search engines tailored to the digital humanities, because the domain involves different types of sources (primary and secondary; text and images). One example of an art history search engine is RemBench, which provides access to information in four

  9. Takifugu rubripes cation independent mannose 6-phosphate receptor: Cloning, expression and functional characterization of the IGF-II binding domain.

    Science.gov (United States)

    A, Ajith Kumar; Nadimpalli, Siva Kumar

    2018-07-01

    Mannose 6-phosphate/IGF-II receptor mediated lysosomal clearance of insulin-like growth factor-II is significantly associated with the evolution of placental mammals. The protein is also referred to as the IGF-II receptor. Earlier studies suggested relatively low binding affinity between the receptor and ligand in prototherian and metatherian mammals. In the present study, we cloned the IGF-II binding domain of the early vertebrate fugu fish and expressed it in bacteria. A 72000Da truncated receptor containing the IGF-II binding domain was obtained. Analysis of this protein (covering domains 11-13 of the CIMPR) for its affinity to fish and human IGF-II by ligand blot assays and ELISA showed that the expressed receptor can specifically bind to both fish and human IGF-II. Additionally, a peptide-specific antibody raised against the region of the IGF-II binding domain also was able to recognize the IGF-II binding regions of mammalian and non-mammalian cation independent MPR protein. These interactions were further characterized by Surface Plasma resonance support that the receptor binds to fish IGF-II, with a dissociation constant of 548nM. Preliminary analysis suggests that the binding mechanism as well as the affinity of the fish and human receptor for IGF-II may have varied according to different evolutionary pressures. Copyright © 2018. Published by Elsevier B.V.

  10. Independent EEG sources are dipolar.

    Directory of Open Access Journals (Sweden)

    Arnaud Delorme

    Full Text Available Independent component analysis (ICA and blind source separation (BSS methods are increasingly used to separate individual brain and non-brain source signals mixed by volume conduction in electroencephalographic (EEG and other electrophysiological recordings. We compared results of decomposing thirteen 71-channel human scalp EEG datasets by 22 ICA and BSS algorithms, assessing the pairwise mutual information (PMI in scalp channel pairs, the remaining PMI in component pairs, the overall mutual information reduction (MIR effected by each decomposition, and decomposition 'dipolarity' defined as the number of component scalp maps matching the projection of a single equivalent dipole with less than a given residual variance. The least well-performing algorithm was principal component analysis (PCA; best performing were AMICA and other likelihood/mutual information based ICA methods. Though these and other commonly-used decomposition methods returned many similar components, across 18 ICA/BSS algorithms mean dipolarity varied linearly with both MIR and with PMI remaining between the resulting component time courses, a result compatible with an interpretation of many maximally independent EEG components as being volume-conducted projections of partially-synchronous local cortical field activity within single compact cortical domains. To encourage further method comparisons, the data and software used to prepare the results have been made available (http://sccn.ucsd.edu/wiki/BSSComparison.

  11. Crystallization and preliminary crystallographic studies of the W2 domain of Drosophila melanogaster eukaryotic translation initiation factor 5C domain-containing protein

    International Nuclear Information System (INIS)

    Zhao, Hui; Wang, Hong; Liu, Huihui; Teng, Maikun; Li, Xu

    2012-01-01

    The crystallization and preliminary crystallographic studies of the carboxy-terminal domain of D. melanogaster eukaryotic translation initiation factor 5C domain-containing protein are reported. The Drosophila melanogaster eukaryotic translation initiation factor 5C domain-containing protein (ECP) is composed of two independently folded domains which belong to the basic leucine-zipper and W2 domain-containing protein (BZW) family. Based on the sequence similarity between the C-terminal W2 domain of ECP and some eukaryotic translation initiation factors (such as eIF2B∊, eIF4γ, eIF5 etc.), ECP has been speculated to participate in the translation initiation process. Structural information on the C-terminal W2 domain of ECP would be helpful in understanding the specific cellular function of this protein. Here, the W2 domain of ECP was expressed and crystallized. Crystals grown by the hanging-drop vapour-diffusion method diffracted to 2.70 Å resolution and belonged to space group I4, with unit-cell parameters a = b = 81.05, c = 57.44 Å. The Matthews coefficient suggested that there was one molecule per asymmetric unit in the crystal

  12. Data Assimilation to Extract Soil Moisture Information from SMAP Observations

    Directory of Open Access Journals (Sweden)

    Jana Kolassa

    2017-11-01

    Full Text Available This study compares different methods to extract soil moisture information through the assimilation of Soil Moisture Active Passive (SMAP observations. Neural network (NN and physically-based SMAP soil moisture retrievals were assimilated into the National Aeronautics and Space Administration (NASA Catchment model over the contiguous United States for April 2015 to March 2017. By construction, the NN retrievals are consistent with the global climatology of the Catchment model soil moisture. Assimilating the NN retrievals without further bias correction improved the surface and root zone correlations against in situ measurements from 14 SMAP core validation sites (CVS by 0.12 and 0.16, respectively, over the model-only skill, and reduced the surface and root zone unbiased root-mean-square error (ubRMSE by 0.005 m 3 m − 3 and 0.001 m 3 m − 3 , respectively. The assimilation reduced the average absolute surface bias against the CVS measurements by 0.009 m 3 m − 3 , but increased the root zone bias by 0.014 m 3 m − 3 . Assimilating the NN retrievals after a localized bias correction yielded slightly lower surface correlation and ubRMSE improvements, but generally the skill differences were small. The assimilation of the physically-based SMAP Level-2 passive soil moisture retrievals using a global bias correction yielded similar skill improvements, as did the direct assimilation of locally bias-corrected SMAP brightness temperatures within the SMAP Level-4 soil moisture algorithm. The results show that global bias correction methods may be able to extract more independent information from SMAP observations compared to local bias correction methods, but without accurate quality control and observation error characterization they are also more vulnerable to adverse effects from retrieval errors related to uncertainties in the retrieval inputs and algorithm. Furthermore, the results show that using global bias correction approaches without a

  13. Music information retrieval in compressed audio files: a survey

    Science.gov (United States)

    Zampoglou, Markos; Malamos, Athanasios G.

    2014-07-01

    In this paper, we present an organized survey of the existing literature on music information retrieval systems in which descriptor features are extracted directly from the compressed audio files, without prior decompression to pulse-code modulation format. Avoiding the decompression step and utilizing the readily available compressed-domain information can significantly lighten the computational cost of a music information retrieval system, allowing application to large-scale music databases. We identify a number of systems relying on compressed-domain information and form a systematic classification of the features they extract, the retrieval tasks they tackle and the degree in which they achieve an actual increase in the overall speed-as well as any resulting loss in accuracy. Finally, we discuss recent developments in the field, and the potential research directions they open toward ultra-fast, scalable systems.

  14. Recognition techniques for extracting information from semistructured documents

    Science.gov (United States)

    Della Ventura, Anna; Gagliardi, Isabella; Zonta, Bruna

    2000-12-01

    Archives of optical documents are more and more massively employed, the demand driven also by the new norms sanctioning the legal value of digital documents, provided they are stored on supports that are physically unalterable. On the supply side there is now a vast and technologically advanced market, where optical memories have solved the problem of the duration and permanence of data at costs comparable to those for magnetic memories. The remaining bottleneck in these systems is the indexing. The indexing of documents with a variable structure, while still not completely automated, can be machine supported to a large degree with evident advantages both in the organization of the work, and in extracting information, providing data that is much more detailed and potentially significant for the user. We present here a system for the automatic registration of correspondence to and from a public office. The system is based on a general methodology for the extraction, indexing, archiving, and retrieval of significant information from semi-structured documents. This information, in our prototype application, is distributed among the database fields of sender, addressee, subject, date, and body of the document.

  15. An Ontology Driven Information Architecture for Big Data and Diverse Domains

    Science.gov (United States)

    Hughes, John S.; Crichton, Dan; Hardman, Sean; Joyner, Ron; Ramirez, Paul

    2013-04-01

    The Planetary Data System's has just released the PDS4 system for first use. Its architecture is comprised of three principle parts, an ontology that captures knowledge from the planetary science domain, a federated registry/repository system for product identification, versioning, tracking, and storage, and a REST-based service layer for search, retrieval, and distribution. An ontology modeling tool is used to prescriptively capture product definitions that adhere to object-oriented principles and that are compliant with specific registry, archive, and data dictionary reference models. The resulting information model is product centric, allowing all information to be packaged into products and tracked in the registry. The flexibility required in a diverse domain is provided through the use of object-oriented extensions and a hierarchical governance scheme with common, discipline, and mission levels. Finally all PDS4 data standards are generated or derived from the information model. The federated registry provides identification, versioning, and tracking functionality across federated repositories and is configured for deployment using configuration files generated from the ontology. Finally a REST-based service layer provides for metadata harvest, product transformation, packaging, and search, and portal hosting. A model driven architecture allows the data and software engineering teams to develop in parallel with minimal team interaction. The resulting software remains relatively stable as the domain evolves. Finally the development of a single shared ontology promotes interoperability and data correlation and helps meet the expectations of modern scientists for science data discovery, access and use. This presentation will provide an overview of PDS4 focusing on the data standards, how they were developed, how they are now being used, and will present some of the lessons learned while developing in a diverse scientific community. Copyright 2013 California

  16. Map as a Service: A Framework for Visualising and Maximising Information Return from Multi-ModalWireless Sensor Networks

    Directory of Open Access Journals (Sweden)

    Mohammad Hammoudeh

    2015-09-01

    Full Text Available This paper presents a distributed information extraction and visualisation service, called the mapping service, for maximising information return from large-scale wireless sensor networks. Such a service would greatly simplify the production of higher-level, information-rich, representations suitable for informing other network services and the delivery of field information visualisations. The mapping service utilises a blend of inductive and deductive models to map sense data accurately using externally available knowledge. It utilises the special characteristics of the application domain to render visualisations in a map format that are a precise reflection of the concrete reality. This service is suitable for visualising an arbitrary number of sense modalities. It is capable of visualising from multiple independent types of the sense data to overcome the limitations of generating visualisations from a single type of sense modality. Furthermore, the mapping service responds dynamically to changes in the environmental conditions, which may affect the visualisation performance by continuously updating the application domain model in a distributed manner. Finally, a distributed self-adaptation function is proposed with the goal of saving more power and generating more accurate data visualisation. We conduct comprehensive experimentation to evaluate the performance of our mapping service and show that it achieves low communication overhead, produces maps of high fidelity, and further minimises the mapping predictive error dynamically through integrating the application domain model in the mapping service.

  17. A Unified Framework for Creating Domain Dependent Polarity Lexicons from User Generated Reviews.

    Directory of Open Access Journals (Sweden)

    Muhammad Zubair Asghar

    Full Text Available The exponential increase in the explosion of Web-based user generated reviews has resulted in the emergence of Opinion Mining (OM applications for analyzing the users' opinions toward products, services, and policies. The polarity lexicons often play a pivotal role in the OM, indicating the positivity and negativity of a term along with the numeric score. However, the commonly available domain independent lexicons are not an optimal choice for all of the domains within the OM applications. The aforementioned is due to the fact that the polarity of a term changes from one domain to other and such lexicons do not contain the correct polarity of a term for every domain. In this work, we focus on the problem of adapting a domain dependent polarity lexicon from set of labeled user reviews and domain independent lexicon to propose a unified learning framework based on the information theory concepts that can assign the terms with correct polarity (+ive, -ive scores. The benchmarking on three datasets (car, hotel, and drug reviews shows that our approach improves the performance of the polarity classification by achieving higher accuracy. Moreover, using the derived domain dependent lexicon changed the polarity of terms, and the experimental results show that our approach is more effective than the base line methods.

  18. Naive Bayes Bearing Fault Diagnosis Based on Enhanced Independence of Data.

    Science.gov (United States)

    Zhang, Nannan; Wu, Lifeng; Yang, Jing; Guan, Yong

    2018-02-05

    The bearing is the key component of rotating machinery, and its performance directly determines the reliability and safety of the system. Data-based bearing fault diagnosis has become a research hotspot. Naive Bayes (NB), which is based on independent presumption, is widely used in fault diagnosis. However, the bearing data are not completely independent, which reduces the performance of NB algorithms. In order to solve this problem, we propose a NB bearing fault diagnosis method based on enhanced independence of data. The method deals with data vector from two aspects: the attribute feature and the sample dimension. After processing, the classification limitation of NB is reduced by the independence hypothesis. First, we extract the statistical characteristics of the original signal of the bearings effectively. Then, the Decision Tree algorithm is used to select the important features of the time domain signal, and the low correlation features is selected. Next, the Selective Support Vector Machine (SSVM) is used to prune the dimension data and remove redundant vectors. Finally, we use NB to diagnose the fault with the low correlation data. The experimental results show that the independent enhancement of data is effective for bearing fault diagnosis.

  19. Naive Bayes Bearing Fault Diagnosis Based on Enhanced Independence of Data

    Science.gov (United States)

    Zhang, Nannan; Wu, Lifeng; Yang, Jing; Guan, Yong

    2018-01-01

    The bearing is the key component of rotating machinery, and its performance directly determines the reliability and safety of the system. Data-based bearing fault diagnosis has become a research hotspot. Naive Bayes (NB), which is based on independent presumption, is widely used in fault diagnosis. However, the bearing data are not completely independent, which reduces the performance of NB algorithms. In order to solve this problem, we propose a NB bearing fault diagnosis method based on enhanced independence of data. The method deals with data vector from two aspects: the attribute feature and the sample dimension. After processing, the classification limitation of NB is reduced by the independence hypothesis. First, we extract the statistical characteristics of the original signal of the bearings effectively. Then, the Decision Tree algorithm is used to select the important features of the time domain signal, and the low correlation features is selected. Next, the Selective Support Vector Machine (SSVM) is used to prune the dimension data and remove redundant vectors. Finally, we use NB to diagnose the fault with the low correlation data. The experimental results show that the independent enhancement of data is effective for bearing fault diagnosis. PMID:29401730

  20. Information rich mapping requirement to product architecture through functional system deployment: The multi entity domain approach

    DEFF Research Database (Denmark)

    Hauksdóttir, Dagný; Mortensen, Niels Henrik

    2017-01-01

    may impede the ability to evolve, maintain or reuse systems. In this paper the Multi Entity Domain Approach (MEDA) is presented. The approach combines different design information within the domain views, incorporates both Software and Hardware design and supports iterative requirements definition...

  1. Domain-restricted mutation analysis to identify novel driver events in human cancer

    Directory of Open Access Journals (Sweden)

    Sanket Desai

    2017-10-01

    Full Text Available Analysis of mutational spectra across various cancer types has given valuable insights into tumorigenesis. Different approaches have been used to identify novel drivers from the set of somatic mutations, including the methods which use sequence conservation, geometric localization and pathway information. Recent computational methods suggest use of protein domain information for analysis and understanding of the functional consequence of non-synonymous mutations. Similarly, evidence suggests recurrence at specific position in proteins is robust indicators of its functional impact. Building on this, we performed a systematic analysis of TCGA exome derived somatic mutations across 6089 PFAM domains and significantly mutated domains were identified using randomization approach. Multiple alignment of individual domain allowed us to prioritize for conserved residues mutated at analogous positions across different proteins in a statistically disciplined manner. In addition to the known frequently mutated genes, this analysis independently identifies low frequency Meprin and TRAF-Homology (MATH domain in Speckle Type BTB/POZ (SPOP protein, in prostate adenocarcinoma. Results from this analysis will help generate hypotheses about the downstream molecular mechanism resulting in cancer phenotypes.

  2. Memory and Comprehension for Health Information among Older Adults: Distinguishing the Effects of Domain-General and Domain-Specific Knowledge

    Science.gov (United States)

    Chin, Jessie; Payne, Brennan; Gao, Xuefei; Conner-Garcia, Thembi; Graumlich, James F.; Murray, Michael D.; Morrow, Daniel G.; Stine-Morrow, Elizabeth A.L.

    2014-01-01

    While there is evidence that knowledge influences understanding of health information, less is known about the processing mechanisms underlying this effect and its impact on memory. We used the moving window paradigm to examine how older adults varying in domain-general crystallized ability (verbal ability) and health knowledge allocate attention to understand health and domain-general texts. Participants (n=107, aged 60 to 88 yrs) read and recalled single sentences about hypertension and about non-health topics. Mixed-effects modeling of word-by-word reading times suggested that domain-general crystallized ability increased conceptual integration regardless of text domain, while health knowledge selectively increased resource allocation to conceptual integration at clause boundaries in health texts. These patterns of attentional allocation were related to subsequent recall performance. Although older adults with lower levels of crystallized ability were less likely to engage in integrative processing, when they did, this strategy had a compensatory effect in improving recall. These findings suggest that semantic integration during reading is an important comprehension process that supports the construction of the memory representation and is engendered by knowledge. Implications of the findings for theories of text processing and memory as well as for designing patient education materials are discussed. PMID:24787361

  3. Memory and comprehension for health information among older adults: distinguishing the effects of domain-general and domain-specific knowledge.

    Science.gov (United States)

    Chin, Jessie; Payne, Brennan; Gao, Xuefei; Conner-Garcia, Thembi; Graumlich, James F; Murray, Michael D; Morrow, Daniel G; Stine-Morrow, Elizabeth A L

    2015-01-01

    While there is evidence that knowledge influences understanding of health information, less is known about the processing mechanisms underlying this effect and its impact on memory. We used the moving window paradigm to examine how older adults varying in domain-general crystallised ability (verbal ability) and health knowledge allocate attention to understand health and domain-general texts. Participants (n = 107, age: 60-88 years) read and recalled single sentences about hypertension and about non-health topics. Mixed-effects modelling of word-by-word reading times suggested that domain-general crystallised ability increased conceptual integration regardless of text domain, while health knowledge selectively increased resource allocation to conceptual integration at clause boundaries in health texts. These patterns of attentional allocation were related to subsequent recall performance. Although older adults with lower levels of crystallised ability were less likely to engage in integrative processing, when they did, this strategy had a compensatory effect in improving recall. These findings suggest that semantic integration during reading is an important comprehension process that supports the construction of the memory representation and is engendered by knowledge. Implications of the findings for theories of text processing and memory as well as for designing patient education materials are discussed.

  4. RESEARCH ON REMOTE SENSING GEOLOGICAL INFORMATION EXTRACTION BASED ON OBJECT ORIENTED CLASSIFICATION

    Directory of Open Access Journals (Sweden)

    H. Gao

    2018-04-01

    Full Text Available The northern Tibet belongs to the Sub cold arid climate zone in the plateau. It is rarely visited by people. The geological working conditions are very poor. However, the stratum exposures are good and human interference is very small. Therefore, the research on the automatic classification and extraction of remote sensing geological information has typical significance and good application prospect. Based on the object-oriented classification in Northern Tibet, using the Worldview2 high-resolution remote sensing data, combined with the tectonic information and image enhancement, the lithological spectral features, shape features, spatial locations and topological relations of various geological information are excavated. By setting the threshold, based on the hierarchical classification, eight kinds of geological information were classified and extracted. Compared with the existing geological maps, the accuracy analysis shows that the overall accuracy reached 87.8561 %, indicating that the classification-oriented method is effective and feasible for this study area and provides a new idea for the automatic extraction of remote sensing geological information.

  5. Terrain Extraction by Integrating Terrestrial Laser Scanner Data and Spectral Information

    Science.gov (United States)

    Lau, C. L.; Halim, S.; Zulkepli, M.; Azwan, A. M.; Tang, W. L.; Chong, A. K.

    2015-10-01

    The extraction of true terrain points from unstructured laser point cloud data is an important process in order to produce an accurate digital terrain model (DTM). However, most of these spatial filtering methods just utilizing the geometrical data to discriminate the terrain points from nonterrain points. The point cloud filtering method also can be improved by using the spectral information available with some scanners. Therefore, the objective of this study is to investigate the effectiveness of using the three-channel (red, green and blue) of the colour image captured from built-in digital camera which is available in some Terrestrial Laser Scanner (TLS) for terrain extraction. In this study, the data acquisition was conducted at a mini replica landscape in Universiti Teknologi Malaysia (UTM), Skudai campus using Leica ScanStation C10. The spectral information of the coloured point clouds from selected sample classes are extracted for spectral analysis. The coloured point clouds which within the corresponding preset spectral threshold are identified as that specific feature point from the dataset. This process of terrain extraction is done through using developed Matlab coding. Result demonstrates that a higher spectral resolution passive image is required in order to improve the output. This is because low quality of the colour images captured by the sensor contributes to the low separability in spectral reflectance. In conclusion, this study shows that, spectral information is capable to be used as a parameter for terrain extraction.

  6. Extracting Semantic Information from Visual Data: A Survey

    Directory of Open Access Journals (Sweden)

    Qiang Liu

    2016-03-01

    Full Text Available The traditional environment maps built by mobile robots include both metric ones and topological ones. These maps are navigation-oriented and not adequate for service robots to interact with or serve human users who normally rely on the conceptual knowledge or semantic contents of the environment. Therefore, the construction of semantic maps becomes necessary for building an effective human-robot interface for service robots. This paper reviews recent research and development in the field of visual-based semantic mapping. The main focus is placed on how to extract semantic information from visual data in terms of feature extraction, object/place recognition and semantic representation methods.

  7. Dynamic properties of independent chromatin domains measured by correlation spectroscopy in living cells.

    NARCIS (Netherlands)

    M. Wachsmuth (Malte); T.A. Knoch (Tobias); K. Rippe (Karsten)

    2016-01-01

    textabstractBackground: Genome organization into subchromosomal topologically associating domains (TADs) is linked to cell-type-specific gene expression programs. However, dynamic properties of such domains remain elusive, and it is unclear how domain plasticity modulates genomic accessibility for

  8. Aggression and Moral Development: Integrating Social Information Processing and Moral Domain Models

    Science.gov (United States)

    Arsenio, William F.; Lemerise, Elizabeth A.

    2004-01-01

    Social information processing and moral domain theories have developed in relative isolation from each other despite their common focus on intentional harm and victimization, and mutual emphasis on social cognitive processes in explaining aggressive, morally relevant behaviors. This article presents a selective summary of these literatures with…

  9. Domain-Specific Thesaurus as a Tool for Information Retrieval and Collection of Knowledge

    Directory of Open Access Journals (Sweden)

    Vladimir N. Boikov

    2013-01-01

    Full Text Available This paper reports basic approaches to constructive creation of an open resource named ”Domain-specified thesaurus of poetics”, which is one of the levels of an information-analytical system of the Russian poetry (IAS RP. The poetics is a group of disciplines focused on a comprehensive theoretical and historical study of poetry. IAS RP will be used as a tool for a wide range of studies allowing to determine the characteristic features of the analyzed works of poetry. Consequently, the thesaurus is the knowledge base from which one can borrow input data for training the system. The aim of our research requires a specific approach to formating the knowledge base. Thesaurus is a web-based resource which includes a domain-specific directory, information retrieval tools and tools for further analyzes. The study of glossary consisting of three thousand terms and a set of semantic fields is reviewed in this paper. Rdf-graph of the domain-specified thesaurus of poetics is presented, containing 9 types of objects and different kinds of relationships among them. Wiki-tecnologies are used for implementing a resource which allows to store data in Semantic Web formats.

  10. Extracting intrinsic functional networks with feature-based group independent component analysis.

    Science.gov (United States)

    Calhoun, Vince D; Allen, Elena

    2013-04-01

    There is increasing use of functional imaging data to understand the macro-connectome of the human brain. Of particular interest is the structure and function of intrinsic networks (regions exhibiting temporally coherent activity both at rest and while a task is being performed), which account for a significant portion of the variance in functional MRI data. While networks are typically estimated based on the temporal similarity between regions (based on temporal correlation, clustering methods, or independent component analysis [ICA]), some recent work has suggested that these intrinsic networks can be extracted from the inter-subject covariation among highly distilled features, such as amplitude maps reflecting regions modulated by a task or even coordinates extracted from large meta analytic studies. In this paper our goal was to explicitly compare the networks obtained from a first-level ICA (ICA on the spatio-temporal functional magnetic resonance imaging (fMRI) data) to those from a second-level ICA (i.e., ICA on computed features rather than on the first-level fMRI data). Convergent results from simulations, task-fMRI data, and rest-fMRI data show that the second-level analysis is slightly noisier than the first-level analysis but yields strikingly similar patterns of intrinsic networks (spatial correlations as high as 0.85 for task data and 0.65 for rest data, well above the empirical null) and also preserves the relationship of these networks with other variables such as age (for example, default mode network regions tended to show decreased low frequency power for first-level analyses and decreased loading parameters for second-level analyses). In addition, the best-estimated second-level results are those which are the most strongly reflected in the input feature. In summary, the use of feature-based ICA appears to be a valid tool for extracting intrinsic networks. We believe it will become a useful and important approach in the study of the macro

  11. Real-time decision support and information gathering system for financial domain

    Science.gov (United States)

    Tseng, Chiu-Che; Gmytrasiewicz, Piotr J.

    2006-05-01

    The challenge of the investment domain is that a large amount of diverse information can be potentially relevant to an investment decision, and that, frequently, the decisions have to be made in a timely manner. This presents the potential for better decision support, but poses the challenge of building a decision support agent that gathers information from different sources and incorporates it for timely decision support. These problems motivate us to investigate ways in which the investors can be equipped with a flexible real-time decision support system to be practical in time-critical situations. The flexible real-time decision support system considers a tradeoff between decision quality and computation cost. For this purpose, we propose a system that uses the object oriented Bayesian knowledge base (OOBKB) design to create a decision model at the most suitable level of detail to guide the information gathering activities, and to produce an investment recommendation within a reasonable length of time. The decision models our system uses are implemented as influence diagrams. We validate our system with experiments in a simplified investment domain. The experiments show that our system produces a quality recommendation under different urgency situations. The contribution of our system is that it provides the flexible decision recommendation for an investor under time constraints in a complex environment.

  12. Care episode retrieval: distributional semantic models for information retrieval in the clinical domain.

    Science.gov (United States)

    Moen, Hans; Ginter, Filip; Marsi, Erwin; Peltonen, Laura-Maria; Salakoski, Tapio; Salanterä, Sanna

    2015-01-01

    Patients' health related information is stored in electronic health records (EHRs) by health service providers. These records include sequential documentation of care episodes in the form of clinical notes. EHRs are used throughout the health care sector by professionals, administrators and patients, primarily for clinical purposes, but also for secondary purposes such as decision support and research. The vast amounts of information in EHR systems complicate information management and increase the risk of information overload. Therefore, clinicians and researchers need new tools to manage the information stored in the EHRs. A common use case is, given a--possibly unfinished--care episode, to retrieve the most similar care episodes among the records. This paper presents several methods for information retrieval, focusing on care episode retrieval, based on textual similarity, where similarity is measured through domain-specific modelling of the distributional semantics of words. Models include variants of random indexing and the semantic neural network model word2vec. Two novel methods are introduced that utilize the ICD-10 codes attached to care episodes to better induce domain-specificity in the semantic model. We report on experimental evaluation of care episode retrieval that circumvents the lack of human judgements regarding episode relevance. Results suggest that several of the methods proposed outperform a state-of-the art search engine (Lucene) on the retrieval task.

  13. Extracting the information backbone in online system.

    Directory of Open Access Journals (Sweden)

    Qian-Ming Zhang

    Full Text Available Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  14. Extracting the information backbone in online system.

    Science.gov (United States)

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such "less can be more" feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency.

  15. Extracting the Information Backbone in Online System

    Science.gov (United States)

    Zhang, Qian-Ming; Zeng, An; Shang, Ming-Sheng

    2013-01-01

    Information overload is a serious problem in modern society and many solutions such as recommender system have been proposed to filter out irrelevant information. In the literature, researchers have been mainly dedicated to improving the recommendation performance (accuracy and diversity) of the algorithms while they have overlooked the influence of topology of the online user-object bipartite networks. In this paper, we find that some information provided by the bipartite networks is not only redundant but also misleading. With such “less can be more” feature, we design some algorithms to improve the recommendation performance by eliminating some links from the original networks. Moreover, we propose a hybrid method combining the time-aware and topology-aware link removal algorithms to extract the backbone which contains the essential information for the recommender systems. From the practical point of view, our method can improve the performance and reduce the computational time of the recommendation system, thus improving both of their effectiveness and efficiency. PMID:23690946

  16. Activatory and Inhibitory Fcγ Receptors Augment Rituximab-mediated Internalization of CD20 Independent of Signaling via the Cytoplasmic Domain*

    Science.gov (United States)

    Vaughan, Andrew T.; Chan, Claude H. T.; Klein, Christian; Glennie, Martin J.; Beers, Stephen A.; Cragg, Mark S.

    2015-01-01

    Type I anti-CD20 mAb such as rituximab and ofatumumab engage with the inhibitory FcγR, FcγRIIb on the surface of B cells, resulting in immunoreceptor tyrosine-based inhibitory motif (ITIM) phosphorylation. Internalization of the CD20·mAb·FcγRIIb complex follows, the rate of which correlates with FcγRIIb expression. In contrast, although type II anti-CD20 mAb such as tositumomab and obinutuzumab also interact with and activate FcγRIIb, this interaction fails to augment the rate of CD20·mAb internalization, raising the question of whether ITIM phosphorylation plays any role in this process. We have assessed the molecular requirements for the internalization process and demonstrate that in contrast to internalization of IgG immune complexes, FcγRIIb-augmented internalization of rituximab-ligated CD20 occurs independently of the FcγRIIb ITIM, indicating that signaling downstream of FcγRIIb is not required. In transfected cells, activatory FcγRI, FcγRIIa, and FcγRIIIa augmented internalization of rituximab-ligated CD20 in a similar manner. However, FcγRIIa mediated a slower rate of internalization than cells expressing equivalent levels of the highly homologous FcγRIIb. The difference was maintained in cells expressing FcγRIIa and FcγRIIb lacking cytoplasmic domains and in which the transmembrane domains had been exchanged. This difference may be due to increased degradation of FcγRIIa, which traffics to lysosomes independently of rituximab. We conclude that the cytoplasmic domain of FcγR is not required for promoting internalization of rituximab-ligated CD20. Instead, we propose that FcγR provides a structural role in augmenting endocytosis that differs from that employed during the endocytosis of immune complexes. PMID:25568316

  17. How ISO/IEC 17799 can be used for base lining information assurance among entities using data mining for defense, homeland security, commercial, and other civilian/commercial domains

    Science.gov (United States)

    Perry, William G.

    2006-04-01

    One goal of database mining is to draw unique and valid perspectives from multiple data sources. Insights that are fashioned from closely-held data stores are likely to possess a high degree of reliability. The degree of information assurance comes into question, however, when external databases are accessed, combined and analyzed to form new perspectives. ISO/IEC 17799, Information technology-Security techniques-Code of practice for information security management, can be used to establish a higher level of information assurance among disparate entities using data mining in the defense, homeland security, commercial and other civilian/commercial domains. Organizations that meet ISO/IEC information security standards have identified and assessed risks, threats and vulnerabilities and have taken significant proactive steps to meet their unique security requirements. The ISO standards address twelve domains: risk assessment and treatment, security policy, organization of information security, asset management, human resources security, physical and environmental security, communications and operations management, access control, information systems acquisition, development and maintenance, information security incident management and business continuity management and compliance. Analysts can be relatively confident that if organizations are ISO 17799 compliant, a high degree of information assurance is likely to be a characteristic of the data sets being used. The reverse may be true. Extracting, fusing and drawing conclusions based upon databases with a low degree of information assurance may be wrought with all of the hazards that come from knowingly using bad data to make decisions. Using ISO/IEC 17799 as a baseline for information assurance can help mitigate these risks.

  18. Cytotoxicity and Proapoptotic Effects of Allium atroviolaceum Flower Extract by Modulating Cell Cycle Arrest and Caspase-Dependent and p53-Independent Pathway in Breast Cancer Cell Lines

    Directory of Open Access Journals (Sweden)

    Somayeh Khazaei

    2017-01-01

    Full Text Available Breast cancer is the second leading cause of cancer death among women and despite significant advances in therapy, it remains a critical health problem worldwide. Allium atroviolaceum is an herbaceous plant, with limited information about the therapeutic capability. We aimed to study the anticancer effect of flower extract and the mechanisms of action in MCF-7 and MDA-MB-231. The extract inhibits the proliferation of the cells in a time- and dose-dependent manner. The underlying mechanism involved the stimulation of S and G2/M phase arrest in MCF-7 and S phase arrest in MDA-MB-231 associated with decreased level of Cdk1, in a p53-independent pathway. Furthermore, the extract induces apoptosis in both cell lines, as indicated by the percentage of sub-G0 population, the morphological changes observed by phase contrast and fluorescent microscopy, and increase in Annexin-V-positive cells. The apoptosis induction was related to downregulation of Bcl-2 and also likely to be caspase-dependent. Moreover, the combination of the extract and tamoxifen exhibits synergistic effect, suggesting that it can complement current chemotherapy. LC-MS analysis displayed 17 major compounds in the extract which might be responsible for the observed effects. Overall, this study demonstrates the potential applications of Allium atroviolaceum extract as an anticancer drug for breast cancer treatment.

  19. Cultural-Historical Activity Theory and Domain Analysis: Metatheoretical Implications for Information Science

    Science.gov (United States)

    Wang, Lin

    2013-01-01

    Background: Cultural-historical activity theory is an important theory in modern psychology. In recent years, it has drawn more attention from related disciplines including information science. Argument: This paper argues that activity theory and domain analysis which uses the theory as one of its bases could bring about some important…

  20. Integrating Semantic Information in Metadata Descriptions for a Geoscience-wide Resource Inventory.

    Science.gov (United States)

    Zaslavsky, I.; Richard, S. M.; Gupta, A.; Valentine, D.; Whitenack, T.; Ozyurt, I. B.; Grethe, J. S.; Schachne, A.

    2016-12-01

    Integrating semantic information into legacy metadata catalogs is a challenging issue and so far has been mostly done on a limited scale. We present experience of CINERGI (Community Inventory of Earthcube Resources for Geoscience Interoperability), an NSF Earthcube Building Block project, in creating a large cross-disciplinary catalog of geoscience information resources to enable cross-domain discovery. The project developed a pipeline for automatically augmenting resource metadata, in particular generating keywords that describe metadata documents harvested from multiple geoscience information repositories or contributed by geoscientists through various channels including surveys and domain resource inventories. The pipeline examines available metadata descriptions using text parsing, vocabulary management and semantic annotation and graph navigation services of GeoSciGraph. GeoSciGraph, in turn, relies on a large cross-domain ontology of geoscience terms, which bridges several independently developed ontologies or taxonomies including SWEET, ENVO, YAGO, GeoSciML, GCMD, SWO, and CHEBI. The ontology content enables automatic extraction of keywords reflecting science domains, equipment used, geospatial features, measured properties, methods, processes, etc. We specifically focus on issues of cross-domain geoscience ontology creation, resolving several types of semantic conflicts among component ontologies or vocabularies, and constructing and managing facets for improved data discovery and navigation. The ontology and keyword generation rules are iteratively improved as pipeline results are presented to data managers for selective manual curation via a CINERGI Annotator user interface. We present lessons learned from applying CINERGI metadata augmentation pipeline to a number of federal agency and academic data registries, in the context of several use cases that require data discovery and integration across multiple earth science data catalogs of varying quality

  1. Providing information on the spot : Using augmented reality for situational awareness in the security domain

    NARCIS (Netherlands)

    Lukosch, S.G.; Lukosch, H.K.; Datcu, D.; Cidota, M.A.

    2015-01-01

    For operational units in the security domain that work together in teams, it is important to quickly and adequately exchange context-related information to ensure well-working collaboration. Currently, most information exchange is based on oral communication. This paper reports on different

  2. Providing Information on the Spot : Using Augmented Reality for Situational Awareness in the Security Domain

    NARCIS (Netherlands)

    Lukosch, S.G.; Lukosch, H.K.; Datcu, D.; Cidota, M.A.

    2015-01-01

    For operational units in the security domain that work together in teams, it is important to quickly and adequately exchange context-related information to ensure well-working collaboration. Currently, most information exchange is based on oral communication. This paper reports on different

  3. Application of wavelet transform for PDZ domain classification.

    Directory of Open Access Journals (Sweden)

    Khaled Daqrouq

    Full Text Available PDZ domains have been identified as part of an array of signaling proteins that are often unrelated, except for the well-conserved structural PDZ domain they contain. These domains have been linked to many disease processes including common Avian influenza, as well as very rare conditions such as Fraser and Usher syndromes. Historically, based on the interactions and the nature of bonds they form, PDZ domains have most often been classified into one of three classes (class I, class II and others - class III, that is directly dependent on their binding partner. In this study, we report on three unique feature extraction approaches based on the bigram and trigram occurrence and existence rearrangements within the domain's primary amino acid sequences in assisting PDZ domain classification. Wavelet packet transform (WPT and Shannon entropy denoted by wavelet entropy (WE feature extraction methods were proposed. Using 115 unique human and mouse PDZ domains, the existence rearrangement approach yielded a high recognition rate (78.34%, which outperformed our occurrence rearrangements based method. The recognition rate was (81.41% with validation technique. The method reported for PDZ domain classification from primary sequences proved to be an encouraging approach for obtaining consistent classification results. We anticipate that by increasing the database size, we can further improve feature extraction and correct classification.

  4. Radio weak lensing shear measurement in the visibility domain - II. Source extraction

    Science.gov (United States)

    Rivi, M.; Miller, L.

    2018-05-01

    This paper extends the method introduced in Rivi et al. (2016b) to measure galaxy ellipticities in the visibility domain for radio weak lensing surveys. In that paper, we focused on the development and testing of the method for the simple case of individual galaxies located at the phase centre, and proposed to extend it to the realistic case of many sources in the field of view by isolating visibilities of each source with a faceting technique. In this second paper, we present a detailed algorithm for source extraction in the visibility domain and show its effectiveness as a function of the source number density by running simulations of SKA1-MID observations in the band 950-1150 MHz and comparing original and measured values of galaxies' ellipticities. Shear measurements from a realistic population of 104 galaxies randomly located in a field of view of 1 \\deg ^2 (i.e. the source density expected for the current radio weak lensing survey proposal with SKA1) are also performed. At SNR ≥ 10, the multiplicative bias is only a factor 1.5 worse than what found when analysing individual sources, and is still comparable to the bias values reported for similar measurement methods at optical wavelengths. The additive bias is unchanged from the case of individual sources, but it is significantly larger than typically found in optical surveys. This bias depends on the shape of the uv coverage and we suggest that a uv-plane weighting scheme to produce a more isotropic shape could reduce and control additive bias.

  5. Domain analysis

    DEFF Research Database (Denmark)

    Hjørland, Birger

    2017-01-01

    The domain-analytic approach to knowledge organization (KO) (and to the broader field of library and information science, LIS) is outlined. The article reviews the discussions and proposals on the definition of domains, and provides an example of a domain-analytic study in the field of art studies....... Varieties of domain analysis as well as criticism and controversies are presented and discussed....

  6. Extraction Method for Earthquake-Collapsed Building Information Based on High-Resolution Remote Sensing

    International Nuclear Information System (INIS)

    Chen, Peng; Wu, Jian; Liu, Yaolin; Wang, Jing

    2014-01-01

    At present, the extraction of earthquake disaster information from remote sensing data relies on visual interpretation. However, this technique cannot effectively and quickly obtain precise and efficient information for earthquake relief and emergency management. Collapsed buildings in the town of Zipingpu after the Wenchuan earthquake were used as a case study to validate two kinds of rapid extraction methods for earthquake-collapsed building information based on pixel-oriented and object-oriented theories. The pixel-oriented method is based on multi-layer regional segments that embody the core layers and segments of the object-oriented method. The key idea is to mask layer by layer all image information, including that on the collapsed buildings. Compared with traditional techniques, the pixel-oriented method is innovative because it allows considerably rapid computer processing. As for the object-oriented method, a multi-scale segment algorithm was applied to build a three-layer hierarchy. By analyzing the spectrum, texture, shape, location, and context of individual object classes in different layers, the fuzzy determined rule system was established for the extraction of earthquake-collapsed building information. We compared the two sets of results using three variables: precision assessment, visual effect, and principle. Both methods can extract earthquake-collapsed building information quickly and accurately. The object-oriented method successfully overcomes the pepper salt noise caused by the spectral diversity of high-resolution remote sensing data and solves the problem of same object, different spectrums and that of same spectrum, different objects. With an overall accuracy of 90.38%, the method achieves more scientific and accurate results compared with the pixel-oriented method (76.84%). The object-oriented image analysis method can be extensively applied in the extraction of earthquake disaster information based on high-resolution remote sensing

  7. Improving information extraction using a probability-based approach

    DEFF Research Database (Denmark)

    Kim, S.; Ahmed, Saeema; Wallace, K.

    2007-01-01

    Information plays a crucial role during the entire life-cycle of a product. It has been shown that engineers frequently consult colleagues to obtain the information they require to solve problems. However, the industrial world is now more transient and key personnel move to other companies...... or retire. It is becoming essential to retrieve vital information from archived product documents, if it is available. There is, therefore, great interest in ways of extracting relevant and sharable information from documents. A keyword-based search is commonly used, but studies have shown...... the recall, while maintaining the high precision, a learning approach that makes identification decisions based on a probability model, rather than simply looking up the presence of the pre-defined variations, looks promising. This paper presents the results of developing such a probability-based entity...

  8. Semi-automatic building extraction in informal settlements from high-resolution satellite imagery

    Science.gov (United States)

    Mayunga, Selassie David

    The extraction of man-made features from digital remotely sensed images is considered as an important step underpinning management of human settlements in any country. Man-made features and buildings in particular are required for varieties of applications such as urban planning, creation of geographical information systems (GIS) databases and Urban City models. The traditional man-made feature extraction methods are very expensive in terms of equipment, labour intensive, need well-trained personnel and cannot cope with changing environments, particularly in dense urban settlement areas. This research presents an approach for extracting buildings in dense informal settlement areas using high-resolution satellite imagery. The proposed system uses a novel strategy of extracting building by measuring a single point at the approximate centre of the building. The fine measurement of the building outlines is then effected using a modified snake model. The original snake model on which this framework is based, incorporates an external constraint energy term which is tailored to preserving the convergence properties of the snake model; its use to unstructured objects will negatively affect their actual shapes. The external constrained energy term was removed from the original snake model formulation, thereby, giving ability to cope with high variability of building shapes in informal settlement areas. The proposed building extraction system was tested on two areas, which have different situations. The first area was Tungi in Dar Es Salaam, Tanzania where three sites were tested. This area is characterized by informal settlements, which are illegally formulated within the city boundaries. The second area was Oromocto in New Brunswick, Canada where two sites were tested. Oromocto area is mostly flat and the buildings are constructed using similar materials. Qualitative and quantitative measures were employed to evaluate the accuracy of the results as well as the performance

  9. State space model extraction of thermohydraulic systems – Part I: A linear graph approach

    International Nuclear Information System (INIS)

    Uren, K.R.; Schoor, G. van

    2013-01-01

    Thermohydraulic simulation codes are increasingly making use of graphical design interfaces. The user can quickly and easily design a thermohydraulic system by placing symbols on the screen resembling system components. These components can then be connected to form a system representation. Such system models may then be used to obtain detailed simulations of the physical system. Usually this kind of simulation models are too complex and not ideal for control system design. Therefore, a need exists for automated techniques to extract lumped parameter models useful for control system design. The goal of this first paper, in a two part series, is to propose a method that utilises a graphical representation of a thermohydraulic system, and a lumped parameter modelling approach, to extract state space models. In this methodology each physical domain of the thermohydraulic system is represented by a linear graph. These linear graphs capture the interaction between all components within and across energy domains – hydraulic, thermal and mechanical. These linear graphs are analysed using a graph-theoretic approach to derive reduced order state space models. These models capture the dominant dynamics of the thermohydraulic system and are ideal for control system design purposes. The proposed state space model extraction method is demonstrated by considering a U-tube system. A non-linear state space model is extracted representing both the hydraulic and thermal domain dynamics of the system. The simulated state space model is compared with a Flownex ® model of the U-tube. Flownex ® is a validated systems thermal-fluid simulation software package. - Highlights: • A state space model extraction methodology based on graph-theoretic concepts. • An energy-based approach to consider multi-domain systems in a common framework. • Allow extraction of transparent (white-box) state space models automatically. • Reduced order models containing only independent state

  10. Sequence complexity and work extraction

    International Nuclear Information System (INIS)

    Merhav, Neri

    2015-01-01

    We consider a simplified version of a solvable model by Mandal and Jarzynski, which constructively demonstrates the interplay between work extraction and the increase of the Shannon entropy of an information reservoir which is in contact with a physical system. We extend Mandal and Jarzynski’s main findings in several directions: first, we allow sequences of correlated bits rather than just independent bits. Secondly, at least for the case of binary information, we show that, in fact, the Shannon entropy is only one measure of complexity of the information that must increase in order for work to be extracted. The extracted work can also be upper bounded in terms of the increase in other quantities that measure complexity, like the predictability of future bits from past ones. Third, we provide an extension to the case of non-binary information (i.e. a larger alphabet), and finally, we extend the scope to the case where the incoming bits (before the interaction) form an individual sequence, rather than a random one. In this case, the entropy before the interaction can be replaced by the Lempel–Ziv (LZ) complexity of the incoming sequence, a fact that gives rise to an entropic meaning of the LZ complexity, not only in information theory, but also in physics. (paper)

  11. Accurate facade feature extraction method for buildings from three-dimensional point cloud data considering structural information

    Science.gov (United States)

    Wang, Yongzhi; Ma, Yuqing; Zhu, A.-xing; Zhao, Hui; Liao, Lixia

    2018-05-01

    Facade features represent segmentations of building surfaces and can serve as a building framework. Extracting facade features from three-dimensional (3D) point cloud data (3D PCD) is an efficient method for 3D building modeling. By combining the advantages of 3D PCD and two-dimensional optical images, this study describes the creation of a highly accurate building facade feature extraction method from 3D PCD with a focus on structural information. The new extraction method involves three major steps: image feature extraction, exploration of the mapping method between the image features and 3D PCD, and optimization of the initial 3D PCD facade features considering structural information. Results show that the new method can extract the 3D PCD facade features of buildings more accurately and continuously. The new method is validated using a case study. In addition, the effectiveness of the new method is demonstrated by comparing it with the range image-extraction method and the optical image-extraction method in the absence of structural information. The 3D PCD facade features extracted by the new method can be applied in many fields, such as 3D building modeling and building information modeling.

  12. 3D refraction correction and extraction of clinical parameters from spectral domain optical coherence tomography of the cornea.

    Science.gov (United States)

    Zhao, Mingtao; Kuo, Anthony N; Izatt, Joseph A

    2010-04-26

    Capable of three-dimensional imaging of the cornea with micrometer-scale resolution, spectral domain-optical coherence tomography (SDOCT) offers potential advantages over Placido ring and Scheimpflug photography based systems for accurate extraction of quantitative keratometric parameters. In this work, an SDOCT scanning protocol and motion correction algorithm were implemented to minimize the effects of patient motion during data acquisition. Procedures are described for correction of image data artifacts resulting from 3D refraction of SDOCT light in the cornea and from non-idealities of the scanning system geometry performed as a pre-requisite for accurate parameter extraction. Zernike polynomial 3D reconstruction and a recursive half searching algorithm (RHSA) were implemented to extract clinical keratometric parameters including anterior and posterior radii of curvature, central cornea optical power, central corneal thickness, and thickness maps of the cornea. Accuracy and repeatability of the extracted parameters obtained using a commercial 859nm SDOCT retinal imaging system with a corneal adapter were assessed using a rigid gas permeable (RGP) contact lens as a phantom target. Extraction of these parameters was performed in vivo in 3 patients and compared to commercial Placido topography and Scheimpflug photography systems. The repeatability of SDOCT central corneal power measured in vivo was 0.18 Diopters, and the difference observed between the systems averaged 0.1 Diopters between SDOCT and Scheimpflug photography, and 0.6 Diopters between SDOCT and Placido topography.

  13. Information retrieval and terminology extraction in online resources for patients with diabetes.

    Science.gov (United States)

    Seljan, Sanja; Baretić, Maja; Kucis, Vlasta

    2014-06-01

    Terminology use, as a mean for information retrieval or document indexing, plays an important role in health literacy. Specific types of users, i.e. patients with diabetes need access to various online resources (on foreign and/or native language) searching for information on self-education of basic diabetic knowledge, on self-care activities regarding importance of dietetic food, medications, physical exercises and on self-management of insulin pumps. Automatic extraction of corpus-based terminology from online texts, manuals or professional papers, can help in building terminology lists or list of "browsing phrases" useful in information retrieval or in document indexing. Specific terminology lists represent an intermediate step between free text search and controlled vocabulary, between user's demands and existing online resources in native and foreign language. The research aiming to detect the role of terminology in online resources, is conducted on English and Croatian manuals and Croatian online texts, and divided into three interrelated parts: i) comparison of professional and popular terminology use ii) evaluation of automatic statistically-based terminology extraction on English and Croatian texts iii) comparison and evaluation of extracted terminology performed on English manual using statistical and hybrid approaches. Extracted terminology candidates are evaluated by comparison with three types of reference lists: list created by professional medical person, list of highly professional vocabulary contained in MeSH and list created by non-medical persons, made as intersection of 15 lists. Results report on use of popular and professional terminology in online diabetes resources, on evaluation of automatically extracted terminology candidates in English and Croatian texts and on comparison of statistical and hybrid extraction methods in English text. Evaluation of automatic and semi-automatic terminology extraction methods is performed by recall

  14. The nuclear localization of low risk HPV11 E7 protein mediated by its zinc binding domain is independent of nuclear import receptors

    International Nuclear Information System (INIS)

    Piccioli, Zachary; McKee, Courtney H.; Leszczynski, Anna; Onder, Zeynep; Hannah, Erin C.; Mamoor, Shahan; Crosby, Lauren; Moroianu, Junona

    2010-01-01

    We investigated the nuclear import of low risk HPV11 E7 protein using 1) transfection assays in HeLa cells with EGFP fusion plasmids containing 11E7 and its domains and 2) nuclear import assays in digitonin-permeabilized HeLa cells with GST fusion proteins containing 11E7 and its domains. The EGFP-11E7 and EGFP-11cE7 39-98 localized mostly to the nucleus. The GST-11E7 and GST-11cE7 39-98 were imported into the nuclei in the presence of either Ran-GDP or RanG19V-GTP mutant and in the absence of nuclear import receptors. This suggests that 11E7 enters the nucleus via a Ran-dependent pathway, independent of nuclear import receptors, mediated by a nuclear localization signal located in its C-terminal domain (cNLS). This cNLS contains the zinc binding domain consisting of two copies of Cys-X-X-Cys motif. Mutagenesis of Cys residues in these motifs changed the localization of the EGFP-11cE7/-11E7 mutants to cytoplasmic, suggesting that the zinc binding domain is essential for nuclear localization of 11E7.

  15. [Extraction of buildings three-dimensional information from high-resolution satellite imagery based on Barista software].

    Science.gov (United States)

    Zhang, Pei-feng; Hu, Yuan-man; He, Hong-shi

    2010-05-01

    The demand for accurate and up-to-date spatial information of urban buildings is becoming more and more important for urban planning, environmental protection, and other vocations. Today's commercial high-resolution satellite imagery offers the potential to extract the three-dimensional information of urban buildings. This paper extracted the three-dimensional information of urban buildings from QuickBird imagery, and validated the precision of the extraction based on Barista software. It was shown that the extraction of three-dimensional information of the buildings from high-resolution satellite imagery based on Barista software had the advantages of low professional level demand, powerful universality, simple operation, and high precision. One pixel level of point positioning and height determination accuracy could be achieved if the digital elevation model (DEM) and sensor orientation model had higher precision and the off-Nadir View Angle was relatively perfect.

  16. Information Extraction of High Resolution Remote Sensing Images Based on the Calculation of Optimal Segmentation Parameters

    Science.gov (United States)

    Zhu, Hongchun; Cai, Lijie; Liu, Haiying; Huang, Wei

    2016-01-01

    Multi-scale image segmentation and the selection of optimal segmentation parameters are the key processes in the object-oriented information extraction of high-resolution remote sensing images. The accuracy of remote sensing special subject information depends on this extraction. On the basis of WorldView-2 high-resolution data, the optimal segmentation parameters methodof object-oriented image segmentation and high-resolution image information extraction, the following processes were conducted in this study. Firstly, the best combination of the bands and weights was determined for the information extraction of high-resolution remote sensing image. An improved weighted mean-variance method was proposed andused to calculatethe optimal segmentation scale. Thereafter, the best shape factor parameter and compact factor parameters were computed with the use of the control variables and the combination of the heterogeneity and homogeneity indexes. Different types of image segmentation parameters were obtained according to the surface features. The high-resolution remote sensing images were multi-scale segmented with the optimal segmentation parameters. Ahierarchical network structure was established by setting the information extraction rules to achieve object-oriented information extraction. This study presents an effective and practical method that can explain expert input judgment by reproducible quantitative measurements. Furthermore the results of this procedure may be incorporated into a classification scheme. PMID:27362762

  17. Language-Agnostic Relation Extraction from Abstracts in Wikis

    Directory of Open Access Journals (Sweden)

    Nicolas Heist

    2018-03-01

    Full Text Available Large-scale knowledge graphs, such as DBpedia, Wikidata, or YAGO, can be enhanced by relation extraction from text, using the data in the knowledge graph as training data, i.e., using distant supervision. While most existing approaches use language-specific methods (usually for English, we present a language-agnostic approach that exploits background knowledge from the graph instead of language-specific techniques and builds machine learning models only from language-independent features. We demonstrate the extraction of relations from Wikipedia abstracts, using the twelve largest language editions of Wikipedia. From those, we can extract 1.6 M new relations in DBpedia at a level of precision of 95%, using a RandomForest classifier trained only on language-independent features. We furthermore investigate the similarity of models for different languages and show an exemplary geographical breakdown of the information extracted. In a second series of experiments, we show how the approach can be transferred to DBkWik, a knowledge graph extracted from thousands of Wikis. We discuss the challenges and first results of extracting relations from a larger set of Wikis, using a less formalized knowledge graph.

  18. Independent measuring station for the GNEJS neutron spectrometer information and measurement system

    International Nuclear Information System (INIS)

    Gorokhov, I.S.; Laptev, A.B.; Marchenkov, V.V.; Tubol'tsev, Yu.V.; Fokin, E.Yu.; Shcherbakov, O.A.

    1984-01-01

    An independent measuring station (IMS) being a part of the information and measuring system of the neutron time-of-flight spectrometer is described. IMS represents the time (time-to-amplitude or time-to-angular) module analyzer assembled on the base of one or several CAMAC crates. The station permits to perform time measurements as well as the time measurements with amplitude weighing in an independent mode with subseqUent data transmission the central PDP-11/05 computer. The station processor unit is an independent controller. The software consists of the programming langUage of the controller from local terminal and programs ensuring measuring and shaping the spectra with their output on TV display devic. communication with the computer and dialogue with the operator

  19. Towards A Model Of Knowledge Extraction Of Text Mining For Palliative Care Patients In Panama.

    Directory of Open Access Journals (Sweden)

    Denis Cedeno Moreno

    2015-08-01

    Full Text Available Solutions using information technology is an innovative way to manage the information hospice patients in hospitals in Panama. The application of techniques of text mining for the domain of medicine especially information from electronic health records of patients in palliative care is one of the most recent and promising research areas for the analysis of textual data. Text mining is based on new knowledge extraction from unstructured natural language data. We may also create ontologies to describe the terminology and knowledge in a given domain. In an ontology conceptualization of a domain that may be general or specific formalized. Knowledge can be used for decision making by health specialists or can help in research topics for improving the health system.

  20. WW domains of the yes-kinase-associated-protein (YAP transcriptional regulator behave as independent units with different binding preferences for PPxY motif-containing ligands.

    Directory of Open Access Journals (Sweden)

    Manuel Iglesias-Bexiga

    Full Text Available YAP is a WW domain-containing effector of the Hippo tumor suppressor pathway, and the object of heightened interest as a potent oncogene and stemness factor. YAP has two major isoforms that differ in the number of WW domains they harbor. Elucidating the degree of co-operation between these WW domains is important for a full understanding of the molecular function of YAP. We present here a detailed biophysical study of the structural stability and binding properties of the two YAP WW domains aimed at investigating the relationship between both domains in terms of structural stability and partner recognition. We have carried out a calorimetric study of the structural stability of the two YAP WW domains, both isolated and in a tandem configuration, and their interaction with a set of functionally relevant ligands derived from PTCH1 and LATS kinases. We find that the two YAP WW domains behave as independent units with different binding preferences, suggesting that the presence of the second WW domain might contribute to modulate target recognition between the two YAP isoforms. Analysis of structural models and phage-display studies indicate that electrostatic interactions play a critical role in binding specificity. Together, these results are relevant to understand of YAP function and open the door to the design of highly specific ligands of interest to delineate the functional role of each WW domain in YAP signaling.

  1. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories.

    Science.gov (United States)

    Yang, Wei; Ai, Tinghua; Lu, Wei

    2018-04-19

    Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT). First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS) traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction) by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  2. A Method for Extracting Road Boundary Information from Crowdsourcing Vehicle GPS Trajectories

    Directory of Open Access Journals (Sweden)

    Wei Yang

    2018-04-01

    Full Text Available Crowdsourcing trajectory data is an important approach for accessing and updating road information. In this paper, we present a novel approach for extracting road boundary information from crowdsourcing vehicle traces based on Delaunay triangulation (DT. First, an optimization and interpolation method is proposed to filter abnormal trace segments from raw global positioning system (GPS traces and interpolate the optimization segments adaptively to ensure there are enough tracking points. Second, constructing the DT and the Voronoi diagram within interpolated tracking lines to calculate road boundary descriptors using the area of Voronoi cell and the length of triangle edge. Then, the road boundary detection model is established integrating the boundary descriptors and trajectory movement features (e.g., direction by DT. Third, using the boundary detection model to detect road boundary from the DT constructed by trajectory lines, and a regional growing method based on seed polygons is proposed to extract the road boundary. Experiments were conducted using the GPS traces of taxis in Beijing, China, and the results show that the proposed method is suitable for extracting the road boundary from low-frequency GPS traces, multi-type road structures, and different time intervals. Compared with two existing methods, the automatically extracted boundary information was proved to be of higher quality.

  3. Environmental monitoring and independent analysis: the experience of the local information commission in Cadarache

    International Nuclear Information System (INIS)

    Foucher, M.

    2010-01-01

    The 2006 Act on 'Open Access to Information and Nuclear Safety' provides free access to nuclear installations data and submits nuclear facilities to an independent administrative authority, the ASN (French Nuclear Safety Authority). A local information committee (CLI - Commission locale d'informations) has been set up for each nuclear facility. It includes all the parties involved and may commission expert studies and environmental measurement tests from independent laboratories. Prompted by environmental organizations, the Cadarache CLI commissioned a CRIIRAD (Commission de Recherche et d'Informations Independantes sur la Radioactivite) fact-finding survey in 2008-2009 to evaluate the environmental impact of the CEA/Cadarache installations since they were created. This study points to a moderate present impact, though a lack of data makes it impossible to reconstruct environmental impact prior to 1979. A set of guidelines are put forward to assess the past impact in greater detail and to lessen the current impact even further. Following this study, additional steps to minimize this impact will be taken by CEA/Cadarache inside the facilities, and by the relevant LCI outside the installations. Developing independent expert skills is a major challenge for CLIs to gain better knowledge of nuclear activities and improve their local management. (author)

  4. A cascade of classifiers for extracting medication information from discharge summaries

    Directory of Open Access Journals (Sweden)

    Halgrim Scott

    2011-07-01

    Full Text Available Abstract Background Extracting medication information from clinical records has many potential applications, and recently published research, systems, and competitions reflect an interest therein. Much of the early extraction work involved rules and lexicons, but more recently machine learning has been applied to the task. Methods We present a hybrid system consisting of two parts. The first part, field detection, uses a cascade of statistical classifiers to identify medication-related named entities. The second part uses simple heuristics to link those entities into medication events. Results The system achieved performance that is comparable to other approaches to the same task. This performance is further improved by adding features that reference external medication name lists. Conclusions This study demonstrates that our hybrid approach outperforms purely statistical or rule-based systems. The study also shows that a cascade of classifiers works better than a single classifier in extracting medication information. The system is available as is upon request from the first author.

  5. Informing General CSCW Product Development through Cooperative Design in Specific Work Domains

    DEFF Research Database (Denmark)

    Grønbæk, Kaj; Mogensen, Preben Holst

    1997-01-01

    sharing of materials in the engineering domain. In our project, a single engineering company (Great Belt Link Ltd.) was chosen as the user organization. The paper summarizes the process from observational studies, over a future workshop and cooperative prototyping activities, to a pilot installation. We...... describe how these activities informed the general hypermedia framework and application design. Use scenarios and prototypes with example data from the users‘ daily work were used as sources both to trigger design ideas and new insights regarding work practice. Common participants in specific activities...... and general development activities supported transfer of work domain knowledge into general features of the product being developed. Mutual challenging characterized the interaction between specific cooperative analysis and design activities and general development activities. Prototypes, scenarios, materials...

  6. Bilateral transaction bargaining between independent utilities under incomplete information

    International Nuclear Information System (INIS)

    David, A. K.; Wen, F. S.

    2001-01-01

    A new approach to designing bilateral power transaction bargaining models between two independent utilities in a deregulated electricity market is proposed. In the paper it is assumed that each utility (a seller or a buyer) knows its own operating costs but does not know those of its opponent. The bilateral power transaction problem is then considered as non-cooperative bargaining under incomplete information. Each participant develops its own bargaining strategy based on estimates of the opponent's operating costs and bargaining strategy. Two bargaining models are developed and examples are employed for demonstration. (Author)

  7. An Effective Fault Feature Extraction Method for Gas Turbine Generator System Diagnosis

    Directory of Open Access Journals (Sweden)

    Jian-Hua Zhong

    2016-01-01

    Full Text Available Fault diagnosis is very important to maintain the operation of a gas turbine generator system (GTGS in power plants, where any abnormal situations will interrupt the electricity supply. The fault diagnosis of the GTGS faces the main challenge that the acquired data, vibration or sound signals, contain a great deal of redundant information which extends the fault identification time and degrades the diagnostic accuracy. To improve the diagnostic performance in the GTGS, an effective fault feature extraction framework is proposed to solve the problem of the signal disorder and redundant information in the acquired signal. The proposed framework combines feature extraction with a general machine learning method, support vector machine (SVM, to implement an intelligent fault diagnosis. The feature extraction method adopts wavelet packet transform and time-domain statistical features to extract the features of faults from the vibration signal. To further reduce the redundant information in extracted features, kernel principal component analysis is applied in this study. Experimental results indicate that the proposed feature extracted technique is an effective method to extract the useful features of faults, resulting in improvement of the performance of fault diagnosis for the GTGS.

  8. Rapid automatic keyword extraction for information retrieval and analysis

    Science.gov (United States)

    Rose, Stuart J [Richland, WA; Cowley,; E, Wendy [Richland, WA; Crow, Vernon L [Richland, WA; Cramer, Nicholas O [Richland, WA

    2012-03-06

    Methods and systems for rapid automatic keyword extraction for information retrieval and analysis. Embodiments can include parsing words in an individual document by delimiters, stop words, or both in order to identify candidate keywords. Word scores for each word within the candidate keywords are then calculated based on a function of co-occurrence degree, co-occurrence frequency, or both. Based on a function of the word scores for words within the candidate keyword, a keyword score is calculated for each of the candidate keywords. A portion of the candidate keywords are then extracted as keywords based, at least in part, on the candidate keywords having the highest keyword scores.

  9. Evaluating, Comparing, and Interpreting Protein Domain Hierarchies

    Science.gov (United States)

    2014-01-01

    Abstract Arranging protein domain sequences hierarchically into evolutionarily divergent subgroups is important for investigating evolutionary history, for speeding up web-based similarity searches, for identifying sequence determinants of protein function, and for genome annotation. However, whether or not a particular hierarchy is optimal is often unclear, and independently constructed hierarchies for the same domain can often differ significantly. This article describes methods for statistically evaluating specific aspects of a hierarchy, for probing the criteria underlying its construction and for direct comparisons between hierarchies. Information theoretical notions are used to quantify the contributions of specific hierarchical features to the underlying statistical model. Such features include subhierarchies, sequence subgroups, individual sequences, and subgroup-associated signature patterns. Underlying properties are graphically displayed in plots of each specific feature's contributions, in heat maps of pattern residue conservation, in “contrast alignments,” and through cross-mapping of subgroups between hierarchies. Together, these approaches provide a deeper understanding of protein domain functional divergence, reveal uncertainties caused by inconsistent patterns of sequence conservation, and help resolve conflicts between competing hierarchies. PMID:24559108

  10. Activatory and inhibitory Fcγ receptors augment rituximab-mediated internalization of CD20 independent of signaling via the cytoplasmic domain.

    Science.gov (United States)

    Vaughan, Andrew T; Chan, Claude H T; Klein, Christian; Glennie, Martin J; Beers, Stephen A; Cragg, Mark S

    2015-02-27

    Type I anti-CD20 mAb such as rituximab and ofatumumab engage with the inhibitory FcγR, FcγRIIb on the surface of B cells, resulting in immunoreceptor tyrosine-based inhibitory motif (ITIM) phosphorylation. Internalization of the CD20·mAb·FcγRIIb complex follows, the rate of which correlates with FcγRIIb expression. In contrast, although type II anti-CD20 mAb such as tositumomab and obinutuzumab also interact with and activate FcγRIIb, this interaction fails to augment the rate of CD20·mAb internalization, raising the question of whether ITIM phosphorylation plays any role in this process. We have assessed the molecular requirements for the internalization process and demonstrate that in contrast to internalization of IgG immune complexes, FcγRIIb-augmented internalization of rituximab-ligated CD20 occurs independently of the FcγRIIb ITIM, indicating that signaling downstream of FcγRIIb is not required. In transfected cells, activatory FcγRI, FcγRIIa, and FcγRIIIa augmented internalization of rituximab-ligated CD20 in a similar manner. However, FcγRIIa mediated a slower rate of internalization than cells expressing equivalent levels of the highly homologous FcγRIIb. The difference was maintained in cells expressing FcγRIIa and FcγRIIb lacking cytoplasmic domains and in which the transmembrane domains had been exchanged. This difference may be due to increased degradation of FcγRIIa, which traffics to lysosomes independently of rituximab. We conclude that the cytoplasmic domain of FcγR is not required for promoting internalization of rituximab-ligated CD20. Instead, we propose that FcγR provides a structural role in augmenting endocytosis that differs from that employed during the endocytosis of immune complexes. © 2015 by The American Society for Biochemistry and Molecular Biology, Inc.

  11. Identification of hidden relationships from the coupling of hydrophobic cluster analysis and domain architecture information.

    Science.gov (United States)

    Faure, Guilhem; Callebaut, Isabelle

    2013-07-15

    Describing domain architecture is a critical step in the functional characterization of proteins. However, some orphan domains do not match any profile stored in dedicated domain databases and are thereby difficult to analyze. We present here an original novel approach, called TREMOLO-HCA, for the analysis of orphan domain sequences and inspired from our experience in the use of Hydrophobic Cluster Analysis (HCA). Hidden relationships between protein sequences can be more easily identified from the PSI-BLAST results, using information on domain architecture, HCA plots and the conservation degree of amino acids that may participate in the protein core. This can lead to reveal remote relationships with known families of domains, as illustrated here with the identification of a hidden Tudor tandem in the human BAHCC1 protein and a hidden ET domain in the Saccharomyces cerevisiae Taf14p and human AF9 proteins. The results obtained in such a way are consistent with those provided by HHPRED, based on pairwise comparisons of HHMs. Our approach can, however, be applied even in absence of domain profiles or known 3D structures for the identification of novel families of domains. It can also be used in a reverse way for refining domain profiles, by starting from known protein domain families and identifying highly divergent members, hitherto considered as orphan. We provide a possible integration of this approach in an open TREMOLO-HCA package, which is fully implemented in python v2.7 and is available on request. Instructions are available at http://www.impmc.upmc.fr/∼callebau/tremolohca.html. isabelle.callebaut@impmc.upmc.fr Supplementary Data are available at Bioinformatics online.

  12. Predicting and analyzing DNA-binding domains using a systematic approach to identifying a set of informative physicochemical and biochemical properties

    Science.gov (United States)

    2011-01-01

    Background Existing methods of predicting DNA-binding proteins used valuable features of physicochemical properties to design support vector machine (SVM) based classifiers. Generally, selection of physicochemical properties and determination of their corresponding feature vectors rely mainly on known properties of binding mechanism and experience of designers. However, there exists a troublesome problem for designers that some different physicochemical properties have similar vectors of representing 20 amino acids and some closely related physicochemical properties have dissimilar vectors. Results This study proposes a systematic approach (named Auto-IDPCPs) to automatically identify a set of physicochemical and biochemical properties in the AAindex database to design SVM-based classifiers for predicting and analyzing DNA-binding domains/proteins. Auto-IDPCPs consists of 1) clustering 531 amino acid indices in AAindex into 20 clusters using a fuzzy c-means algorithm, 2) utilizing an efficient genetic algorithm based optimization method IBCGA to select an informative feature set of size m to represent sequences, and 3) analyzing the selected features to identify related physicochemical properties which may affect the binding mechanism of DNA-binding domains/proteins. The proposed Auto-IDPCPs identified m=22 features of properties belonging to five clusters for predicting DNA-binding domains with a five-fold cross-validation accuracy of 87.12%, which is promising compared with the accuracy of 86.62% of the existing method PSSM-400. For predicting DNA-binding sequences, the accuracy of 75.50% was obtained using m=28 features, where PSSM-400 has an accuracy of 74.22%. Auto-IDPCPs and PSSM-400 have accuracies of 80.73% and 82.81%, respectively, applied to an independent test data set of DNA-binding domains. Some typical physicochemical properties discovered are hydrophobicity, secondary structure, charge, solvent accessibility, polarity, flexibility, normalized Van Der

  13. XIAP BIR domain suppresses miR-200a expression and subsequently promotes EGFR protein translation and anchorage-independent growth of bladder cancer cell

    Directory of Open Access Journals (Sweden)

    Chao Huang

    2017-01-01

    Full Text Available Abstract Background The X-linked inhibitor of apoptosis protein (XIAP is a well-known potent apoptosis suppressor and also participates in cancer cell biological behaviors, therefore attracting great attentions as a potential antineoplastic therapeutic target for past years. Anti-IAP therapy is reported to be closely related to epidermal growth factor receptor (EGFR expression level. However, whether and how XIAP modulates EGFR expression remains largely unknown. Methods Human XIAP was knockdown with short-hairpin RNA in two different bladder cancer cell lines, T24T and UMUC3. Two XIAP mutants, XIAP ∆BIR (deletion of N-terminal three BIR domains and XIAP ∆RING (deletion of C-terminal RING domain and keeping the function of BIR domains, were generated to determine which domain is involved in regulating EGFR. Results We found here that lacking of XIAP expression resulted in a remarkable suppression of EGFR expression, consequently leading to the deficiency of anchorage-independent cell growth. Further study demonstrated that BIR domain of XIAP was crucial for regulating the EGFR translation by suppressing the transcription and expression of miR-200a. Mechanistic studies indicated that BIR domain activated the protein phosphatase 2 (PP2A activity by decreasing the phosphorylation of PP2A at Tyr307 in its catalytic subunit, PP2A-C. Such activated PP2A prevented the deviant phosphorylation and activation of MAPK kinases/MAPKs, their downstream effector c-Jun, and in turn inhibiting transcription of c-Jun-regulated the miR-200a. Conclusions Our study uncovered a novel function of BIR domain of XIAP in regulating the EGFR translation, providing significant insight into the understanding of the XIAP overexpression in the cancer development and progression, further offering a new theoretical support for using XIAP BIR domain and EGFR as targets for cancer therapy.

  14. Robust Vehicle and Traffic Information Extraction for Highway Surveillance

    Directory of Open Access Journals (Sweden)

    Yeh Chia-Hung

    2005-01-01

    Full Text Available A robust vision-based traffic monitoring system for vehicle and traffic information extraction is developed in this research. It is challenging to maintain detection robustness at all time for a highway surveillance system. There are three major problems in detecting and tracking a vehicle: (1 the moving cast shadow effect, (2 the occlusion effect, and (3 nighttime detection. For moving cast shadow elimination, a 2D joint vehicle-shadow model is employed. For occlusion detection, a multiple-camera system is used to detect occlusion so as to extract the exact location of each vehicle. For vehicle nighttime detection, a rear-view monitoring technique is proposed to maintain tracking and detection accuracy. Furthermore, we propose a method to improve the accuracy of background extraction, which usually serves as the first step in any vehicle detection processing. Experimental results are given to demonstrate that the proposed techniques are effective and efficient for vision-based highway surveillance.

  15. Independent and arbitrary generation of spots in the 3D space domain with computer generated holograms written on a phase-only liquid crystal spatial light modulator

    International Nuclear Information System (INIS)

    Wang, Dong; Zhang, Jian; Xia, Yang; Wang, Hao

    2012-01-01

    An improved multiple independent iterative plane algorithm, based on a projection optimization idea, is proposed for the independent and arbitrary generation of one spot or multiple spots in a speckle-suppressed 3D work-area. Details of the mathematical expressions of the algorithm are given to theoretically show how it is improved for 3D spot generation. Both simulations and experiments are conducted to investigate the performance of the algorithm for independent and arbitrary 3D spot generation in several different cases. Simulation results agree well with experimental results, which validates the effectiveness of the algorithm proposed. Several additional experiments are demonstrated for fast and independent generation of four or more spots in the 3D space domain, which confirms the capabilities and practicalities of the algorithm further. (paper)

  16. Calibration of TAMA300 in time domain

    International Nuclear Information System (INIS)

    Telada, Souichi; Tatsumi, Daisuke; Akutsu, Tomomi; Ando, Masaki; Kanda, Nobuyuki

    2005-01-01

    We could reconstruct the strain of gravitational wave signals from acquired data in the time domain by using the infinite impulse response filter technique in TAMA300. We would like to analyse the waveform in the time domain for burst-like signal, merger phase waveform of binary neutron stars, and so on. We established the way to make a continuous time-series gravitational wave strain signal. We compared the time-domain reconstruction with the Fourier-space reconstruction. Both coincided within 3% in the observation range. We could also produce the voltage signal which would be recorded by the data-acquisition system from a simulated gravitational wave. This is useful for some analyses of simulations and signal injections. We could extract the waveform of the hardware injection signal in an observational run in the time domain. The extracted waveform was similar to the injection signal

  17. Domain decomposition methods for solving an image problem

    Energy Technology Data Exchange (ETDEWEB)

    Tsui, W.K.; Tong, C.S. [Hong Kong Baptist College (Hong Kong)

    1994-12-31

    The domain decomposition method is a technique to break up a problem so that ensuing sub-problems can be solved on a parallel computer. In order to improve the convergence rate of the capacitance systems, pre-conditioned conjugate gradient methods are commonly used. In the last decade, most of the efficient preconditioners are based on elliptic partial differential equations which are particularly useful for solving elliptic partial differential equations. In this paper, the authors apply the so called covering preconditioner, which is based on the information of the operator under investigation. Therefore, it is good for various kinds of applications, specifically, they shall apply the preconditioned domain decomposition method for solving an image restoration problem. The image restoration problem is to extract an original image which has been degraded by a known convolution process and additive Gaussian noise.

  18. Research of building information extraction and evaluation based on high-resolution remote-sensing imagery

    Science.gov (United States)

    Cao, Qiong; Gu, Lingjia; Ren, Ruizhi; Wang, Lang

    2016-09-01

    Building extraction currently is important in the application of high-resolution remote sensing imagery. At present, quite a few algorithms are available for detecting building information, however, most of them still have some obvious disadvantages, such as the ignorance of spectral information, the contradiction between extraction rate and extraction accuracy. The purpose of this research is to develop an effective method to detect building information for Chinese GF-1 data. Firstly, the image preprocessing technique is used to normalize the image and image enhancement is used to highlight the useful information in the image. Secondly, multi-spectral information is analyzed. Subsequently, an improved morphological building index (IMBI) based on remote sensing imagery is proposed to get the candidate building objects. Furthermore, in order to refine building objects and further remove false objects, the post-processing (e.g., the shape features, the vegetation index and the water index) is employed. To validate the effectiveness of the proposed algorithm, the omission errors (OE), commission errors (CE), the overall accuracy (OA) and Kappa are used at final. The proposed method can not only effectively use spectral information and other basic features, but also avoid extracting excessive interference details from high-resolution remote sensing images. Compared to the original MBI algorithm, the proposed method reduces the OE by 33.14% .At the same time, the Kappa increase by 16.09%. In experiments, IMBI achieved satisfactory results and outperformed other algorithms in terms of both accuracies and visual inspection

  19. Domain Differences in the Weights of Perceptual and Conceptual Information in Children's Categorization

    Science.gov (United States)

    Diesendruck, Gil; Peretz, Shimon

    2013-01-01

    Visual appearance is one of the main cues children rely on when categorizing novel objects. In 3 studies, testing 128 3-year-olds and 192 5-year-olds, we investigated how various kinds of information may differentially lead children to overlook visual appearance in their categorization decisions across domains. Participants saw novel animals or…

  20. Extracting Social Networks and Contact Information From Email and the Web

    National Research Council Canada - National Science Library

    Culotta, Aron; Bekkerman, Ron; McCallum, Andrew

    2005-01-01

    ...-suited for such information extraction tasks. By recursively calling itself on new people discovered on the Web, the system builds a social network with multiple degrees of separation from the user...

  1. Information Systems Foundations - Karl Popper's third world

    Directory of Open Access Journals (Sweden)

    Craig McDonald

    2002-11-01

    Full Text Available The various information professions have matured separately over the years, developing different bodies of theory and practice to meet their evolving purposes and needs. A problem arises however, when different information professions address the same knowledge domain and there is no explicit correspondence between the conceptual structures embedded independently in each. In this situation, a knowledge worker involved in the domain is faced with a range of possibly incompatible structures presented in different forms by a range of information professions. This is a common problem that is being exacerbated by the explosion in information production and the widening access to information distribution technology, notably the World Wide Web. Information Systems now need to combine the best of what the information professions in a domain have to offer the domain’s knowledge workers. This paper examines the problem by exploring one of the foundations of the information disciplines - Karl Popper’s 3 Worlds theory, applying it to a case study and suggesting that the Information Systems discipline alone has a sufficiently broad agenda to integrate the various Informatics themes needed to support today’s knowledge workers.

  2. Information Extraction of High-Resolution Remotely Sensed Image Based on Multiresolution Segmentation

    Directory of Open Access Journals (Sweden)

    Peng Shao

    2014-08-01

    Full Text Available The principle of multiresolution segmentation was represented in detail in this study, and the canny algorithm was applied for edge-detection of a remotely sensed image based on this principle. The target image was divided into regions based on object-oriented multiresolution segmentation and edge-detection. Furthermore, object hierarchy was created, and a series of features (water bodies, vegetation, roads, residential areas, bare land and other information were extracted by the spectral and geometrical features. The results indicate that the edge-detection has a positive effect on multiresolution segmentation, and overall accuracy of information extraction reaches to 94.6% by the confusion matrix.

  3. BioCause: Annotating and analysing causality in the biomedical domain.

    Science.gov (United States)

    Mihăilă, Claudiu; Ohta, Tomoko; Pyysalo, Sampo; Ananiadou, Sophia

    2013-01-16

    Biomedical corpora annotated with event-level information represent an important resource for domain-specific information extraction (IE) systems. However, bio-event annotation alone cannot cater for all the needs of biologists. Unlike work on relation and event extraction, most of which focusses on specific events and named entities, we aim to build a comprehensive resource, covering all statements of causal association present in discourse. Causality lies at the heart of biomedical knowledge, such as diagnosis, pathology or systems biology, and, thus, automatic causality recognition can greatly reduce the human workload by suggesting possible causal connections and aiding in the curation of pathway models. A biomedical text corpus annotated with such relations is, hence, crucial for developing and evaluating biomedical text mining. We have defined an annotation scheme for enriching biomedical domain corpora with causality relations. This schema has subsequently been used to annotate 851 causal relations to form BioCause, a collection of 19 open-access full-text biomedical journal articles belonging to the subdomain of infectious diseases. These documents have been pre-annotated with named entity and event information in the context of previous shared tasks. We report an inter-annotator agreement rate of over 60% for triggers and of over 80% for arguments using an exact match constraint. These increase significantly using a relaxed match setting. Moreover, we analyse and describe the causality relations in BioCause from various points of view. This information can then be leveraged for the training of automatic causality detection systems. Augmenting named entity and event annotations with information about causal discourse relations could benefit the development of more sophisticated IE systems. These will further influence the development of multiple tasks, such as enabling textual inference to detect entailments, discovering new facts and providing new

  4. Overview of ImageCLEF 2017: information extraction from images

    OpenAIRE

    Ionescu, Bogdan; Müller, Henning; Villegas, Mauricio; Arenas, Helbert; Boato, Giulia; Dang Nguyen, Duc Tien; Dicente Cid, Yashin; Eickhoff, Carsten; Seco de Herrera, Alba G.; Gurrin, Cathal; Islam, Bayzidul; Kovalev, Vassili; Liauchuk, Vitali; Mothe, Josiane; Piras, Luca

    2017-01-01

    This paper presents an overview of the ImageCLEF 2017 evaluation campaign, an event that was organized as part of the CLEF (Conference and Labs of the Evaluation Forum) labs 2017. ImageCLEF is an ongoing initiative (started in 2003) that promotes the evaluation of technologies for annotation, indexing and retrieval for providing information access to collections of images in various usage scenarios and domains. In 2017, the 15th edition of ImageCLEF, three main tasks were proposed and one pil...

  5. Design study on an independently-tunable-cells thermionic RF gun

    International Nuclear Information System (INIS)

    Hama, H.; Tanaka, T.; Hinode, F.; Kawai, M.

    2006-01-01

    Characteristics of a thermionic RF gun have been studied by a 3-D simulation code developed using an FDTD (Finite Difference Time Domain) method as a Maxwell's equations solver. The gun is consists of two independent power feeding cavities, so that we call it independently-tunable-cells (ITC)'-RF gun. The first cell is the cathode cell and the second one is an accelerating cell. The ITC gun can be operated at various modes of different RF-power ratio and phase between two cavities. Simulation study shows a velocity-bunching like effect may be occurred in the gun, so that the short pulse beam from the thermionic RF gun is a better candidate to produce the coherent THz synchrotron radiation. Expected bunch length with a total charge of ∼20 pC (1% energy width from the top energy) is around 200 fs (fwhm). Even the beam energy extracted from the gun is varied by which the input powers are changed, almost same shape of the longitudinal phase space can be produced by tuning the phase. (author)

  6. Extracting information from multiplex networks

    Science.gov (United States)

    Iacovacci, Jacopo; Bianconi, Ginestra

    2016-06-01

    Multiplex networks are generalized network structures that are able to describe networks in which the same set of nodes are connected by links that have different connotations. Multiplex networks are ubiquitous since they describe social, financial, engineering, and biological networks as well. Extending our ability to analyze complex networks to multiplex network structures increases greatly the level of information that is possible to extract from big data. For these reasons, characterizing the centrality of nodes in multiplex networks and finding new ways to solve challenging inference problems defined on multiplex networks are fundamental questions of network science. In this paper, we discuss the relevance of the Multiplex PageRank algorithm for measuring the centrality of nodes in multilayer networks and we characterize the utility of the recently introduced indicator function Θ ˜ S for describing their mesoscale organization and community structure. As working examples for studying these measures, we consider three multiplex network datasets coming for social science.

  7. OpenCV-Based Nanomanipulation Information Extraction and the Probe Operation in SEM

    Directory of Open Access Journals (Sweden)

    Dongjie Li

    2015-02-01

    Full Text Available Aimed at the established telenanomanipulation system, the method of extracting location information and the strategies of probe operation were studied in this paper. First, the machine learning algorithm of OpenCV was used to extract location information from SEM images. Thus nanowires and probe in SEM images can be automatically tracked and the region of interest (ROI can be marked quickly. Then the location of nanowire and probe can be extracted from the ROI. To study the probe operation strategy, the Van der Waals force between probe and a nanowire was computed; thus relevant operating parameters can be obtained. With these operating parameters, the nanowire in 3D virtual environment can be preoperated and an optimal path of the probe can be obtained. The actual probe runs automatically under the telenanomanipulation system's control. Finally, experiments were carried out to verify the above methods, and results show the designed methods have achieved the expected effect.

  8. A method for automating the extraction of specialized information from the web

    NARCIS (Netherlands)

    Lin, L.; Liotta, A.; Hippisley, A.; Hao, Y.; Liu, J.; Wang, Y.; Cheung, Y-M.; Yin, H.; Jiao, L.; Ma, j.; Jiao, Y-C.

    2005-01-01

    The World Wide Web can be viewed as a gigantic distributed database including millions of interconnected hosts some of which publish information via web servers or peer-to-peer systems. We present here a novel method for the extraction of semantically rich information from the web in a fully

  9. Chemical name extraction based on automatic training data generation and rich feature set.

    Science.gov (United States)

    Yan, Su; Spangler, W Scott; Chen, Ying

    2013-01-01

    The automation of extracting chemical names from text has significant value to biomedical and life science research. A major barrier in this task is the difficulty of getting a sizable and good quality data to train a reliable entity extraction model. Another difficulty is the selection of informative features of chemical names, since comprehensive domain knowledge on chemistry nomenclature is required. Leveraging random text generation techniques, we explore the idea of automatically creating training sets for the task of chemical name extraction. Assuming the availability of an incomplete list of chemical names, called a dictionary, we are able to generate well-controlled, random, yet realistic chemical-like training documents. We statistically analyze the construction of chemical names based on the incomplete dictionary, and propose a series of new features, without relying on any domain knowledge. Compared to state-of-the-art models learned from manually labeled data and domain knowledge, our solution shows better or comparable results in annotating real-world data with less human effort. Moreover, we report an interesting observation about the language for chemical names. That is, both the structural and semantic components of chemical names follow a Zipfian distribution, which resembles many natural languages.

  10. Informal Learning on "YouTube": Exploring Digital Literacy in Independent Online Learning

    Science.gov (United States)

    Tan, Elaine

    2013-01-01

    The focus of this paper is a project conducted in 2011, exploring the use of "YouTube" in the classroom. The project conducted a number of focus groups for which highlighted a number of issues surrounding independent informal learning environments. The questions posed by this research are concerned with what constitutes learning in these…

  11. Extraction of land cover change information from ENVISAT-ASAR data in Chengdu Plain

    Science.gov (United States)

    Xu, Wenbo; Fan, Jinlong; Huang, Jianxi; Tian, Yichen; Zhang, Yong

    2006-10-01

    Land cover data are essential to most global change research objectives, including the assessment of current environmental conditions and the simulation of future environmental scenarios that ultimately lead to public policy development. Chinese Academy of Sciences generated a nationwide land cover database in order to carry out the quantification and spatial characterization of land use/cover changes (LUCC) in 1990s. In order to improve the reliability of the database, we will update the database anytime. But it is difficult to obtain remote sensing data to extract land cover change information in large-scale. It is hard to acquire optical remote sensing data in Chengdu plain, so the objective of this research was to evaluate multitemporal ENVISAT advanced synthetic aperture radar (ASAR) data for extracting land cover change information. Based on the fieldwork and the nationwide 1:100000 land cover database, the paper assesses several land cover changes in Chengdu plain, for example: crop to buildings, forest to buildings, and forest to bare land. The results show that ENVISAT ASAR data have great potential for the applications of extracting land cover change information.

  12. Frequency domain analysis of knock images

    Science.gov (United States)

    Qi, Yunliang; He, Xin; Wang, Zhi; Wang, Jianxin

    2014-12-01

    High speed imaging-based knock analysis has mainly focused on time domain information, e.g. the spark triggered flame speed, the time when end gas auto-ignition occurs and the end gas flame speed after auto-ignition. This study presents a frequency domain analysis on the knock images recorded using a high speed camera with direct photography in a rapid compression machine (RCM). To clearly visualize the pressure wave oscillation in the combustion chamber, the images were high-pass-filtered to extract the luminosity oscillation. The luminosity spectrum was then obtained by applying fast Fourier transform (FFT) to three basic colour components (red, green and blue) of the high-pass-filtered images. Compared to the pressure spectrum, the luminosity spectra better identify the resonant modes of pressure wave oscillation. More importantly, the resonant mode shapes can be clearly visualized by reconstructing the images based on the amplitudes of luminosity spectra at the corresponding resonant frequencies, which agree well with the analytical solutions for mode shapes of gas vibration in a cylindrical cavity.

  13. A Domain Independent Framework for Extracting Linked Semantic Data from Tables

    Science.gov (United States)

    2012-01-01

    queda, J ., Ezzat, A.: A survey of current approaches for mapping of relational databases to rdf. Tech. rep., W3C (2009) 24. Salton , G., Mcgill, M.J...Once again ψ2 will assign a score to each entity which can be used to rank the entities. Thus, ψ2 = exp(w T 2 .f2(Ri, j , Ei, j )) where w2 is the weight...vector, Ei, j is the candidate entity and Ri, j is string value in column i and row j . The feature vector f2 is composed as follows: f2

  14. Independent component analysis for the extraction of reliable protein signal profiles from MALDI-TOF mass spectra.

    Science.gov (United States)

    Mantini, Dante; Petrucci, Francesca; Del Boccio, Piero; Pieragostino, Damiana; Di Nicola, Marta; Lugaresi, Alessandra; Federici, Giorgio; Sacchetta, Paolo; Di Ilio, Carmine; Urbani, Andrea

    2008-01-01

    Independent component analysis (ICA) is a signal processing technique that can be utilized to recover independent signals from a set of their linear mixtures. We propose ICA for the analysis of signals obtained from large proteomics investigations such as clinical multi-subject studies based on MALDI-TOF MS profiling. The method is validated on simulated and experimental data for demonstrating its capability of correctly extracting protein profiles from MALDI-TOF mass spectra. The comparison on peak detection with an open-source and two commercial methods shows its superior reliability in reducing the false discovery rate of protein peak masses. Moreover, the integration of ICA and statistical tests for detecting the differences in peak intensities between experimental groups allows to identify protein peaks that could be indicators of a diseased state. This data-driven approach demonstrates to be a promising tool for biomarker-discovery studies based on MALDI-TOF MS technology. The MATLAB implementation of the method described in the article and both simulated and experimental data are freely available at http://www.unich.it/proteomica/bioinf/.

  15. Towards a robust algorithm to determine topological domains from colocalization data

    Directory of Open Access Journals (Sweden)

    Alexander P. Moscalets

    2015-09-01

    Full Text Available One of the most important tasks in understanding the complex spatial organization of the genome consists in extracting information about this spatial organization, the function and structure of chromatin topological domains from existing experimental data, in particular, from genome colocalization (Hi-C matrices. Here we present an algorithm allowing to reveal the underlying hierarchical domain structure of a polymer conformation from analyzing the modularity of colocalization matrices. We also test this algorithm on several model polymer structures: equilibrium globules, random fractal globules and regular fractal (Peano conformations. We define what we call a spectrum of cluster borders, and show that these spectra behave strikingly di erently for equilibrium and fractal conformations, allowing us to suggest an additional criterion to identify fractal polymer conformations.

  16. Fault Diagnosis of Rotating Machinery Based on Multisensor Information Fusion Using SVM and Time-Domain Features

    Directory of Open Access Journals (Sweden)

    Ling-li Jiang

    2014-01-01

    Full Text Available Multisensor information fusion, when applied to fault diagnosis, the time-space scope, and the quantity of information are expanded compared to what could be acquired by a single sensor, so the diagnostic object can be described more comprehensively. This paper presents a methodology of fault diagnosis in rotating machinery using multisensor information fusion that all the features are calculated using vibration data in time domain to constitute fusional vector and the support vector machine (SVM is used for classification. The effectiveness of the presented methodology is tested by three case studies: diagnostic of faulty gear, rolling bearing, and identification of rotor crack. For each case study, the sensibilities of the features are analyzed. The results indicate that the peak factor is the most sensitive feature in the twelve time-domain features for identifying gear defect, and the mean, amplitude square, root mean square, root amplitude, and standard deviation are all sensitive for identifying gear, rolling bearing, and rotor crack defect comparatively.

  17. Evolving spectral transformations for multitemporal information extraction using evolutionary computation

    Science.gov (United States)

    Momm, Henrique; Easson, Greg

    2011-01-01

    Remote sensing plays an important role in assessing temporal changes in land features. The challenge often resides in the conversion of large quantities of raw data into actionable information in a timely and cost-effective fashion. To address this issue, research was undertaken to develop an innovative methodology integrating biologically-inspired algorithms with standard image classification algorithms to improve information extraction from multitemporal imagery. Genetic programming was used as the optimization engine to evolve feature-specific candidate solutions in the form of nonlinear mathematical expressions of the image spectral channels (spectral indices). The temporal generalization capability of the proposed system was evaluated by addressing the task of building rooftop identification from a set of images acquired at different dates in a cross-validation approach. The proposed system generates robust solutions (kappa values > 0.75 for stage 1 and > 0.4 for stage 2) despite the statistical differences between the scenes caused by land use and land cover changes coupled with variable environmental conditions, and the lack of radiometric calibration between images. Based on our results, the use of nonlinear spectral indices enhanced the spectral differences between features improving the clustering capability of standard classifiers and providing an alternative solution for multitemporal information extraction.

  18. Extraction of Graph Information Based on Image Contents and the Use of Ontology

    Science.gov (United States)

    Kanjanawattana, Sarunya; Kimura, Masaomi

    2016-01-01

    A graph is an effective form of data representation used to summarize complex information. Explicit information such as the relationship between the X- and Y-axes can be easily extracted from a graph by applying human intelligence. However, implicit knowledge such as information obtained from other related concepts in an ontology also resides in…

  19. Impact of school-based health promotion interventions aimed at different behavioral domains: a systematic review.

    Science.gov (United States)

    Lima-Serrano, Marta; Lima-Rodríguez, Joaquín S

    2014-01-01

    Given that lifestyleshave similar determinants and that school-based interventions are usually targeted at all the risks that affect adolescents, the objective of this systematic review was to summarize the characteristics and effects of school-based interventions acting on different behavioral domains of adolescent health promotion. The review process was conducted by two independent reviewers who searched PubMed, Scopus, PsycINFO, and ERIC databases for experimental or observational studies with at least two measures of results published from 2007 to 2011, given that the research information available doubles every 5 years. Methodological quality was assessed with a standardized tool. Information was extracted from 35 studies aiming to prevent risk behaviors and promote healthy nutrition, physical activity, and mental and holistic health. Activities were based on theoretical models and were classified into interactive lessons, peer mediation, environmental changes, parents' and community activities, and tailored messages by computer-assisted training or other resources, usually including multiple components. In some cases, we identified some moderate to large, short- and long-term effects on behavioral and intermediate variable. This exhaustive review found that well-implemented interventions can promote adolescent health. These findings are consistent with recent reviews. Implications for practice, public health, and research are discussed. Copyright © 2014 SESPAS. Published by Elsevier Espana. All rights reserved.

  20. Modeling Network Traffic in Wavelet Domain

    Directory of Open Access Journals (Sweden)

    Sheng Ma

    2004-12-01

    Full Text Available This work discovers that although network traffic has the complicated short- and long-range temporal dependence, the corresponding wavelet coefficients are no longer long-range dependent. Therefore, a "short-range" dependent process can be used to model network traffic in the wavelet domain. Both independent and Markov models are investigated. Theoretical analysis shows that the independent wavelet model is sufficiently accurate in terms of the buffer overflow probability for Fractional Gaussian Noise traffic. Any model, which captures additional correlations in the wavelet domain, only improves the performance marginally. The independent wavelet model is then used as a unified approach to model network traffic including VBR MPEG video and Ethernet data. The computational complexity is O(N for developing such wavelet models and generating synthesized traffic of length N, which is among the lowest attained.

  1. Abscisic Acid Regulates Inflammation via Ligand-binding Domain-independent Activation of Peroxisome Proliferator-activated Receptor γ*

    Science.gov (United States)

    Bassaganya-Riera, Josep; Guri, Amir J.; Lu, Pinyi; Climent, Montse; Carbo, Adria; Sobral, Bruno W.; Horne, William T.; Lewis, Stephanie N.; Bevan, David R.; Hontecillas, Raquel

    2011-01-01

    Abscisic acid (ABA) has shown efficacy in the treatment of diabetes and inflammation; however, its molecular targets and the mechanisms of action underlying its immunomodulatory effects remain unclear. This study investigates the role of peroxisome proliferator-activated receptor γ (PPAR γ) and lanthionine synthetase C-like 2 (LANCL2) as molecular targets for ABA. We demonstrate that ABA increases PPAR γ reporter activity in RAW 264.7 macrophages and increases ppar γ expression in vivo, although it does not bind to the ligand-binding domain of PPAR γ. LANCL2 knockdown studies provide evidence that ABA-mediated activation of macrophage PPAR γ is dependent on lancl2 expression. Consistent with the association of LANCL2 with G proteins, we provide evidence that ABA increases cAMP accumulation in immune cells. ABA suppresses LPS-induced prostaglandin E2 and MCP-1 production via a PPAR γ-dependent mechanism possibly involving activation of PPAR γ and suppression of NF-κB and nuclear factor of activated T cells. LPS challenge studies in PPAR γ-expressing and immune cell-specific PPAR γ null mice demonstrate that ABA down-regulates toll-like receptor 4 expression in macrophages and T cells in vivo through a PPAR γ-dependent mechanism. Global transcriptomic profiling and confirmatory quantitative RT-PCR suggest novel candidate targets and demonstrate that ABA treatment mitigates the effect of LPS on the expression of genes involved in inflammation, metabolism, and cell signaling, in part, through PPAR γ. In conclusion, ABA decreases LPS-mediated inflammation and regulates innate immune responses through a bifurcating pathway involving LANCL2 and an alternative, ligand-binding domain-independent mechanism of PPAR γ activation. PMID:21088297

  2. Abscisic acid regulates inflammation via ligand-binding domain-independent activation of peroxisome proliferator-activated receptor gamma.

    Science.gov (United States)

    Bassaganya-Riera, Josep; Guri, Amir J; Lu, Pinyi; Climent, Montse; Carbo, Adria; Sobral, Bruno W; Horne, William T; Lewis, Stephanie N; Bevan, David R; Hontecillas, Raquel

    2011-01-28

    Abscisic acid (ABA) has shown efficacy in the treatment of diabetes and inflammation; however, its molecular targets and the mechanisms of action underlying its immunomodulatory effects remain unclear. This study investigates the role of peroxisome proliferator-activated receptor γ (PPAR γ) and lanthionine synthetase C-like 2 (LANCL2) as molecular targets for ABA. We demonstrate that ABA increases PPAR γ reporter activity in RAW 264.7 macrophages and increases ppar γ expression in vivo, although it does not bind to the ligand-binding domain of PPAR γ. LANCL2 knockdown studies provide evidence that ABA-mediated activation of macrophage PPAR γ is dependent on lancl2 expression. Consistent with the association of LANCL2 with G proteins, we provide evidence that ABA increases cAMP accumulation in immune cells. ABA suppresses LPS-induced prostaglandin E(2) and MCP-1 production via a PPAR γ-dependent mechanism possibly involving activation of PPAR γ and suppression of NF-κB and nuclear factor of activated T cells. LPS challenge studies in PPAR γ-expressing and immune cell-specific PPAR γ null mice demonstrate that ABA down-regulates toll-like receptor 4 expression in macrophages and T cells in vivo through a PPAR γ-dependent mechanism. Global transcriptomic profiling and confirmatory quantitative RT-PCR suggest novel candidate targets and demonstrate that ABA treatment mitigates the effect of LPS on the expression of genes involved in inflammation, metabolism, and cell signaling, in part, through PPAR γ. In conclusion, ABA decreases LPS-mediated inflammation and regulates innate immune responses through a bifurcating pathway involving LANCL2 and an alternative, ligand-binding domain-independent mechanism of PPAR γ activation.

  3. Source-independent time-domain waveform inversion using convolved wavefields: Application to the encoded multisource waveform inversion

    KAUST Repository

    Choi, Yun Seok

    2011-09-01

    Full waveform inversion requires a good estimation of the source wavelet to improve our chances of a successful inversion. This is especially true for an encoded multisource time-domain implementation, which, conventionally, requires separate-source modeling, as well as the Fourier transform of wavefields. As an alternative, we have developed the source-independent time-domain waveform inversion using convolved wavefields. Specifically, the misfit function consists of the convolution of the observed wavefields with a reference trace from the modeled wavefield, plus the convolution of the modeled wavefields with a reference trace from the observed wavefield. In this case, the source wavelet of the observed and the modeled wavefields are equally convolved with both terms in the misfit function, and thus, the effects of the source wavelets are eliminated. Furthermore, because the modeled wavefields play a role of low-pass filtering, the observed wavefields in the misfit function, the frequency-selection strategy from low to high can be easily adopted just by setting the maximum frequency of the source wavelet of the modeled wavefields; and thus, no filtering is required. The gradient of the misfit function is computed by back-propagating the new residual seismograms and applying the imaging condition, similar to reverse-time migration. In the synthetic data evaluations, our waveform inversion yields inverted models that are close to the true model, but demonstrates, as predicted, some limitations when random noise is added to the synthetic data. We also realized that an average of traces is a better choice for the reference trace than using a single trace. © 2011 Society of Exploration Geophysicists.

  4. Automated concept and relationship extraction for the semi-automated ontology management (SEAM) system.

    Science.gov (United States)

    Doing-Harris, Kristina; Livnat, Yarden; Meystre, Stephane

    2015-01-01

    We develop medical-specialty specific ontologies that contain the settled science and common term usage. We leverage current practices in information and relationship extraction to streamline the ontology development process. Our system combines different text types with information and relationship extraction techniques in a low overhead modifiable system. Our SEmi-Automated ontology Maintenance (SEAM) system features a natural language processing pipeline for information extraction. Synonym and hierarchical groups are identified using corpus-based semantics and lexico-syntactic patterns. The semantic vectors we use are term frequency by inverse document frequency and context vectors. Clinical documents contain the terms we want in an ontology. They also contain idiosyncratic usage and are unlikely to contain the linguistic constructs associated with synonym and hierarchy identification. By including both clinical and biomedical texts, SEAM can recommend terms from those appearing in both document types. The set of recommended terms is then used to filter the synonyms and hierarchical relationships extracted from the biomedical corpus. We demonstrate the generality of the system across three use cases: ontologies for acute changes in mental status, Medically Unexplained Syndromes, and echocardiogram summary statements. Across the three uses cases, we held the number of recommended terms relatively constant by changing SEAM's parameters. Experts seem to find more than 300 recommended terms to be overwhelming. The approval rate of recommended terms increased as the number and specificity of clinical documents in the corpus increased. It was 60% when there were 199 clinical documents that were not specific to the ontology domain and 90% when there were 2879 documents very specific to the target domain. We found that fewer than 100 recommended synonym groups were also preferred. Approval rates for synonym recommendations remained low varying from 43% to 25% as the

  5. Identifying APT Malware Domain Based on Mobile DNS Logging

    Directory of Open Access Journals (Sweden)

    Weina Niu

    2017-01-01

    Full Text Available Advanced Persistent Threat (APT is a serious threat against sensitive information. Current detection approaches are time-consuming since they detect APT attack by in-depth analysis of massive amounts of data after data breaches. Specifically, APT attackers make use of DNS to locate their command and control (C&C servers and victims’ machines. In this paper, we propose an efficient approach to detect APT malware C&C domain with high accuracy by analyzing DNS logs. We first extract 15 features from DNS logs of mobile devices. According to Alexa ranking and the VirusTotal’s judgement result, we give each domain a score. Then, we select the most normal domains by the score metric. Finally, we utilize our anomaly detection algorithm, called Global Abnormal Forest (GAF, to identify malware C&C domains. We conduct a performance analysis to demonstrate that our approach is more efficient than other existing works in terms of calculation efficiency and recognition accuracy. Compared with Local Outlier Factor (LOF, k-Nearest Neighbor (KNN, and Isolation Forest (iForest, our approach obtains more than 99% F-M and R for the detection of C&C domains. Our approach not only can reduce data volume that needs to be recorded and analyzed but also can be applicable to unsupervised learning.

  6. Knowledge discovery: Extracting usable information from large amounts of data

    International Nuclear Information System (INIS)

    Whiteson, R.

    1998-01-01

    The threat of nuclear weapons proliferation is a problem of world wide concern. Safeguards are the key to nuclear nonproliferation and data is the key to safeguards. The safeguards community has access to a huge and steadily growing volume of data. The advantages of this data rich environment are obvious, there is a great deal of information which can be utilized. The challenge is to effectively apply proven and developing technologies to find and extract usable information from that data. That information must then be assessed and evaluated to produce the knowledge needed for crucial decision making. Efficient and effective analysis of safeguards data will depend on utilizing technologies to interpret the large, heterogeneous data sets that are available from diverse sources. With an order-of-magnitude increase in the amount of data from a wide variety of technical, textual, and historical sources there is a vital need to apply advanced computer technologies to support all-source analysis. There are techniques of data warehousing, data mining, and data analysis that can provide analysts with tools that will expedite their extracting useable information from the huge amounts of data to which they have access. Computerized tools can aid analysts by integrating heterogeneous data, evaluating diverse data streams, automating retrieval of database information, prioritizing inputs, reconciling conflicting data, doing preliminary interpretations, discovering patterns or trends in data, and automating some of the simpler prescreening tasks that are time consuming and tedious. Thus knowledge discovery technologies can provide a foundation of support for the analyst. Rather than spending time sifting through often irrelevant information, analysts could use their specialized skills in a focused, productive fashion. This would allow them to make their analytical judgments with more confidence and spend more of their time doing what they do best

  7. Extraction process

    International Nuclear Information System (INIS)

    Rendall, J.S.; Cahalan, M.J.

    1979-01-01

    A process is described for extracting at least two desired constituents from a mineral, using a liquid reagent which produces the constituents, or compounds thereof, in separable form and independently extracting those constituents, or compounds. The process is especially valuable for the extraction of phosphoric acid and metal values from acidulated phosphate rock, the slurry being contacted with selective extractants for phosphoric acid and metal (e.g. uranium) values. In an example, uranium values are oxidized to uranyl form and extracted using an ion exchange resin. (U.K.)

  8. A Multi-Level Model of Information Seeking in the Clinical Domain

    Science.gov (United States)

    Hung, Peter W.; Johnson, Stephen B.; Kaufman, David R.; Mendonça, Eneida A.

    2008-01-01

    Objective: Clinicians often have difficulty translating information needs into effective search strategies to find appropriate answers. Information retrieval systems employing an intelligent search agent that generates adaptive search strategies based on human search expertise could be helpful in meeting clinician information needs. A prerequisite for creating such systems is an information seeking model that facilitates the representation of human search expertise. The purpose of developing such a model is to provide guidance to information seeking system development and to shape an empirical research program. Design: The information seeking process was modeled as a complex problem-solving activity. After considering how similarly complex activities had been modeled in other domains, we determined that modeling context-initiated information seeking across multiple problem spaces allows the abstraction of search knowledge into functionally consistent layers. The knowledge layers were identified in the information science literature and validated through our observations of searches performed by health science librarians. Results: A hierarchical multi-level model of context-initiated information seeking is proposed. Each level represents (1) a problem space that is traversed during the online search process, and (2) a distinct layer of knowledge that is required to execute a successful search. Grand strategy determines what information resources will be searched, for what purpose, and in what order. The strategy level represents an overall approach for searching a single resource. Tactics are individual moves made to further a strategy. Operations are mappings of abstract intentions to information resource-specific concrete input. Assessment is the basis of interaction within the strategic hierarchy, influencing the direction of the search. Conclusion: The described multi-level model provides a framework for future research and the foundation for development of an

  9. Extraction of temporal information in functional MRI

    Science.gov (United States)

    Singh, M.; Sungkarat, W.; Jeong, Jeong-Won; Zhou, Yongxia

    2002-10-01

    The temporal resolution of functional MRI (fMRI) is limited by the shape of the haemodynamic response function (hrf) and the vascular architecture underlying the activated regions. Typically, the temporal resolution of fMRI is on the order of 1 s. We have developed a new data processing approach to extract temporal information on a pixel-by-pixel basis at the level of 100 ms from fMRI data. Instead of correlating or fitting the time-course of each pixel to a single reference function, which is the common practice in fMRI, we correlate each pixel's time-course to a series of reference functions that are shifted with respect to each other by 100 ms. The reference function yielding the highest correlation coefficient for a pixel is then used as a time marker for that pixel. A Monte Carlo simulation and experimental study of this approach were performed to estimate the temporal resolution as a function of signal-to-noise ratio (SNR) in the time-course of a pixel. Assuming a known and stationary hrf, the simulation and experimental studies suggest a lower limit in the temporal resolution of approximately 100 ms at an SNR of 3. The multireference function approach was also applied to extract timing information from an event-related motor movement study where the subjects flexed a finger on cue. The event was repeated 19 times with the event's presentation staggered to yield an approximately 100-ms temporal sampling of the haemodynamic response over the entire presentation cycle. The timing differences among different regions of the brain activated by the motor task were clearly visualized and quantified by this method. The results suggest that it is possible to achieve a temporal resolution of /spl sim/200 ms in practice with this approach.

  10. Information Retrieval Using Hadoop Big Data Analysis

    Science.gov (United States)

    Motwani, Deepak; Madan, Madan Lal

    This paper concern on big data analysis which is the cognitive operation of probing huge amounts of information in an attempt to get uncovers unseen patterns. Through Big Data Analytics Applications such as public and private organization sectors have formed a strategic determination to turn big data into cut throat benefit. The primary occupation of extracting value from big data give rise to a process applied to pull information from multiple different sources; this process is known as extract transforms and lode. This paper approach extract information from log files and Research Paper, awareness reduces the efforts for blueprint finding and summarization of document from several positions. The work is able to understand better Hadoop basic concept and increase the user experience for research. In this paper, we propose an approach for analysis log files for finding concise information which is useful and time saving by using Hadoop. Our proposed approach will be applied on different research papers on a specific domain and applied for getting summarized content for further improvement and make the new content.

  11. Towards Domain-specific Flow-based Languages

    DEFF Research Database (Denmark)

    Zarrin, Bahram; Baumeister, Hubert; Sarjoughian, Hessam S.

    2018-01-01

    describe their problems and solutions, instead of using general purpose programming languages. The goal of these languages is to improve the productivity and efficiency of the development and simulation of concurrent scientific models and systems. Moreover, they help to expose parallelism and to specify...... the concurrency within a component or across different independent components. In this paper, we introduce the concept of domain-specific flowbased languages which allows domain experts to use flow-based languages adapted to a particular problem domain. Flow-based programming is used to support concurrency, while......Due to the significant growth of the demand for data-intensive computing, in addition to the emergence of new parallel and distributed computing technologies, scientists and domain experts are leveraging languages specialized for their problem domain, i.e., domain-specific languages, to help them...

  12. The SACSESS hydrometallurgy domain - an overview

    Energy Technology Data Exchange (ETDEWEB)

    Geist, A. [Karlsruhe Institute of Technology - KIT, Institute for Nuclear Wsaste Disposal - INE, Karlsruhe (Germany); Taylor, R. [National Nuclear Laboratory, Central Laboratory, Sellafield, Seascale, CA20 1PG (United Kingdom); Ekberg, C. [Chalmers University of Technology, Nuclear Chemistry/Industrial Materials Recycling, SE-412 96 Goeteborg (Sweden); Guilbaud, P.; Bourg, S. [CEA, Centre de Marcoule, Nuclear Energy Division, F-30207 Bagnols-sur-Ceze (France); Modolo, G. [Forschungszentrum Juelich GmbH - FZJ, Institut fuer Energie- und Klimaforschung - IEK-6, Juelich (Germany)

    2016-07-01

    The EURATOM FP7 project SACSESS (Safety of Actinide Separation Processes) is in continuity of a long line of preceding EURATOM projects. SACSESS is organised along four domains, one of them related to the development of hydrometallurgical (i.e. solvent extraction based) actinide separations processes. Within this domain, the most promising processes developed in previous projects are further developed, improving their technology readiness level (TRL) towards the point at which safe industrial implementation will be achievable. The SACSESS reference compounds are: TODGA, CyMe{sub 4}-BTBP, SO{sub 3}-Ph-BTP, HEDTA and DTPA. TODGA is used to co-extract actinides and lanthanides from high-acidity raffinate solutions, separating from the non-lanthanide fission products. TODGA is also used to accelerate the extraction kinetics of CyMe{sub 4}-BTBP. CyMe{sub 4}-BTBP extracts actinides selectively over lanthanides and many other fission products. HEDTA and DTPA are used to strip actinides selectively over lanthanides from an organic phase containing both actinides and lanthanides. SO{sub 3}-Ph-BTP was developed to overcome some of the drawbacks of HEDTA and DTPA, such as the narrow pH window they are effective in.

  13. Independent component analysis using prior information for signal detection in a functional imaging system of the retina

    NARCIS (Netherlands)

    Barriga, E. Simon; Pattichis, Marios; Ts’o, Dan; Abramoff, Michael; Kardon, Randy; Kwon, Young; Soliz, Peter

    2011-01-01

    Independent component analysis (ICA) is a statistical technique that estimates a set of sources mixed by an unknown mixing matrix using only a set of observations. For this purpose, the only assumption is that the sources are statistically independent. In many applications, some information about

  14. Platelet-rich plasma: updating of extraction devices

    Directory of Open Access Journals (Sweden)

    Raquel Moreno

    2016-12-01

    Full Text Available Propose: To describe PRP extraction devices, through a review of kits available in Spain, taking into account AEMPS and SEFH working groups (GPS, Farmacotecnia, Hemoderivados groups contributions. Methods: Three independent searches about PRP extraction devices were carried out. Device suppliers were contacted and an individually meeting was called with each one. Characteristics of each device was reviewed by virtual demonstration. A kits comparison chart was made with all the information acquired. Kits were classified as Closed-Technique and Opened- Technique in accordance with the AEMPS technical committee report. Results: Ten devices were found: ACP®; Angel®, Cascade®, Endoret ®, GPS®, Magellan®, Minos®, Ortho-pras®, Smart-prepr® and Tricell®. However, we could found out the mechanism in detail of seven of them. Information about Cascade®, Magellan ® and Smart-prepr® kits was not enough. Conclusion: The review provided the main PRP extraction devices available with CE marking and its distinguishing characteristics, however, it is crucial to pay attention to PRP extraction procedure and administration, to guarantee the final product quality. Pharmacy Department must get involved in the device selections due to the close link with the manufactured drug quality. Working together with the AEMPS will contribute to defining extraction procedure specifically.

  15. INFORMATION EXTRACTION IN TOMB PIT USING HYPERSPECTRAL DATA

    Directory of Open Access Journals (Sweden)

    X. Yang

    2018-04-01

    Full Text Available Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  16. Information Extraction in Tomb Pit Using Hyperspectral Data

    Science.gov (United States)

    Yang, X.; Hou, M.; Lyu, S.; Ma, S.; Gao, Z.; Bai, S.; Gu, M.; Liu, Y.

    2018-04-01

    Hyperspectral data has characteristics of multiple bands and continuous, large amount of data, redundancy, and non-destructive. These characteristics make it possible to use hyperspectral data to study cultural relics. In this paper, the hyperspectral imaging technology is adopted to recognize the bottom images of an ancient tomb located in Shanxi province. There are many black remains on the bottom surface of the tomb, which are suspected to be some meaningful texts or paintings. Firstly, the hyperspectral data is preprocessing to get the reflectance of the region of interesting. For the convenient of compute and storage, the original reflectance value is multiplied by 10000. Secondly, this article uses three methods to extract the symbols at the bottom of the ancient tomb. Finally we tried to use morphology to connect the symbols and gave fifteen reference images. The results show that the extraction of information based on hyperspectral data can obtain a better visual experience, which is beneficial to the study of ancient tombs by researchers, and provides some references for archaeological research findings.

  17. Robust Digital Image Watermarking Against Cropping Using Sudoku Puzzle in Spatial and Transform Domain

    Directory of Open Access Journals (Sweden)

    shadi saneie

    2016-10-01

    Full Text Available With rapid development of digital technology, protecting information such as copyright, content ownership confirmation has become more important. In image watermarking, information of the image is inserted such that the visual quality of the image is not reduced and the receiver is able to get the required information. Some attacks such as image cropping, destroy the watermark’s information. In this article, a new watermarking scheme is proposed which is robust against tough cropping. In the proposed scheme, classic Sudoku table which is a 9*9 table, has been used. One feature of Sudoku table is that Sudoku's limitations cause uniform scattering of symbols or numbers throughout the table. In the proposed scheme, Sudoku table and both watermarking approaches based on spatial domain and transform domain such as DCT and DWT are used. Lack of using of soduko solution at the stage of extraction and finding correct solution to obtain watermark, is innovation of this scheme. Robustness of watermarking against cropping attack is up to 92%, which shows good and effective performance of the proposed scheme.

  18. Toward Multi-Domain Battle: Combined Arms Precedents to Inform Today’s Joint Force

    Science.gov (United States)

    2017-03-31

    physical and abstract] domains ... the electromagnetic spectrum [ EMS ], the information environment, and the cognitive dimension of warfare.”8 The...of war. Americans brought about several technical and operational innovations that supported the strategic bombing campaign. Combined-Systems...defense, strategic bombing , and amphibious warfare as examples of “combined-systems revolutions.”’8 Each of these innovations took decades to develop

  19. Dual-wavelength phase-shifting digital holography selectively extracting wavelength information from wavelength-multiplexed holograms.

    Science.gov (United States)

    Tahara, Tatsuki; Mori, Ryota; Kikunaga, Shuhei; Arai, Yasuhiko; Takaki, Yasuhiro

    2015-06-15

    Dual-wavelength phase-shifting digital holography that selectively extracts wavelength information from five wavelength-multiplexed holograms is presented. Specific phase shifts for respective wavelengths are introduced to remove the crosstalk components and extract only the object wave at the desired wavelength from the holograms. Object waves in multiple wavelengths are selectively extracted by utilizing 2π ambiguity and the subtraction procedures based on phase-shifting interferometry. Numerical results show the validity of the proposed technique. The proposed technique is also experimentally demonstrated.

  20. The Financial Resilience of Independent Colleges and Universities

    Science.gov (United States)

    Chessman, Hollie M.; Hartley, Harold V., III; Williams, Michael

    2017-01-01

    This report explores the financial resilience of independent colleges and universities through an analysis of 14 years (fiscal years 2001-2014) of financial data from 559 private nondoctoral institutions. Through analysis of financial performance in several domains, the data reveal that the majority of independent baccalaureate and master's-level…

  1. Combined failure acoustical diagnosis based on improved frequency domain blind deconvolution

    International Nuclear Information System (INIS)

    Pan, Nan; Wu, Xing; Chi, YiLin; Liu, Xiaoqin; Liu, Chang

    2012-01-01

    According to gear box combined failure extraction in complex sound field, an acoustic fault detection method based on improved frequency domain blind deconvolution was proposed. Follow the frequency-domain blind deconvolution flow, the morphological filtering was firstly used to extract modulation features embedded in the observed signals, then the CFPA algorithm was employed to do complex-domain blind separation, finally the J-Divergence of spectrum was employed as distance measure to resolve the permutation. Experiments using real machine sound signals was carried out. The result demonstrate this algorithm can be efficiently applied to gear box combined failure detection in practice.

  2. A Risk Assessment System with Automatic Extraction of Event Types

    Science.gov (United States)

    Capet, Philippe; Delavallade, Thomas; Nakamura, Takuya; Sandor, Agnes; Tarsitano, Cedric; Voyatzi, Stavroula

    In this article we describe the joint effort of experts in linguistics, information extraction and risk assessment to integrate EventSpotter, an automatic event extraction engine, into ADAC, an automated early warning system. By detecting as early as possible weak signals of emerging risks ADAC provides a dynamic synthetic picture of situations involving risk. The ADAC system calculates risk on the basis of fuzzy logic rules operated on a template graph whose leaves are event types. EventSpotter is based on a general purpose natural language dependency parser, XIP, enhanced with domain-specific lexical resources (Lexicon-Grammar). Its role is to automatically feed the leaves with input data.

  3. Non-slipping domains of a pulled spool

    International Nuclear Information System (INIS)

    Wagner, Clemens; Vaterlaus, Andreas

    2014-01-01

    We have investigated the pulled spool by considering pulling angles up to 360 ∘ . Our focus was on downward pulling forces with pulling angles in the range of 180 ∘ to 360 ∘ . In this range we have found a domain of pulling angles where the spool never starts to slip independent of the strength of the pulling force. The size of the domain depends on the static friction coefficient and on the moment of inertia of the spool. The non-slipping domain is mainly formed around the critical angle where the static friction force becomes zero. For low static friction the non-slipping domain decays into two different domains. We have determined the limiting angles of the non-slipping domains and explored the transitions from a single domain to two separated domains in parameter space. (paper)

  4. A Novel Sub-pixel Measurement Algorithm Based on Mixed the Fractal and Digital Speckle Correlation in Frequency Domain

    Directory of Open Access Journals (Sweden)

    Zhangfang Hu

    2014-10-01

    Full Text Available The digital speckle correlation is a non-contact in-plane displacement measurement method based on machine vision. Motivated by the facts that the low accuracy and large amount of calculation produced by the traditional digital speckle correlation method in spatial domain, we introduce a sub-pixel displacement measurement algorithm which employs a fast interpolation method based on fractal theory and digital speckle correlation in frequency domain. This algorithm can overcome either the blocking effect or the blurring caused by the traditional interpolation methods, and the frequency domain processing also avoids the repeated searching in the correlation recognition of the spatial domain, thus the operation quantity is largely reduced and the information extracting speed is improved. The comparative experiment is given to verify that the proposed algorithm in this paper is effective.

  5. Domain Wall Motion in Magnetic Nanostrips under the Influence of Rashba Field

    Directory of Open Access Journals (Sweden)

    Vito Puliafito

    2012-01-01

    Full Text Available Spin-orbit Rashba effect applies a torque on the magnetization of a ferromagnetic nanostrip in the case of structural inversion asymmetry, also affecting the steady domain wall motion induced by a spin-polarized current. This influence is here analytically studied in the framework of the extended Landau-Lifshitz-Gilbert equation, including the Rashba effect as an additive term of the effective field. Results of previous micromagnetic simulations and experiments have shown that this field yields an increased value of the Walker breakdown current together with an enlargement of the domain wall width. In order to analytically describe these results, the standard travelling wave ansatz for the steady domain wall motion is here adopted. Results of our investigations reveal the impossibility to reproduce, at the same time, the previous features and suggest the need of a more sophisticated model whose development requires, in turn, additional information to be extracted from ad hoc micromagnetic simulations.

  6. Ferroelectric negative capacitance domain dynamics

    Science.gov (United States)

    Hoffmann, Michael; Khan, Asif Islam; Serrao, Claudy; Lu, Zhongyuan; Salahuddin, Sayeef; Pešić, Milan; Slesazeck, Stefan; Schroeder, Uwe; Mikolajick, Thomas

    2018-05-01

    Transient negative capacitance effects in epitaxial ferroelectric Pb(Zr0.2Ti0.8)O3 capacitors are investigated with a focus on the dynamical switching behavior governed by domain nucleation and growth. Voltage pulses are applied to a series connection of the ferroelectric capacitor and a resistor to directly measure the ferroelectric negative capacitance during switching. A time-dependent Ginzburg-Landau approach is used to investigate the underlying domain dynamics. The transient negative capacitance is shown to originate from reverse domain nucleation and unrestricted domain growth. However, with the onset of domain coalescence, the capacitance becomes positive again. The persistence of the negative capacitance state is therefore limited by the speed of domain wall motion. By changing the applied electric field, capacitor area or external resistance, this domain wall velocity can be varied predictably over several orders of magnitude. Additionally, detailed insights into the intrinsic material properties of the ferroelectric are obtainable through these measurements. A new method for reliable extraction of the average negative capacitance of the ferroelectric is presented. Furthermore, a simple analytical model is developed, which accurately describes the negative capacitance transient time as a function of the material properties and the experimental boundary conditions.

  7. A test of the domain-specific acculturation strategy hypothesis.

    Science.gov (United States)

    Miller, Matthew J; Yang, Minji; Lim, Robert H; Hui, Kayi; Choi, Na-Yeun; Fan, Xiaoyan; Lin, Li-Ling; Grome, Rebekah E; Farrell, Jerome A; Blackmon, Sha'kema

    2013-01-01

    Acculturation literature has evolved over the past several decades and has highlighted the dynamic ways in which individuals negotiate experiences in multiple cultural contexts. The present study extends this literature by testing M. J. Miller and R. H. Lim's (2010) domain-specific acculturation strategy hypothesis-that individuals might use different acculturation strategies (i.e., assimilated, bicultural, separated, and marginalized strategies; J. W. Berry, 2003) across behavioral and values domains-in 3 independent cluster analyses with Asian American participants. Present findings supported the domain-specific acculturation strategy hypothesis as 67% to 72% of participants from 3 independent samples using different strategies across behavioral and values domains. Consistent with theory, a number of acculturation strategy cluster group differences emerged across generational status, acculturative stress, mental health symptoms, and attitudes toward seeking professional psychological help. Study limitations and future directions for research are discussed.

  8. Extracting information from two-dimensional electrophoresis gels by partial least squares regression

    DEFF Research Database (Denmark)

    Jessen, Flemming; Lametsch, R.; Bendixen, E.

    2002-01-01

    of all proteins/spots in the gels. In the present study it is demonstrated how information can be extracted by multivariate data analysis. The strategy is based on partial least squares regression followed by variable selection to find proteins that individually or in combination with other proteins vary......Two-dimensional gel electrophoresis (2-DE) produces large amounts of data and extraction of relevant information from these data demands a cautious and time consuming process of spot pattern matching between gels. The classical approach of data analysis is to detect protein markers that appear...... or disappear depending on the experimental conditions. Such biomarkers are found by comparing the relative volumes of individual spots in the individual gels. Multivariate statistical analysis and modelling of 2-DE data for comparison and classification is an alternative approach utilising the combination...

  9. A model-driven approach to information security compliance

    Science.gov (United States)

    Correia, Anacleto; Gonçalves, António; Teodoro, M. Filomena

    2017-06-01

    The availability, integrity and confidentiality of information are fundamental to the long-term survival of any organization. Information security is a complex issue that must be holistically approached, combining assets that support corporate systems, in an extended network of business partners, vendors, customers and other stakeholders. This paper addresses the conception and implementation of information security systems, conform the ISO/IEC 27000 set of standards, using the model-driven approach. The process begins with the conception of a domain level model (computation independent model) based on information security vocabulary present in the ISO/IEC 27001 standard. Based on this model, after embedding in the model mandatory rules for attaining ISO/IEC 27001 conformance, a platform independent model is derived. Finally, a platform specific model serves the base for testing the compliance of information security systems with the ISO/IEC 27000 set of standards.

  10. Chaotic spectra: How to extract dynamic information

    International Nuclear Information System (INIS)

    Taylor, H.S.; Gomez Llorente, J.M.; Zakrzewski, J.; Kulander, K.C.

    1988-10-01

    Nonlinear dynamics is applied to chaotic unassignable atomic and molecular spectra with the aim of extracting detailed information about regular dynamic motions that exist over short intervals of time. It is shown how this motion can be extracted from high resolution spectra by doing low resolution studies or by Fourier transforming limited regions of the spectrum. These motions mimic those of periodic orbits (PO) and are inserts into the dominant chaotic motion. Considering these inserts and the PO as a dynamically decoupled region of space, resonant scattering theory and stabilization methods enable us to compute ladders of resonant states which interact with the chaotic quasi-continuum computed in principle from basis sets placed off the PO. The interaction of the resonances with the quasicontinuum explains the low resolution spectra seen in such experiments. It also allows one to associate low resolution features with a particular PO. The motion on the PO thereby supplies the molecular movements whose quantization causes the low resolution spectra. Characteristic properties of the periodic orbit based resonances are discussed. The method is illustrated on the photoabsorption spectrum of the hydrogen atom in a strong magnetic field and on the photodissociation spectrum of H 3 + . Other molecular systems which are currently under investigation using this formalism are also mentioned. 53 refs., 10 figs., 2 tabs

  11. Infrared and visible fusion face recognition based on NSCT domain

    Science.gov (United States)

    Xie, Zhihua; Zhang, Shuai; Liu, Guodong; Xiong, Jinquan

    2018-01-01

    Visible face recognition systems, being vulnerable to illumination, expression, and pose, can not achieve robust performance in unconstrained situations. Meanwhile, near infrared face images, being light- independent, can avoid or limit the drawbacks of face recognition in visible light, but its main challenges are low resolution and signal noise ratio (SNR). Therefore, near infrared and visible fusion face recognition has become an important direction in the field of unconstrained face recognition research. In this paper, a novel fusion algorithm in non-subsampled contourlet transform (NSCT) domain is proposed for Infrared and visible face fusion recognition. Firstly, NSCT is used respectively to process the infrared and visible face images, which exploits the image information at multiple scales, orientations, and frequency bands. Then, to exploit the effective discriminant feature and balance the power of high-low frequency band of NSCT coefficients, the local Gabor binary pattern (LGBP) and Local Binary Pattern (LBP) are applied respectively in different frequency parts to obtain the robust representation of infrared and visible face images. Finally, the score-level fusion is used to fuse the all the features for final classification. The visible and near infrared face recognition is tested on HITSZ Lab2 visible and near infrared face database. Experiments results show that the proposed method extracts the complementary features of near-infrared and visible-light images and improves the robustness of unconstrained face recognition.

  12. Human-computer interface incorporating personal and application domains

    Science.gov (United States)

    Anderson, Thomas G [Albuquerque, NM

    2011-03-29

    The present invention provides a human-computer interface. The interface includes provision of an application domain, for example corresponding to a three-dimensional application. The user is allowed to navigate and interact with the application domain. The interface also includes a personal domain, offering the user controls and interaction distinct from the application domain. The separation into two domains allows the most suitable interface methods in each: for example, three-dimensional navigation in the application domain, and two- or three-dimensional controls in the personal domain. Transitions between the application domain and the personal domain are under control of the user, and the transition method is substantially independent of the navigation in the application domain. For example, the user can fly through a three-dimensional application domain, and always move to the personal domain by moving a cursor near one extreme of the display.

  13. Network and Ensemble Enabled Entity Extraction in Informal Text (NEEEEIT) final report

    Energy Technology Data Exchange (ETDEWEB)

    Kegelmeyer, Philip W. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Shead, Timothy M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Dunlavy, Daniel M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2013-09-01

    This SAND report summarizes the activities and outcomes of the Network and Ensemble Enabled Entity Extraction in Information Text (NEEEEIT) LDRD project, which addressed improving the accuracy of conditional random fields for named entity recognition through the use of ensemble methods.

  14. Transform domain Wyner-Ziv video coding with refinement of noise residue and side information

    DEFF Research Database (Denmark)

    Huang, Xin; Forchhammer, Søren

    2010-01-01

    are successively updating the estimated noise residue for noise modeling and side information frame quality during decoding. Experimental results show that the proposed decoder can improve the Rate- Distortion (RD) performance of a state-of-the-art Wyner Ziv video codec for the set of test sequences.......Distributed Video Coding (DVC) is a video coding paradigm which mainly exploits the source statistics at the decoder based on the availability of side information at the decoder. This paper considers feedback channel based Transform Domain Wyner-Ziv (TDWZ) DVC. The coding efficiency of TDWZ video...... coding does not match that of conventional video coding yet, mainly due to the quality of side information and inaccurate noise estimation. In this context, a novel TDWZ video decoder with noise residue refinement (NRR) and side information refinement (SIR) is proposed. The proposed refinement schemes...

  15. An activation domain within the walleye dermal sarcoma virus retroviral cyclin protein is essential for inhibition of the viral promoter

    International Nuclear Information System (INIS)

    Rovnak, Joel; Hronek, Brett W.; Ryan, Sean O.; Cai, Sumin; Quackenbush, Sandra L.

    2005-01-01

    Walleye dermal sarcoma virus (WDSV) is a complex retrovirus associated with seasonal dermal sarcomas. Developing tumors have low levels of accessory gene transcripts, A1 and B, and regressing tumors have high levels of full-length and spliced transcripts. Transcript A1 encodes a retroviral cyclin (rv-cyclin) with limited homology to host cyclins. The rv-cyclin is physically linked to components of the transcriptional co-activator complex, Mediator, and regulates transcription. In walleye fibroblasts, it inhibits the WDSV promoter independently of cis-acting DNA sequences. The rv-cyclin activates transcription from GAL4 promoters when fused to the GAL4 DNA binding domain. A 30 a.a. activation domain in the carboxy region can be inactivated by single point mutations, and these mutations diminish the ability of the rv-cyclin to inhibit the WDSV promoter. When fused to glutathione S-transferase, the rv-cyclin, its carboxy region, and the activation domain pull down components of transcription complexes from nuclear extracts, and pulldown is lost by mutation of the activation domain

  16. The Role of Domain and System Knowledge on Text Comprehension and Information Search in Hypermedia

    Science.gov (United States)

    Waniek, Jacqueline; Schafer, Thomas

    2009-01-01

    The goal of this study was to examine the role of domain and system knowledge on learner performance in reading and information search in hypermedia. Previous studies have shown that prior knowledge is an important individual factor for effective hypermedia use. However, current research lacks a full understanding of how these two aspects of prior…

  17. A Qualitative and Quantitative Comparative Analysis of Commercial and Independent Online Information for Hip Surgery: A Bias in Online Information Targeting Patients?

    Science.gov (United States)

    Kelly, Martin J; Feeley, Iain H; O'Byrne, John M

    2016-10-01

    Direct to consumer (DTC) advertising, targeting the public over the physician, is an increasingly pervasive presence in medical clinics. It is trending toward a format of online interaction rather than that of traditional print and television advertising. We analyze patient-focused Web pages from the top 5 companies supplying prostheses for total hip arthroplasties, comparing them to the top 10 independent medical websites. Quantitative comparison is performed using the Journal of American Medical Association benchmark and DISCERN criteria, and for comparative readability, we use the Flesch-Kincaid grade level, the Flesch reading ease, and the Gunning fog index. Content is analyzed for information on type of surgery and surgical approach. There is a statistically significant difference between the independent and DTC websites in both the mean DISCERN score (independent 74.6, standard deviation [SD] = 4.77; DTC 32.2, SD = 10.28; P = .0022) and the mean Journal of American Medical Association score (Independent 3.45, SD = 0.49; DTC 1.9, SD = 0.74; P = .004). The difference between the readability scores is not statistically significantly. The commercial content is found to be heavily biased in favor of the direct anterior approach and minimally invasive surgical techniques. We demonstrate that the quality of information on commercial websites is inferior to that of the independent sites. The advocacy of surgical approaches by industry to the patient group is a concern. This study underlines the importance of future regulation of commercial patient education Web pages. Copyright © 2016 Elsevier Inc. All rights reserved.

  18. Earth Science Data Analytics: Preparing for Extracting Knowledge from Information

    Science.gov (United States)

    Kempler, Steven; Barbieri, Lindsay

    2016-01-01

    Data analytics is the process of examining large amounts of data of a variety of types to uncover hidden patterns, unknown correlations and other useful information. Data analytics is a broad term that includes data analysis, as well as an understanding of the cognitive processes an analyst uses to understand problems and explore data in meaningful ways. Analytics also include data extraction, transformation, and reduction, utilizing specific tools, techniques, and methods. Turning to data science, definitions of data science sound very similar to those of data analytics (which leads to a lot of the confusion between the two). But the skills needed for both, co-analyzing large amounts of heterogeneous data, understanding and utilizing relevant tools and techniques, and subject matter expertise, although similar, serve different purposes. Data Analytics takes on a practitioners approach to applying expertise and skills to solve issues and gain subject knowledge. Data Science, is more theoretical (research in itself) in nature, providing strategic actionable insights and new innovative methodologies. Earth Science Data Analytics (ESDA) is the process of examining, preparing, reducing, and analyzing large amounts of spatial (multi-dimensional), temporal, or spectral data using a variety of data types to uncover patterns, correlations and other information, to better understand our Earth. The large variety of datasets (temporal spatial differences, data types, formats, etc.) invite the need for data analytics skills that understand the science domain, and data preparation, reduction, and analysis techniques, from a practitioners point of view. The application of these skills to ESDA is the focus of this presentation. The Earth Science Information Partners (ESIP) Federation Earth Science Data Analytics (ESDA) Cluster was created in recognition of the practical need to facilitate the co-analysis of large amounts of data and information for Earth science. Thus, from a to

  19. Remaining useful life estimation based on discriminating shapelet extraction

    International Nuclear Information System (INIS)

    Malinowski, Simon; Chebel-Morello, Brigitte; Zerhouni, Noureddine

    2015-01-01

    In the Prognostics and Health Management domain, estimating the remaining useful life (RUL) of critical machinery is a challenging task. Various research topics including data acquisition, fusion, diagnostics and prognostics are involved in this domain. This paper presents an approach, based on shapelet extraction, to estimate the RUL of equipment. This approach extracts, in an offline step, discriminative rul-shapelets from an history of run-to-failure data. These rul-shapelets are patterns that are selected for their correlation with the remaining useful life of the equipment. In other words, every selected rul-shapelet conveys its own information about the RUL of the equipment. In an online step, these rul-shapelets are compared to testing units and the ones that match these units are used to estimate their RULs. Therefore, RUL estimation is based on patterns that have been selected for their high correlation with the RUL. This approach is different from classical similarity-based approaches that attempt to match complete testing units (or only late instants of testing units) with training ones to estimate the RUL. The performance of our approach is evaluated on a case study on the remaining useful life estimation of turbofan engines and performance is compared with other similarity-based approaches. - Highlights: • A data-driven RUL estimation technique based on pattern extraction is proposed. • Patterns are extracted for their correlation with the RUL. • The proposed method shows good performance compared to other techniques

  20. Access to the kinematic information for the velocity model determination by 3-D reflexion tomography; Acces a l'information cinematique pour la determination du modele de vitesse par tomographie de reflexion 3D

    Energy Technology Data Exchange (ETDEWEB)

    Broto, K.

    1999-04-01

    The access to a reliable image of the subsurface requires a kinematically correct velocity depth model.Reflection tomography allows to meet this requirement if a complete and coherent pre-stack kinematic database can be provided. However, in case of complex sub-surfaces, wave propagation may lead to hardly interpretable seismic events in the time data. The SMART method is a sequential method that relies on reflection tomography for updating the velocity model and on the pre-stack depth migrated domain for extracting kinematic information that is not readily accessible in the time domain. For determining 3-D subsurface velocity models in case of complex structures, we propose the seriated SMART 2-D method as an alternative to the currently inconceivable SMART 3-D method. In order to extract kinematic information from a 3-D pre-stack data set, we combine detours through the 2-D pre-stack depth domain for a number of selected lines of the studied 3-D survey and 3-D reflection tomography for updating the velocity model. The travel-times from the SMART method being independent of the velocity model used for passing through the pre-stack depth migrated domain, the access to 3-D travel-times is ensured, even if they have been obtained via a 2-D domain. Besides, we propose to build a kinematical guide for ensuring the coherency of the seriated 2-D pre-stack depth interpretations and the access to a complete 3-D pre-stack kinematic database when dealing with structures associated with 3-D wave propagation. We opt for a blocky representation of the velocity model in order to be able to cope with complex structures. This representation leads us to define specific methodological rules for carrying out the different steps of the seriated SMART 2-D method. We also define strategies, built from the analysis of first inversion results, for an efficient application of reflection tomography. Besides, we discuss the problem of uncertainties to be assigned to travel-times obtained

  1. Three-dimensional information extraction from GaoFen-1 satellite images for landslide monitoring

    Science.gov (United States)

    Wang, Shixin; Yang, Baolin; Zhou, Yi; Wang, Futao; Zhang, Rui; Zhao, Qing

    2018-05-01

    To more efficiently use GaoFen-1 (GF-1) satellite images for landslide emergency monitoring, a Digital Surface Model (DSM) can be generated from GF-1 across-track stereo image pairs to build a terrain dataset. This study proposes a landslide 3D information extraction method based on the terrain changes of slope objects. The slope objects are mergences of segmented image objects which have similar aspects; and the terrain changes are calculated from the post-disaster Digital Elevation Model (DEM) from GF-1 and the pre-disaster DEM from GDEM V2. A high mountain landslide that occurred in Wenchuan County, Sichuan Province is used to conduct a 3D information extraction test. The extracted total area of the landslide is 22.58 ha; the displaced earth volume is 652,100 m3; and the average sliding direction is 263.83°. The accuracies of them are 0.89, 0.87 and 0.95, respectively. Thus, the proposed method expands the application of GF-1 satellite images to the field of landslide emergency monitoring.

  2. Lithium NLP: A System for Rich Information Extraction from Noisy User Generated Text on Social Media

    OpenAIRE

    Bhargava, Preeti; Spasojevic, Nemanja; Hu, Guoning

    2017-01-01

    In this paper, we describe the Lithium Natural Language Processing (NLP) system - a resource-constrained, high- throughput and language-agnostic system for information extraction from noisy user generated text on social media. Lithium NLP extracts a rich set of information including entities, topics, hashtags and sentiment from text. We discuss several real world applications of the system currently incorporated in Lithium products. We also compare our system with existing commercial and acad...

  3. Inferring domain-domain interactions from protein-protein interactions with formal concept analysis.

    Directory of Open Access Journals (Sweden)

    Susan Khor

    Full Text Available Identifying reliable domain-domain interactions will increase our ability to predict novel protein-protein interactions, to unravel interactions in protein complexes, and thus gain more information about the function and behavior of genes. One of the challenges of identifying reliable domain-domain interactions is domain promiscuity. Promiscuous domains are domains that can occur in many domain architectures and are therefore found in many proteins. This becomes a problem for a method where the score of a domain-pair is the ratio between observed and expected frequencies because the protein-protein interaction network is sparse. As such, many protein-pairs will be non-interacting and domain-pairs with promiscuous domains will be penalized. This domain promiscuity challenge to the problem of inferring reliable domain-domain interactions from protein-protein interactions has been recognized, and a number of work-arounds have been proposed. This paper reports on an application of Formal Concept Analysis to this problem. It is found that the relationship between formal concepts provides a natural way for rare domains to elevate the rank of promiscuous domain-pairs and enrich highly ranked domain-pairs with reliable domain-domain interactions. This piggybacking of promiscuous domain-pairs onto less promiscuous domain-pairs is possible only with concept lattices whose attribute-labels are not reduced and is enhanced by the presence of proteins that comprise both promiscuous and rare domains.

  4. Inferring Domain-Domain Interactions from Protein-Protein Interactions with Formal Concept Analysis

    Science.gov (United States)

    Khor, Susan

    2014-01-01

    Identifying reliable domain-domain interactions will increase our ability to predict novel protein-protein interactions, to unravel interactions in protein complexes, and thus gain more information about the function and behavior of genes. One of the challenges of identifying reliable domain-domain interactions is domain promiscuity. Promiscuous domains are domains that can occur in many domain architectures and are therefore found in many proteins. This becomes a problem for a method where the score of a domain-pair is the ratio between observed and expected frequencies because the protein-protein interaction network is sparse. As such, many protein-pairs will be non-interacting and domain-pairs with promiscuous domains will be penalized. This domain promiscuity challenge to the problem of inferring reliable domain-domain interactions from protein-protein interactions has been recognized, and a number of work-arounds have been proposed. This paper reports on an application of Formal Concept Analysis to this problem. It is found that the relationship between formal concepts provides a natural way for rare domains to elevate the rank of promiscuous domain-pairs and enrich highly ranked domain-pairs with reliable domain-domain interactions. This piggybacking of promiscuous domain-pairs onto less promiscuous domain-pairs is possible only with concept lattices whose attribute-labels are not reduced and is enhanced by the presence of proteins that comprise both promiscuous and rare domains. PMID:24586450

  5. An investigation of time-frequency domain phase-weighted stacking and its application to phase-velocity extraction from ambient noise's empirical Green's functions

    Science.gov (United States)

    Li, Guoliang; Niu, Fenglin; Yang, Yingjie; Xie, Jun

    2018-02-01

    The time-frequency domain phase-weighted stacking (tf-PWS) technique based on the S transform has been employed in stacking empirical Green's functions (EGFs) derived from ambient noise data, mainly due to its superior power in enhancing weak signals. Questions such as the induced waveform distortion and the feasibility of phase-velocity extraction are yet to be thoroughly explored. In this study, we investigate these issues by conducting extensive numerical tests with both synthetic data and USArray transportable array (TA) ambient noise data. We find that the errors in the measured phase velocities associated with waveform distortion caused by the tf-PWS depend largely on the way of how the inverse S transform (IST) is implemented. If frequency IST is employed in tf-PWS, the corresponding errors are generally less than 0.1 per cent, sufficiently small that the measured phase velocities can be safely used in regular surface wave tomography. On the other hand, if a time IST is used in tf-PWS, then the extracted phase velocities are systematically larger than those measured from linearly stacked ones, and the discrepancy can reach as much as ˜0.4 per cent at some periods. Therefore, if tf-PWS is used in stacking EGFs, then frequency IST is preferred to transform the stacked S spectra back to the time domain for the stacked EGFs.

  6. Mutual information, perceptual independence, and holistic face perception.

    Science.gov (United States)

    Fitousi, Daniel

    2013-07-01

    The concept of perceptual independence is ubiquitous in psychology. It addresses the question of whether two (or more) dimensions are perceived independently. Several authors have proposed perceptual independence (or its lack thereof) as a viable measure of holistic face perception (Loftus, Oberg, & Dillon, Psychological Review 111:835-863, 2004; Wenger & Ingvalson, Learning, Memory, and Cognition 28:872-892, 2002). According to this notion, the processing of facial features occurs in an interactive manner. Here, I examine this idea from the perspective of two theories of perceptual independence: the multivariate uncertainty analysis (MUA; Garner & Morton, Definitions, models, and experimental paradigms. Psychological Bulletin 72:233-259, 1969), and the general recognition theory (GRT; Ashby & Townsend, Psychological Review 93:154-179, 1986). The goals of the study were to (1) introduce the MUA, (2) examine various possible relations between MUA and GRT using numerical simulations, and (3) apply the MUA to two consensual markers of holistic face perception(-)recognition of facial features (Farah, Wilson, Drain, & Tanaka, Psychological Review 105:482-498, 1998) and the composite face effect (Young, Hellawell, & Hay, Perception 16:747-759, 1987). The results suggest that facial holism is generated by violations of several types of perceptual independence. They highlight the important theoretical role played by converging operations in the study of holistic face perception.

  7. Cultural orientations, parental beliefs and practices, and latino adolescents' autonomy and independence.

    Science.gov (United States)

    Roche, Kathleen M; Caughy, Margaret O; Schuster, Mark A; Bogart, Laura M; Dittus, Patricia J; Franzini, Luisa

    2014-08-01

    Despite the salience of behavioral autonomy and independence to parent-child interactions during middle adolescence, little is known about parenting processes pertinent to youth autonomy development for Latino families. Among a diverse sample of 684 Latino-origin parent-adolescent dyads in Houston, Texas, this study examines how parents' cultural orientations are associated directly and indirectly, through parental beliefs, with parenting practices giving youth behavioral autonomy and independence. Informed by social domain theory, the study's parenting constructs pertain to youth behaviors in an "ambiguously personal" domain-activities that adolescents believe are up to youth to decide, but which parents might argue require parents' supervision, knowledge, and/or decision-making. Results for latent profile analyses of parents' cultural identity across various facets of acculturation indicate considerable cultural heterogeneity among Latino parents. Although 43% of parents have a Latino cultural orientation, others represent Spanish-speaking/bicultural (21%), bilingual/bicultural (15%), English-speaking/bicultural (15%), or US (6%) cultural orientations. Structural equation modeling results indicate that bilingual/bicultural, English-speaking/bicultural, and US-oriented parents report less emphasis on the legitimacy of parental authority and younger age expectations for youth to engage in independent behaviors than do Latino-oriented parents. Parental beliefs endorsing youth's behavioral independence and autonomy, in turn, are associated with less stringent parental rules (parental report), less parental supervision (parental and youth report), and more youth autonomy in decision-making (parental and youth report). Evidence thus supports the idea that the diverse cultural orientations of Latino parents in the US may result in considerable variations in parenting processes pertinent to Latino adolescents' development.

  8. System and method for extracting physiological information from remotely detected electromagnetic radiation

    NARCIS (Netherlands)

    2016-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  9. System and method for extracting physiological information from remotely detected electromagnetic radiation

    NARCIS (Netherlands)

    2015-01-01

    The present invention relates to a device and a method for extracting physiological information indicative of at least one health symptom from remotely detected electromagnetic radiation. The device comprises an interface (20) for receiving a data stream comprising remotely detected image data

  10. Image segmentation-based robust feature extraction for color image watermarking

    Science.gov (United States)

    Li, Mianjie; Deng, Zeyu; Yuan, Xiaochen

    2018-04-01

    This paper proposes a local digital image watermarking method based on Robust Feature Extraction. The segmentation is achieved by Simple Linear Iterative Clustering (SLIC) based on which an Image Segmentation-based Robust Feature Extraction (ISRFE) method is proposed for feature extraction. Our method can adaptively extract feature regions from the blocks segmented by SLIC. This novel method can extract the most robust feature region in every segmented image. Each feature region is decomposed into low-frequency domain and high-frequency domain by Discrete Cosine Transform (DCT). Watermark images are then embedded into the coefficients in the low-frequency domain. The Distortion-Compensated Dither Modulation (DC-DM) algorithm is chosen as the quantization method for embedding. The experimental results indicate that the method has good performance under various attacks. Furthermore, the proposed method can obtain a trade-off between high robustness and good image quality.

  11. Are Independent Fiscal Institutions Really Independent?

    Directory of Open Access Journals (Sweden)

    Slawomir Franek

    2015-08-01

    Full Text Available In the last decade the number of independent fiscal institutions (known also as fiscal councils has tripled. They play an important oversight role over fiscal policy-making in democratic societies, especially as they seek to restore public finance stability in the wake of the recent financial crisis. Although common functions of such institutions include a role in analysis of fiscal policy, forecasting, monitoring compliance with fiscal rules or costing of spending proposals, their roles, resources and structures vary considerably across countries. The aim of the article is to determine the degree of independence of such institutions based on the analysis of the independence index of independent fiscal institutions. The analysis of this index values may be useful to determine the relations between the degree of independence of fiscal councils and fiscal performance of particular countries. The data used to calculate the index values will be derived from European Commission and IMF, which collect sets of information about characteristics of activity of fiscal councils.

  12. A Time-Frequency Approach to Feature Extraction for a Brain-Computer Interface with a Comparative Analysis of Performance Measures

    Directory of Open Access Journals (Sweden)

    T. M. McGinnity

    2005-11-01

    Full Text Available The paper presents an investigation into a time-frequency (TF method for extracting features from the electroencephalogram (EEG recorded from subjects performing imagination of left- and right-hand movements. The feature extraction procedure (FEP extracts frequency domain information to form features whilst time-frequency resolution is attained by localising the fast Fourier transformations (FFTs of the signals to specific windows localised in time. All features are extracted at the rate of the signal sampling interval from a main feature extraction (FE window through which all data passes. Subject-specific frequency bands are selected for optimal feature extraction and intraclass variations are reduced by smoothing the spectra for each signal by an interpolation (IP process. The TF features are classified using linear discriminant analysis (LDA. The FE window has potential advantages for the FEP to be applied in an online brain-computer interface (BCI. The approach achieves good performance when quantified by classification accuracy (CA rate, information transfer (IT rate, and mutual information (MI. The information that these performance measures provide about a BCI system is analysed and the importance of this is demonstrated through the results.

  13. Epithelium-Stroma Classification via Convolutional Neural Networks and Unsupervised Domain Adaptation in Histopathological Images.

    Science.gov (United States)

    Huang, Yue; Zheng, Han; Liu, Chi; Ding, Xinghao; Rohde, Gustavo K

    2017-11-01

    Epithelium-stroma classification is a necessary preprocessing step in histopathological image analysis. Current deep learning based recognition methods for histology data require collection of large volumes of labeled data in order to train a new neural network when there are changes to the image acquisition procedure. However, it is extremely expensive for pathologists to manually label sufficient volumes of data for each pathology study in a professional manner, which results in limitations in real-world applications. A very simple but effective deep learning method, that introduces the concept of unsupervised domain adaptation to a simple convolutional neural network (CNN), has been proposed in this paper. Inspired by transfer learning, our paper assumes that the training data and testing data follow different distributions, and there is an adaptation operation to more accurately estimate the kernels in CNN in feature extraction, in order to enhance performance by transferring knowledge from labeled data in source domain to unlabeled data in target domain. The model has been evaluated using three independent public epithelium-stroma datasets by cross-dataset validations. The experimental results demonstrate that for epithelium-stroma classification, the proposed framework outperforms the state-of-the-art deep neural network model, and it also achieves better performance than other existing deep domain adaptation methods. The proposed model can be considered to be a better option for real-world applications in histopathological image analysis, since there is no longer a requirement for large-scale labeled data in each specified domain.

  14. A scalable architecture for extracting, aligning, linking, and visualizing multi-Int data

    Science.gov (United States)

    Knoblock, Craig A.; Szekely, Pedro

    2015-05-01

    An analyst today has a tremendous amount of data available, but each of the various data sources typically exists in their own silos, so an analyst has limited ability to see an integrated view of the data and has little or no access to contextual information that could help in understanding the data. We have developed the Domain-Insight Graph (DIG) system, an innovative architecture for extracting, aligning, linking, and visualizing massive amounts of domain-specific content from unstructured sources. Under the DARPA Memex program we have already successfully applied this architecture to multiple application domains, including the enormous international problem of human trafficking, where we extracted, aligned and linked data from 50 million online Web pages. DIG builds on our Karma data integration toolkit, which makes it easy to rapidly integrate structured data from a variety of sources, including databases, spreadsheets, XML, JSON, and Web services. The ability to integrate Web services allows Karma to pull in live data from the various social media sites, such as Twitter, Instagram, and OpenStreetMaps. DIG then indexes the integrated data and provides an easy to use interface for query, visualization, and analysis.

  15. Answer Extraction Based on Merging Score Strategy of Hot Terms

    Institute of Scientific and Technical Information of China (English)

    LE Juan; ZHANG Chunxia; NIU Zhendong

    2016-01-01

    Answer extraction (AE) is one of the key technologies in developing the open domain Question&an-swer (Q&A) system . Its task is to yield the highest score to the expected answer based on an effective answer score strategy. We introduce an answer extraction method by Merging score strategy (MSS) based on hot terms. The hot terms are defined according to their lexical and syn-tactic features to highlight the role of the question terms. To cope with the syntactic diversities of the corpus, we propose four improved candidate answer score algorithms. Each of them is based on the lexical function of hot terms and their syntactic relationships with the candidate an-swers. Two independent corpus score algorithms are pro-posed to tap the role of the corpus in ranking the candi-date answers. Six algorithms are adopted in MSS to tap the complementary action among the corpus, the candi-date answers and the questions. Experiments demonstrate the effectiveness of the proposed strategy.

  16. Pumpkin seed extract: Cell growth inhibition of hyperplastic and cancer cells, independent of steroid hormone receptors.

    Science.gov (United States)

    Medjakovic, Svjetlana; Hobiger, Stefanie; Ardjomand-Woelkart, Karin; Bucar, Franz; Jungbauer, Alois

    2016-04-01

    Pumpkin seeds have been known in folk medicine as remedy for kidney, bladder and prostate disorders since centuries. Nevertheless, pumpkin research provides insufficient data to back up traditional beliefs of ethnomedical practice. The bioactivity of a hydro-ethanolic extract of pumpkin seeds from the Styrian pumpkin, Cucurbita pepo L. subsp. pepo var. styriaca, was investigated. As pumpkin seed extracts are standardized to cucurbitin, this compound was also tested. Transactivational activity was evaluated for human androgen receptor, estrogen receptor and progesterone receptor with in vitro yeast assays. Cell viability tests with prostate cancer cells, breast cancer cells, colorectal adenocarcinoma cells and a hyperplastic cell line from benign prostate hyperplasia tissue were performed. As model for non-hyperplastic cells, effects on cell viability were tested with a human dermal fibroblast cell line (HDF-5). No transactivational activity was found for human androgen receptor, estrogen receptor and progesterone receptor, for both, extract and cucurbitin. A cell growth inhibition of ~40-50% was observed for all cell lines, with the exception of HDF-5, which showed with ~20% much lower cell growth inhibition. Given the receptor status of some cell lines, a steroid-hormone receptor independent growth inhibiting effect can be assumed. The cell growth inhibition for fast growing cells together with the cell growth inhibition of prostate-, breast- and colon cancer cells corroborates the ethnomedical use of pumpkin seeds for a treatment of benign prostate hyperplasia. Moreover, due to the lack of androgenic activity, pumpkin seed applications can be regarded as safe for the prostate. Copyright © 2016 The Authors. Published by Elsevier B.V. All rights reserved.

  17. Quality indicators to compare accredited independent pharmacies and accredited chain pharmacies in Thailand.

    Science.gov (United States)

    Arkaravichien, Wiwat; Wongpratat, Apichaya; Lertsinudom, Sunee

    2016-08-01

    Background Quality indicators determine the quality of actual practice in reference to standard criteria. The Community Pharmacy Association (Thailand), with technical support from the International Pharmaceutical Federation, developed a tool for quality assessment and quality improvement at community pharmacies. This tool has passed validity and reliability tests, but has not yet had feasibility testing. Objective (1) To test whether this quality tool could be used in routine settings. (2) To compare quality scores between accredited independent and accredited chain pharmacies. Setting Accredited independent pharmacies and accredited chain pharmacies in the north eastern region of Thailand. Methods A cross sectional study was conducted in 34 accredited independent pharmacies and accredited chain pharmacies. Quality scores were assessed by observation and by interviewing the responsible pharmacists. Data were collected and analyzed by independent t-test and Mann-Whitney U test as appropriate. Results were plotted by histogram and spider chart. Main outcome measure Domain's assessable scores, possible maximum scores, mean and median of measured scores. Results Domain's assessable scores were close to domain's possible maximum scores. This meant that most indicators could be assessed in most pharmacies. The spider chart revealed that measured scores in the personnel, drug inventory and stocking, and patient satisfaction and health promotion domains of chain pharmacies were significantly higher than those of independent pharmacies (p pharmacies and chain pharmacies in the premise and facility or dispensing and patient care domains. Conclusion Quality indicators developed by the Community Pharmacy Association (Thailand) could be used to assess quality of practice in pharmacies in routine settings. It is revealed that the quality scores of chain pharmacies were higher than those of independent pharmacies.

  18. Extracting of implicit information in English advertising texts with phonetic and lexical-morphological means

    Directory of Open Access Journals (Sweden)

    Traikovskaya Natalya Petrovna

    2015-12-01

    Full Text Available The article deals with phonetic and lexical-morphological language means participating in the process of extracting implicit information in English-speaking advertising texts for men and women. The functioning of phonetic means of the English language is not the basis for implication of information in advertising texts. Lexical and morphological means play the role of markers of relevant information, playing the role of the activator ofimplicit information in the texts of advertising.

  19. Towards a Relation Extraction Framework for Cyber-Security Concepts

    Energy Technology Data Exchange (ETDEWEB)

    Jones, Corinne L [ORNL; Bridges, Robert A [ORNL; Huffer, Kelly M [ORNL; Goodall, John R [ORNL

    2015-01-01

    In order to assist security analysts in obtaining information pertaining to their network, such as novel vulnerabilities, exploits, or patches, information retrieval methods tailored to the security domain are needed. As labeled text data is scarce and expensive, we follow developments in semi-supervised NLP and implement a bootstrapping algorithm for extracting security entities and their relationships from text. The algorithm requires little input data, specifically, a few relations or patterns (heuristics for identifying relations), and incorporates an active learning component which queries the user on the most important decisions to prevent drifting the desired relations. Preliminary testing on a small corpus shows promising results, obtaining precision of .82.

  20. THE USE OF ENGLISH-LOAN WORDS IN THE INFORMATION TECHNOLOGY DOMAIN

    Directory of Open Access Journals (Sweden)

    Abrudan Caciora Simona Veronica

    2015-07-01

    Full Text Available This paper examines the phenomenon of linguistic borrowings in general and the use of anglicisms in particular, in the contexts of globalization and the progress made in different domains, especially in the Information Technology domain. Borrowing is the phenomenon which undertakes the word transmission process from one language into another among the connected communities. The language that lends is the donor or the borrower language, and the language that borrows is known as the receptor or the borrowing language. Anglicisms are the English-loan words that are present in many European languages. The phenomenon of globalisation is the main topic of several contradictory and polarization discussions and undoubtedly globalisation is necessary and it is a fact in today’s world but the question is: what will be the long-term language consequences? There might be two situations: the consequences may be positive, such as ease of communication and perception, language economics and also negative: not understanding anglicisms resulting in disturbed communication, rejection of anglicisms, loss of semantic diversity of your own language, less and less possibilities of expressing yourself in your own language. We can mention that responsible for the impact of English is the prestige associated with this language. This language is generally held in highest regard by the younger generation of Romanians, who are easily influenced by the television, radio, movies, and other media that are supplied by British and American entertainment, which are considered to be trendy and cool. Each of those resources has influenced the Romanian population on numerous levels, creating new words and phrases in the Romanian language. In order to illustrate and discuss the pertinent or, on the contrary, the unjustified use of English-loan words in the field of Information Technology, the study will focus on a selection of articles published in the electronic Romanian issues

  1. Methods from Information Extraction from LIDAR Intensity Data and Multispectral LIDAR Technology

    Science.gov (United States)

    Scaioni, M.; Höfle, B.; Baungarten Kersting, A. P.; Barazzetti, L.; Previtali, M.; Wujanz, D.

    2018-04-01

    LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on `Information Extraction from LiDAR Intensity Data' has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  2. METHODS FROM INFORMATION EXTRACTION FROM LIDAR INTENSITY DATA AND MULTISPECTRAL LIDAR TECHNOLOGY

    Directory of Open Access Journals (Sweden)

    M. Scaioni

    2018-04-01

    Full Text Available LiDAR is a consolidated technology for topographic mapping and 3D reconstruction, which is implemented in several platforms On the other hand, the exploitation of the geometric information has been coupled by the use of laser intensity, which may provide additional data for multiple purposes. This option has been emphasized by the availability of sensors working on different wavelength, thus able to provide additional information for classification of surfaces and objects. Several applications ofmonochromatic and multi-spectral LiDAR data have been already developed in different fields: geosciences, agriculture, forestry, building and cultural heritage. The use of intensity data to extract measures of point cloud quality has been also developed. The paper would like to give an overview on the state-of-the-art of these techniques, and to present the modern technologies for the acquisition of multispectral LiDAR data. In addition, the ISPRS WG III/5 on ‘Information Extraction from LiDAR Intensity Data’ has collected and made available a few open data sets to support scholars to do research on this field. This service is presented and data sets delivered so far as are described.

  3. A novel nuclear DnaJ protein, DNAJC8, can suppress the formation of spinocerebellar ataxia 3 polyglutamine aggregation in a J-domain independent manner

    Energy Technology Data Exchange (ETDEWEB)

    Ito, Norie [Department of Pathology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan); Department of Neurology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan); Kamiguchi, Kenjiro; Nakanishi, Katsuya; Sokolovskya, Alice; Hirohashi, Yoshihiko; Tamura, Yasuaki; Murai, Aiko; Yamamoto, Eri; Kanaseki, Takayuki; Tsukahara, Tomohide; Kochin, Vitaly [Department of Pathology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan); Chiba, Susumu [Department of Neurology, Clinical Brain Research Laboratory, Toyokura Memorial Hall, Sapporo Yamano-ue Hospital (Japan); Shimohama, Shun [Department of Neurology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan); Sato, Noriyuki [Department of Pathology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan); Torigoe, Toshihiko, E-mail: torigoe@sapmed.ac.jp [Department of Pathology, Sapporo Medical University School of Medicine, South-1, West-17, Chuo-ku, Sapporo 060-8556 (Japan)

    2016-06-10

    Polyglutamine (polyQ) diseases comprise neurodegenerative disorders caused by expression of expanded polyQ-containing proteins. The cytotoxicity of the expanded polyQ-containing proteins is closely associated with aggregate formation. In this study, we report that a novel J-protein, DNAJ (HSP40) Homolog, Subfamily C, Member 8 (DNAJC8), suppresses the aggregation of polyQ-containing protein in a cellular model of spinocerebellar ataxia type 3 (SCA3), which is also known as Machado-Joseph disease. Overexpression of DNAJC8 in SH-SY5Y neuroblastoma cells significantly reduced the polyQ aggregation and apoptosis, and DNAJC8 was co-localized with the polyQ aggregation in the cell nucleus. Deletion mutants of DNAJC8 revealed that the C-terminal domain of DNAJC8 was essential for the suppression of polyQ aggregation, whereas the J-domain was dispensable. Furthermore, 22-mer oligopeptide derived from C-termilal domain could suppress the polyQ aggregation. These results indicate that DNAJC8 can suppress the polyQ aggregation via a distinct mechanism independent of HSP70-based chaperone machinery and have a unique protective role against the aggregation of expanded polyQ-containing proteins such as pathogenic ataxin-3 proteins.

  4. Yeast lipids can phase separate into micrometer-scale membrane domains

    DEFF Research Database (Denmark)

    Klose, Christian; Ejsing, Christer S; Garcia-Saez, Ana J

    2010-01-01

    The lipid raft concept proposes that biological membranes have the potential to form functional domains based on a selective interaction between sphingolipids and sterols. These domains seem to be involved in signal transduction and vesicular sorting of proteins and lipids. Although there is bioc......The lipid raft concept proposes that biological membranes have the potential to form functional domains based on a selective interaction between sphingolipids and sterols. These domains seem to be involved in signal transduction and vesicular sorting of proteins and lipids. Although...... there is biochemical evidence for lipid raft-dependent protein and lipid sorting in the yeast Saccharomyces cerevisiae, direct evidence for an interaction between yeast sphingolipids and the yeast sterol ergosterol, resulting in membrane domain formation, is lacking. Here we show that model membranes formed from yeast...... total lipid extracts possess an inherent self-organization potential resulting in Ld-Lo phase coexistence at physiologically relevant temperature. Analyses of lipid extracts from mutants defective in sphingolipid metabolism as well as reconstitution of purified yeast lipids in model membranes of defined...

  5. Phylogeny of the TRAF/MATH domain.

    Science.gov (United States)

    Zapata, Juan M; Martínez-García, Vanesa; Lefebvre, Sophie

    2007-01-01

    The TNF-receptor associated factor (TRAF) domain (TD), also known as the meprin and TRAF-C homology (MATH) domain is a fold of seven anti-parallel p-helices that participates in protein-protein interactions. This fold is broadly represented among eukaryotes, where it is found associated with a discrete set of protein-domains. Virtually all protein families encompassing a TRAF/MATH domain seem to be involved in the regulation of protein processing and ubiquitination, strongly suggesting a parallel evolution of the TRAF/MATH domain and certain proteolysis pathways in eukaryotes. The restricted number of living organisms for which we have information of their genetic and protein make-up limits the scope and analysis of the MATH domain in evolution. However, the available information allows us to get a glimpse on the origins, distribution and evolution of the TRAF/MATH domain, which will be overviewed in this chapter.

  6. Terminology extraction from medical texts in Polish.

    Science.gov (United States)

    Marciniak, Małgorzata; Mykowiecka, Agnieszka

    2014-01-01

    Hospital documents contain free text describing the most important facts relating to patients and their illnesses. These documents are written in specific language containing medical terminology related to hospital treatment. Their automatic processing can help in verifying the consistency of hospital documentation and obtaining statistical data. To perform this task we need information on the phrases we are looking for. At the moment, clinical Polish resources are sparse. The existing terminologies, such as Polish Medical Subject Headings (MeSH), do not provide sufficient coverage for clinical tasks. It would be helpful therefore if it were possible to automatically prepare, on the basis of a data sample, an initial set of terms which, after manual verification, could be used for the purpose of information extraction. Using a combination of linguistic and statistical methods for processing over 1200 children hospital discharge records, we obtained a list of single and multiword terms used in hospital discharge documents written in Polish. The phrases are ordered according to their presumed importance in domain texts measured by the frequency of use of a phrase and the variety of its contexts. The evaluation showed that the automatically identified phrases cover about 84% of terms in domain texts. At the top of the ranked list, only 4% out of 400 terms were incorrect while out of the final 200, 20% of expressions were either not domain related or syntactically incorrect. We also observed that 70% of the obtained terms are not included in the Polish MeSH. Automatic terminology extraction can give results which are of a quality high enough to be taken as a starting point for building domain related terminological dictionaries or ontologies. This approach can be useful for preparing terminological resources for very specific subdomains for which no relevant terminologies already exist. The evaluation performed showed that none of the tested ranking procedures were

  7. Model-independent study of light cone current commutators

    International Nuclear Information System (INIS)

    Gautam, S.R.; Dicus, D.A.

    1974-01-01

    An attempt is made to extract information on the nature of light cone current commutators (L. C. C.) in a model independent manner. Using simple assumptions on the validity of the DGS representation for the structure functions of deep inelastic scattering and using the Bjorken--Johnston--Low theorem it is shown that in principle the L. C. C. may be constructed knowing the experimental electron--proton scattering data. On the other hand the scaling behavior of the structure functions is utilized to study the consistency of a vanishing value for various L. C. C. under mild assumptions on the behavior of the DGS spectral moments. (U.S.)

  8. Evaluation of Keyphrase Extraction Algorithm and Tiling Process for a Document/Resource Recommender within E-Learning Environments

    Science.gov (United States)

    Mangina, Eleni; Kilbride, John

    2008-01-01

    The research presented in this paper is an examination of the applicability of IUI techniques in an online e-learning environment. In particular we make use of user modeling techniques, information retrieval and extraction mechanisms and collaborative filtering methods. The domains of e-learning, web-based training and instruction and intelligent…

  9. An Effective Experimental Optimization Method for Wireless Power Transfer System Design Using Frequency Domain Measurement

    Directory of Open Access Journals (Sweden)

    Sangyeong Jeong

    2017-10-01

    Full Text Available This paper proposes an experimental optimization method for a wireless power transfer (WPT system. The power transfer characteristics of a WPT system with arbitrary loads and various types of coupling and compensation networks can be extracted by frequency domain measurements. The various performance parameters of the WPT system, such as input real/imaginary/apparent power, power factor, efficiency, output power and voltage gain, can be accurately extracted in a frequency domain by a single passive measurement. Subsequently, the design parameters can be efficiently tuned by separating the overall design steps into two parts. The extracted performance parameters of the WPT system were validated with time-domain experiments.

  10. Testing the reliability of information extracted from ancient zircon

    Science.gov (United States)

    Kielman, Ross; Whitehouse, Martin; Nemchin, Alexander

    2015-04-01

    Studies combining zircon U-Pb chronology, trace element distribution as well as O and Hf isotope systematics are a powerful way to gain understanding of the processes shaping Earth's evolution, especially in detrital populations where constraints from the original host are missing. Such studies of the Hadean detrital zircon population abundant in sedimentary rocks in Western Australia have involved analysis of an unusually large number of individual grains, but also highlighted potential problems with the approach, only apparent when multiple analyses are obtained from individual grains. A common feature of the Hadean as well as many early Archaean zircon populations is their apparent inhomogeneity, which reduces confidence in conclusions based on studies combining chemistry and isotopic characteristics of zircon. In order to test the reliability of information extracted from early Earth zircon, we report results from one of the first in-depth multi-method study of zircon from a relatively simple early Archean magmatic rock, used as an analogue to ancient detrital zircon. The approach involves making multiple SIMS analyses in individual grains in order to be comparable to the most advanced studies of detrital zircon populations. The investigated sample is a relatively undeformed, non-migmatitic ca. 3.8 Ga tonalite collected a few kms south of the Isua Greenstone Belt, southwest Greenland. Extracted zircon grains can be combined into three different groups based on the behavior of their U-Pb systems: (i) grains that show internally consistent and concordant ages and define an average age of 3805±15 Ma, taken to be the age of the rock, (ii) grains that are distributed close to the concordia line, but with significant variability between multiple analyses, suggesting an ancient Pb loss and (iii) grains that have multiple analyses distributed along a discordia pointing towards a zero intercept, indicating geologically recent Pb-loss. This overall behavior has

  11. NovelFam3000 – Uncharacterized human protein domains conserved across model organisms

    Science.gov (United States)

    Kemmer, Danielle; Podowski, Raf M; Arenillas, David; Lim, Jonathan; Hodges, Emily; Roth, Peggy; Sonnhammer, Erik LL; Höög, Christer; Wasserman, Wyeth W

    2006-01-01

    Background Despite significant efforts from the research community, an extensive portion of the proteins encoded by human genes lack an assigned cellular function. Most metazoan proteins are composed of structural and/or functional domains, of which many appear in multiple proteins. Once a domain is characterized in one protein, the presence of a similar sequence in an uncharacterized protein serves as a basis for inference of function. Thus knowledge of a domain's function, or the protein within which it arises, can facilitate the analysis of an entire set of proteins. Description From the Pfam domain database, we extracted uncharacterized protein domains represented in proteins from humans, worms, and flies. A data centre was created to facilitate the analysis of the uncharacterized domain-containing proteins. The centre both provides researchers with links to dispersed internet resources containing gene-specific experimental data and enables them to post relevant experimental results or comments. For each human gene in the system, a characterization score is posted, allowing users to track the progress of characterization over time or to identify for study uncharacterized domains in well-characterized genes. As a test of the system, a subset of 39 domains was selected for analysis and the experimental results posted to the NovelFam3000 system. For 25 human protein members of these 39 domain families, detailed sub-cellular localizations were determined. Specific observations are presented based on the analysis of the integrated information provided through the online NovelFam3000 system. Conclusion Consistent experimental results between multiple members of a domain family allow for inferences of the domain's functional role. We unite bioinformatics resources and experimental data in order to accelerate the functional characterization of scarcely annotated domain families. PMID:16533400

  12. NovelFam3000 – Uncharacterized human protein domains conserved across model organisms

    Directory of Open Access Journals (Sweden)

    Sonnhammer Erik LL

    2006-03-01

    Full Text Available Abstract Background Despite significant efforts from the research community, an extensive portion of the proteins encoded by human genes lack an assigned cellular function. Most metazoan proteins are composed of structural and/or functional domains, of which many appear in multiple proteins. Once a domain is characterized in one protein, the presence of a similar sequence in an uncharacterized protein serves as a basis for inference of function. Thus knowledge of a domain's function, or the protein within which it arises, can facilitate the analysis of an entire set of proteins. Description From the Pfam domain database, we extracted uncharacterized protein domains represented in proteins from humans, worms, and flies. A data centre was created to facilitate the analysis of the uncharacterized domain-containing proteins. The centre both provides researchers with links to dispersed internet resources containing gene-specific experimental data and enables them to post relevant experimental results or comments. For each human gene in the system, a characterization score is posted, allowing users to track the progress of characterization over time or to identify for study uncharacterized domains in well-characterized genes. As a test of the system, a subset of 39 domains was selected for analysis and the experimental results posted to the NovelFam3000 system. For 25 human protein members of these 39 domain families, detailed sub-cellular localizations were determined. Specific observations are presented based on the analysis of the integrated information provided through the online NovelFam3000 system. Conclusion Consistent experimental results between multiple members of a domain family allow for inferences of the domain's functional role. We unite bioinformatics resources and experimental data in order to accelerate the functional characterization of scarcely annotated domain families.

  13. Warfighter information services: lessons learned in the intelligence domain

    Science.gov (United States)

    Bray, S. E.

    2014-05-01

    A vision was presented in a previous paper of how a common set of services within a framework could be used to provide all the information processing needs of Warfighters. Central to that vision was the concept of a "Virtual Knowledge Base". The paper presents an implementation of these ideas in the intelligence domain. Several innovative technologies were employed in the solution, which are presented and their benefits explained. The project was successful, validating many of the design principles for such a system which had been proposed in earlier work. Many of these principles are discussed in detail, explaining lessons learned. The results showed that it is possible to make vast improvements in the ability to exploit available data, making it discoverable and queryable wherever it is from anywhere within a participating network; and to exploit machine reasoning to make faster and better inferences from available data, enabling human analysts to spend more of their time doing more difficult analytical tasks rather than searching for relevant data. It was also demonstrated that a small number of generic Information Processing services can be combined and configured in a variety of ways (without changing any software code) to create "fact-processing" workflows, in this case to create different intelligence analysis capabilities. It is yet to be demonstrated that the same generic services can be reused to create analytical/situational awareness capabilities for logistics, operations, planning or other military functions but this is considered likely.

  14. Layered and Laterally Constrained 2D Inversion of Time Domain Induced Polarization Data

    DEFF Research Database (Denmark)

    Fiandaca, Gianluca; Ramm, James; Auken, Esben

    description of the transmitter waveform and of the receiver transfer function allowing for a quantitative interpretation of the parameters. The code has been optimized for parallel computation and the inversion time is comparable to codes inverting just for direct current resistivity. The new inversion......In a sedimentary environment, quasi-layered models often represent the actual geology more accurately than smooth minimum-structure models. We have developed a new layered and laterally constrained inversion algorithm for time domain induced polarization data. The algorithm is based on the time...... transform of a complex resistivity forward response and the inversion extracts the spectral information of the time domain measures in terms of the Cole-Cole parameters. The developed forward code and inversion algorithm use the full time decay of the induced polarization response, together with an accurate...

  15. Definition of an automatic information retrieval system independent from the data base used

    International Nuclear Information System (INIS)

    Cunha, E.R.

    1983-04-01

    A bibliographic information retrieval system using data stored at the standardized interchange format ISO 2709 or ANSI Z39.2, is specified. A set of comands for interchange format manipulation wich allows the data access at the logical level, achieving the data independence, are used. A data base description language, a storage structure and data base manipulation comands are specified, using retrieval techniques which consider the applications needs. (Author) [pt

  16. Knowledge-driven information mining in remote-sensing image archives

    Science.gov (United States)

    Datcu, M.; Seidel, K.; D'Elia, S.; Marchetti, P. G.

    2002-05-01

    Users in all domains require information or information-related services that are focused, concise, reliable, low cost and timely and which are provided in forms and formats compatible with the user's own activities. In the current Earth Observation (EO) scenario, the archiving centres generally only offer data, images and other "low level" products. The user's needs are being only partially satisfied by a number of, usually small, value-adding companies applying time-consuming (mostly manual) and expensive processes relying on the knowledge of experts to extract information from those data or images.

  17. Automated analysis of information processing, kinetic independence and modular architecture in biochemical networks using MIDIA.

    Science.gov (United States)

    Bowsher, Clive G

    2011-02-15

    Understanding the encoding and propagation of information by biochemical reaction networks and the relationship of such information processing properties to modular network structure is of fundamental importance in the study of cell signalling and regulation. However, a rigorous, automated approach for general biochemical networks has not been available, and high-throughput analysis has therefore been out of reach. Modularization Identification by Dynamic Independence Algorithms (MIDIA) is a user-friendly, extensible R package that performs automated analysis of how information is processed by biochemical networks. An important component is the algorithm's ability to identify exact network decompositions based on both the mass action kinetics and informational properties of the network. These modularizations are visualized using a tree structure from which important dynamic conditional independence properties can be directly read. Only partial stoichiometric information needs to be used as input to MIDIA, and neither simulations nor knowledge of rate parameters are required. When applied to a signalling network, for example, the method identifies the routes and species involved in the sequential propagation of information between its multiple inputs and outputs. These routes correspond to the relevant paths in the tree structure and may be further visualized using the Input-Output Path Matrix tool. MIDIA remains computationally feasible for the largest network reconstructions currently available and is straightforward to use with models written in Systems Biology Markup Language (SBML). The package is distributed under the GNU General Public License and is available, together with a link to browsable Supplementary Material, at http://code.google.com/p/midia. Further information is at www.maths.bris.ac.uk/~macgb/Software.html.

  18. Domain-Specific Control of Selective Attention

    Science.gov (United States)

    Lin, Szu-Hung; Yeh, Yei-Yu

    2014-01-01

    Previous research has shown that loading information on working memory affects selective attention. However, whether the load effect on selective attention is domain-general or domain-specific remains unresolved. The domain-general effect refers to the findings that load in one content (e.g. phonological) domain in working memory influences processing in another content (e.g., visuospatial) domain. Attentional control supervises selection regardless of information domain. The domain-specific effect refers to the constraint of influence only when maintenance and processing operate in the same domain. Selective attention operates in a specific content domain. This study is designed to resolve this controversy. Across three experiments, we manipulated the type of representation maintained in working memory and the type of representation upon which the participants must exert control to resolve conflict and select a target into the focus of attention. In Experiments 1a and 1b, participants maintained digits and nonverbalized objects, respectively, in working memory while selecting a target in a letter array. In Experiment 2, we presented auditory digits with a letter flanker task to exclude the involvement of resource competition within the same input modality. In Experiments 3a and 3b, we replaced the letter flanker task with an object flanker task while manipulating the memory load on object and digit representation, respectively. The results consistently showed that memory load modulated distractibility only when the stimuli of the two tasks were represented in the same domain. The magnitude of distractor interference was larger under high load than under low load, reflecting a lower efficacy of information prioritization. When the stimuli of the two tasks were represented in different domains, memory load did not modulate distractibility. Control of processing priority in selective attention demands domain-specific resources. PMID:24866977

  19. Statistical techniques to extract information during SMAP soil moisture assimilation

    Science.gov (United States)

    Kolassa, J.; Reichle, R. H.; Liu, Q.; Alemohammad, S. H.; Gentine, P.

    2017-12-01

    Statistical techniques permit the retrieval of soil moisture estimates in a model climatology while retaining the spatial and temporal signatures of the satellite observations. As a consequence, the need for bias correction prior to an assimilation of these estimates is reduced, which could result in a more effective use of the independent information provided by the satellite observations. In this study, a statistical neural network (NN) retrieval algorithm is calibrated using SMAP brightness temperature observations and modeled soil moisture estimates (similar to those used to calibrate the SMAP Level 4 DA system). Daily values of surface soil moisture are estimated using the NN and then assimilated into the NASA Catchment model. The skill of the assimilation estimates is assessed based on a comprehensive comparison to in situ measurements from the SMAP core and sparse network sites as well as the International Soil Moisture Network. The NN retrieval assimilation is found to significantly improve the model skill, particularly in areas where the model does not represent processes related to agricultural practices. Additionally, the NN method is compared to assimilation experiments using traditional bias correction techniques. The NN retrieval assimilation is found to more effectively use the independent information provided by SMAP resulting in larger model skill improvements than assimilation experiments using traditional bias correction techniques.

  20. A Joint Method of Envelope Inversion Combined with Hybrid-domain Full Waveform Inversion

    Science.gov (United States)

    CUI, C.; Hou, W.

    2017-12-01

    Full waveform inversion (FWI) aims to construct high-precision subsurface models by fully using the information in seismic records, including amplitude, travel time, phase and so on. However, high non-linearity and the absence of low frequency information in seismic data lead to the well-known cycle skipping problem and make inversion easily fall into local minima. In addition, those 3D inversion methods that are based on acoustic approximation ignore the elastic effects in real seismic field, and make inversion harder. As a result, the accuracy of final inversion results highly relies on the quality of initial model. In order to improve stability and quality of inversion results, multi-scale inversion that reconstructs subsurface model from low to high frequency are applied. But, the absence of very low frequencies (time domain and inversion in the frequency domain. To accelerate the inversion, we adopt CPU/GPU heterogeneous computing techniques. There were two levels of parallelism. In the first level, the inversion tasks are decomposed and assigned to each computation node by shot number. In the second level, GPU multithreaded programming is used for the computation tasks in each node, including forward modeling, envelope extraction, DFT (discrete Fourier transform) calculation and gradients calculation. Numerical tests demonstrated that the combined envelope inversion + hybrid-domain FWI could obtain much faithful and accurate result than conventional hybrid-domain FWI. The CPU/GPU heterogeneous parallel computation could improve the performance speed.

  1. A Reference Architecture for Space Information Management

    Science.gov (United States)

    Mattmann, Chris A.; Crichton, Daniel J.; Hughes, J. Steven; Ramirez, Paul M.; Berrios, Daniel C.

    2006-01-01

    We describe a reference architecture for space information management systems that elegantly overcomes the rigid design of common information systems in many domains. The reference architecture consists of a set of flexible, reusable, independent models and software components that function in unison, but remain separately managed entities. The main guiding principle of the reference architecture is to separate the various models of information (e.g., data, metadata, etc.) from implemented system code, allowing each to evolve independently. System modularity, systems interoperability, and dynamic evolution of information system components are the primary benefits of the design of the architecture. The architecture requires the use of information models that are substantially more advanced than those used by the vast majority of information systems. These models are more expressive and can be more easily modularized, distributed and maintained than simpler models e.g., configuration files and data dictionaries. Our current work focuses on formalizing the architecture within a CCSDS Green Book and evaluating the architecture within the context of the C3I initiative.

  2. Dominant integration locus drives continuous diversification of plant immune receptors with exogenous domain fusions.

    Science.gov (United States)

    Bailey, Paul C; Schudoma, Christian; Jackson, William; Baggs, Erin; Dagdas, Gulay; Haerty, Wilfried; Moscou, Matthew; Krasileva, Ksenia V

    2018-02-19

    The plant immune system is innate and encoded in the germline. Using it efficiently, plants are capable of recognizing a diverse range of rapidly evolving pathogens. A recently described phenomenon shows that plant immune receptors are able to recognize pathogen effectors through the acquisition of exogenous protein domains from other plant genes. We show that plant immune receptors with integrated domains are distributed unevenly across their phylogeny in grasses. Using phylogenetic analysis, we uncover a major integration clade, whose members underwent repeated independent integration events producing diverse fusions. This clade is ancestral in grasses with members often found on syntenic chromosomes. Analyses of these fusion events reveals that homologous receptors can be fused to diverse domains. Furthermore, we discover a 43 amino acid long motif associated with this dominant integration clade which is located immediately upstream of the fusion site. Sequence analysis reveals that DNA transposition and/or ectopic recombination are the most likely mechanisms of formation for nucleotide binding leucine rich repeat proteins with integrated domains. The identification of this subclass of plant immune receptors that is naturally adapted to new domain integration will inform biotechnological approaches for generating synthetic receptors with novel pathogen "baits."

  3. Automated concept-level information extraction to reduce the need for custom software and rules development.

    Science.gov (United States)

    D'Avolio, Leonard W; Nguyen, Thien M; Goryachev, Sergey; Fiore, Louis D

    2011-01-01

    Despite at least 40 years of promising empirical performance, very few clinical natural language processing (NLP) or information extraction systems currently contribute to medical science or care. The authors address this gap by reducing the need for custom software and rules development with a graphical user interface-driven, highly generalizable approach to concept-level retrieval. A 'learn by example' approach combines features derived from open-source NLP pipelines with open-source machine learning classifiers to automatically and iteratively evaluate top-performing configurations. The Fourth i2b2/VA Shared Task Challenge's concept extraction task provided the data sets and metrics used to evaluate performance. Top F-measure scores for each of the tasks were medical problems (0.83), treatments (0.82), and tests (0.83). Recall lagged precision in all experiments. Precision was near or above 0.90 in all tasks. Discussion With no customization for the tasks and less than 5 min of end-user time to configure and launch each experiment, the average F-measure was 0.83, one point behind the mean F-measure of the 22 entrants in the competition. Strong precision scores indicate the potential of applying the approach for more specific clinical information extraction tasks. There was not one best configuration, supporting an iterative approach to model creation. Acceptable levels of performance can be achieved using fully automated and generalizable approaches to concept-level information extraction. The described implementation and related documentation is available for download.

  4. Extraction of quantifiable information from complex systems

    CERN Document Server

    Dahmen, Wolfgang; Griebel, Michael; Hackbusch, Wolfgang; Ritter, Klaus; Schneider, Reinhold; Schwab, Christoph; Yserentant, Harry

    2014-01-01

    In April 2007, the  Deutsche Forschungsgemeinschaft (DFG) approved the  Priority Program 1324 “Mathematical Methods for Extracting Quantifiable Information from Complex Systems.” This volume presents a comprehensive overview of the most important results obtained over the course of the program.   Mathematical models of complex systems provide the foundation for further technological developments in science, engineering and computational finance.  Motivated by the trend toward steadily increasing computer power, ever more realistic models have been developed in recent years. These models have also become increasingly complex, and their numerical treatment poses serious challenges.   Recent developments in mathematics suggest that, in the long run, much more powerful numerical solution strategies could be derived if the interconnections between the different fields of research were systematically exploited at a conceptual level. Accordingly, a deeper understanding of the mathematical foundations as w...

  5. THE EXTRACTION OF INDOOR BUILDING INFORMATION FROM BIM TO OGC INDOORGML

    Directory of Open Access Journals (Sweden)

    T.-A. Teo

    2017-07-01

    Full Text Available Indoor Spatial Data Infrastructure (indoor-SDI is an important SDI for geosptial analysis and location-based services. Building Information Model (BIM has high degree of details in geometric and semantic information for building. This study proposed direct conversion schemes to extract indoor building information from BIM to OGC IndoorGML. The major steps of the research include (1 topological conversion from building model into indoor network model; and (2 generation of IndoorGML. The topological conversion is a major process of generating and mapping nodes and edges from IFC to indoorGML. Node represents every space (e.g. IfcSpace and objects (e.g. IfcDoor in the building while edge shows the relationships between nodes. According to the definition of IndoorGML, the topological model in the dual space is also represented as a set of nodes and edges. These definitions of IndoorGML are the same as in the indoor network. Therefore, we can extract the necessary data in the indoor network and easily convert them into IndoorGML based on IndoorGML Schema. The experiment utilized a real BIM model to examine the proposed method. The experimental results indicated that the 3D indoor model (i.e. IndoorGML model can be automatically imported from IFC model by the proposed procedure. In addition, the geometric and attribute of building elements are completely and correctly converted from BIM to indoor-SDI.

  6. Emotion Recognition from EEG Signals Using Multidimensional Information in EMD Domain.

    Science.gov (United States)

    Zhuang, Ning; Zeng, Ying; Tong, Li; Zhang, Chi; Zhang, Hanming; Yan, Bin

    2017-01-01

    This paper introduces a method for feature extraction and emotion recognition based on empirical mode decomposition (EMD). By using EMD, EEG signals are decomposed into Intrinsic Mode Functions (IMFs) automatically. Multidimensional information of IMF is utilized as features, the first difference of time series, the first difference of phase, and the normalized energy. The performance of the proposed method is verified on a publicly available emotional database. The results show that the three features are effective for emotion recognition. The role of each IMF is inquired and we find that high frequency component IMF1 has significant effect on different emotional states detection. The informative electrodes based on EMD strategy are analyzed. In addition, the classification accuracy of the proposed method is compared with several classical techniques, including fractal dimension (FD), sample entropy, differential entropy, and discrete wavelet transform (DWT). Experiment results on DEAP datasets demonstrate that our method can improve emotion recognition performance.

  7. A Domain Standard for Land Administration

    NARCIS (Netherlands)

    Lemmen, C.; Van Oosterom, P.; Van der Molen, P.

    2013-01-01

    This paper presents the design of a Domain Model for Land Administration (LA). As a result a formal International Standard is available: ISO 19152 Geographic Information – Land Administration Domain Model (LADM) (ISO, 2012). Domain specific standardisation is needed to capture the semantics of the

  8. Flexible time domain averaging technique

    Science.gov (United States)

    Zhao, Ming; Lin, Jing; Lei, Yaguo; Wang, Xiufeng

    2013-09-01

    Time domain averaging(TDA) is essentially a comb filter, it cannot extract the specified harmonics which may be caused by some faults, such as gear eccentric. Meanwhile, TDA always suffers from period cutting error(PCE) to different extent. Several improved TDA methods have been proposed, however they cannot completely eliminate the waveform reconstruction error caused by PCE. In order to overcome the shortcomings of conventional methods, a flexible time domain averaging(FTDA) technique is established, which adapts to the analyzed signal through adjusting each harmonic of the comb filter. In this technique, the explicit form of FTDA is first constructed by frequency domain sampling. Subsequently, chirp Z-transform(CZT) is employed in the algorithm of FTDA, which can improve the calculating efficiency significantly. Since the signal is reconstructed in the continuous time domain, there is no PCE in the FTDA. To validate the effectiveness of FTDA in the signal de-noising, interpolation and harmonic reconstruction, a simulated multi-components periodic signal that corrupted by noise is processed by FTDA. The simulation results show that the FTDA is capable of recovering the periodic components from the background noise effectively. Moreover, it can improve the signal-to-noise ratio by 7.9 dB compared with conventional ones. Experiments are also carried out on gearbox test rigs with chipped tooth and eccentricity gear, respectively. It is shown that the FTDA can identify the direction and severity of the eccentricity gear, and further enhances the amplitudes of impulses by 35%. The proposed technique not only solves the problem of PCE, but also provides a useful tool for the fault symptom extraction of rotating machinery.

  9. Recurrent adenylation domain replacement in the microcystin synthetase gene cluster

    Directory of Open Access Journals (Sweden)

    Laakso Kati

    2007-10-01

    Full Text Available Abstract Background Microcystins are small cyclic heptapeptide toxins produced by a range of distantly related cyanobacteria. Microcystins are synthesized on large NRPS-PKS enzyme complexes. Many structural variants of microcystins are produced simulatenously. A recombination event between the first module of mcyB (mcyB1 and mcyC in the microcystin synthetase gene cluster is linked to the simultaneous production of microcystin variants in strains of the genus Microcystis. Results Here we undertook a phylogenetic study to investigate the order and timing of recombination between the mcyB1 and mcyC genes in a diverse selection of microcystin producing cyanobacteria. Our results provide support for complex evolutionary processes taking place at the mcyB1 and mcyC adenylation domains which recognize and activate the amino acids found at X and Z positions. We find evidence for recent recombination between mcyB1 and mcyC in strains of the genera Anabaena, Microcystis, and Hapalosiphon. We also find clear evidence for independent adenylation domain conversion of mcyB1 by unrelated peptide synthetase modules in strains of the genera Nostoc and Microcystis. The recombination events replace only the adenylation domain in each case and the condensation domains of mcyB1 and mcyC are not transferred together with the adenylation domain. Our findings demonstrate that the mcyB1 and mcyC adenylation domains are recombination hotspots in the microcystin synthetase gene cluster. Conclusion Recombination is thought to be one of the main mechanisms driving the diversification of NRPSs. However, there is very little information on how recombination takes place in nature. This study demonstrates that functional peptide synthetases are created in nature through transfer of adenylation domains without the concomitant transfer of condensation domains.

  10. MedEx: a medication information extraction system for clinical narratives

    Science.gov (United States)

    Stenner, Shane P; Doan, Son; Johnson, Kevin B; Waitman, Lemuel R; Denny, Joshua C

    2010-01-01

    Medication information is one of the most important types of clinical data in electronic medical records. It is critical for healthcare safety and quality, as well as for clinical research that uses electronic medical record data. However, medication data are often recorded in clinical notes as free-text. As such, they are not accessible to other computerized applications that rely on coded data. We describe a new natural language processing system (MedEx), which extracts medication information from clinical notes. MedEx was initially developed using discharge summaries. An evaluation using a data set of 50 discharge summaries showed it performed well on identifying not only drug names (F-measure 93.2%), but also signature information, such as strength, route, and frequency, with F-measures of 94.5%, 93.9%, and 96.0% respectively. We then applied MedEx unchanged to outpatient clinic visit notes. It performed similarly with F-measures over 90% on a set of 25 clinic visit notes. PMID:20064797

  11. Enzymatic vegetable organic extracts as soil biochemical biostimulants and atrazine extenders.

    Science.gov (United States)

    García-Martínez, Ana María; Tejada, Manuel; Díaz, Ana Isabel; Rodríguez-Morgado, Bruno; Bautista, Juan; Parrado, Juan

    2010-09-08

    The purpose of this study was to gather information on the potential effects of organic biostimulants on soil activity and atrazine biodegradation. Carob germ enzymatic extract (CGEE) and wheat condensed distiller solubles enzymatic extract (WCDS-EE) have been obtained using an enzymatic process; their main organic components are soluble carbohydrates and proteins in the form of peptides and free amino acids. Their application to soil results in high biostimulation, rapidly increased dehydrogenase, phosphatase and glucosidase activities, and an observed atrazine extender capacity due to inhibition of its mineralization. The extender capacity of both extracts is proportional to the protein/carbohydrate ratio content. As a result, these enzymatic extracts are highly microbially available, leading to two independent phenomena, fertility and an atrazine persistence that is linked to increased soil activity.

  12. Methods to extract information on the atomic and molecular states from scientific abstracts

    International Nuclear Information System (INIS)

    Sasaki, Akira; Ueshima, Yutaka; Yamagiwa, Mitsuru; Murata, Masaki; Kanamaru, Toshiyuki; Shirado, Tamotsu; Isahara, Hitoshi

    2005-01-01

    We propose a new application of information technology to recognize and extract expressions of atomic and molecular states from electrical forms of scientific abstracts. Present results will help scientists to understand atomic states as well as the physics discussed in the articles. Combining with the internet search engines, it will make one possible to collect not only atomic and molecular data but broader scientific information over a wide range of research fields. (author)

  13. Trust in online prescription drug information among internet users: the impact on information search behavior after exposure to direct-to-consumer advertising.

    Science.gov (United States)

    Menon, Ajit M; Deshpande, Aparna D; Perri, Matthew; Zinkhan, George M

    2002-01-01

    The proliferation of both manufacturer-controlled and independent medication-related websites has aroused concern among consumers and policy-makers concerning the trustworthiness of Web-based drug information. The authors examine consumers' trust in on-line prescription drug information and its influence on information search behavior. The study design involves a retrospective analysis of data from a 1998 national survey. The findings reveal that trust in drug information from traditional media sources such as television and newspapers transfers to the domain of the Internet. Furthermore, a greater trust in on-line prescription drug information stimulates utilization of the Internet for information search after exposure to prescription drug advertising.

  14. Specialized Information Processing Deficits and Distinct Metabolomic Profiles Following TM-Domain Disruption of Nrg1

    OpenAIRE

    O'Tuathaigh, Colm MP; Mathur, Naina; O'Callaghan, Matthew J; MacIntyre, Lynsey; Harvey, Richard; Lai, Dona; Waddington, John L; Pickard, Bemjamom S; Watson, David D; Moran, Paula M

    2017-01-01

    While there is considerable genetic and pathologic evidence for an association between neuregulin 1 (NRG1) dysregulation and schizophrenia, the underlying molecular and cellular mechanisms remain unclear. Mutant mice containing disruption of the transmembrane (TM) domain of the NRG1 gene constitute a heuristic model for dysregulation of NRG1-ErbB4 signalling in schizophrenia. The present study focused on specialised behavioural and characterisation of hitherto un-characterised information pro...

  15. Influence of the extracted solute on the aggregation of malonamide extractant in organic phases: Consequences for phase stability

    International Nuclear Information System (INIS)

    Berthon, L.; Martinet, L.; Testard, F.; Madic, Ch.; Zem, Th.

    2010-01-01

    Due to their amphiphilic properties, malonamide molecules in alkane are organized in reverse micelle type aggregates, composed of a polar core formed by the malonamide polar heads and the extracted solutes, and surrounded by a hydrophobic shell made up of the extractant alkyl chains. The aggregates interact with one another through an attractive potential, leading to the formation of a third phase. This occurs with the splitting of the organic phase into a light phase composed mostly of diluent, and a heavy third phase containing highly concentrated extractant and solutes. In this article, we show that the aggregation (monomer concentration, domain of stability, and attractive potential between micelles) greatly depends on the nature of the extracted solute, whereas the size of aggregate (aggregation number) is only slightly influenced by this. We describe the extraction of water, nitric acid, neodymium nitrate and uranyl nitrate. Strongly polarizable species induce consistently large attraction potentials and a small stability domain for the dispersion of nano-droplets in the solvent. Highly polarizable ions such as lanthanides or uranyl induce more long-range attractive interactions than do protons. (authors)

  16. The Usher lifestyle survey: maintaining independence: a multi-centre study.

    Science.gov (United States)

    Damen, Godelieve W J A; Krabbe, Paul F M; Kilsby, M; Mylanus, Emmanuel A M

    2005-12-01

    Patients with Usher syndrome face a special set of challenges in order to maintain their independence when their sight and hearing worsen. Three different types of Usher (I, II and III) are distinguished by differences in onset, progression and severity of hearing loss, and by the presence or absence of balance problems. In this study 93 Usher patients from seven European countries filled out a questionnaire on maintaining independence (60 patients type I, 25 patients type II, four patients type III and four patients type unknown). Results of Usher type I and II patients are presented. Following the Nordic definition of maintaining independence in deaf-blindness, three domains are investigated: access to information, communication and mobility. Research variables in this study are: age and type of Usher, considered hearing loss- and the number of retinitis pigmentosa-related sight problems. Usher type I patients tend to need more help than Usher type II patients and the amount of help that they need grows when patients get older or when considered hearing loss worsens. No patterns in results were seen for the number of retinitis pigmentosa related sight problems.

  17. FacetGist: Collective Extraction of Document Facets in Large Technical Corpora.

    Science.gov (United States)

    Siddiqui, Tarique; Ren, Xiang; Parameswaran, Aditya; Han, Jiawei

    2016-10-01

    Given the large volume of technical documents available, it is crucial to automatically organize and categorize these documents to be able to understand and extract value from them. Towards this end, we introduce a new research problem called Facet Extraction. Given a collection of technical documents, the goal of Facet Extraction is to automatically label each document with a set of concepts for the key facets ( e.g. , application, technique, evaluation metrics, and dataset) that people may be interested in. Facet Extraction has numerous applications, including document summarization, literature search, patent search and business intelligence. The major challenge in performing Facet Extraction arises from multiple sources: concept extraction, concept to facet matching, and facet disambiguation. To tackle these challenges, we develop FacetGist, a framework for facet extraction. Facet Extraction involves constructing a graph-based heterogeneous network to capture information available across multiple local sentence-level features, as well as global context features. We then formulate a joint optimization problem, and propose an efficient algorithm for graph-based label propagation to estimate the facet of each concept mention. Experimental results on technical corpora from two domains demonstrate that Facet Extraction can lead to an improvement of over 25% in both precision and recall over competing schemes.

  18. Nora: A Vocabulary Discovery Tool for Concept Extraction.

    Science.gov (United States)

    Divita, Guy; Carter, Marjorie E; Durgahee, B S Begum; Pettey, Warren E; Redd, Andrew; Samore, Matthew H; Gundlapalli, Adi V

    2015-01-01

    Coverage of terms in domain-specific terminologies and ontologies is often limited in controlled medical vocabularies. Creating and augmenting such terminologies is resource intensive. We developed Nora as an interactive tool to discover terminology from text corpora; the output can then be employed to refine and enhance natural language processing-based concept extraction tasks. Nora provides a visualization of chains of words foraged from word frequency indexes from a text corpus. Domain experts direct and curate chains that contain relevant terms, which are further curated to identify lexical variants. A test of Nora demonstrated an increase of a domain lexicon in homelessness and related psychosocial factors by 38%, yielding an additional 10% extracted concepts.

  19. Analysis Methods for Extracting Knowledge from Large-Scale WiFi Monitoring to Inform Building Facility Planning

    DEFF Research Database (Denmark)

    Ruiz-Ruiz, Antonio; Blunck, Henrik; Prentow, Thor Siiger

    2014-01-01

    realistic data to inform facility planning. In this paper, we propose analysis methods to extract knowledge from large sets of network collected WiFi traces to better inform facility management and planning in large building complexes. The analysis methods, which build on a rich set of temporal and spatial......The optimization of logistics in large building com- plexes with many resources, such as hospitals, require realistic facility management and planning. Current planning practices rely foremost on manual observations or coarse unverified as- sumptions and therefore do not properly scale or provide....... Spatio-temporal visualization tools built on top of these methods enable planners to inspect and explore extracted information to inform facility-planning activities. To evaluate the methods, we present results for a large hospital complex covering more than 10 hectares. The evaluation is based on Wi...

  20. Extending the Effective Ranging Depth of Spectral Domain Optical Coherence Tomography by Spatial Frequency Domain Multiplexing

    Directory of Open Access Journals (Sweden)

    Tong Wu

    2016-11-01

    Full Text Available We present a spatial frequency domain multiplexing method for extending the imaging depth range of a spectral domain optical coherence tomography (SDOCT system without any expensive device. This method uses two galvo scanners with different pivot-offset distances in two independent reference arms for spatial frequency modulation and multiplexing. The spatial frequency contents corresponding to different depth regions of the sample can be shifted to different frequency bands. The spatial frequency domain multiplexing SDOCT system provides an approximately 1.9-fold increase in the effective ranging depth compared with that of a conventional full-range SDOCT system. The reconstructed images of phantom and biological tissue demonstrate the expected increase in ranging depth. The parameters choice criterion for this method is discussed.

  1. Specialized Information Processing Deficits and Distinct Metabolomic Profiles Following TM-Domain Disruption of Nrg1.

    Science.gov (United States)

    O'Tuathaigh, Colm M P; Mathur, Naina; O'Callaghan, Matthew J; MacIntyre, Lynsey; Harvey, Richard; Lai, Donna; Waddington, John L; Pickard, Benjamin S; Watson, David G; Moran, Paula M

    2017-09-01

    Although there is considerable genetic and pathologic evidence for an association between neuregulin 1 (NRG1) dysregulation and schizophrenia, the underlying molecular and cellular mechanisms remain unclear. Mutant mice containing disruption of the transmembrane (TM) domain of the NRG1 gene constitute a heuristic model for dysregulation of NRG1-ErbB4 signaling in schizophrenia. The present study focused on hitherto uncharacterized information processing phenotypes in this mutant line. Using a mass spectrometry-based metabolomics approach, we also quantified levels of unique metabolites in brain. Across 2 different sites and protocols, Nrg1 mutants demonstrated deficits in prepulse inhibition, a measure of sensorimotor gating, that is, disrupted in schizophrenia; these deficits were partially reversed by acute treatment with second, but not first-, generation antipsychotic drugs. However, Nrg1 mutants did not show a specific deficit in latent inhibition, a measure of selective attention that is also disrupted in schizophrenia. In contrast, in a "what-where-when" object recognition memory task, Nrg1 mutants displayed sex-specific (males only) disruption of "what-when" performance, indicative of impaired temporal aspects of episodic memory. Differential metabolomic profiling revealed that these behavioral phenotypes were accompanied, most prominently, by alterations in lipid metabolism pathways. This study is the first to associate these novel physiological mechanisms, previously independently identified as being abnormal in schizophrenia, with disruption of NRG1 function. These data suggest novel mechanisms by which compromised neuregulin function from birth might lead to schizophrenia-relevant behavioral changes in adulthood. © The Author 2017. Published by Oxford University Press on behalf of the Maryland Psychiatric Research Center.

  2. Torsion Dependence of Domain Transition and MI Effect of Melt-Extracted Co68.15Fe4.35Si12.25B13.25Nb1Cu1 Microwires

    Directory of Open Access Journals (Sweden)

    Dawei Xing

    2015-01-01

    Full Text Available We present the torsional stress induced magnetoimpedance (MI effect and surface domain structure evolution of magnetostrictive melt-extracted Co68.15Fe4.35Si12.25B13.25Nb1Cu1 microwires. Experimental results indicate that the surface domain structures observed by magnetic force microscope (MFM transform from the weak circumferential domain of as-cast state to the helical domain under large torsional strain of 81.6 (2π rad/m. Domain wall movement distorts at torsional strain ξ=20.4 (2π rad/m and forms a helical anisotropy with an angle of around 30° versus axial direction of wire. At 15 MHz, the maximum of GMI ratio ΔZ/Z(% increases to 194.4% at ξ=20.4 (2π rad/m from 116.3% of the as-cast state and then decreases to 134.9% at ξ=102.0 (2π rad/m. The torsion magnetoimpedance (TMI ratio ΔZ/Zξ(% is up to 290%. Based on this large torsional strain and high MI ratio, the microwire can be as an referred candidate for high-performance TMI sensor application.

  3. Dissecting empathy: high levels of psychopathic and autistic traits are characterised by difficulties in different social information processing domains

    Directory of Open Access Journals (Sweden)

    Patricia L Lockwood

    2013-11-01

    Full Text Available Individuals with psychopathy or autism spectrum disorder (ASD can behave in ways that suggest lack of empathy towards others. However, many different cognitive and affective processes may lead to unempathic behavior and the social processing profiles of individuals with high psychopathic vs. ASD traits are likely different. Whilst psychopathy appears characterized by problems with resonating with others’ emotions, ASD appears characterized by problems with cognitive perspective-taking. In addition, alexithymia has previously been associated with both disorders, but the contribution of alexithymia needs further exploration. In a community sample (N=110 we show for the first time that although affective resonance and cognitive perspective-taking are related, high psychopathic traits relate to problems with resonating with others’ emotions, but not cognitive perspective taking. Conversely, high ASD traits relate to problems with cognitive perspective-taking but not resonating with others’ emotions. Alexithymia was associated with problems with affective resonance independently of psychopathic traits, suggesting that different component processes (reduced tendency to feel what others feel and reduced ability to identify and describe feelings comprise affective resonance. Alexithymia was not associated with the reduced cognitive perspective-taking in high ASD traits. Our data suggest that (1 elevated psychopathic and ASD traits are characterized by difficulties in different social information processing domains and (2 reduced affective resonance in individuals with elevated psychopathic traits and the reduced cognitive perspective taking in individuals with elevated ASD traits are not explained by co-occurring alexithymia. (3 Alexithymia is independently associated with reduced affective resonance. Consequently, our data point to different component processes within the construct of empathy that are suggestive of partially separable cognitive

  4. Machinery running state identification based on discriminant semi-supervised local tangent space alignment for feature fusion and extraction

    International Nuclear Information System (INIS)

    Su, Zuqiang; Xiao, Hong; Zhang, Yi; Tang, Baoping; Jiang, Yonghua

    2017-01-01

    Extraction of sensitive features is a challenging but key task in data-driven machinery running state identification. Aimed at solving this problem, a method for machinery running state identification that applies discriminant semi-supervised local tangent space alignment (DSS-LTSA) for feature fusion and extraction is proposed. Firstly, in order to extract more distinct features, the vibration signals are decomposed by wavelet packet decomposition WPD, and a mixed-domain feature set consisted of statistical features, autoregressive (AR) model coefficients, instantaneous amplitude Shannon entropy and WPD energy spectrum is extracted to comprehensively characterize the properties of machinery running state(s). Then, the mixed-dimension feature set is inputted into DSS-LTSA for feature fusion and extraction to eliminate redundant information and interference noise. The proposed DSS-LTSA can extract intrinsic structure information of both labeled and unlabeled state samples, and as a result the over-fitting problem of supervised manifold learning and blindness problem of unsupervised manifold learning are overcome. Simultaneously, class discrimination information is integrated within the dimension reduction process in a semi-supervised manner to improve sensitivity of the extracted fusion features. Lastly, the extracted fusion features are inputted into a pattern recognition algorithm to achieve the running state identification. The effectiveness of the proposed method is verified by a running state identification case in a gearbox, and the results confirm the improved accuracy of the running state identification. (paper)

  5. Spatial frequency domain spectroscopy of two layer media

    Science.gov (United States)

    Yudovsky, Dmitry; Durkin, Anthony J.

    2011-10-01

    Monitoring of tissue blood volume and oxygen saturation using biomedical optics techniques has the potential to inform the assessment of tissue health, healing, and dysfunction. These quantities are typically estimated from the contribution of oxyhemoglobin and deoxyhemoglobin to the absorption spectrum of the dermis. However, estimation of blood related absorption in superficial tissue such as the skin can be confounded by the strong absorption of melanin in the epidermis. Furthermore, epidermal thickness and pigmentation varies with anatomic location, race, gender, and degree of disease progression. This study describes a technique for decoupling the effect of melanin absorption in the epidermis from blood absorption in the dermis for a large range of skin types and thicknesses. An artificial neural network was used to map input optical properties to spatial frequency domain diffuse reflectance of two layer media. Then, iterative fitting was used to determine the optical properties from simulated spatial frequency domain diffuse reflectance. Additionally, an artificial neural network was trained to directly map spatial frequency domain reflectance to sets of optical properties of a two layer medium, thus bypassing the need for iteration. In both cases, the optical thickness of the epidermis and absorption and reduced scattering coefficients of the dermis were determined independently. The accuracy and efficiency of the iterative fitting approach was compared with the direct neural network inversion.

  6. The Yin and Yang of SagS: Distinct Residues in the HmsP Domain of SagS Independently Regulate Biofilm Formation and Biofilm Drug Tolerance

    Science.gov (United States)

    Dingemans, Jozef; Poudyal, Bandita

    2018-01-01

    ABSTRACT The formation of inherently drug-tolerant biofilms by the opportunistic pathogen Pseudomonas aeruginosa requires the sensor-regulator hybrid SagS, with ΔsagS biofilms being unstructured and exhibiting increased antimicrobial susceptibility. Recent findings indicated SagS to function as a switch to control biofilm formation and drug tolerance independently. Moreover, findings suggested the periplasmic sensory HmsP domain of SagS is likely to be the control point in the regulation of biofilm formation and biofilm cells transitioning to a drug-tolerant state. We thus asked whether specific amino acid residues present in the HmsP domain contribute to the switch function of SagS. HmsP domain residues were therefore subjected to alanine replacement mutagenesis to identify substitutions that block the sensory function(s) of SagS, which is apparent by attached cells being unable to develop mature biofilms and/or prevent transition to an antimicrobial-resistant state. Mutant analyses revealed 32 residues that only contribute to blocking one sensory function. Moreover, amino acid residues affecting attachment and subsequent biofilm formation but not biofilm tolerance also impaired histidine kinase signaling via BfiS. In contrast, residues affecting biofilm drug tolerance but not attachment and subsequent biofilm formation negatively impacted BrlR transcription factor levels. Structure prediction suggested the two sets of residues affecting sensory functions are located in distinct areas that were previously described as being involved in ligand binding interactions. Taken together, these studies identify the molecular basis for the dual regulatory function of SagS. IMPORTANCE The membrane-bound sensory protein SagS plays a pivotal role in P. aeruginosa biofilm formation and biofilm cells gaining their heightened resistance to antimicrobial agents, with SagS being the control point at which both pathways diverge. Here, we demonstrate for the first time that the two

  7. Hybrid ontology for semantic information retrieval model using keyword matching indexing system.

    Science.gov (United States)

    Uthayan, K R; Mala, G S Anandha

    2015-01-01

    Ontology is the process of growth and elucidation of concepts of an information domain being common for a group of users. Establishing ontology into information retrieval is a normal method to develop searching effects of relevant information users require. Keywords matching process with historical or information domain is significant in recent calculations for assisting the best match for specific input queries. This research presents a better querying mechanism for information retrieval which integrates the ontology queries with keyword search. The ontology-based query is changed into a primary order to predicate logic uncertainty which is used for routing the query to the appropriate servers. Matching algorithms characterize warm area of researches in computer science and artificial intelligence. In text matching, it is more dependable to study semantics model and query for conditions of semantic matching. This research develops the semantic matching results between input queries and information in ontology field. The contributed algorithm is a hybrid method that is based on matching extracted instances from the queries and information field. The queries and information domain is focused on semantic matching, to discover the best match and to progress the executive process. In conclusion, the hybrid ontology in semantic web is sufficient to retrieve the documents when compared to standard ontology.

  8. Executive dysfunction is independently associated with reduced functional independence in heart failure.

    Science.gov (United States)

    Alosco, Michael L; Spitznagel, Mary Beth; Raz, Naftali; Cohen, Ronald; Sweet, Lawrence H; Colbert, Lisa H; Josephson, Richard; van Dulmen, Manfred; Hughes, Joel; Rosneck, Jim; Gunstad, John

    2014-03-01

    To examine the independent association between executive function with instrumental activities of daily living and health behaviours in older adults with heart failure. Executive function is an important contributor to functional independence as it consists of cognitive processes needed for decision-making, planning, organising and behavioural monitoring. Impairment in this domain is common in heart failure patients and associated with reduced performance of instrumental activities of daily living in many medical and neurological populations. However, the contribution of executive functions to functional independence and healthy lifestyle choices in heart failure patients has not been fully examined. Cross-sectional analyses. One hundred and seventy-five heart failure patients completed a neuropsychological battery and echocardiogram. Participants also completed the Lawton-Brody Instrumental Activities of Daily Living Scale and reported current cigarette use. Hierarchical regressions revealed that reduced executive function was independently associated with worse instrumental activity of daily living performance with a specific association for decreased ability to manage medications. Partial correlations showed that executive dysfunction was associated with current cigarette use. Our findings suggest that executive dysfunction is associated with poorer functional independence and contributes to unhealthy behaviours in heart failure. Future studies should examine whether heart failure patients benefit from formal organisation schema (i.e. pill organisers) to maintain independence. Screening of executive function in heart failure patients may provide key insight into their ability to perform daily tasks, including the management of treatment recommendations. © 2013 John Wiley & Sons Ltd.

  9. Individual Building Extraction from TerraSAR-X Images Based on Ontological Semantic Analysis

    Directory of Open Access Journals (Sweden)

    Rong Gui

    2016-08-01

    Full Text Available Accurate building information plays a crucial role for urban planning, human settlements and environmental management. Synthetic aperture radar (SAR images, which deliver images with metric resolution, allow for analyzing and extracting detailed information on urban areas. In this paper, we consider the problem of extracting individual buildings from SAR images based on domain ontology. By analyzing a building scattering model with different orientations and structures, the building ontology model is set up to express multiple characteristics of individual buildings. Under this semantic expression framework, an object-based SAR image segmentation method is adopted to provide homogeneous image objects, and three categories of image object features are extracted. Semantic rules are implemented by organizing image object features, and the individual building objects expression based on an ontological semantic description is formed. Finally, the building primitives are used to detect buildings among the available image objects. Experiments on TerraSAR-X images of Foshan city, China, with a spatial resolution of 1.25 m × 1.25 m, have shown the total extraction rates are above 84%. The results indicate the ontological semantic method can exactly extract flat-roof and gable-roof buildings larger than 250 pixels with different orientations.

  10. A framework to evaluate information quality in Public Administration websites

    OpenAIRE

    Geraci, Filippo; Martinelli, Maurizio; Pellegrini, Marco; Serrecchia, Michela

    2013-01-01

    The paper presents a framework aimed at assessing the capacity of Public Administration bodies (PA) to offer a good quality of information and service on their web portals. Our framework is based on the extraction of ".it? domain names registered by Italian public institutions and the subsequent analysis of their relative websites. The analysis foresees an automatic gathering of the web pages of PA portals by means of web crawling and an assessment of the quality of their online information s...

  11. Comparison of environmental and socio-economic domains of vulnerability to flood hazards

    Science.gov (United States)

    Leidel, M.; Kienberger, S.; Lang, S.; Zeil, P.

    2009-04-01

    Socio-economic and environmental based vulnerability models have been developed within the research context of the FP6 project BRAHMATWINN. The conceptualisation of vulnerability has been defined in the project and is characterised as a function of sensitivity and adaptive capacity, where sensitivity is used to refer to systems that are susceptible to the impacts of environmental stress. Adaptive capacity is used to refer to systems or resources available to communities that could help them adapt or cope with the adverse consequences of environmental stresses in the recovery phase. In a wider context the approach reflects the wider objective and conceptualizations of the IPCC (Intergovernmental Panel on Climate Change) framework, where vulnerability is characterized as a component of overall risk. A methodology has been developed which delineates spatial units of vulnerability (VULNUS). These units share a specific common characteristic and allow the independent spatial modelling of a complex phenomena independent from administrative units and raster based approaches. An increasing detail of spatial data and complex decision problems require flexible means for scaled spatial representations, for mapping the dynamics and constant changes, and delivering the crucial information. Automated techniques of object-based image analysis (OBIA, Lang & Blaschke, 2006), capable of integrating a virtually unlimited set of spatial data sets, try to match the information extraction with our world view. To account for that, a flexible concept of manageable units is required. The term geon was proposed by Lang (2008) to describe generic spatial objects that are homogenous in terms of a varying spatial phenomena under the influence of, and partly controlled by, policy actions. The geon concept acts as a framework for the regionalization of continuous spatial information according to defined parameters of homogeneity. It is flexible in terms of a certain perception of a problem

  12. Age related differences in individual quality of life domains in youth with type 1 diabetes

    Directory of Open Access Journals (Sweden)

    Lett Syretta

    2004-09-01

    Full Text Available Abstract Background Investigating individual, as opposed to predetermined, quality of life domains may yield important information about quality of life. This study investigated the individual quality of life domains nominated by youth with type 1 diabetes. Methods Eighty young people attending a diabetes summer camp completed the Schedule for the Evaluation of Individual Quality of Life-Direct Weighting interview, which allows respondents to nominate and evaluate their own quality of life domains. Results The most frequently nominated life domains were 'family', 'friends', 'diabetes', 'school', and 'health' respectively; ranked in terms of importance, domains were 'religion', 'family', 'diabetes', 'health', and 'the golden rule'; ranked in order of satisfaction, domains were 'camp', 'religion', 'pets', and 'family' and 'a special person' were tied for fifth. Respondent age was significantly positively associated with the importance of 'friends', and a significantly negatively associated with the importance of 'family'. Nearly all respondents nominated a quality of life domain relating to physical status, however, the specific physical status domain and the rationale for its nomination varied. Some respondents nominated 'diabetes' as a domain and emphasized diabetes 'self-care behaviors' in order to avoid negative health consequences such as hospitalization. Other respondents nominated 'health' and focused more generally on 'living well with diabetes'. In an ANOVA with physical status domain as the independent variable and age as the dependent variable, participants who nominated 'diabetes' were younger (M = 12.9 years than those who nominated 'health' (M = 15.9 years. In a second ANOVA, with rationale for nomination the physical status domain as the independent variable, and age as the dependent variable, those who emphasized 'self care behaviors' were younger (M = 11.8 years than those who emphasized 'living well with diabetes' (M = 14.6 years

  13. A new NMIS characteristic signature acquisition method based on time-domain fission correlation spectrum

    International Nuclear Information System (INIS)

    Wei Biao; Feng Peng; Yang Fan; Ren Yong

    2014-01-01

    To deal with the disadvantages of the homogeneous signature of the nuclear material identification system (NMIS) and limited methods to extract the characteristic parameters of the nuclear materials, an enhanced method using the combination of the Time-of-Flight (TOF) and the Pulse Shape Discrimination (PSD) was introduced into the traditional characteristic parameters extraction and recognition system of the NMIS. With the help of the PSD, the γ signal and the neutron signal can be discriminated. Further based on the differences of the neutron-γ flight time of the detectors in various positions, a new time-domain signature reflecting the position information of unknown nuclear material was investigated. The simulation result showed that the algorithm is feasible and helpful to identify the relative position of unknown nuclear material. (authors)

  14. The Feature Extraction Based on Texture Image Information for Emotion Sensing in Speech

    Directory of Open Access Journals (Sweden)

    Kun-Ching Wang

    2014-09-01

    Full Text Available In this paper, we present a novel texture image feature for Emotion Sensing in Speech (ESS. This idea is based on the fact that the texture images carry emotion-related information. The feature extraction is derived from time-frequency representation of spectrogram images. First, we transform the spectrogram as a recognizable image. Next, we use a cubic curve to enhance the image contrast. Then, the texture image information (TII derived from the spectrogram image can be extracted by using Laws’ masks to characterize emotional state. In order to evaluate the effectiveness of the proposed emotion recognition in different languages, we use two open emotional databases including the Berlin Emotional Speech Database (EMO-DB and eNTERFACE corpus and one self-recorded database (KHUSC-EmoDB, to evaluate the performance cross-corpora. The results of the proposed ESS system are presented using support vector machine (SVM as a classifier. Experimental results show that the proposed TII-based feature extraction inspired by visual perception can provide significant classification for ESS systems. The two-dimensional (2-D TII feature can provide the discrimination between different emotions in visual expressions except for the conveyance pitch and formant tracks. In addition, the de-noising in 2-D images can be more easily completed than de-noising in 1-D speech.

  15. Clinical Research Informatics: Challenges, Opportunities and Definition for an Emerging Domain

    Science.gov (United States)

    Embi, Peter J.; Payne, Philip R.O.

    2009-01-01

    Objectives Clinical Research Informatics, an emerging sub-domain of Biomedical Informatics, is currently not well defined. A formal description of CRI including major challenges and opportunities is needed to direct progress in the field. Design Given the early stage of CRI knowledge and activity, we engaged in a series of qualitative studies with key stakeholders and opinion leaders to determine the range of challenges and opportunities facing CRI. These phases employed complimentary methods to triangulate upon our findings. Measurements Study phases included: 1) a group interview with key stakeholders, 2) an email follow-up survey with a larger group of self-identified CRI professionals, and 3) validation of our results via electronic peer-debriefing and member-checking with a group of CRI-related opinion leaders. Data were collected, transcribed, and organized for formal, independent content analyses by experienced qualitative investigators, followed by an iterative process to identify emergent categorizations and thematic descriptions of the data. Results We identified a range of challenges and opportunities facing the CRI domain. These included 13 distinct themes spanning academic, practical, and organizational aspects of CRI. These findings also informed the development of a formal definition of CRI and supported further representations that illustrate areas of emphasis critical to advancing the domain. Conclusions CRI has emerged as a distinct discipline that faces multiple challenges and opportunities. The findings presented summarize those challenges and opportunities and provide a framework that should help inform next steps to advance this important new discipline. PMID:19261934

  16. Emergence of CD134 cysteine-rich domain 2 (CRD2)-independent strains of feline immunodeficiency virus (FIV) is associated with disease progression in naturally infected cats.

    Science.gov (United States)

    Bęczkowski, Paweł M; Techakriengkrai, Navapon; Logan, Nicola; McMonagle, Elizabeth; Litster, Annette; Willett, Brian J; Hosie, Margaret J

    2014-11-28

    Feline immunodeficiency virus (FIV) infection is mediated by sequential interactions with CD134 and CXCR4. Field strains of virus vary in their dependence on cysteine-rich domain 2 (CRD2) of CD134 for infection. Here, we analyse the receptor usage of viral variants in the blood of 39 naturally infected cats, revealing that CRD2-dependent viral variants dominate in early infection, evolving towards CRD2-independence with disease progression. These findings are consistent with a shift in CRD2 of CD134 usage with disease progression.

  17. The carboxy-terminal domain of Dictyostelium C-module-binding factor is an independent gene regulatory entity.

    Directory of Open Access Journals (Sweden)

    Jörg Lucas

    Full Text Available The C-module-binding factor (CbfA is a multidomain protein that belongs to the family of jumonji-type (JmjC transcription regulators. In the social amoeba Dictyostelium discoideum, CbfA regulates gene expression during the unicellular growth phase and multicellular development. CbfA and a related D. discoideum CbfA-like protein, CbfB, share a paralogous domain arrangement that includes the JmjC domain, presumably a chromatin-remodeling activity, and two zinc finger-like (ZF motifs. On the other hand, the CbfA and CbfB proteins have completely different carboxy-terminal domains, suggesting that the plasticity of such domains may have contributed to the adaptation of the CbfA-like transcription factors to the rapid genome evolution in the dictyostelid clade. To support this hypothesis we performed DNA microarray and real-time RT-PCR measurements and found that CbfA regulates at least 160 genes during the vegetative growth of D. discoideum cells. Functional annotation of these genes revealed that CbfA predominantly controls the expression of gene products involved in housekeeping functions, such as carbohydrate, purine nucleoside/nucleotide, and amino acid metabolism. The CbfA protein displays two different mechanisms of gene regulation. The expression of one set of CbfA-dependent genes requires at least the JmjC/ZF domain of the CbfA protein and thus may depend on chromatin modulation. Regulation of the larger group of genes, however, does not depend on the entire CbfA protein and requires only the carboxy-terminal domain of CbfA (CbfA-CTD. An AT-hook motif located in CbfA-CTD, which is known to mediate DNA binding to A+T-rich sequences in vitro, contributed to CbfA-CTD-dependent gene regulatory functions in vivo.

  18. Thundercloud: Domain specific information security training for the smart grid

    Science.gov (United States)

    Stites, Joseph

    In this paper, we describe a cloud-based virtual smart grid test bed: ThunderCloud, which is intended to be used for domain-specific security training applicable to the smart grid environment. The test bed consists of virtual machines connected using a virtual internal network. ThunderCloud is remotely accessible, allowing students to undergo educational exercises online. We also describe a series of practical exercises that we have developed for providing the domain-specific training using ThunderCloud. The training exercises and attacks are designed to be realistic and to reflect known vulnerabilities and attacks reported in the smart grid environment. We were able to use ThunderCloud to offer practical domain-specific security training for smart grid environment to computer science students at little or no cost to the department and no risk to any real networks or systems.

  19. Barriers and facilitators to preventing pressure ulcers in nursing home residents: A qualitative analysis informed by the Theoretical Domains Framework.

    Science.gov (United States)

    Lavallée, Jacqueline F; Gray, Trish A; Dumville, Jo; Cullum, Nicky

    2018-06-01

    Pressure ulcers are areas of localised damage to the skin and underlying tissue; and can cause pain, immobility, and delay recovery, impacting on health-related quality of life. The individuals who are most at risk of developing a pressure ulcer are those who are seriously ill, elderly, have impaired mobility and/or poor nutrition; thus, many nursing home residents are at risk. To understand the context of pressure ulcer prevention in nursing homes and to explore the potential barriers and facilitators to evidence-informed practices. Semi-structured interviews were conducted with nursing home nurses, healthcare assistants and managers, National Health Service community-based wound specialist nurses (known in the UK as tissue viability nurses) and a nurse manager in the North West of England. The interview guide was developed using the Theoretical Domains Framework to explore the barriers and facilitators to pressure ulcer prevention in nursing home residents. Data were analysed using a framework analysis and domains were identified as salient based on their frequency and the potential strength of their impact. 25 participants (nursing home: 2 managers, 7 healthcare assistants, 11 qualified nurses; National Health Service community services: 4 tissue viability nurses, 1 manager) were interviewed. Depending upon the behaviours reported and the context, the same domain could be classified as both a barrier and a facilitator. We identified seven domains as relevant in the prevention of pressure ulcers in nursing home residents mapping to four "barrier" domains and six "facilitator" domains. The four "barrier" domains were knowledge, physical skills, social influences and environmental context and resources and the six "facilitator" domains were interpersonal skills, environmental context and resources, social influences, beliefs about capabilities, beliefs about consequences and social/professional role and identity). Knowledge and insight into these barriers and

  20. YAdumper: extracting and translating large information volumes from relational databases to structured flat files.

    Science.gov (United States)

    Fernández, José M; Valencia, Alfonso

    2004-10-12

    Downloading the information stored in relational databases into XML and other flat formats is a common task in bioinformatics. This periodical dumping of information requires considerable CPU time, disk and memory resources. YAdumper has been developed as a purpose-specific tool to deal with the integral structured information download of relational databases. YAdumper is a Java application that organizes database extraction following an XML template based on an external Document Type Declaration. Compared with other non-native alternatives, YAdumper substantially reduces memory requirements and considerably improves writing performance.

  1. A semantic-based method for extracting concept definitions from scientific publications: evaluation in the autism phenotype domain

    OpenAIRE

    Hassanpour, Saeed; O?Connor, Martin J; Das, Amar K

    2013-01-01

    Background A variety of informatics approaches have been developed that use information retrieval, NLP and text-mining techniques to identify biomedical concepts and relations within scientific publications or their sentences. These approaches have not typically addressed the challenge of extracting more complex knowledge such as biomedical definitions. In our efforts to facilitate knowledge acquisition of rule-based definitions of autism phenotypes, we have developed a novel semantic-based t...

  2. Extraction of Independent Structural Images for Principal Component Thermography

    Directory of Open Access Journals (Sweden)

    Dmitry Gavrilov

    2018-03-01

    Full Text Available Thermography is a powerful tool for non-destructive testing of a wide range of materials. Thermography has a number of approaches differing in both experiment setup and the way the collected data are processed. Among such approaches is the Principal Component Thermography (PCT method, which is based on the statistical processing of raw thermal images collected by thermal camera. The processed images (principal components or empirical orthogonal functions form an orthonormal basis, and often look like a superposition of all possible structural features found in the object under inspection—i.e., surface heating non-uniformity, internal defects and material structure. At the same time, from practical point of view it is desirable to have images representing independent structural features. The work presented in this paper proposes an approach for separation of independent image patterns (archetypes from a set of principal component images. The approach is demonstrated in the application of inspection of composite materials as well as the non-invasive analysis of works of art.

  3. From Specific Information Extraction to Inferences: A Hierarchical Framework of Graph Comprehension

    Science.gov (United States)

    2004-09-01

    The skill to interpret the information displayed in graphs is so important to have, the National Council of Teachers of Mathematics has created...guidelines to ensure that students learn these skills ( NCTM : Standards for Mathematics , 2003). These guidelines are based primarily on the extraction of...graphical perception. Human Computer Interaction, 8, 353-388. NCTM : Standards for Mathematics . (2003, 2003). Peebles, D., & Cheng, P. C.-H. (2002

  4. Src binds cortactin through an SH2 domain cystine-mediated linkage

    Science.gov (United States)

    Evans, Jason V.; Ammer, Amanda G.; Jett, John E.; Bolcato, Chris A.; Breaux, Jason C.; Martin, Karen H.; Culp, Mark V.; Gannett, Peter M.; Weed, Scott A.

    2012-01-01

    Summary Tyrosine-kinase-based signal transduction mediated by modular protein domains is critical for cellular function. The Src homology (SH)2 domain is an important conductor of intracellular signaling that binds to phosphorylated tyrosines on acceptor proteins, producing molecular complexes responsible for signal relay. Cortactin is a cytoskeletal protein and tyrosine kinase substrate that regulates actin-based motility through interactions with SH2-domain-containing proteins. The Src kinase SH2 domain mediates cortactin binding and tyrosine phosphorylation, but how Src interacts with cortactin is unknown. Here we demonstrate that Src binds cortactin through cystine bonding between Src C185 in the SH2 domain within the phosphotyrosine binding pocket and cortactin C112/246 in the cortactin repeats domain, independent of tyrosine phosphorylation. Interaction studies show that the presence of reducing agents ablates Src-cortactin binding, eliminates cortactin phosphorylation by Src, and prevents Src SH2 domain binding to cortactin. Tandem MS/MS sequencing demonstrates cystine bond formation between Src C185 and cortactin C112/246. Mutational studies indicate that an intact cystine binding interface is required for Src-mediated cortactin phosphorylation, cell migration, and pre-invadopodia formation. Our results identify a novel phosphotyrosine-independent binding mode between the Src SH2 domain and cortactin. Besides Src, one quarter of all SH2 domains contain cysteines at or near the analogous Src C185 position. This provides a potential alternative mechanism to tyrosine phosphorylation for cysteine-containing SH2 domains to bind cognate ligands that may be widespread in propagating signals regulating diverse cellular functions. PMID:23097045

  5. Src binds cortactin through an SH2 domain cystine-mediated linkage.

    Science.gov (United States)

    Evans, Jason V; Ammer, Amanda G; Jett, John E; Bolcato, Chris A; Breaux, Jason C; Martin, Karen H; Culp, Mark V; Gannett, Peter M; Weed, Scott A

    2012-12-15

    Tyrosine-kinase-based signal transduction mediated by modular protein domains is critical for cellular function. The Src homology (SH)2 domain is an important conductor of intracellular signaling that binds to phosphorylated tyrosines on acceptor proteins, producing molecular complexes responsible for signal relay. Cortactin is a cytoskeletal protein and tyrosine kinase substrate that regulates actin-based motility through interactions with SH2-domain-containing proteins. The Src kinase SH2 domain mediates cortactin binding and tyrosine phosphorylation, but how Src interacts with cortactin is unknown. Here we demonstrate that Src binds cortactin through cystine bonding between Src C185 in the SH2 domain within the phosphotyrosine binding pocket and cortactin C112/246 in the cortactin repeats domain, independent of tyrosine phosphorylation. Interaction studies show that the presence of reducing agents ablates Src-cortactin binding, eliminates cortactin phosphorylation by Src, and prevents Src SH2 domain binding to cortactin. Tandem MS/MS sequencing demonstrates cystine bond formation between Src C185 and cortactin C112/246. Mutational studies indicate that an intact cystine binding interface is required for Src-mediated cortactin phosphorylation, cell migration, and pre-invadopodia formation. Our results identify a novel phosphotyrosine-independent binding mode between the Src SH2 domain and cortactin. Besides Src, one quarter of all SH2 domains contain cysteines at or near the analogous Src C185 position. This provides a potential alternative mechanism to tyrosine phosphorylation for cysteine-containing SH2 domains to bind cognate ligands that may be widespread in propagating signals regulating diverse cellular functions.

  6. Vertical selection in the information domain of children

    NARCIS (Netherlands)

    Duarte Torres, Sergio; Hiemstra, Djoerd; Huibers, Theo W.C.

    In this paper we explore the vertical selection methods in aggregated search in the specific domain of topics for children between 7 and 12 years old. A test collection consisting of 25 verticals, 3.8K queries and relevant assessments for a large sample of these queries mapping relevant verticals to

  7. Natural-Annotation-based Unsupervised Construction of Korean-Chinese Domain Dictionary

    Science.gov (United States)

    Liu, Wuying; Wang, Lin

    2018-03-01

    The large-scale bilingual parallel resource is significant to statistical learning and deep learning in natural language processing. This paper addresses the automatic construction issue of the Korean-Chinese domain dictionary, and presents a novel unsupervised construction method based on the natural annotation in the raw corpus. We firstly extract all Korean-Chinese word pairs from Korean texts according to natural annotations, secondly transform the traditional Chinese characters into the simplified ones, and finally distill out a bilingual domain dictionary after retrieving the simplified Chinese words in an extra Chinese domain dictionary. The experimental results show that our method can automatically build multiple Korean-Chinese domain dictionaries efficiently.

  8. Domain-Specific and Unspecific Reaction Times in Experienced Team Handball Goalkeepers and Novices.

    Science.gov (United States)

    Helm, Fabian; Reiser, Mathias; Munzert, Jörn

    2016-01-01

    In our everyday environments, we are constantly having to adapt our behavior to changing conditions. Hence, processing information is a fundamental cognitive activity, especially the linking together of perceptual and action processes. In this context, expertise research in the sport domain has concentrated on arguing that superior processing performance is driven by an advantage to be found in anticipatory processes (see Williams et al., 2011, for a review). This has resulted in less attention being paid to the benefits coming from basic internal perceptual-motor processing. In general, research on reaction time (RT) indicates that practicing a RT task leads to an increase in processing speed (Mowbray and Rhoades, 1959; Rabbitt and Banerji, 1989). Against this background, the present study examined whether the speed of internal processing is dependent on or independent from domain-specific motor expertise in unpredictable stimulus-response tasks and in a double stimulus-response paradigm. Thirty male participants (15 team handball goalkeepers and 15 novices) performed domain-unspecific simple or choice stimulus-response (CSR) tasks as well as CSR tasks that were domain-specific only for goalkeepers. As expected, results showed significantly faster RTs for goalkeepers on domain-specific tasks, whereas novices' RTs were more frequently excessively long. However, differences between groups in the double stimulus-response paradigm were not significant. It is concluded that the reported expertise advantage might be due to recalling stored perceptual-motor representations for the domain-specific tasks, implying that experience with (practice of) a motor task explicitly enhances the internal processing of other related domain-specific tasks.

  9. Integrating the Supervised Information into Unsupervised Learning

    Directory of Open Access Journals (Sweden)

    Ping Ling

    2013-01-01

    Full Text Available This paper presents an assembling unsupervised learning framework that adopts the information coming from the supervised learning process and gives the corresponding implementation algorithm. The algorithm consists of two phases: extracting and clustering data representatives (DRs firstly to obtain labeled training data and then classifying non-DRs based on labeled DRs. The implementation algorithm is called SDSN since it employs the tuning-scaled Support vector domain description to collect DRs, uses spectrum-based method to cluster DRs, and adopts the nearest neighbor classifier to label non-DRs. The validation of the clustering procedure of the first-phase is analyzed theoretically. A new metric is defined data dependently in the second phase to allow the nearest neighbor classifier to work with the informed information. A fast training approach for DRs’ extraction is provided to bring more efficiency. Experimental results on synthetic and real datasets verify that the proposed idea is of correctness and performance and SDSN exhibits higher popularity in practice over the traditional pure clustering procedure.

  10. Survey of Applications of Complex Event Processing (CEP in Health Domain

    Directory of Open Access Journals (Sweden)

    Nadeem Mahmood

    2017-12-01

    Full Text Available It is always difficult to manipulate the production of huge amount of data which comes from multiple sources and to extract meaningful information to make appropriate decisions. When data comes from various input resources, to get required streams of events form this complex input network, the one of the strong functionality of Business Intelligence (BI the Complex Event Processing (CEP is the appropriate solution for the above mention problems. Real time processing, pattern matching, stream processing, big data management, sensor data processing and many more are the application areas of CEP. Health domain itself is a multi-dimension domain such as hospital supply chain, OPD management, disease diagnostic, In-patient, out-patient management, and emergency care etc. In this paper, the main focus is to discuss the application areas of Complex Event Processing (CEP in health domain by using sensor device, such that how CEP manipulate health data set events coming from sensor devices such as blood pressure, heart rate, fall detection, sugar level, temperature or any other vital signs and how this systems respond to these events as quickly as possible. Different existing models and application using CEP are discussed and summarized according to different characteristics.

  11. Data Model Management for Space Information Systems

    Science.gov (United States)

    Hughes, J. Steven; Crichton, Daniel J.; Ramirez, Paul; Mattmann, chris

    2006-01-01

    The Reference Architecture for Space Information Management (RASIM) suggests the separation of the data model from software components to promote the development of flexible information management systems. RASIM allows the data model to evolve independently from the software components and results in a robust implementation that remains viable as the domain changes. However, the development and management of data models within RASIM are difficult and time consuming tasks involving the choice of a notation, the capture of the model, its validation for consistency, and the export of the model for implementation. Current limitations to this approach include the lack of ability to capture comprehensive domain knowledge, the loss of significant modeling information during implementation, the lack of model visualization and documentation capabilities, and exports being limited to one or two schema types. The advent of the Semantic Web and its demand for sophisticated data models has addressed this situation by providing a new level of data model management in the form of ontology tools. In this paper we describe the use of a representative ontology tool to capture and manage a data model for a space information system. The resulting ontology is implementation independent. Novel on-line visualization and documentation capabilities are available automatically, and the ability to export to various schemas can be added through tool plug-ins. In addition, the ingestion of data instances into the ontology allows validation of the ontology and results in a domain knowledge base. Semantic browsers are easily configured for the knowledge base. For example the export of the knowledge base to RDF/XML and RDFS/XML and the use of open source metadata browsers provide ready-made user interfaces that support both text- and facet-based search. This paper will present the Planetary Data System (PDS) data model as a use case and describe the import of the data model into an ontology tool

  12. Automatic Extraction of Urban Built-Up Area Based on Object-Oriented Method and Remote Sensing Data

    Science.gov (United States)

    Li, L.; Zhou, H.; Wen, Q.; Chen, T.; Guan, F.; Ren, B.; Yu, H.; Wang, Z.

    2018-04-01

    Built-up area marks the use of city construction land in the different periods of the development, the accurate extraction is the key to the studies of the changes of urban expansion. This paper studies the technology of automatic extraction of urban built-up area based on object-oriented method and remote sensing data, and realizes the automatic extraction of the main built-up area of the city, which saves the manpower cost greatly. First, the extraction of construction land based on object-oriented method, the main technical steps include: (1) Multi-resolution segmentation; (2) Feature Construction and Selection; (3) Information Extraction of Construction Land Based on Rule Set, The characteristic parameters used in the rule set mainly include the mean of the red band (Mean R), Normalized Difference Vegetation Index (NDVI), Ratio of residential index (RRI), Blue band mean (Mean B), Through the combination of the above characteristic parameters, the construction site information can be extracted. Based on the degree of adaptability, distance and area of the object domain, the urban built-up area can be quickly and accurately defined from the construction land information without depending on other data and expert knowledge to achieve the automatic extraction of the urban built-up area. In this paper, Beijing city as an experimental area for the technical methods of the experiment, the results show that: the city built-up area to achieve automatic extraction, boundary accuracy of 2359.65 m to meet the requirements. The automatic extraction of urban built-up area has strong practicality and can be applied to the monitoring of the change of the main built-up area of city.

  13. Temporary formation of highly conducting domain walls for non-destructive read-out of ferroelectric domain-wall resistance switching memories

    Science.gov (United States)

    Jiang, Jun; Bai, Zi Long; Chen, Zhi Hui; He, Long; Zhang, David Wei; Zhang, Qing Hua; Shi, Jin An; Park, Min Hyuk; Scott, James F.; Hwang, Cheol Seong; Jiang, An Quan

    2018-01-01

    Erasable conductive domain walls in insulating ferroelectric thin films can be used for non-destructive electrical read-out of the polarization states in ferroelectric memories. Still, the domain-wall currents extracted by these devices have not yet reached the intensity and stability required to drive read-out circuits operating at high speeds. This study demonstrated non-destructive read-out of digital data stored using specific domain-wall configurations in epitaxial BiFeO3 thin films formed in mesa-geometry structures. Partially switched domains, which enable the formation of conductive walls during the read operation, spontaneously retract when the read voltage is removed, reducing the accumulation of mobile defects at the domain walls and potentially improving the device stability. Three-terminal memory devices produced 14 nA read currents at an operating voltage of 5 V, and operated up to T = 85 °C. The gap length can also be smaller than the film thickness, allowing the realization of ferroelectric memories with device dimensions far below 100 nm.

  14. Calibrations of a tritium extraction facility

    International Nuclear Information System (INIS)

    Bretscher, M.M.; Oliver, B.M.; Farrar, H. IV.

    1983-01-01

    A tritium extraction facility has been built for the purpose of measuring the absolute tritium concentration in neutron-irradiated lithium metal samples. Two independent calibration procedures have been used to determine what fraction, if any, of tritium is lost during the extraction process. The first procedure compares independently measured 4 He and 3 H concentrations from the 6 Li(n,α)T reaction. The second procedure compared measured 6 Li(n,α)T/ 197 Au (n,γ) 198 Au thermal neutron reaction rate ratios with those obtained from Monte Carlo calculations using well-known cross sections. Both calibration methods show that within experimental errors (approx. 1.5%) no tritium is lost during the extraction process

  15. Functional Independence and Interdependence of the Src Homology Domains of Phospholipase C-γ1 in B-Cell Receptor Signal Transduction

    Science.gov (United States)

    DeBell, Karen E.; Stoica, Bogdan A.; Verí, Maria-Concetta; Di Baldassarre, Angela; Miscia, Sebastiano; Graham, Laurie J.; Rellahan, Barbara L.; Ishiai, Masamichi; Kurosaki, Tomohiro; Bonvini, Ezio

    1999-01-01

    B-cell receptor (BCR)-induced activation of phospholipase C-γ1 (PLCγ1) and PLCγ2 is crucial for B-cell function. While several signaling molecules have been implicated in PLCγ activation, the mechanism coupling PLCγ to the BCR remains undefined. The role of PLCγ1 SH2 and SH3 domains at different steps of BCR-induced PLCγ1 activation was examined by reconstitution in a PLCγ-negative B-cell line. PLCγ1 membrane translocation required a functional SH2 N-terminal [SH2(N)] domain, was decreased by mutation of the SH3 domain, but was unaffected by mutation of the SH2(C) domain. Tyrosine phosphorylation did not require the SH2(C) or SH3 domains but depended exclusively on a functional SH2(N) domain, which mediated the association of PLCγ1 with the adapter protein, BLNK. Forcing PLCγ1 to the membrane via a myristoylation signal did not bypass the SH2(N) domain requirement for phosphorylation, indicating that the phosphorylation mediated by this domain is not due to membrane anchoring alone. Mutation of the SH2(N) or the SH2(C) domain abrogated BCR-stimulated phosphoinositide hydrolysis and signaling events, while mutation of the SH3 domain partially decreased signaling. PLCγ1 SH domains, therefore, have interrelated but distinct roles in BCR-induced PLCγ1 activation. PMID:10523627

  16. CFAI-Plus: Adding cognitive frailty as a new domain to the comprehensive frailty assessment instrument.

    Science.gov (United States)

    De Roeck, Ellen Elisa; Dury, Sarah; De Witte, Nico; De Donder, Liesbeth; Bjerke, Maria; De Deyn, Peter Paul; Engelborghs, Sebastiaan; Dierckx, Eva

    2018-07-01

    Cognitive frailty is characterized by the presence of cognitive impairment in exclusion of dementia. In line with other frailty domains, cognitive frailty is associated with negative outcomes. The Comprehensive Frailty Assessment Instrument (CFAI) measures 4 domains of frailty, namely physical, psychological, social, and environmental frailty. The absence of cognitive frailty is a limitation. An expert panel selected 6 questions from the Informant Questionnaire on Cognitive Decline that were, together with the CFAI and the Montreal cognitive assessment administered to 355 older community dwelling adults (mean age = 77). After multivariate analysis, 2 questions were excluded. All the questions from the original CFAI were implemented in a principal component analysis together with the 4 cognitive questions, showing that the 4 cognitive questions all load on 1 factor, representing the cognitive domain of frailty. By adding the cognitive domain to the CFAI, the reliability of the adapted CFAI (CFAI-Plus), remains good (Cronbach's alpha: .767). This study showed that cognitive frailty can be added to the CFAI without affecting its good psychometric properties. In the future, the CFAI-Plus needs to be validated in an independent cohort, and the interaction with the other frailty domains needs to be studied. Copyright © 2018 John Wiley & Sons, Ltd.

  17. Alternative to domain wall fermions

    International Nuclear Information System (INIS)

    Neuberger, H.

    2002-01-01

    An alternative to commonly used domain wall fermions is presented. Some rigorous bounds on the condition number of the associated linear problem are derived. On the basis of these bounds and some experimentation it is argued that domain wall fermions will in general be associated with a condition number that is of the same order of magnitude as the product of the condition number of the linear problem in the physical dimensions by the inverse bare quark mass. Thus, the computational cost of implementing true domain wall fermions using a single conjugate gradient algorithm is of the same order of magnitude as that of implementing the overlap Dirac operator directly using two nested conjugate gradient algorithms. At a cost of about a factor of two in operation count it is possible to make the memory usage of direct implementations of the overlap Dirac operator independent of the accuracy of the approximation to the sign function and of the same order as that of standard Wilson fermions

  18. Using context to improve protein domain identification

    Directory of Open Access Journals (Sweden)

    Llinás Manuel

    2011-03-01

    Full Text Available Abstract Background Identifying domains in protein sequences is an important step in protein structural and functional annotation. Existing domain recognition methods typically evaluate each domain prediction independently of the rest. However, the majority of proteins are multidomain, and pairwise domain co-occurrences are highly specific and non-transitive. Results Here, we demonstrate how to exploit domain co-occurrence to boost weak domain predictions that appear in previously observed combinations, while penalizing higher confidence domains if such combinations have never been observed. Our framework, Domain Prediction Using Context (dPUC, incorporates pairwise "context" scores between domains, along with traditional domain scores and thresholds, and improves domain prediction across a variety of organisms from bacteria to protozoa and metazoa. Among the genomes we tested, dPUC is most successful at improving predictions for the poorly-annotated malaria parasite Plasmodium falciparum, for which over 38% of the genome is currently unannotated. Our approach enables high-confidence annotations in this organism and the identification of orthologs to many core machinery proteins conserved in all eukaryotes, including those involved in ribosomal assembly and other RNA processing events, which surprisingly had not been previously known. Conclusions Overall, our results demonstrate that this new context-based approach will provide significant improvements in domain and function prediction, especially for poorly understood genomes for which the need for additional annotations is greatest. Source code for the algorithm is available under a GPL open source license at http://compbio.cs.princeton.edu/dpuc/. Pre-computed results for our test organisms and a web server are also available at that location.

  19. Independent component analysis of high-resolution imaging data identifies distinct functional domains

    DEFF Research Database (Denmark)

    Reidl, Juergen; Starke, Jens; Omer, David

    2007-01-01

    be automatically detected. In the visual cortex orientation columns can be extracted. In all cases artifacts due to movement, heartbeat or respiration were separated from the functional signal by sICA and could be removed from the data set. sICA is therefore a powerful technique for data compression, unbiased...

  20. Fault feature extraction method based on local mean decomposition Shannon entropy and improved kernel principal component analysis model

    Directory of Open Access Journals (Sweden)

    Jinlu Sheng

    2016-07-01

    Full Text Available To effectively extract the typical features of the bearing, a new method that related the local mean decomposition Shannon entropy and improved kernel principal component analysis model was proposed. First, the features are extracted by time–frequency domain method, local mean decomposition, and using the Shannon entropy to process the original separated product functions, so as to get the original features. However, the features been extracted still contain superfluous information; the nonlinear multi-features process technique, kernel principal component analysis, is introduced to fuse the characters. The kernel principal component analysis is improved by the weight factor. The extracted characteristic features were inputted in the Morlet wavelet kernel support vector machine to get the bearing running state classification model, bearing running state was thereby identified. Cases of test and actual were analyzed.

  1. 5W1H Information Extraction with CNN-Bidirectional LSTM

    Science.gov (United States)

    Nurdin, A.; Maulidevi, N. U.

    2018-03-01

    In this work, information about who, did what, when, where, why, and how on Indonesian news articles were extracted by combining Convolutional Neural Network and Bidirectional Long Short-Term Memory. Convolutional Neural Network can learn semantically meaningful representations of sentences. Bidirectional LSTM can analyze the relations among words in the sequence. We also use word embedding word2vec for word representation. By combining these algorithms, we obtained F-measure 0.808. Our experiments show that CNN-BLSTM outperforms other shallow methods, namely IBk, C4.5, and Naïve Bayes with the F-measure 0.655, 0.645, and 0.595, respectively.

  2. Respiratory Information Extraction from Electrocardiogram Signals

    KAUST Repository

    Amin, Gamal El Din Fathy

    2010-12-01

    The Electrocardiogram (ECG) is a tool measuring the electrical activity of the heart, and it is extensively used for diagnosis and monitoring of heart diseases. The ECG signal reflects not only the heart activity but also many other physiological processes. The respiratory activity is a prominent process that affects the ECG signal due to the close proximity of the heart and the lungs. In this thesis, several methods for the extraction of respiratory process information from the ECG signal are presented. These methods allow an estimation of the lung volume and the lung pressure from the ECG signal. The potential benefit of this is to eliminate the corresponding sensors used to measure the respiration activity. A reduction of the number of sensors connected to patients will increase patients’ comfort and reduce the costs associated with healthcare. As a further result, the efficiency of diagnosing respirational disorders will increase since the respiration activity can be monitored with a common, widely available method. The developed methods can also improve the detection of respirational disorders that occur while patients are sleeping. Such disorders are commonly diagnosed in sleeping laboratories where the patients are connected to a number of different sensors. Any reduction of these sensors will result in a more natural sleeping environment for the patients and hence a higher sensitivity of the diagnosis.

  3. Comparison of different soil water extraction systems for the prognoses of solute transport at the field scale using numerical simulations, field and lysimeter experiments

    Energy Technology Data Exchange (ETDEWEB)

    Weihermueller, L

    2005-07-01

    To date, the understanding of processes, factors, and interactions that influence the amount of extracted water and the solute composition sampled with suction cups is limited. But this information is required for process description of solute transport in natural soils. Improved system understanding can lead to a low cost and easy to install water sampling system which can help to predict solute transport in natural soils for the benefit of environmental protection. The main objectives of this work were to perform numerical simulations with different boundary conditions and to implement the findings in the interpretation of the lysimeter and field experiments. In a first part of this thesis, theoretical considerations on the processes affecting the spatial influence of a suction cup in soil and changes in solute transport initiated by the suction cups are presented, including testing and validation of available model and experimental approaches. In the second part, a detailed experimental study was conducted to obtain data for the comparison of the different soil water sampling systems. Finally, the numerical experiments of the suction cup influence were used for the interpretation of the experimental data. The main goals are summarized as follows: - Characterization of the suction cup activity domain (SCAD), suction cup extraction domain (SCED) and suction cup sampling area (SCSA) of active suction cups (definitions are given in Chapter 6). - Determination of the boundary conditions and soil properties [e.g. infiltration, applied suction, duration of water extraction, soil hydraulic properties and soil heterogeneity] affecting the activity domain, extraction domain and sampling area of a suction cup. - Identification of processes that change the travel time and travel time variance of solutes extracted by suction cups. - Validation of the numerically derived data with analytical and experimental data from literature. - Comparison of the experimental data obtained

  4. The extraction and integration framework: a two-process account of statistical learning.

    Science.gov (United States)

    Thiessen, Erik D; Kronstein, Alexandra T; Hufnagle, Daniel G

    2013-07-01

    The term statistical learning in infancy research originally referred to sensitivity to transitional probabilities. Subsequent research has demonstrated that statistical learning contributes to infant development in a wide array of domains. The range of statistical learning phenomena necessitates a broader view of the processes underlying statistical learning. Learners are sensitive to a much wider range of statistical information than the conditional relations indexed by transitional probabilities, including distributional and cue-based statistics. We propose a novel framework that unifies learning about all of these kinds of statistical structure. From our perspective, learning about conditional relations outputs discrete representations (such as words). Integration across these discrete representations yields sensitivity to cues and distributional information. To achieve sensitivity to all of these kinds of statistical structure, our framework combines processes that extract segments of the input with processes that compare across these extracted items. In this framework, the items extracted from the input serve as exemplars in long-term memory. The similarity structure of those exemplars in long-term memory leads to the discovery of cues and categorical structure, which guides subsequent extraction. The extraction and integration framework provides a way to explain sensitivity to both conditional statistical structure (such as transitional probabilities) and distributional statistical structure (such as item frequency and variability), and also a framework for thinking about how these different aspects of statistical learning influence each other. 2013 APA, all rights reserved

  5. Information extraction and knowledge graph construction from geoscience literature

    Science.gov (United States)

    Wang, Chengbin; Ma, Xiaogang; Chen, Jianguo; Chen, Jingwen

    2018-03-01

    Geoscience literature published online is an important part of open data, and brings both challenges and opportunities for data analysis. Compared with studies of numerical geoscience data, there are limited works on information extraction and knowledge discovery from textual geoscience data. This paper presents a workflow and a few empirical case studies for that topic, with a focus on documents written in Chinese. First, we set up a hybrid corpus combining the generic and geology terms from geology dictionaries to train Chinese word segmentation rules of the Conditional Random Fields model. Second, we used the word segmentation rules to parse documents into individual words, and removed the stop-words from the segmentation results to get a corpus constituted of content-words. Third, we used a statistical method to analyze the semantic links between content-words, and we selected the chord and bigram graphs to visualize the content-words and their links as nodes and edges in a knowledge graph, respectively. The resulting graph presents a clear overview of key information in an unstructured document. This study proves the usefulness of the designed workflow, and shows the potential of leveraging natural language processing and knowledge graph technologies for geoscience.

  6. A Time-Space Domain Information Fusion Method for Specific Emitter Identification Based on Dempster-Shafer Evidence Theory.

    Science.gov (United States)

    Jiang, Wen; Cao, Ying; Yang, Lin; He, Zichang

    2017-08-28

    Specific emitter identification plays an important role in contemporary military affairs. However, most of the existing specific emitter identification methods haven't taken into account the processing of uncertain information. Therefore, this paper proposes a time-space domain information fusion method based on Dempster-Shafer evidence theory, which has the ability to deal with uncertain information in the process of specific emitter identification. In this paper, radars will generate a group of evidence respectively based on the information they obtained, and our main task is to fuse the multiple groups of evidence to get a reasonable result. Within the framework of recursive centralized fusion model, the proposed method incorporates a correlation coefficient, which measures the relevance between evidence and a quantum mechanical approach, which is based on the parameters of radar itself. The simulation results of an illustrative example demonstrate that the proposed method can effectively deal with uncertain information and get a reasonable recognition result.

  7. The DNA translocase RAD5A acts independently of the other main DNA repair pathways, and requires both its ATPase and RING domain for activity in Arabidopsis thaliana.

    Science.gov (United States)

    Klemm, Tobias; Mannuß, Anja; Kobbe, Daniela; Knoll, Alexander; Trapp, Oliver; Dorn, Annika; Puchta, Holger

    2017-08-01

    Multiple pathways exist to repair DNA damage induced by methylating and crosslinking agents in Arabidopsis thaliana. The SWI2/SNF2 translocase RAD5A, the functional homolog of budding yeast Rad5 that is required for the error-free branch of post-replicative repair, plays a surprisingly prominent role in the repair of both kinds of lesions in Arabidopsis. Here we show that both the ATPase domain and the ubiquitination function of the RING domain of the Arabidopsis protein are essential for the cellular response to different forms of DNA damage. To define the exact role of RAD5A within the complex network of DNA repair pathways, we crossed the rad5a mutant line with mutants of different known repair factors of Arabidopsis. We had previously shown that RAD5A acts independently of two main pathways of replication-associated DNA repair defined by the helicase RECQ4A and the endonuclease MUS81. The enhanced sensitivity of all double mutants tested in this study indicates that the repair of damaged DNA by RAD5A also occurs independently of nucleotide excision repair (AtRAD1), single-strand break repair (AtPARP1), as well as microhomology-mediated double-strand break repair (AtTEB). Moreover, RAD5A can partially complement for a deficient AtATM-mediated DNA damage response in plants, as the double mutant shows phenotypic growth defects. © 2017 The Authors The Plant Journal © 2017 John Wiley & Sons Ltd.

  8. Extraction of neutron spectral information from Bonner-Sphere data

    CERN Document Server

    Haney, J H; Zaidins, C S

    1999-01-01

    We have extended a least-squares method of extracting neutron spectral information from Bonner-Sphere data which was previously developed by Zaidins et al. (Med. Phys. 5 (1978) 42). A pulse-height analysis with background stripping is employed which provided a more accurate count rate for each sphere. Newer response curves by Mares and Schraube (Nucl. Instr. and Meth. A 366 (1994) 461) were included for the moderating spheres and the bare detector which comprise the Bonner spectrometer system. Finally, the neutron energy spectrum of interest was divided using the philosophy of fuzzy logic into three trapezoidal regimes corresponding to slow, moderate, and fast neutrons. Spectral data was taken using a PuBe source in two different environments and the analyzed data is presented for these cases as slow, moderate, and fast neutron fluences. (author)

  9. Sieve-based relation extraction of gene regulatory networks from biological literature.

    Science.gov (United States)

    Žitnik, Slavko; Žitnik, Marinka; Zupan, Blaž; Bajec, Marko

    2015-01-01

    Relation extraction is an essential procedure in literature mining. It focuses on extracting semantic relations between parts of text, called mentions. Biomedical literature includes an enormous amount of textual descriptions of biological entities, their interactions and results of related experiments. To extract them in an explicit, computer readable format, these relations were at first extracted manually from databases. Manual curation was later replaced with automatic or semi-automatic tools with natural language processing capabilities. The current challenge is the development of information extraction procedures that can directly infer more complex relational structures, such as gene regulatory networks. We develop a computational approach for extraction of gene regulatory networks from textual data. Our method is designed as a sieve-based system and uses linear-chain conditional random fields and rules for relation extraction. With this method we successfully extracted the sporulation gene regulation network in the bacterium Bacillus subtilis for the information extraction challenge at the BioNLP 2013 conference. To enable extraction of distant relations using first-order models, we transform the data into skip-mention sequences. We infer multiple models, each of which is able to extract different relationship types. Following the shared task, we conducted additional analysis using different system settings that resulted in reducing the reconstruction error of bacterial sporulation network from 0.73 to 0.68, measured as the slot error rate between the predicted and the reference network. We observe that all relation extraction sieves contribute to the predictive performance of the proposed approach. Also, features constructed by considering mention words and their prefixes and suffixes are the most important features for higher accuracy of extraction. Analysis of distances between different mention types in the text shows that our choice of transforming

  10. Remote-sensing image encryption in hybrid domains

    Science.gov (United States)

    Zhang, Xiaoqiang; Zhu, Guiliang; Ma, Shilong

    2012-04-01

    Remote-sensing technology plays an important role in military and industrial fields. Remote-sensing image is the main means of acquiring information from satellites, which always contain some confidential information. To securely transmit and store remote-sensing images, we propose a new image encryption algorithm in hybrid domains. This algorithm makes full use of the advantages of image encryption in both spatial domain and transform domain. First, the low-pass subband coefficients of image DWT (discrete wavelet transform) decomposition are sorted by a PWLCM system in transform domain. Second, the image after IDWT (inverse discrete wavelet transform) reconstruction is diffused with 2D (two-dimensional) Logistic map and XOR operation in spatial domain. The experiment results and algorithm analyses show that the new algorithm possesses a large key space and can resist brute-force, statistical and differential attacks. Meanwhile, the proposed algorithm has the desirable encryption efficiency to satisfy requirements in practice.

  11. MPEG-2 Compressed-Domain Algorithms for Video Analysis

    Directory of Open Access Journals (Sweden)

    Hesseler Wolfgang

    2006-01-01

    Full Text Available This paper presents new algorithms for extracting metadata from video sequences in the MPEG-2 compressed domain. Three algorithms for efficient low-level metadata extraction in preprocessing stages are described. The first algorithm detects camera motion using the motion vector field of an MPEG-2 video. The second method extends the idea of motion detection to a limited region of interest, yielding an efficient algorithm to track objects inside video sequences. The third algorithm performs a cut detection using macroblock types and motion vectors.

  12. Protein domain organisation: adding order.

    Science.gov (United States)

    Kummerfeld, Sarah K; Teichmann, Sarah A

    2009-01-29

    reverse orientation in different proteins relative to random graphs with identical degree distributions. While these features were statistically over-represented, they are still fairly rare. Looking in detail at the proteins involved, we found strong functional relationships within each cluster. In addition, the domains tended to be involved in protein-protein interaction and are able to function as independent structural units. A particularly striking example was the human Jak-STAT signalling pathway which makes use of a set of domains in a range of orders and orientations to provide nuanced signaling functionality. This illustrated the importance of functional and structural constraints (or lack thereof) on domain organisation.

  13. Protein domain organisation: adding order

    Directory of Open Access Journals (Sweden)

    Kummerfeld Sarah K

    2009-01-01

    degree of clustering and more domain pairs in forward and reverse orientation in different proteins relative to random graphs with identical degree distributions. While these features were statistically over-represented, they are still fairly rare. Looking in detail at the proteins involved, we found strong functional relationships within each cluster. In addition, the domains tended to be involved in protein-protein interaction and are able to function as independent structural units. A particularly striking example was the human Jak-STAT signalling pathway which makes use of a set of domains in a range of orders and orientations to provide nuanced signaling functionality. This illustrated the importance of functional and structural constraints (or lack thereof on domain organisation.

  14. Europlexus: a domain decomposition method in explicit dynamics

    International Nuclear Information System (INIS)

    Faucher, V.; Hariddh, Bung; Combescure, A.

    2003-01-01

    Explicit time integration methods are used in structural dynamics to simulate fast transient phenomena, such as impacts or explosions. A very fine analysis is required in the vicinity of the loading areas but extending the same method, and especially the same small time-step, to the whole structure frequently yields excessive calculation times. We thus perform a dual Schur domain decomposition, to divide the global problem into several independent ones, to which is added a reduced size interface problem, to ensure connections between sub-domains. Each sub-domain is given its own time-step and its own mesh fineness. Non-matching meshes at the interfaces are handled. An industrial example demonstrates the interest of our approach. (authors)

  15. Curcuma longa extract exerts a myorelaxant effect on the ileum and colon in a mouse experimental colitis model, independent of the anti-inflammatory effect.

    Science.gov (United States)

    Aldini, Rita; Budriesi, Roberta; Roda, Giulia; Micucci, Matteo; Ioan, Pierfranco; D'Errico-Grigioni, Antonia; Sartini, Alessandro; Guidetti, Elena; Marocchi, Margherita; Cevenini, Monica; Rosini, Francesca; Montagnani, Marco; Chiarini, Alberto; Mazzella, Giuseppe

    2012-01-01

    Curcuma has long been used as an anti-inflammatory agent in inflammatory bowel disease. Since gastrointestinal motility is impaired in inflammatory states, the aim of this work was to evaluate if Curcuma Longa had any effect on intestinal motility. The biological activity of Curcuma extract was evaluated against Carbachol induced contraction in isolated mice intestine. Acute and chronic colitis were induced in Balb/c mice by Dextran Sulphate Sodium administration (5% and 2.5% respectively) and either Curcuma extract (200 mg/kg/day) or placebo was thereafter administered for 7 and 21 days respectively. Spontaneous contractions and the response to Carbachol and Atropine of ileum and colon were studied after colitis induction and Curcuma administration. Curcuma extract reduced the spontaneous contractions in the ileum and colon; the maximal response to Carbachol was inhibited in a non-competitive and reversible manner. Similar results were obtained in ileum and colon from Curcuma fed mice. DSS administration decreased the motility, mainly in the colon and Curcuma almost restored both the spontaneous contractions and the response to Carbachol after 14 days assumption, compared to standard diet, but a prolonged assumption of Curcuma decreased the spontaneous and Carbachol-induced contractions. Curcuma extract has a direct and indirect myorelaxant effect on mouse ileum and colon, independent of the anti-inflammatory effect. The indirect effect is reversible and non-competitive with the cholinergic agent. These results suggest the use of curcuma extract as a spasmolytic agent.

  16. Information Extraction for Clinical Data Mining: A Mammography Case Study.

    Science.gov (United States)

    Nassif, Houssam; Woods, Ryan; Burnside, Elizabeth; Ayvaci, Mehmet; Shavlik, Jude; Page, David

    2009-01-01

    Breast cancer is the leading cause of cancer mortality in women between the ages of 15 and 54. During mammography screening, radiologists use a strict lexicon (BI-RADS) to describe and report their findings. Mammography records are then stored in a well-defined database format (NMD). Lately, researchers have applied data mining and machine learning techniques to these databases. They successfully built breast cancer classifiers that can help in early detection of malignancy. However, the validity of these models depends on the quality of the underlying databases. Unfortunately, most databases suffer from inconsistencies, missing data, inter-observer variability and inappropriate term usage. In addition, many databases are not compliant with the NMD format and/or solely consist of text reports. BI-RADS feature extraction from free text and consistency checks between recorded predictive variables and text reports are crucial to addressing this problem. We describe a general scheme for concept information retrieval from free text given a lexicon, and present a BI-RADS features extraction algorithm for clinical data mining. It consists of a syntax analyzer, a concept finder and a negation detector. The syntax analyzer preprocesses the input into individual sentences. The concept finder uses a semantic grammar based on the BI-RADS lexicon and the experts' input. It parses sentences detecting BI-RADS concepts. Once a concept is located, a lexical scanner checks for negation. Our method can handle multiple latent concepts within the text, filtering out ultrasound concepts. On our dataset, our algorithm achieves 97.7% precision, 95.5% recall and an F 1 -score of 0.97. It outperforms manual feature extraction at the 5% statistical significance level.

  17. Magnetization reversal in ferromagnetic spirals via domain wall motion

    Science.gov (United States)

    Schumm, Ryan D.; Kunz, Andrew

    2016-11-01

    Domain wall dynamics have been investigated in a variety of ferromagnetic nanostructures for potential applications in logic, sensing, and recording. We present a combination of analytic and simulated results describing the reliable field driven motion of a domain wall through the arms of a ferromagnetic spiral nanowire. The spiral geometry is capable of taking advantage of the benefits of both straight and circular wires. Measurements of the in-plane components of the spirals' magnetization can be used to determine the angular location of the domain wall, impacting the magnetoresistive applications dependent on the domain wall location. The spirals' magnetization components are found to depend on the spiral parameters: the initial radius and spacing between spiral arms, along with the domain wall location. The magnetization is independent of the parameters of the rotating field used to move the domain wall, and therefore the model is valid for current induced domain wall motion as well. The speed of the domain wall is found to depend on the frequency of the rotating driving field, and the domain wall speeds can be reliably varied over several orders of magnitude. We further demonstrate a technique capable of injecting multiple domain walls and show the reliable and unidirectional motion of domain walls through the arms of the spiral.

  18. 30 CFR 45.4 - Independent contractor register.

    Science.gov (United States)

    2010-07-01

    ... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Independent contractor register. 45.4 Section... ADMINISTRATIVE REQUIREMENTS INDEPENDENT CONTRACTORS § 45.4 Independent contractor register. (a) Each independent contractor shall provide the production-operator in writing the following information: (1) The independent...

  19. Structure determination and analysis of a haemolytic gingipain adhesin domain from Porphyromonas gingivalis

    Energy Technology Data Exchange (ETDEWEB)

    Li, N.; Yun, P.; Nadkarni, M.A.; Ghadikolaee, N.B.; Nguyen, K.A.; Lee, M.; Hunter, N.; Collyer, C.A. (Sydney)

    2010-08-27

    Porphyromonas gingivalis is an obligately anaerobic bacterium recognized as an aetiological agent of adult periodontitis. P. gingivalis produces cysteine proteinases, the gingipains. The crystal structure of a domain within the haemagglutinin region of the lysine gingipain (Kgp) is reported here. The domain was named K2 as it is the second of three homologous structural modules in Kgp. The K2 domain structure is a 'jelly-roll' fold with two anti-parallel {beta}-sheets. This fold topology is shared with adhesive domains from functionally diverse receptors such as MAM domains, ephrin receptor ligand binding domains and a number of carbohydrate binding modules. Possible functions of K2 were investigated. K2 induced haemolysis of erythrocytes in a dose-dependent manner that was augmented by the blocking of anion transport. Further, cysteine-activated arginine gingipain RgpB, which degrades glycophorin A, sensitized erythrocytes to the haemolytic effect of K2. Cleaved K2, similar to that found in extracted Kgp, lacks the haemolytic activity indicating that autolysis of Kgp may be a staged process which is artificially enhanced by extraction of the protein. The data indicate a functional role for K2 in the integrated capacity conferred by Kgp to enable the porphyrin auxotroph P. gingivalis to capture essential haem from erythrocytes.

  20. Interaction Between the Biotin Carboxyl Carrier Domain and the Biotin Carboxylase Domain in Pyruvate Carboxylase from Rhizobium etli†

    Science.gov (United States)

    Lietzan, Adam D.; Menefee, Ann L.; Zeczycki, Tonya N.; Kumar, Sudhanshu; Attwood, Paul V.; Wallace, John C.; Cleland, W. Wallace; Maurice, Martin St.

    2011-01-01

    Pyruvate carboxylase (PC) catalyzes the ATP-dependent carboxylation of pyruvate to oxaloacetate, an important anaplerotic reaction in mammalian tissues. To effect catalysis, the tethered biotin of PC must gain access to active sites in both the biotin carboxylase domain and the carboxyl transferase domain. Previous studies have demonstrated that a mutation of threonine 882 to alanine in PC from Rhizobium etli renders the carboxyl transferase domain inactive and favors the positioning of biotin in the biotin carboxylase domain. We report the 2.4 Å resolution X-ray crystal structure of the Rhizobium etli PC T882A mutant which reveals the first high-resolution description of the domain interaction between the biotin carboxyl carrier protein domain and the biotin carboxylase domain. The overall quaternary arrangement of Rhizobium etli PC remains highly asymmetrical and is independent of the presence of allosteric activator. While biotin is observed in the biotin carboxylase domain, its access to the active site is precluded by the interaction between Arg353 and Glu248, revealing a mechanism for regulating carboxybiotin access to the BC domain active site. The binding location for the biotin carboxyl carrier protein domain demonstrates that tethered biotin cannot bind in the biotin carboxylase domain active site in the same orientation as free biotin, helping to explain the difference in catalysis observed between tethered biotin and free biotin substrates in biotin carboxylase enzymes. Electron density located in the biotin carboxylase domain active site is assigned to phosphonoacetate, offering a probable location for the putative carboxyphosphate intermediate formed during biotin carboxylation. The insights gained from the T882A Rhizobium etli PC crystal structure provide a new series of catalytic snapshots in PC and offer a revised perspective on catalysis in the biotin-dependent enzyme family. PMID:21958016

  1. Datacube Interoperability, Encoding Independence, and Analytics

    Science.gov (United States)

    Baumann, Peter; Hirschorn, Eric; Maso, Joan

    2017-04-01

    Datacubes are commonly accepted as an enabling paradigm which provides a handy abstraction for accessing and analyzing the zillions of image files delivered by the manifold satellite instruments and climate simulations, among others. Additionally, datacubes are the classic model for statistical and OLAP datacubes, so a further information category can be integrated. From a standards perspective, spatio-temporal datacubes naturally are included in the concept of coverages which encompass regular and irregular grids, point clouds, and general meshes - or, more abstractly, digital representations of spatio-temporally varying phenomena. ISO 19123, which is identical to OGC Abstract Topic 6, gives a high-level abstract definition which is complemented by the OGC Coverage Implementation Schema (CIS) which is an interoperable, yet format independent concretization of the abstract model. Currently, ISO is working on adopting OGC CIS as ISO 19123-2; the existing ISO 19123 standard is under revision by one of the abstract authors and will become ISO 19123-1. The roadmap agreed by ISO further foresees adoption of the OGC Web Coverage Service (WCS) as an ISO standard so that a complete data and service model will exist. In 2016, INSPIRE has adopted WCS as Coverage Download Service, including the datacube analytics language Web Coverage Processing Service (WCPS). The rasdaman technology (www.rasdaman.org) is both OGC and INSPIRE Reference Implementation. In the global EarthServer initiative rasdaman database sizes are exceeding 250 TB today, heading for the Petabyte frontier well in 2017. Technically, CIS defines a compact, efficient model for representing multi-dimensional datacubes in several ways. The classical coverage cube defines a domain set (where are values?), a range set (what are these values?), and range type (what do the values mean?), as well as a "bag" for arbitrary metadata. With CIS 1.1, coordinate/value pair sequences have been added, as well as tiled

  2. Domain requirements for the Dock adapter protein in growth- cone signaling.

    Science.gov (United States)

    Rao, Y; Zipursky, S L

    1998-03-03

    Tyrosine phosphorylation has been implicated in growth-cone guidance through genetic, biochemical, and pharmacological studies. Adapter proteins containing src homology 2 (SH2) domains and src homology 3 (SH3) domains provide a means of linking guidance signaling through phosphotyrosine to downstream effectors regulating growth-cone motility. The Drosophila adapter, Dreadlocks (Dock), the homolog of mammalian Nck containing three N-terminal SH3 domains and a single SH2 domain, is highly specialized for growth-cone guidance. In this paper, we demonstrate that Dock can couple signals in either an SH2-dependent or an SH2-independent fashion in photoreceptor (R cell) growth cones, and that Dock displays different domain requirements in different neurons.

  3. WW or WoW: the WW domains in a union of bliss.

    Science.gov (United States)

    Sudol, Marius; Recinos, Claudia C; Abraczinskas, Jennifer; Humbert, Jasper; Farooq, Amjad

    2005-12-01

    WW domains are small protein modules that recognize proline-rich peptide motifs or phosphorylated-serine/threonine proline sites in cognate proteins. Within host proteins these modules are joined to other protein domains or to a variety of catalytic domains acting together as adaptors or targeting anchors of enzymes. An important aspect of signaling by WW domains is their ability to recognize their cognate ligands in tandem. Tandem WW domains not only act in a synergistic manner but also appear to chaperone the function of each other. In this review, we focus on structure, function, and mechanism of the tandem WW domains co-operativity as well as independent actions. We emphasize here the implications of tandem arrangement and cooperative function of the domains for signaling pathways.

  4. Casimir forces in the time domain: Theory

    International Nuclear Information System (INIS)

    Rodriguez, Alejandro W.; McCauley, Alexander P.; Joannopoulos, John D.; Johnson, Steven G.

    2009-01-01

    We present a method to compute Casimir forces in arbitrary geometries and for arbitrary materials based on the finite-difference time-domain (FDTD) scheme. The method involves the time evolution of electric and magnetic fields in response to a set of current sources, in a modified medium with frequency-independent conductivity. The advantage of this approach is that it allows one to exploit existing FDTD software, without modification, to compute Casimir forces. In this paper, we focus on the derivation, implementation choices, and essential properties of the time-domain algorithm, both considered analytically and illustrated in the simplest parallel-plate geometry.

  5. Entity recognition in the biomedical domain using a hybrid approach.

    Science.gov (United States)

    Basaldella, Marco; Furrer, Lenz; Tasso, Carlo; Rinaldi, Fabio

    2017-11-09

    This article describes a high-recall, high-precision approach for the extraction of biomedical entities from scientific articles. The approach uses a two-stage pipeline, combining a dictionary-based entity recognizer with a machine-learning classifier. First, the OGER entity recognizer, which has a bias towards high recall, annotates the terms that appear in selected domain ontologies. Subsequently, the Distiller framework uses this information as a feature for a machine learning algorithm to select the relevant entities only. For this step, we compare two different supervised machine-learning algorithms: Conditional Random Fields and Neural Networks. In an in-domain evaluation using the CRAFT corpus, we test the performance of the combined systems when recognizing chemicals, cell types, cellular components, biological processes, molecular functions, organisms, proteins, and biological sequences. Our best system combines dictionary-based candidate generation with Neural-Network-based filtering. It achieves an overall precision of 86% at a recall of 60% on the named entity recognition task, and a precision of 51% at a recall of 49% on the concept recognition task. These results are to our knowledge the best reported so far in this particular task.

  6. Extraction of prospecting information of uranium deposit based on high spatial resolution satellite data. Taking bashibulake region as an example

    International Nuclear Information System (INIS)

    Yang Xu; Liu Dechang; Zhang Jielin

    2008-01-01

    In this study, the signification and content of prospecting information of uranium deposit are expounded. Quickbird high spatial resolution satellite data are used to extract the prospecting information of uranium deposit in Bashibulake area in the north of Tarim Basin. By using the pertinent methods of image processing, the information of ore-bearing bed, ore-control structure and mineralized alteration have been extracted. The results show a high consistency with the field survey. The aim of this study is to explore practicability of high spatial resolution satellite data for prospecting minerals, and to broaden the thinking of prospectation at similar area. (authors)

  7. An automatic method to generate domain-specific investigator networks using PubMed abstracts

    Directory of Open Access Journals (Sweden)

    Gwinn Marta

    2007-06-01

    Full Text Available Abstract Background Collaboration among investigators has become critical to scientific research. This includes ad hoc collaboration established through personal contacts as well as formal consortia established by funding agencies. Continued growth in online resources for scientific research and communication has promoted the development of highly networked research communities. Extending these networks globally requires identifying additional investigators in a given domain, profiling their research interests, and collecting current contact information. We present a novel strategy for building investigator networks dynamically and producing detailed investigator profiles using data available in PubMed abstracts. Results We developed a novel strategy to obtain detailed investigator information by automatically parsing the affiliation string in PubMed records. We illustrated the results by using a published literature database in human genome epidemiology (HuGE Pub Lit as a test case. Our parsing strategy extracted country information from 92.1% of the affiliation strings in a random sample of PubMed records and in 97.0% of HuGE records, with accuracies of 94.0% and 91.0%, respectively. Institution information was parsed from 91.3% of the general PubMed records (accuracy 86.8% and from 94.2% of HuGE PubMed records (accuracy 87.0. We demonstrated the application of our approach to dynamic creation of investigator networks by creating a prototype information system containing a large database of PubMed abstracts relevant to human genome epidemiology (HuGE Pub Lit, indexed using PubMed medical subject headings converted to Unified Medical Language System concepts. Our method was able to identify 70–90% of the investigators/collaborators in three different human genetics fields; it also successfully identified 9 of 10 genetics investigators within the PREBIC network, an existing preterm birth research network. Conclusion We successfully created a

  8. An automatic method to generate domain-specific investigator networks using PubMed abstracts

    Science.gov (United States)

    Yu, Wei; Yesupriya, Ajay; Wulf, Anja; Qu, Junfeng; Gwinn, Marta; Khoury, Muin J

    2007-01-01

    Background Collaboration among investigators has become critical to scientific research. This includes ad hoc collaboration established through personal contacts as well as formal consortia established by funding agencies. Continued growth in online resources for scientific research and communication has promoted the development of highly networked research communities. Extending these networks globally requires identifying additional investigators in a given domain, profiling their research interests, and collecting current contact information. We present a novel strategy for building investigator networks dynamically and producing detailed investigator profiles using data available in PubMed abstracts. Results We developed a novel strategy to obtain detailed investigator information by automatically parsing the affiliation string in PubMed records. We illustrated the results by using a published literature database in human genome epidemiology (HuGE Pub Lit) as a test case. Our parsing strategy extracted country information from 92.1% of the affiliation strings in a random sample of PubMed records and in 97.0% of HuGE records, with accuracies of 94.0% and 91.0%, respectively. Institution information was parsed from 91.3% of the general PubMed records (accuracy 86.8%) and from 94.2% of HuGE PubMed records (accuracy 87.0). We demonstrated the application of our approach to dynamic creation of investigator networks by creating a prototype information system containing a large database of PubMed abstracts relevant to human genome epidemiology (HuGE Pub Lit), indexed using PubMed medical subject headings converted to Unified Medical Language System concepts. Our method was able to identify 70–90% of the investigators/collaborators in three different human genetics fields; it also successfully identified 9 of 10 genetics investigators within the PREBIC network, an existing preterm birth research network. Conclusion We successfully created a web-based prototype

  9. Extraction of drainage networks from large terrain datasets using high throughput computing

    Science.gov (United States)

    Gong, Jianya; Xie, Jibo

    2009-02-01

    Advanced digital photogrammetry and remote sensing technology produces large terrain datasets (LTD). How to process and use these LTD has become a big challenge for GIS users. Extracting drainage networks, which are basic for hydrological applications, from LTD is one of the typical applications of digital terrain analysis (DTA) in geographical information applications. Existing serial drainage algorithms cannot deal with large data volumes in a timely fashion, and few GIS platforms can process LTD beyond the GB size. High throughput computing (HTC), a distributed parallel computing mode, is proposed to improve the efficiency of drainage networks extraction from LTD. Drainage network extraction using HTC involves two key issues: (1) how to decompose the large DEM datasets into independent computing units and (2) how to merge the separate outputs into a final result. A new decomposition method is presented in which the large datasets are partitioned into independent computing units using natural watershed boundaries instead of using regular 1-dimensional (strip-wise) and 2-dimensional (block-wise) decomposition. Because the distribution of drainage networks is strongly related to watershed boundaries, the new decomposition method is more effective and natural. The method to extract natural watershed boundaries was improved by using multi-scale DEMs instead of single-scale DEMs. A HTC environment is employed to test the proposed methods with real datasets.

  10. A High-Spin Rate Measurement Method for Projectiles Using a Magnetoresistive Sensor Based on Time-Frequency Domain Analysis.

    Science.gov (United States)

    Shang, Jianyu; Deng, Zhihong; Fu, Mengyin; Wang, Shunting

    2016-06-16

    Traditional artillery guidance can significantly improve the attack accuracy and overall combat efficiency of projectiles, which makes it more adaptable to the information warfare of the future. Obviously, the accurate measurement of artillery spin rate, which has long been regarded as a daunting task, is the basis of precise guidance and control. Magnetoresistive (MR) sensors can be applied to spin rate measurement, especially in the high-spin and high-g projectile launch environment. In this paper, based on the theory of a MR sensor measuring spin rate, the mathematical relationship model between the frequency of MR sensor output and projectile spin rate was established through a fundamental derivation. By analyzing the characteristics of MR sensor output whose frequency varies with time, this paper proposed the Chirp z-Transform (CZT) time-frequency (TF) domain analysis method based on the rolling window of a Blackman window function (BCZT) which can accurately extract the projectile spin rate. To put it into practice, BCZT was applied to measure the spin rate of 155 mm artillery projectile. After extracting the spin rate, the impact that launch rotational angular velocity and aspect angle have on the extraction accuracy of the spin rate was analyzed. Simulation results show that the BCZT TF domain analysis method can effectively and accurately measure the projectile spin rate, especially in a high-spin and high-g projectile launch environment.

  11. A Survey: Framework of an Information Retrieval for Malay Translated Hadith Document

    Directory of Open Access Journals (Sweden)

    Zulkefli Nurul Syeilla Syazhween

    2017-01-01

    Full Text Available This paper reviews and analyses the limitation of the existing method used in the IR process in retrieving Malay Translated Hadith documents related to the search request. Traditional Malay Translated Hadith retrieval system has not focused on semantic extraction from text. The bag-of-words representation ignores the conceptual similarity of information in the query text and documents, which produce unsatisfactory retrieval results. Therefore, a more efficient IR framework is needed. This paper claims that the significant information extraction and subject-related information are actually important because the clues from this information can be used to search and find the relevance document to a query. Also, unimportant information can be discarded to represent the document content. So, semantic understanding of query and document is necessary to improve the effectiveness and accuracy of retrieval results for this domain study. Therefore, advance research is needed and it will be experimented in the future work. It is hoped that it will help users to search and find information regarding to the Malay Translated Hadith document.

  12. Independent alignment of RNA for dynamic studies using residual dipolar couplings

    Energy Technology Data Exchange (ETDEWEB)

    Bardaro, Michael F.; Varani, Gabriele, E-mail: varani@chem.washington.edu [University of Washington, Department of Chemistry (United States)

    2012-09-15

    Molecular motion and dynamics play an essential role in the biological function of many RNAs. An important source of information on biomolecular motion can be found in residual dipolar couplings which contain dynamics information over the entire ms-ps timescale. However, these methods are not fully applicable to RNA because nucleic acid molecules tend to align in a highly collinear manner in different alignment media. As a consequence, information on dynamics that can be obtained with this method is limited. In order to overcome this limitation, we have generated a chimeric RNA containing both the wild type TAR RNA, the target of our investigation of dynamics, as well as the binding site for U1A protein. When U1A protein was bound to the portion of the chimeric RNA containing its binding site, we obtained independent alignment of TAR by exploiting the physical chemical characteristics of this protein. This technique can allow the extraction of new information on RNA dynamics, which is particularly important for time scales not covered by relaxation methods where important RNA motions occur.

  13. Extraction of fast neuronal changes from multichannel functional near-infrared spectroscopy signals using independent component analysis

    Science.gov (United States)

    Morren, Geert; Wolf, Martin; Lemmerling, Philippe; Wolf, Ursula; Choi, Jee H.; Gratton, Enrico; De Lathauwer, Lieven; Van Huffel, Sabine

    2002-06-01

    Fast changes in the range of milliseconds in the optical properties of cerebral tissue, which are associated with brain activity, can be detected using non-invasive near-infrared spectroscopy (NIRS). These changes in light scattering are due to an alteration in the refractive index at neuronal membranes. The aim of this study was to develop highly sensitive data analysis algorithms to detect this fast signal, which is small compared to other physiological signals. A frequency-domain tissue oximeter, whose laser diodes were modulated at 110MHz was used. The amplitude, mean intensity and phase of the modulated optical signal was measured at 96Hz sample rate. The probe consisting of 4 crossed source detector pairs was placed above the motor cortex, contralateral to the hand performing a tapping exercise consisting of alternating rest- and tapping periods of 20s each. The tapping frequency, which was set to 3.55Hz or 2.5 times the heart rate of the subject to avoid the influence of harmonics on the signal, could not be observed in any of the individual signals measured by the detectors. An adaptive filter was used to remove the arterial pulsatility from the optical signals. Independent Component Analysis allowed to separate signal components in which the tapping frequency was clearly visible.

  14. Fuzzy concept analysis for semantic knowledge extraction

    OpenAIRE

    De Maio, Carmen

    2012-01-01

    2010 - 2011 Availability of controlled vocabularies, ontologies, and so on is enabling feature to provide some added values in terms of knowledge management. Nevertheless, the design, maintenance and construction of domain ontologies are a human intensive and time consuming task. The Knowledge Extraction consists of automatic techniques aimed to identify and to define relevant concepts and relations of the domain of interest by analyzing structured (relational databases, XML) and unstructu...

  15. Identification of a novel splice variant of human PD-L1 mRNA encoding an isoform-lacking Igv-like domain.

    Science.gov (United States)

    He, Xian-hui; Xu, Li-hui; Liu, Yi

    2005-04-01

    To investigate the expression and regulation of PD-1 ligand 1 (PD-L1) in peripheral blood mononuclear cells (PBMC). The cDNA encoding human PD-L1 precursor was cloned from the total RNA extracted from the resting and phorbol dibutyrate plus ionomycin- or phytohemagglutinin-activated PBMC, by reverse transcription polymerase chain reaction (RT-PCR), and independent clones were sequenced and analyzed. The expression and subcellular localization were examined in transiently transfected cells. The PD-L1 gene expression in different PBMC was also analyzed by RT-PCR. A novel human PD-L1 splice variant was identified from the activated PBMC. It was generated by splicing out exon? encoding an immunoglobulin variable domain (Igv)-like domain but retaining all other exons without a frame-shift. Consequently, the putative translated protein contained all other domains including the transmembrane region except for the Igv-like domain. Furthermore, the conventional isoform was expressed on the plasma surface whereas the novel isoform showed a pattern of intracellular membrane distribution in transiently transfected K562 cells. In addition, the expression pattern of the PD-L1 splice variant was variable in different individuals and in different cellular status. PD-L1 expression may be regulated at the posttranscriptional level through alternative splicing, and modulation of the PD-L1 isoform expression may influence the outcome of specific immune responses in the peripheral tissues.

  16. Comparison of methods of extracting information for meta-analysis of observational studies in nutritional epidemiology

    Directory of Open Access Journals (Sweden)

    Jong-Myon Bae

    2016-01-01

    Full Text Available OBJECTIVES: A common method for conducting a quantitative systematic review (QSR for observational studies related to nutritional epidemiology is the “highest versus lowest intake” method (HLM, in which only the information concerning the effect size (ES of the highest category of a food item is collected on the basis of its lowest category. However, in the interval collapsing method (ICM, a method suggested to enable a maximum utilization of all available information, the ES information is collected by collapsing all categories into a single category. This study aimed to compare the ES and summary effect size (SES between the HLM and ICM. METHODS: A QSR for evaluating the citrus fruit intake and risk of pancreatic cancer and calculating the SES by using the HLM was selected. The ES and SES were estimated by performing a meta-analysis using the fixed-effect model. The directionality and statistical significance of the ES and SES were used as criteria for determining the concordance between the HLM and ICM outcomes. RESULTS: No significant differences were observed in the directionality of SES extracted by using the HLM or ICM. The application of the ICM, which uses a broader information base, yielded more-consistent ES and SES, and narrower confidence intervals than the HLM. CONCLUSIONS: The ICM is advantageous over the HLM owing to its higher statistical accuracy in extracting information for QSR on nutritional epidemiology. The application of the ICM should hence be recommended for future studies.

  17. GlycoDomainViewer

    DEFF Research Database (Denmark)

    Joshi, Hiren J; Jørgensen, Anja; Schjoldager, Katrine T

    2018-01-01

    features, which enhances visibility and accessibility of the wealth of glycoproteomic data being generated. The GlycoDomainViewer enables visual exploration of glycoproteomic data, incorporating information from recent N- and O-glycoproteome studies on human and animal cell lines and some organs and body...

  18. Protein domain recurrence and order can enhance prediction of protein functions

    KAUST Repository

    Abdel Messih, Mario A.

    2012-09-07

    Motivation: Burgeoning sequencing technologies have generated massive amounts of genomic and proteomic data. Annotating the functions of proteins identified in this data has become a big and crucial problem. Various computational methods have been developed to infer the protein functions based on either the sequences or domains of proteins. The existing methods, however, ignore the recurrence and the order of the protein domains in this function inference. Results: We developed two new methods to infer protein functions based on protein domain recurrence and domain order. Our first method, DRDO, calculates the posterior probability of the Gene Ontology terms based on domain recurrence and domain order information, whereas our second method, DRDO-NB, relies on the nave Bayes methodology using the same domain architecture information. Our large-scale benchmark comparisons show strong improvements in the accuracy of the protein function inference achieved by our new methods, demonstrating that domain recurrence and order can provide important information for inference of protein functions. The Author(s) 2012. Published by Oxford University Press.

  19. Institutionalizing the Human Domain: Achieving Cross Domain Synergy for Every Day Missions

    Science.gov (United States)

    2017-04-06

    others rely on more primitive and traditional sources of information or are characterized by strong central government censorship . Technology can play...many societies, others rely on more primitive and traditional sources of information or are characterized by strong central government censorship and...Capabilities for Operations in the Human Domain 92 31 Notes 1 The National Military Strategy of the United States of America , 2015

  20. Evaluation of needle trap micro-extraction and solid-phase micro-extraction: Obtaining comprehensive information on volatile emissions from in vitro cultures.

    Science.gov (United States)

    Oertel, Peter; Bergmann, Andreas; Fischer, Sina; Trefz, Phillip; Küntzel, Anne; Reinhold, Petra; Köhler, Heike; Schubert, Jochen K; Miekisch, Wolfram

    2018-05-14

    Volatile organic compounds (VOCs) emitted from in vitro cultures may reveal information on species and metabolism. Owing to low nmol L -1 concentration ranges, pre-concentration techniques are required for gas chromatography-mass spectrometry (GC-MS) based analyses. This study was intended to compare the efficiency of established micro-extraction techniques - solid-phase micro-extraction (SPME) and needle-trap micro-extraction (NTME) - for the analysis of complex VOC patterns. For SPME, a 75 μm Carboxen®/polydimethylsiloxane fiber was used. The NTME needle was packed with divinylbenzene, Carbopack X and Carboxen 1000. The headspace was sampled bi-directionally. Seventy-two VOCs were calibrated by reference standard mixtures in the range of 0.041-62.24 nmol L -1 by means of GC-MS. Both pre-concentration methods were applied to profile VOCs from cultures of Mycobacterium avium ssp. paratuberculosis. Limits of detection ranged from 0.004 to 3.93 nmol L -1 (median = 0.030 nmol L -1 ) for NTME and from 0.001 to 5.684 nmol L -1 (median = 0.043 nmol L -1 ) for SPME. NTME showed advantages in assessing polar compounds such as alcohols. SPME showed advantages in reproducibility but disadvantages in sensitivity for N-containing compounds. Micro-extraction techniques such as SPME and NTME are well suited for trace VOC profiling over cultures if the limitations of each technique is taken into account. Copyright © 2018 John Wiley & Sons, Ltd.

  1. Exploring quality of life of children with cerebral palsy and intellectual disability: What are the important domains of life?

    Science.gov (United States)

    Davis, E; Reddihough, D; Murphy, N; Epstein, A; Reid, S M; Whitehouse, A; Williams, K; Leonard, H; Downs, J

    2017-11-01

    Although it is estimated that half of all children with cerebral palsy also have comorbid intellectual disability, the domains of quality of life (QOL) important for these children are not well understood. The aim of this study was to identify important domains of QOL for these children and adolescents. Due to the children's communication impairments, qualitative semi-structured interviews were conducted with 18 parents. The children (9 males) had a median age of 12 (range 7 to 17) years at interview and nearly two thirds were classified as Gross Motor Function Classification System IV or V. A grounded theory approach was used to identify domains of QOL. The 11 domains identified as important to QOL were physical health, body comfort, behaviour and emotion, communication, predictability and routine, movement and physical activity, nature and outdoors, variety of activity, independence and autonomy, social connectedness, and access to services. The domains of QOL that emerged from this study will be useful for professionals who support children with cerebral palsy and their families. They will also be important for developing a QOL instrument essential for informing the development of interventions and their monitoring and evaluation. © 2017 John Wiley & Sons Ltd.

  2. Quantifying the mechanisms of domain gain in animal proteins.

    Science.gov (United States)

    Buljan, Marija; Frankish, Adam; Bateman, Alex

    2010-01-01

    Protein domains are protein regions that are shared among different proteins and are frequently functionally and structurally independent from the rest of the protein. Novel domain combinations have a major role in evolutionary innovation. However, the relative contributions of the different molecular mechanisms that underlie domain gains in animals are still unknown. By using animal gene phylogenies we were able to identify a set of high confidence domain gain events and by looking at their coding DNA investigate the causative mechanisms. Here we show that the major mechanism for gains of new domains in metazoan proteins is likely to be gene fusion through joining of exons from adjacent genes, possibly mediated by non-allelic homologous recombination. Retroposition and insertion of exons into ancestral introns through intronic recombination are, in contrast to previous expectations, only minor contributors to domain gains and have accounted for less than 1% and 10% of high confidence domain gain events, respectively. Additionally, exonization of previously non-coding regions appears to be an important mechanism for addition of disordered segments to proteins. We observe that gene duplication has preceded domain gain in at least 80% of the gain events. The interplay of gene duplication and domain gain demonstrates an important mechanism for fast neofunctionalization of genes.

  3. A domain-based approach to predict protein-protein interactions

    Directory of Open Access Journals (Sweden)

    Resat Haluk

    2007-06-01

    Full Text Available Abstract Background Knowing which proteins exist in a certain organism or cell type and how these proteins interact with each other are necessary for the understanding of biological processes at the whole cell level. The determination of the protein-protein interaction (PPI networks has been the subject of extensive research. Despite the development of reasonably successful methods, serious technical difficulties still exist. In this paper we present DomainGA, a quantitative computational approach that uses the information about the domain-domain interactions to predict the interactions between proteins. Results DomainGA is a multi-parameter optimization method in which the available PPI information is used to derive a quantitative scoring scheme for the domain-domain pairs. Obtained domain interaction scores are then used to predict whether a pair of proteins interacts. Using the yeast PPI data and a series of tests, we show the robustness and insensitivity of the DomainGA method to the selection of the parameter sets, score ranges, and detection rules. Our DomainGA method achieves very high explanation ratios for the positive and negative PPIs in yeast. Based on our cross-verification tests on human PPIs, comparison of the optimized scores with the structurally observed domain interactions obtained from the iPFAM database, and sensitivity and specificity analysis; we conclude that our DomainGA method shows great promise to be applicable across multiple organisms. Conclusion We envision the DomainGA as a first step of a multiple tier approach to constructing organism specific PPIs. As it is based on fundamental structural information, the DomainGA approach can be used to create potential PPIs and the accuracy of the constructed interaction template can be further improved using complementary methods. Explanation ratios obtained in the reported test case studies clearly show that the false prediction rates of the template networks constructed

  4. Approaching the largest ‘API’: extracting information from the Internet with Python

    Directory of Open Access Journals (Sweden)

    Jonathan E. Germann

    2018-02-01

    Full Text Available This article explores the need for libraries to algorithmically access and manipulate the world’s largest API: the Internet. The billions of pages on the ‘Internet API’ (HTTP, HTML, CSS, XPath, DOM, etc. are easily accessible and manipulable. Libraries can assist in creating meaning through the datafication of information on the world wide web. Because most information is created for human consumption, some programming is required for automated extraction. Python is an easy-to-learn programming language with extensive packages and community support for web page automation. Four packages (Urllib, Selenium, BeautifulSoup, Scrapy in Python can automate almost any web page for all sized projects. An example warrant data project is explained to illustrate how well Python packages can manipulate web pages to create meaning through assembling custom datasets.

  5. BRIDG: a domain information model for translational and clinical protocol-driven research.

    Science.gov (United States)

    Becnel, Lauren B; Hastak, Smita; Ver Hoef, Wendy; Milius, Robert P; Slack, MaryAnn; Wold, Diane; Glickman, Michael L; Brodsky, Boris; Jaffe, Charles; Kush, Rebecca; Helton, Edward

    2017-09-01

    It is critical to integrate and analyze data from biological, translational, and clinical studies with data from health systems; however, electronic artifacts are stored in thousands of disparate systems that are often unable to readily exchange data. To facilitate meaningful data exchange, a model that presents a common understanding of biomedical research concepts and their relationships with health care semantics is required. The Biomedical Research Integrated Domain Group (BRIDG) domain information model fulfills this need. Software systems created from BRIDG have shared meaning "baked in," enabling interoperability among disparate systems. For nearly 10 years, the Clinical Data Standards Interchange Consortium, the National Cancer Institute, the US Food and Drug Administration, and Health Level 7 International have been key stakeholders in developing BRIDG. BRIDG is an open-source Unified Modeling Language-class model developed through use cases and harmonization with other models. With its 4+ releases, BRIDG includes clinical and now translational research concepts in its Common, Protocol Representation, Study Conduct, Adverse Events, Regulatory, Statistical Analysis, Experiment, Biospecimen, and Molecular Biology subdomains. The model is a Clinical Data Standards Interchange Consortium, Health Level 7 International, and International Standards Organization standard that has been utilized in national and international standards-based software development projects. It will continue to mature and evolve in the areas of clinical imaging, pathology, ontology, and vocabulary support. BRIDG 4.1.1 and prior releases are freely available at https://bridgmodel.nci.nih.gov . © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com

  6. A Methodology to Develop Ontologies for Emerging Domains

    Science.gov (United States)

    Meenorngwar, Chai

    2013-01-01

    The characteristic of complex, dynamic domains, such as an emerging domain, is that the information necessary to describe them is not fully established. Standards are not yet established for these domains, and hence they are difficult to describe and present, and methods are needed that will reflect the changes that will occur as the domains…

  7. Adaptation of interoperability standards for cross domain usage

    Science.gov (United States)

    Essendorfer, B.; Kerth, Christian; Zaschke, Christian

    2017-05-01

    As globalization affects most aspects of modern life, challenges of quick and flexible data sharing apply to many different domains. To protect a nation's security for example, one has to look well beyond borders and understand economical, ecological, cultural as well as historical influences. Most of the time information is produced and stored digitally and one of the biggest challenges is to receive relevant readable information applicable to a specific problem out of a large data stock at the right time. These challenges to enable data sharing across national, organizational and systems borders are known to other domains (e.g., ecology or medicine) as well. Solutions like specific standards have been worked on for the specific problems. The question is: what can the different domains learn from each other and do we have solutions when we need to interlink the information produced in these domains? A known problem is to make civil security data available to the military domain and vice versa in collaborative operations. But what happens if an environmental crisis leads to the need to quickly cooperate with civil or military security in order to save lives? How can we achieve interoperability in such complex scenarios? The paper introduces an approach to adapt standards from one domain to another and lines out problems that have to be overcome and limitations that may apply.

  8. Structure Crack Identification Based on Surface-mounted Active Sensor Network with Time-Domain Feature Extraction and Neural Network

    Directory of Open Access Journals (Sweden)

    Chunling DU

    2012-03-01

    Full Text Available In this work the condition of metallic structures are classified based on the acquired sensor data from a surface-mounted piezoelectric sensor/actuator network. The structures are aluminum plates with riveted holes and possible crack damage at these holes. A 400 kHz sine wave burst is used as diagnostic signals. The combination of time-domain S0 waves from received sensor signals is directly used as features and preprocessing is not needed for the dam age detection. Since the time sequence of the extracted S0 has a high dimension, principal component estimation is applied to reduce its dimension before entering NN (neural network training for classification. An LVQ (learning vector quantization NN is used to classify the conditions as healthy or damaged. A number of FEM (finite element modeling results are taken as inputs to the NN for training, since the simulated S0 waves agree well with the experimental results on real plates. The performance of the classification is then validated by using these testing results.

  9. A state-of-the-art review of built environment information modelling (BeIM

    Directory of Open Access Journals (Sweden)

    Tah J.H.M.

    2017-12-01

    Full Text Available Elements that constitute the built environment are vast and so are the independent systems developed to model its various aspects. Many of these systems have been developed under various assumptions and approaches to execute functions that are distinct, complementary or sometimes similar. Furthermore, these systems are ever increasing in number and often assume similar nomenclatures and acronyms, thereby exacerbating the challenges of understanding their peculiar functions, definitions and differences. The current societal demand to improve sustainability per­formance through collaboration as well as whole-system and through-life thinking is driving the need to integrate independent systems associated with different aspects and scales of the built environment to deliver smart solutions and services that improve the well-being of citizens. The contemporary object-oriented digitization of real-world elements appears to provide a leeway for amalgamating the modelling systems of various domains in the built environment which we termed as built environment information modelling (BeIM. These domains include architecture, engineering, construction as well as urban planning and design. Applications such as building information modelling, geographic information systems and 3D city modelling systems are now being integrated for city modelling purposes. The various works directed at integrating these systems are examined, revealing that current research efforts on integration fall into three categories: (1 data/file conversion systems, (2 semantic mapping systems and (3 the hybrid of both. The review outcome suggests that good knowledge of these domains and how their respective systems operate is vital to pursuing holistic systems integration in the built environment.

  10. A hybrid sales forecasting scheme by combining independent component analysis with K-means clustering and support vector regression.

    Science.gov (United States)

    Lu, Chi-Jie; Chang, Chi-Chang

    2014-01-01

    Sales forecasting plays an important role in operating a business since it can be used to determine the required inventory level to meet consumer demand and avoid the problem of under/overstocking. Improving the accuracy of sales forecasting has become an important issue of operating a business. This study proposes a hybrid sales forecasting scheme by combining independent component analysis (ICA) with K-means clustering and support vector regression (SVR). The proposed scheme first uses the ICA to extract hidden information from the observed sales data. The extracted features are then applied to K-means algorithm for clustering the sales data into several disjoined clusters. Finally, the SVR forecasting models are applied to each group to generate final forecasting results. Experimental results from information technology (IT) product agent sales data reveal that the proposed sales forecasting scheme outperforms the three comparison models and hence provides an efficient alternative for sales forecasting.

  11. Shape of isolated domains in lithium tantalate single crystals at elevated temperatures

    International Nuclear Information System (INIS)

    Shur, V. Ya.; Akhmatkhanov, A. R.; Baturin, I. S.; Chezganov, D. S.; Lobov, A. I.; Smirnov, M. M.

    2013-01-01

    The shape of isolated domains has been investigated in congruent lithium tantalate (CLT) single crystals at elevated temperatures and analyzed in terms of kinetic approach. The obtained temperature dependence of the growing domain shape in CLT including circular shape at temperatures above 190 °C has been attributed to increase of relative input of isotropic ionic conductivity. The observed nonstop wall motion and independent domain growth after merging in CLT as opposed to stoichiometric lithium tantalate have been attributed to difference in wall orientation. The computer simulation has confirmed applicability of the kinetic approach to the domain shape explanation

  12. Applying Different Independent Component Analysis Algorithms and Support Vector Regression for IT Chain Store Sales Forecasting

    Science.gov (United States)

    Dai, Wensheng

    2014-01-01

    Sales forecasting is one of the most important issues in managing information technology (IT) chain store sales since an IT chain store has many branches. Integrating feature extraction method and prediction tool, such as support vector regression (SVR), is a useful method for constructing an effective sales forecasting scheme. Independent component analysis (ICA) is a novel feature extraction technique and has been widely applied to deal with various forecasting problems. But, up to now, only the basic ICA method (i.e., temporal ICA model) was applied to sale forecasting problem. In this paper, we utilize three different ICA methods including spatial ICA (sICA), temporal ICA (tICA), and spatiotemporal ICA (stICA) to extract features from the sales data and compare their performance in sales forecasting of IT chain store. Experimental results from a real sales data show that the sales forecasting scheme by integrating stICA and SVR outperforms the comparison models in terms of forecasting error. The stICA is a promising tool for extracting effective features from branch sales data and the extracted features can improve the prediction performance of SVR for sales forecasting. PMID:25165740

  13. Applying different independent component analysis algorithms and support vector regression for IT chain store sales forecasting.

    Science.gov (United States)

    Dai, Wensheng; Wu, Jui-Yu; Lu, Chi-Jie

    2014-01-01

    Sales forecasting is one of the most important issues in managing information technology (IT) chain store sales since an IT chain store has many branches. Integrating feature extraction method and prediction tool, such as support vector regression (SVR), is a useful method for constructing an effective sales forecasting scheme. Independent component analysis (ICA) is a novel feature extraction technique and has been widely applied to deal with various forecasting problems. But, up to now, only the basic ICA method (i.e., temporal ICA model) was applied to sale forecasting problem. In this paper, we utilize three different ICA methods including spatial ICA (sICA), temporal ICA (tICA), and spatiotemporal ICA (stICA) to extract features from the sales data and compare their performance in sales forecasting of IT chain store. Experimental results from a real sales data show that the sales forecasting scheme by integrating stICA and SVR outperforms the comparison models in terms of forecasting error. The stICA is a promising tool for extracting effective features from branch sales data and the extracted features can improve the prediction performance of SVR for sales forecasting.

  14. Applying Different Independent Component Analysis Algorithms and Support Vector Regression for IT Chain Store Sales Forecasting

    Directory of Open Access Journals (Sweden)

    Wensheng Dai

    2014-01-01

    Full Text Available Sales forecasting is one of the most important issues in managing information technology (IT chain store sales since an IT chain store has many branches. Integrating feature extraction method and prediction tool, such as support vector regression (SVR, is a useful method for constructing an effective sales forecasting scheme. Independent component analysis (ICA is a novel feature extraction technique and has been widely applied to deal with various forecasting problems. But, up to now, only the basic ICA method (i.e., temporal ICA model was applied to sale forecasting problem. In this paper, we utilize three different ICA methods including spatial ICA (sICA, temporal ICA (tICA, and spatiotemporal ICA (stICA to extract features from the sales data and compare their performance in sales forecasting of IT chain store. Experimental results from a real sales data show that the sales forecasting scheme by integrating stICA and SVR outperforms the comparison models in terms of forecasting error. The stICA is a promising tool for extracting effective features from branch sales data and the extracted features can improve the prediction performance of SVR for sales forecasting.

  15. Extracting and Using Photon Polarization Information in Radiative B Decays

    Energy Technology Data Exchange (ETDEWEB)

    Grossman, Yuval

    2000-05-09

    The authors discuss the uses of conversion electron pairs for extracting photon polarization information in weak radiative B decays. Both cases of leptons produced through a virtual and real photon are considered. Measurements of the angular correlation between the (K-pi) and (e{sup +}e{sup {minus}}) decay planes in B --> K*(--> K-pi)gamma (*)(--> e{sup +}e{sup {minus}}) decays can be used to determine the helicity amplitudes in the radiative B --> K*gamma decays. A large right-handed helicity amplitude in B-bar decays is a signal of new physics. The time-dependent CP asymmetry in the B{sup 0} decay angular correlation is shown to measure sin 2-beta and cos 2-beta with little hadronic uncertainty.

  16. Media independence and dividend policy

    DEFF Research Database (Denmark)

    Farooq, Omar; Dandoune, Salma

    2012-01-01

    independence and dividend policies in emerging markets. Using a dataset from twenty three emerging markets, we show a significantly negative relationship between dividend policies (payout ratio and decision to pay dividend) and media independence. We argue that independent media reduces information asymmetries...... for stock market participants. Consequently, stock market participants in emerging markets with more independent media do not demand as high and as much dividends as their counterparts in emerging markets with less independent media. We also show that press independence is more important in defining......Can media pressurize managers to disgorge excess cash to shareholders? Do firms in countries with more independent media follow different dividend policies than firms with less independent media? This paper seeks to answer these questions and aims to document the relationship between media...

  17. Discoidin domain receptor 1 is activated independently of beta(1) integrin

    DEFF Research Database (Denmark)

    Vogel, W; Brakebusch, C; Fässler, R

    2000-01-01

    independent of the epidermal growth factor (EGF) receptor. In cells that endogenously express both DDR1 and the EGF receptor, stimulation with EGF does not induce DDR activation. Third, we detected full DDR1 activation after collagen stimulation in cells that have been treated with blocking antibodies...... for alpha(2)beta(1) integrin or in cells with a targeted deletion of the beta(1) integrin gene. Finally, we show that overexpression of dominant negative DDR1 in the myoblast cell line C2C12 blocks cellular differentiation and the formation of myofibers....

  18. Functional Implications of Domain Organization Within Prokaryotic Rhomboid Proteases.

    Science.gov (United States)

    Panigrahi, Rashmi; Lemieux, M Joanne

    2015-01-01

    Intramembrane proteases are membrane embedded enzymes that cleave transmembrane substrates. This interesting class of enzyme and its water mediated substrate cleavage mechanism occurring within the hydrophobic lipid bilayer has drawn the attention of researchers. Rhomboids are a family of ubiquitous serine intramembrane proteases. Bacterial forms of rhomboid proteases are mainly composed of six transmembrane helices that are preceded by a soluble N-terminal domain. Several crystal structures of the membrane domain of the E. coli rhomboid protease ecGlpG have been solved. Independently, the ecGlpG N-terminal cytoplasmic domain structure was solved using both NMR and protein crystallography. Despite these structures, we still do not know the structure of the full-length protein, nor do we know the functional role of these domains in the cell. This chapter will review the structural and functional roles of the different domains associated with prokaryotic rhomboid proteases. Lastly, we will address questions remaining in the field.

  19. SAR matrices: automated extraction of information-rich SAR tables from large compound data sets.

    Science.gov (United States)

    Wassermann, Anne Mai; Haebel, Peter; Weskamp, Nils; Bajorath, Jürgen

    2012-07-23

    We introduce the SAR matrix data structure that is designed to elucidate SAR patterns produced by groups of structurally related active compounds, which are extracted from large data sets. SAR matrices are systematically generated and sorted on the basis of SAR information content. Matrix generation is computationally efficient and enables processing of large compound sets. The matrix format is reminiscent of SAR tables, and SAR patterns revealed by different categories of matrices are easily interpretable. The structural organization underlying matrix formation is more flexible than standard R-group decomposition schemes. Hence, the resulting matrices capture SAR information in a comprehensive manner.

  20. Extracting chemical information from high-resolution Kβ X-ray emission spectroscopy

    Science.gov (United States)

    Limandri, S.; Robledo, J.; Tirao, G.

    2018-06-01

    High-resolution X-ray emission spectroscopy allows studying the chemical environment of a wide variety of materials. Chemical information can be obtained by fitting the X-ray spectra and observing the behavior of some spectral features. Spectral changes can also be quantified by means of statistical parameters calculated by considering the spectrum as a probability distribution. Another possibility is to perform statistical multivariate analysis, such as principal component analysis. In this work the performance of these procedures for extracting chemical information in X-ray emission spectroscopy spectra for mixtures of Mn2+ and Mn4+ oxides are studied. A detail analysis of the parameters obtained, as well as the associated uncertainties is shown. The methodologies are also applied for Mn oxidation state characterization of double perovskite oxides Ba1+xLa1-xMnSbO6 (with 0 ≤ x ≤ 0.7). The results show that statistical parameters and multivariate analysis are the most suitable for the analysis of this kind of spectra.

  1. Feature extraction and learning using context cue and Rényi entropy based mutual information

    DEFF Research Database (Denmark)

    Pan, Hong; Olsen, Søren Ingvor; Zhu, Yaping

    2015-01-01

    information. In particular, for feature extraction, we develop a new set of kernel descriptors−Context Kernel Descriptors (CKD), which enhance the original KDES by embedding the spatial context into the descriptors. Context cues contained in the context kernel enforce some degree of spatial consistency, thus...... improving the robustness of CKD. For feature learning and reduction, we propose a novel codebook learning method, based on a Rényi quadratic entropy based mutual information measure called Cauchy-Schwarz Quadratic Mutual Information (CSQMI), to learn a compact and discriminative CKD codebook. Projecting...... as the information about the underlying labels of the CKD using CSQMI. Thus the resulting codebook and reduced CKD are discriminative. We verify the effectiveness of our method on several public image benchmark datasets such as YaleB, Caltech-101 and CIFAR-10, as well as a challenging chicken feet dataset of our own...

  2. Accumulation of macular xanthophylls in unsaturated membrane domains.

    Science.gov (United States)

    Wisniewska, Anna; Subczynski, Witold K

    2006-05-15

    The distribution of macular xanthophylls, lutein and zeaxanthin, between domains formed in membranes made from an equimolar ternary mixture of dioleoylphosphatidylcholine/sphingomyelin/cholesterol, called a raft-forming mixture, was investigated. In these membranes, two domains are formed: the raft domain enriched in saturated lipids and cholesterol (detergent-resistant membranes, DRM), and the bulk domain enriched in unsaturated lipids (detergent-soluble membranes, DSM). These membrane domains have been separated using cold Triton X-100 extraction from membranes containing 1 mol% of either lutein or zeaxanthin. The results indicated that xanthophylls are substantially excluded from DRM and remain concentrated in DSM. Concentrations of xanthophylls in DRM and DSM calculated as the mole ratio of either xanthophyll to phospholipid were 0.005 and 0.03, respectively, and calculated as the mole ratio of either xanthophyll to total lipid (phospholipid + cholesterol) were 0.003 and 0.025, respectively. Thus, xanthophylls are over eight times more concentrated in DSM than in DRM. No significant difference in the distribution of lutein and zeaxanthin was found. It was also demonstrated using saturation-recovery EPR that at 1 mol%, neither lutein nor zeaxanthin affect the formation of membrane domains. The location of xanthophylls in domains formed from unsaturated lipids is ideal if they are to act as a lipid antioxidant, which is the most accepted mechanism through which lutein and zeaxanthin protect the retina from age-related macular diseases.

  3. Gimme Context – towards New Domain-Specific Collocational Dictionaries

    Directory of Open Access Journals (Sweden)

    Sylvana Krausse

    2011-04-01

    Full Text Available The days of traditional drudgery-filled lexicography are long gone. Fortunately today, computers help in the enormous task of storing and analysing language in order to condense and store the found information in the form of dictionaries. In this paper, the way from a corpus to a small domain-specific collocational dictionary will be described and thus exemplified based on the example of the domain-specific language of mining reclamation, which can be duplicated for other specific languages too. So far, domain-specific dictionaries are mostly rare as their creation is very labour- and thus cost-effective and all too often they are just a collection of terms plus translation without any information on how to use them in speech. Particular small domains which do not involve a lot of users have been disregarded by lexicographers as there is also always the question of how well it sells afterwards. Following this, I will describe the creation of a small collocational dictionary on mining reclamation language which is based on the consequent use of corpus information. It is relatively quick to realize in the design phase and is thought to provide the sort of linguistic information engineering experts need when they communicate in English or read specialist texts in the specific domain.

  4. Co-occurrence graphs for word sense disambiguation in the biomedical domain.

    Science.gov (United States)

    Duque, Andres; Stevenson, Mark; Martinez-Romo, Juan; Araujo, Lourdes

    2018-05-01

    Word sense disambiguation is a key step for many natural language processing tasks (e.g. summarization, text classification, relation extraction) and presents a challenge to any system that aims to process documents from the biomedical domain. In this paper, we present a new graph-based unsupervised technique to address this problem. The knowledge base used in this work is a graph built with co-occurrence information from medical concepts found in scientific abstracts, and hence adapted to the specific domain. Unlike other unsupervised approaches based on static graphs such as UMLS, in this work the knowledge base takes the context of the ambiguous terms into account. Abstracts downloaded from PubMed are used for building the graph and disambiguation is performed using the personalized PageRank algorithm. Evaluation is carried out over two test datasets widely explored in the literature. Different parameters of the system are also evaluated to test robustness and scalability. Results show that the system is able to outperform state-of-the-art knowledge-based systems, obtaining more than 10% of accuracy improvement in some cases, while only requiring minimal external resources. Copyright © 2018 Elsevier B.V. All rights reserved.

  5. Independent association of glucocorticoids with damage accrual in SLE.

    Science.gov (United States)

    Apostolopoulos, Diane; Kandane-Rathnayake, Rangi; Raghunath, Sudha; Hoi, Alberta; Nikpour, Mandana; Morand, Eric F

    2016-01-01

    To determine factors associated with damage accrual in a prospective cohort of patients with SLE. Patients with SLE who attended the Lupus Clinic at Monash Health, Australia, between 2007 and 2013 were studied. Clinical variables included disease activity (Systemic Lupus Erythematosus Disease Activity Index-2K, SLEDAI-2K), time-adjusted mean SLEDAI, cumulative glucocorticoid dose and organ damage (Systemic Lupus International Collaborating Clinics Damage Index (SDI)). Multivariate logistic regression analyses were performed to identify factors associated with damage accrual. A total of 162 patients were observed over a median (IQR) 3.6 (2.0-4.7) years. Seventy-five per cent (n=121) of patients received glucocorticoids. Damage accrual was significantly more frequent in glucocorticoid-exposed patients (42% vs 15%, p<0.01). Higher glucocorticoid exposure was independently associated with overall damage accrual after controlling for factors including ethnicity and disease activity and was significant at time-adjusted mean doses above 4.42 mg prednisolone/day; the OR of damage accrual in patients in the highest quartile of cumulative glucocorticoid exposure was over 10. Glucocorticoid exposure was independently associated with damage accrual in glucocorticoid-related and non-glucocorticoid related domains of the SDI. Glucocorticoid use is independently associated with the accrual of damage in SLE, including in non-glucocorticoid related domains.

  6. Semantic feature extraction for interior environment understanding and retrieval

    Science.gov (United States)

    Lei, Zhibin; Liang, Yufeng

    1998-12-01

    In this paper, we propose a novel system of semantic feature extraction and retrieval for interior design and decoration application. The system, V2ID(Virtual Visual Interior Design), uses colored texture and spatial edge layout to obtain simple information about global room environment. We address the domain-specific segmentation problem in our application and present techniques for obtaining semantic features from a room environment. We also discuss heuristics for making use of these features (color, texture, edge layout, and shape), to retrieve objects from an existing database. The final resynthesized room environment, with the original scene and objects from the database, is created for the purpose of animation and virtual walk-through.

  7. Time-domain modeling for shielding effectiveness of materials against electromagnetic pulse based on system identification

    International Nuclear Information System (INIS)

    Chen, Xiang; Chen, Yong Guang; Wei, Ming; Hu, Xiao Feng

    2013-01-01

    Shielding effectiveness (SE) of materials against electromagnetic pulse (EMP) cannot be well estimated by traditional test method of SE of materials which only consider the amplitude-frequency characteristic of materials, but ignore the phase-frequency ones. In order to solve this problem, the model of SE of materials against EMP was established based on system identification (SI) method with time-domain linear cosine frequency sweep signal. The feasibility of the method in this paper was examined depending on infinite planar material and the simulation research of coaxial test method and windowed semi-anechoic box of materials. The results show that the amplitude-frequency and phase-frequency information of each frequency can be fully extracted with this method. SE of materials against strong EMP can be evaluated with time-domain low field strength (voltage) of cosine frequency sweep signal. And SE of materials against a variety EMP will be predicted by the model.

  8. Stress induced magnetic-domain evolution in magnetoelectric composites

    Science.gov (United States)

    Trivedi, Harsh; Shvartsman, Vladimir V.; Lupascu, Doru C.; Medeiros, Marco S. A.; Pullar, Robert C.

    2018-06-01

    Local observation of the stress mediated magnetoelectric (ME) effect in composites has gained a great deal of interest over the last decades. However, there is an apparent lack of rigorous methods for a quantitative characterization of the ME effect at the local scale, especially in polycrystalline microstructures. In the present work, we address this issue by locally probing the surface magnetic state of barium titante–hexagonal barium ferrite (BaTiO3–BaFe12O19) ceramic composites using magnetic force microscopy (MFM). The effect of the piezoelectrically induced local stress on the magnetostrictive component (BaFe12O19, BaM) was observed in the form of the evolution of the magnetic domains. The local piezoelectric stress was induced by applying a voltage to the neighboring BaTiO3 grains, using a conductive atomic force microscopy tip. The resulting stochastic evolution of magnetic domains was studied in the context of the induced magnetoelastic anisotropy. In order to overcome the ambiguity in the domain changes observed by MFM, certain generalizations about the observed MFM contrast are put forward, followed by application of an algorithm for extracting the average micromagnetic changes. An average change in domain wall thickness of 50 nm was extracted, giving a lower limit on the corresponding induced magnetoelastic anisotropy energy. Furthermore, we demonstrate that this induced magnetomechanical energy is approximately equal to the K1 magnetocrystalline anisotropy constant of BaM, and compare it with a modeled value of applied elastic energy density. The comparison allowed us to judge the quality of the interfaces in the composite system, by roughly gauging the energy conversion ratio.

  9. Tiered Auctions for Multi-Agent Coordination in Domains with Precedence Constraints

    National Research Council Canada - National Science Library

    Jones, E. G; Dias, M. B; Stentz, Anthony

    2008-01-01

    .... Among these domains are some that require robots to interact closely at particular times and locations to accomplish some task components, but otherwise allow the team members to act independently...

  10. Adaptation of machine translation for multilingual information retrieval in the medical domain.

    Science.gov (United States)

    Pecina, Pavel; Dušek, Ondřej; Goeuriot, Lorraine; Hajič, Jan; Hlaváčová, Jaroslava; Jones, Gareth J F; Kelly, Liadh; Leveling, Johannes; Mareček, David; Novák, Michal; Popel, Martin; Rosa, Rudolf; Tamchyna, Aleš; Urešová, Zdeňka

    2014-07-01

    We investigate machine translation (MT) of user search queries in the context of cross-lingual information retrieval (IR) in the medical domain. The main focus is on techniques to adapt MT to increase translation quality; however, we also explore MT adaptation to improve effectiveness of cross-lingual IR. Our MT system is Moses, a state-of-the-art phrase-based statistical machine translation system. The IR system is based on the BM25 retrieval model implemented in the Lucene search engine. The MT techniques employed in this work include in-domain training and tuning, intelligent training data selection, optimization of phrase table configuration, compound splitting, and exploiting synonyms as translation variants. The IR methods include morphological normalization and using multiple translation variants for query expansion. The experiments are performed and thoroughly evaluated on three language pairs: Czech-English, German-English, and French-English. MT quality is evaluated on data sets created within the Khresmoi project and IR effectiveness is tested on the CLEF eHealth 2013 data sets. The search query translation results achieved in our experiments are outstanding - our systems outperform not only our strong baselines, but also Google Translate and Microsoft Bing Translator in direct comparison carried out on all the language pairs. The baseline BLEU scores increased from 26.59 to 41.45 for Czech-English, from 23.03 to 40.82 for German-English, and from 32.67 to 40.82 for French-English. This is a 55% improvement on average. In terms of the IR performance on this particular test collection, a significant improvement over the baseline is achieved only for French-English. For Czech-English and German-English, the increased MT quality does not lead to better IR results. Most of the MT techniques employed in our experiments improve MT of medical search queries. Especially the intelligent training data selection proves to be very successful for domain adaptation of

  11. Interleukin-4- and NACHT, LRR and PYD domains-containing protein 3-independent mechanisms of alum enhanced T helper type 2 responses on basophils.

    Science.gov (United States)

    Huang, Feng-Juan; Ma, Yi-Lei; Tang, Ruo-Yu; Gong, Wen-Ci; Li, Jun; Chen, Chun-Xia; Yin, Lan; Chen, Xiao-Ping

    2016-10-01

    Aluminium hydroxide (alum), the most widely used adjuvant in human and animal vaccines, has long been known to promote T helper type 2 (Th2) responses and Th2-associated humoral responses, but the mechanisms have remained poorly understood. In this study, we explored whether alum is able to directly modulate antigen-presenting cells to enhance their potency for Th2 polarization. We found that alum treatment of dendritic cells failed to show any Th2-promoting activities. In contrast, alum was able to enhance the capacity of basophils to induce Th2 cells. When basophils from interleukin-4 (IL-4) knockout mice were examined, the intrinsic Th2-promoting activities by basophils were largely abrogated, but the alum-enhanced Th2-promoting activities on basophils were still detectable. More importantly, Th2-promoting adjuvant activities by alum found in IL-4 knockout mice were also largely reduced when basophils were depleted by antibody administration. Therefore, basophils can mediate Th2-promoting activities by alum both in vitro and in vivo through IL-4-independent mechanisms. Further studies revealed that secreted soluble molecules from alum-treated basophils were able to confer the Th2-promoting activities, and neutralization of thymic stromal lymphopoietin or IL-25 attenuated the IL-4-independent development of Th2 cells elicited by alum-treated basophils. Finally, alum was able to activate NACHT, LRR and PYD domains-containing protein 3 (NLRP3) inflammasome in murine basophils in the same way as alum in professional antigen-presenting cells, but NLRP3 was not required for Th2-promoting activities on basophils by alum in vitro. These results demonstrated that alum can enhance the capacities of basophils to polarize Th2 cells via IL-4- and NLRP3-independent pathways. © 2016 John Wiley & Sons Ltd.

  12. Evaluating use of the Siebens Domain Management Model during inpatient rehabilitation to increase functional independence and discharge rate to home in stroke patients.

    Science.gov (United States)

    Kushner, David S; Peters, Kenneth M; Johnson-Greene, Doug

    2015-04-01

    To evaluate use of the Siebens Domain Management Model (SDMM) during stroke inpatient rehabilitation (IR) to increase functional independence and rate of discharge to home. Before and after study. IR facility. Before the intervention: 154 patients with ischemic/hemorrhagic strokes who were admitted to an IR facility in 2010; on average, they were admitted 9.1 days after receiving acute care. After the intervention: 151 patients with ischemic/hemorrhagic strokes who were admitted to an IR facility in 2012; on average they were admitted 7.3 days after receiving acute care. The comorbidity tier severity and prestroke living setting and living support appeared to be similar in both the preintervention and postintervention groups. Use of the SDMM involving weekly adjustments of IR care focused on potential barriers to discharge home including medical/surgical issues, cognitive/emotional coping issues, physical function, and living environment/community re-entry needs. Use of Functional Independence Measure (FIM) score change during IR length of stay (LOS; FIM-LOS efficiency) and rates of discharge to community/home, acute care, and long-term care (LTC) to compare 2010/preintervention data with postintervention data from 2012, along with comparison of facility data to national aggregate data from the Uniform Data System for Medical Rehabilitation (UDSMR) for both years. Preintervention 2010 FIM-LOS efficiency was 1.44 compared with a 2012 postintervention FIM-LOS efficiency of 2.24, which was significant (t = 4.3; P stroke IR may convey improvement in functional independence and is associated with an increased discharge rate to home/community and a reduction in institutionalization and acute-care transfers. Copyright © 2015 American Academy of Physical Medicine and Rehabilitation. Published by Elsevier Inc. All rights reserved.

  13. Synthesis of High-Frequency Ground Motion Using Information Extracted from Low-Frequency Ground Motion

    Science.gov (United States)

    Iwaki, A.; Fujiwara, H.

    2012-12-01

    Broadband ground motion computations of scenario earthquakes are often based on hybrid methods that are the combinations of deterministic approach in lower frequency band and stochastic approach in higher frequency band. Typical computation methods for low-frequency and high-frequency (LF and HF, respectively) ground motions are the numerical simulations, such as finite-difference and finite-element methods based on three-dimensional velocity structure model, and the stochastic Green's function method, respectively. In such hybrid methods, LF and HF wave fields are generated through two different methods that are completely independent of each other, and are combined at the matching frequency. However, LF and HF wave fields are essentially not independent as long as they are from the same event. In this study, we focus on the relation among acceleration envelopes at different frequency bands, and attempt to synthesize HF ground motion using the information extracted from LF ground motion, aiming to propose a new method for broad-band strong motion prediction. Our study area is Kanto area, Japan. We use the K-NET and KiK-net surface acceleration data and compute RMS envelope at four frequency bands: 0.5-1.0 Hz, 1.0-2.0 Hz, 2.0-4.0 Hz, .0-8.0 Hz, and 8.0-16.0 Hz. Taking the ratio of the envelopes of adjacent bands, we find that the envelope ratios have stable shapes at each site. The empirical envelope-ratio characteristics are combined with low-frequency envelope of the target earthquake to synthesize HF ground motion. We have applied the method to M5-class earthquakes and a M7 target earthquake that occurred in the vicinity of Kanto area, and successfully reproduced the observed HF ground motion of the target earthquake. The method can be applied to a broad band ground motion simulation for a scenario earthquake by combining numerically-computed low-frequency (~1 Hz) ground motion with the empirical envelope ratio characteristics to generate broadband ground motion

  14. On thick domain walls in general relativity

    Science.gov (United States)

    Goetz, Guenter; Noetzold, Dirk

    1989-01-01

    Planar scalar field configurations in general relativity differ considerably from those in flat space. It is shown that static domain walls of finite thickness in curved space-time do not possess a reflection symmetry. At infinity, the space-time tends to the Taub vacuum on one side of the wall and to the Minkowski vacuum (Rindler space-time) on the other. Massive test particles are always accelerated towards the Minkowski side, i.e., domain walls are attractive on the Taub side, but repulsive on the Minkowski side (Taub-vacuum cleaner). It is also proved that the pressure in all directions is always negative. Finally, a brief comment is made concerning the possibility of infinite, i.e., bigger than horizon size, domain walls in our universe. All of the results are independent of the form of the potential V(phi) greater than or equal to 0 of the scalar field phi.

  15. Scholarly Information Extraction Is Going to Make a Quantum Leap with PubMed Central (PMC).

    Science.gov (United States)

    Matthies, Franz; Hahn, Udo

    2017-01-01

    With the increasing availability of complete full texts (journal articles), rather than their surrogates (titles, abstracts), as resources for text analytics, entirely new opportunities arise for information extraction and text mining from scholarly publications. Yet, we gathered evidence that a range of problems are encountered for full-text processing when biomedical text analytics simply reuse existing NLP pipelines which were developed on the basis of abstracts (rather than full texts). We conducted experiments with four different relation extraction engines all of which were top performers in previous BioNLP Event Extraction Challenges. We found that abstract-trained engines loose up to 6.6% F-score points when run on full-text data. Hence, the reuse of existing abstract-based NLP software in a full-text scenario is considered harmful because of heavy performance losses. Given the current lack of annotated full-text resources to train on, our study quantifies the price paid for this short cut.

  16. Exploiting the information revolution: call for independent evaluation of the latest English national experiment.

    Science.gov (United States)

    Scott, Philip

    2015-02-19

    The English National Health Service (NHS) has a long history of national experiments with information technology; some successful, others less so. The NHS England Five Year Forward View aspires to 'Exploit the information revolution' through the transformational work of the National Information Board (NIB). NIB has published a 'Framework for Action' that promotes citizen empowerment, information availability, transparency, public trust, innovation, informatics skills and societal value. The framework sets out many laudable and common sense ambitions, but is light on evidence to support its aspirations, or plans for its evaluation. Considerable resource was invested in evaluation studies in the later stages of the National Programme for IT in England, but the analyses do not seem to have been included. Most of the cited evidence is from an unpublished report by management consultants rather than independent peer-reviewed work. National experiments of this importance should be evidence based and properly evaluated so that each iteration of 'information revolution' produces an evidence base to inform subsequent generations of care provision models and technology innovation. Evaluation should be planned from the very start rather than added in as an afterthought. Like any good business plan there should be defined critical success factors for health and social care, and a declaration of how they might be measured. Unintended consequences should be qualitatively explored. Evaluation should also consider critical-interpretive social perspectives to understand the human factors in technology deployment and should seek a theoretically informed insight into the mechanisms of change. The NHS Five Year Forward View and the NIB framework set out challenging ambitions based on transformative use of information technology and collaborative partnerships with commissioners and providers. The framework stands up very well against the Hayes principles, but the learning opportunities

  17. The effect of informed consent on stress levels associated with extraction of impacted mandibular third molars.

    Science.gov (United States)

    Casap, Nardy; Alterman, Michael; Sharon, Guy; Samuni, Yuval

    2008-05-01

    To evaluate the effect of informed consent on stress levels associated with removal of impacted mandibular third molars. A total of 60 patients scheduled for extraction of impacted mandibular third molars participated in this study. The patients were unaware of the study's objectives. Data from 20 patients established the baseline levels of electrodermal activity (EDA). The remaining 40 patients were randomly assigned into 2 equal groups receiving either a detailed document of informed consent, disclosing the possible risks involved with the surgery, or a simplified version. Pulse, blood pressure, and EDA were monitored before, during, and after completion of the consent document. Changes in EDA, but not in blood pressure, were measured on completion of either version of the consent document. A greater increase in EDA was associated with the detailed version of the consent document (P = .004). A similar concomitant increase (although nonsignificant) in pulse values was monitored on completion of both versions. Completion of overdisclosed document of informed consent is associated with changes in physiological parameters. The results suggest that overdetailed listing and disclosure before extraction of impacted mandibular third molars can increase patient stress.

  18. Astrocyte mega-domain hypothesis of the autistic savantism.

    Science.gov (United States)

    Mitterauer, Bernhard J

    2013-01-01

    Individuals with autism who show high abilities are called savants. Whereas in their brains a disconnection in and between neural networks has been identified, savantism is yet poorly understood. Focusing on astrocyte domain organization, it is hypothesized that local astrocyte mega-organizations may be responsible for exerting high capabilities in brains of autistic savants. Astrocytes, the dominant glial cell type, modulate synaptic information transmission. Each astrocyte is organized in non-overlapping domains. Formally, each astrocyte contacting n-neurons with m-synapses via its processes generates dynamic domains of synaptic interactions based on qualitative computation criteria, and hereby it structures neuronal information processing. If the number of processes is genetically significantly increased, these astrocytes operate in a mega-domain with a higher complexitiy of computation. From this model savant abilities are deduced. Copyright © 2012 Elsevier Ltd. All rights reserved.

  19. Method of extracting significant trouble information of nuclear power plants using probabilistic analysis technique

    International Nuclear Information System (INIS)

    Shimada, Yoshio; Miyazaki, Takamasa

    2005-01-01

    In order to analyze and evaluate large amounts of trouble information of overseas nuclear power plants, it is necessary to select information that is significant in terms of both safety and reliability. In this research, a method of efficiently and simply classifying degrees of importance of components in terms of safety and reliability while paying attention to root-cause components appearing in the information was developed. Regarding safety, the reactor core damage frequency (CDF), which is used in the probabilistic analysis of a reactor, was used. Regarding reliability, the automatic plant trip probability (APTP), which is used in the probabilistic analysis of automatic reactor trips, was used. These two aspects were reflected in the development of criteria for classifying degrees of importance of components. By applying these criteria, a simple method of extracting significant trouble information of overseas nuclear power plants was developed. (author)

  20. Unidirectional Magnon-Driven Domain Wall Motion due to Interfacial Dzyaloshinskii-Moriya Interaction

    KAUST Repository

    Lee, Seo-Won

    2018-03-28

    We theoretically study magnon-driven motion of a tranverse domain wall in the presence of interfacial Dzyaloshinskii-Moriya interaction (DMI). Contrary to previous studies, the domain wall moves along the same direction regardless of the magnon-flow direction. Our symmetry analysis reveals that the odd order DMI contributions to the domain wall velocity are independent of the magnon-flow direction. Corresponding DMI-induced asymmetric transitions from a spin-wave state to another give rise to a large momentum transfer to the domain wall without nonreciprocity and much reflection. This counterintuitive unidirectional motion occurs not only for a spin wave with a single wavevector but also for thermal magnons with distributed wavevectors.