WorldWideScience

Sample records for web engineering wilga

  1. Photonics and Web Engineering: WILGA 2009

    CERN Document Server

    Romaniuk, Ryszard

    2009-01-01

    The paper is a digest of work presented during a cyclic Ph.D. student symposium on Photonics and Web Engineering WILGA 2009. The subject of WILGA are Photonics Applications in Astronomy, Communications, Industry and High-Energy Physics Experiments. WILGA is sponsored by EuCARD Project. Symposium is organized by ISE PW in cooperation with professional organizations IEEE, SPIE, PSP and KEiT PAN. There are presented mainly Ph.D. and M.Sc. theses as well as achievements of young researchers. These papers, presented in such a big number, more than 250 in some years, are in certain sense a good digest of the condition of academic research capabilities in this branch of science and technology. The undertaken research subjects for Ph.D. theses in electronics is determined by the interest and research capacity (financial, laboratory and intellectual) of the young researchers and their tutors. Basically, the condition of academic electronics research depends on financing coming from applications areas. During Wilga 200...

  2. Photonics applications and web engineering: WILGA Summer 2016

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2016-09-01

    Wilga Summer 2016 Symposium on Photonics Applications and Web Engineering was held on 29 May - 06 June. The Symposium gathered over 350 participants, mainly young researchers active in optics, optoelectronics, photonics, electronics technologies and applications. There were presented around 300 presentations in a few main topical tracks including: bio-photonics, optical sensory networks, photonics-electronics-mechatronics co-design and integration, large functional system design and maintenance, Internet of Thins, and other. The paper is an introduction the 2016 WILGA Summer Symposium Proceedings, and digests some of the Symposium chosen key presentations.

  3. Photonics and web engineering in Poland, WILGA 2009

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2009-06-01

    The paper is a digest of work presented during a cyclic Ph.D. student symposium on Photonics and Web Engineering WILGA 2009. Symposium is organized by ISE PW in cooperation with professional organizations IEEE, SPIE, PSP and KEiT PAN. There are presented mainly Ph.D. and M.Sc. theses as well as achievements of young researchers. These papers, presented in such a big number, more than 250 in some years, are in certain sense a good digest of the condition of academic research capabilities in this branch of science and technology. The undertaken research subjects for Ph.D. theses in electronics is determined by the interest and research capacity (financial, laboratory and intellectual) of the young researchers and their tutors. Basically, the condition of academic electronics research depends on financing coming from applications areas. During Wilga 2009 there were organized, and thus the paper debates, the following topical sessions concerning applications of advanced electronics and photonics systems: merging of electronic systems and photonics, Internet engineering, distributed measurement systems, security in information technology, astronomy and cosmic technology, HEP experiments, environment protection, image processing and biometry. The paper contains also more general remarks concerning the workshops organized by and for the Ph.D. students in advanced photonics and electronics systems.

  4. Photonics applications and web engineering: WILGA Winter 2016

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2016-09-01

    Since twenty years, young researchers form the Institute of Electronic Systems, Warsaw University of Technology, organize two times a year, under only a marginal supervision of the senior faculty members, under the patronage of WEiTI PW, KEiT PAN, SPIE, IEEE, PKOpto SEP and PSF, the WILGA Symposium on advanced, integrated functional electronic, photonic and mechatronic systems [1-5]. All aspects are considered like: research and development, theory and design, technology - material and construction, software and hardware, commissioning and tests, as well as pilot and practical applications. The applications concern mostly, which turned after several years to be a proud specialization of the WILGA Symposium, Internet engineering, high energy physics experiments, new power industry including fusion, nuclear industry, space and satellite technologies, telecommunications, smart municipal environment, as well as biology and medicine [6-8]. XXXVIIth WILGA Symposium was held on 29-31 January 2016 and gathered a few tens of young researchers active in the mentioned research areas. There were presented a few tens of technical papers which will be published in Proc.SPIE together with the accepted articles from the Summer Edition of the WILGA Symposium scheduled for 29.05-06.06.2016. This article is a digest of chosen presentations from WILGA Symposium 2016 Winter Edition. The survey is narrowed to a few chosen and main topical tracks, like electronics and photonics design using industrial standards like ATCA/MTCA, also particular designs of functional systems using this series of industrial standards. The paper, summarizing traditionally since many years the accomplished WILGA Symposium organized by young researchers from Warsaw University of Technology, is also the following part of a cycle of papers concerning their participation in design of new generations of electronic systems used in discovery experiments in Poland and in leading research laboratories of the world.

  5. Advanced photonic, electronic, and web engineering systems: WILGA Symposium, January 2013

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2013-10-01

    The cycle of WILGA Symposia [wilga.ise.pw.edu.pl] on Photonics and Web Engineering, Advanced Electronic Systems, under the auspices of SPIE, IEEE, KEiT PAN and WEiTI PW was initiated in 1998 by a Research Team PERG/ELHEP ISE PW. The WILGA conferences take place two times a year and the participants are young scientists from this country and abroad. This paper debates chosen topical tracks and some papers presented during the 31 WILGA Multi-Conference, which took place on 8-10 February 2013 at the Faculty of WEiTI PW. The January conference was attended by around 100 persons. Here we discuss closer the subjects of biomedical photonics, electronics and informatics, as well as chosen aspects of applications of advanced photonic, electronic circuits and systems. The 32 nd WILGA Symposium took place on 27 May - 02 June 2013 in WUT WILGA resort near Warsaw. These two editions of WILGA Conferences - January and May have generated more than 250 articles, from which around 100 were chosen by the Symposium and Conference Committees to be published in this volume of Proc.SPIE. WILGA Symposium papers are traditionally submitted via the WILGA web page [wilga.ise.pw.edu.pl] to the SPIE Proceedings publishing system [spie.org]. Email for the correspondence is: photonics@ise.pw.edu.pl. All Wilga papers are published in journals Elektronika, IJET-PAN and in Proc.SPIE. Topical tracks of the symposium usually embrace, among others, new technologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JET and pi-of-the sky experiments development. The symposium In its two editions a year is a summary of the development of numerable Ph.D. theses carried out in this country and this geographical region in the area of advanced electronic and photonic systems. It is also

  6. WILGA Photonics and Web Engineering, January 2012; EuCARD Sessions on HEP and Accelerator Technology

    CERN Document Server

    Romaniuk, R S

    2012-01-01

    Wilga Sessions on HEP experiments and accelerator technology were organized under the umbrella of the EU FP7 Project EuCARD – European Coordination for Accelerator Research and Development. The paper presents a digest of chosen technical work results shown by young researchers from technical universities during the SPIE-IEEE Wilga January 2012 Symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, new technologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JET and pi-of-the sky experiments development. The symposium held two times a year is a summary in the development of numerable Ph.D. theses carried out in this country in the area of advanced electronic and photonic systems. It is also a great occasion for SPIE, IEEE, OSA and PSP st...

  7. Biomedical, Artificial Intelligence, and DNA Computing Photonics Applications and Web Engineering, Wilga, May 2012

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2012-05-01

    This paper is the fifth part (out of five) of the research survey of WILGA Symposium work, May 2012 Edition, concerned with Biomedical, Artificial Intelligence and DNA Computing technologies. It presents a digest of chosen technical work results shown by young researchers from different technical universities from this country during the Jubilee XXXth SPIE-IEEE Wilga 2012, May Edition, symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, nanomaterials and nanotechnologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JET tokamak and pi-of-the sky experiments development. The symposium is an annual summary in the development of numerable Ph.D. theses carried out in this country in the area of advanced electronic and photonic systems. It is also a great occasion for SPIE, IEEE, OSA and PSP students to meet together in a large group spanning the whole country with guests from this part of Europe. A digest of Wilga references is presented [1-270].

  8. Advanced Photonic and Electronic Systems WILGA 2010

    CERN Document Server

    Romaniuk, R S

    2010-01-01

    SPIE – PSP WILGA Symposium gathers two times a year in January and in May new adepts of advanced photonic and electronic systems. The event is oriented on components and applications. WILGA Symposium on Photonics and Web Engineering is well known on the web for its devotion to “young research” promotion under the eminent sponsorship of international engineering associations like SPIE and IEEE and their Poland Sections or Counterparts. WILGA is supported by the most important national professional organizations like KEiT PAN and PSP-Photonics Society of Poland. The Symposium is organized since 1998 twice a year. It has gathered over 4000 young researchers and published over 2000 papers mainly internationally, including more than 900 in 10 published so far volumes of Proc. SPIE. This paper is a digest of WILGA Symposium Series and WILGA 2010 summary. Introductory part treats WILGA Photonics Applications characteristics over the period 1998-2010. Following part presents a short report on the XXVth and XXVI...

  9. Astronomy and Space Technologies, WILGA 2012; EuCARD Sessions

    CERN Document Server

    Romaniuk, R S

    2012-01-01

    Wilga Sessions on HEP experiments, astroparticle physics and accelerator technology were organized under the umbrella of the EU FP7 Project EuCARD – European Coordination for Accelerator Research and Development. This paper is the first part (out of five) of the research survey of WILGA Symposium work, May 2012 Edition, concerned with photonics and electronics applications in astronomy and space technologies. It presents a digest of chosen technical work results shown by young researchers from different technical universities from this country during the Jubilee XXXth SPIE-IEEE Wilga 2012, May Edition, symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, nanomaterials and nanotechnologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JE...

  10. Optical fibers and photonics applications: topical tracks at Wilga conferences

    Science.gov (United States)

    Romaniuk, Ryszard S.

    2013-01-01

    This paper is a research survey of WILGA Symposium work, 2010-2012 Editions, concerned with Optical Fibers, Optoelectronic Devices, Sensors, Communication and Photonics Applications. It presents a digest of chosen technical work results shown by young researchers from different technical universities from this country during the three recent Wilga Symposia on Photonics and Web Engineering. Topical tracks of the symposia embraced, among others, nanomaterials and nanotechnologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, photonic equipment for JET tokamak and pi-of-the sky experiments development. The symposium is an annual summary in the development of numerable Ph.D. theses carried out in this country in the area of advanced electronic and photonic systems. It is also a great occasion for SPIE, IEEE, OSA and PSP students to meet together in a large group spanning the whole country with guests from this part of Europe. A digest of Wilga references is presented [1-274]. Wilga Symposia play a role of an universal integrator of young science in photonics and related areas in this country and also in this part of Europe. More than 5000 young scientists participated in scientific Wilga meetings and discussions during the last nearly two decades. Over 2500 papers were published, including over 1000 in Proc. SPIE.

  11. Advanced Electronic Systems for HEP Experiments, Astroparticle Physics, Accelerator Technology, FELs and Fusion; 2013 WILGA January Symposium (in Polish)

    CERN Document Server

    Romaniuk, R S

    2013-01-01

    The cycle of WILGA conferences [wilga.ise.pw.edu.pl] on Photonics and Web Engineering, Advanced Electronic Systems, under the auspices of IEEE, SPIE, KEiT PAN and WEiTI PW was initiated in 1998 by a Research Team PERG/ELHEP ISE PW. The WILGA conferences take place two times a year and the participants are young scientists from this country and abroad. This paper debates chosen topical tracks and some papers presented during the 31 WILGA Conference, which took place on 8-10 February 2013 at the Faculty of WEiTI PW. The conference was attended by over 60 persons. Here we discuss closer the subjects of biomedical electronics and informatics, as well as chosen aspects of applications of advanced electronic circuits and systems. The next 32 WILGA Conference will take place on 27 May – 02 June 2013 in WUT WILGA resort near Warsaw. Proposed conference papers are submitted via the WILGA Conference web page. Email for the correspondence is: photonics@ise.pw.edu.pl. The papers are published in journals Elektronika, I...

  12. Accelerator Technology and High Energy Physic Experiments, WILGA 2012; EuCARD Sessions

    CERN Document Server

    Romaniuk, R S

    2012-01-01

    Wilga Sessions on HEP experiments, astroparticle physica and accelerator technology were organized under the umbrella of the EU FP7 Project EuCARD – European Coordination for Accelerator Research and Development. The paper is the second part (out of five) of the research survey of WILGA Symposium work, May 2012 Edition, concerned with accelerator technology and high energy physics experiments. It presents a digest of chosen technical work results shown by young researchers from different technical universities from this country during the XXXth Jubilee SPIE-IEEE Wilga 2012, May Edition, symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, nanomaterials and nanotechnologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JET and pi-of-the ...

  13. Photon Physics and Plasma Research, WILGA 2012; EuCARD Sessions

    CERN Document Server

    Romaniuk, R S

    2012-01-01

    Wilga Sessions on HEP experiments, astroparticle physica and accelerator technology were organized under the umbrella of the EU FP7 Project EuCARD – European Coordination for Accelerator Research and Development. This paper is the third part (out of five) of the research survey of WILGA Symposium work, May 2012 Edition, concerned with Photon Physics and Plasma Research. It presents a digest of chosen technical work results shown by young researchers from different technical universities from this country during the Jubilee XXXth SPIE-IEEE Wilga 2012, May Edition, symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, nanomaterials and nanotechnologies for photonics, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonics-electronics co-design, optoelectronic and electronic systems for astronomy and high energy physics experiments, JET tokamak and pi-of-the sky experiments ...

  14. Web Engineering

    Energy Technology Data Exchange (ETDEWEB)

    White, Bebo

    2003-06-23

    Web Engineering is the application of systematic, disciplined and quantifiable approaches to development, operation, and maintenance of Web-based applications. It is both a pro-active approach and a growing collection of theoretical and empirical research in Web application development. This paper gives an overview of Web Engineering by addressing the questions: (a) why is it needed? (b) what is its domain of operation? (c) how does it help and what should it do to improve Web application development? and (d) how should it be incorporated in education and training? The paper discusses the significant differences that exist between Web applications and conventional software, the taxonomy of Web applications, the progress made so far and the research issues and experience of creating a specialization at the master's level. The paper reaches a conclusion that Web Engineering at this stage is a moving target since Web technologies are constantly evolving, making new types of applications possible, which in turn may require innovations in how they are built, deployed and maintained.

  15. Engineering Web Applications

    DEFF Research Database (Denmark)

    Casteleyn, Sven; Daniel, Florian; Dolog, Peter

    Nowadays, Web applications are almost omnipresent. The Web has become a platform not only for information delivery, but also for eCommerce systems, social networks, mobile services, and distributed learning environments. Engineering Web applications involves many intrinsic challenges due...

  16. Photonics and Web Engineering 2011, International Journal of Electronics and Telecommunication, vol.57, no 3, pp.421-428, September 2011

    CERN Document Server

    Romaniuk, R S

    2011-01-01

    The paper presents a digest of chosen technical work results shown by young researchers from different technical universities in this country during the SPIE-IEEE Wilga 2011 symposium on Photonics and Web Engineering. Topical tracks of the symposium embraced, among others, nanomaterials and nanotechnologies for photonics and telecom, sensory and nonlinear optical fibers, object oriented design of hardware, photonic metrology, optoelectronics and photonics applications, photonicselectronics co-design, optoelectronic and electronic systems for telecom, astronomy and high energy physics experiments, JET and pi-of-the sky experiments development. The symposium is an annual summary in the development of numerable Ph.D. theses carried out in this country in the area of advanced electronic and photonic systems. It is also an occasion for young researchers to meet together in a large group (under the patronage of IEEE) spanning the whole country with guests from this part of Europe. A digest of Wilga references is pr...

  17. Web Search Engines

    OpenAIRE

    Rajashekar, TB

    1998-01-01

    The World Wide Web is emerging as an all-in-one information source. Tools for searching Web-based information include search engines, subject directories and meta search tools. We take a look at key features of these tools and suggest practical hints for effective Web searching.

  18. Engineering Web Applications

    DEFF Research Database (Denmark)

    Casteleyn, Sven; Daniel, Florian; Dolog, Peter

    Nowadays, Web applications are almost omnipresent. The Web has become a platform not only for information delivery, but also for eCommerce systems, social networks, mobile services, and distributed learning environments. Engineering Web applications involves many intrinsic challenges due to their......Nowadays, Web applications are almost omnipresent. The Web has become a platform not only for information delivery, but also for eCommerce systems, social networks, mobile services, and distributed learning environments. Engineering Web applications involves many intrinsic challenges due...... to their distributed nature, content orientation, and the requirement to make them available to a wide spectrum of users who are unknown in advance. The authors discuss these challenges in the context of well-established engineering processes, covering the whole product lifecycle from requirements engineering through...

  19. Evaluating web serch engines

    CERN Document Server

    Lewandowski, Dirk

    2011-01-01

    Every month, more than 130 billion queries worldwide are entered into the search boxes of general-purpose web search engines (ComScore, 2010). This enormous number shows that web searching is not only a large business, but also that many people rely on the search engines' results when researching information. A goal of all search engine evaluation efforts is to generate better systems. This goal is of major importance to the search engine vendors who can directly apply evaluation results to develop better ranking algorithms.

  20. Engineering Adaptive Web Applications

    DEFF Research Database (Denmark)

    Dolog, Peter

    2007-01-01

    suit the user profile the most. This paper summarizes the domain engineering framework for such adaptive web applications. The framework provides guidelines to develop adaptive web applications as members of a family. It suggests how to utilize the design artifacts as knowledge which can be used......Information and services on the web are accessible for everyone. Users of the web differ in their background, culture, political and social environment, interests and so on. Ambient intelligence was envisioned as a concept for systems which are able to adapt to user actions and needs....... With the growing amount of information and services, the web applications become natural candidates to adopt the concepts of ambient intelligence. Such applications can deal with divers user intentions and actions based on the user profile and can suggest the combination of information content and services which...

  1. Credibility in Web Search Engines

    OpenAIRE

    Lewandowski, Dirk

    2012-01-01

    Web search engines apply a variety of ranking signals to achieve user satisfaction, i.e., results pages that provide the best-possible results to the user. While these ranking signals implicitly consider credibility (e.g., by measuring popularity), explicit measures of credibility are not applied. In this chapter, credibility in Web search engines is discussed in a broad context: credibility as a measure for including documents in a search engine's index, credibility as a ranking signal, cred...

  2. Web Engineering: A Survey%WebE研究综述

    Institute of Scientific and Technical Information of China (English)

    王昌达; 鞠时光

    2003-01-01

    Web engineering is the establishment and use of sound scientific,engineering and management principlesand disciplined and systematic approaches to the successful development ,deployment and maintenance of high qualityWeb-based systems and applications. This paper presents the history of WebE and its relationship with software engi-neering,furthermore ,we give the popular Web development process and the areas of further study.

  3. Optimization of web pages for search engines

    OpenAIRE

    Harej, Anže

    2011-01-01

    The thesis describes the most important elements of a Web Page and outside factors that affect Search Engine Optimization. The basic structure of a Web page, structure and functionality of a modern Search Engine is described at the beginning. The first section deals with the start of Search Engine Optimization, including planning, analysis of web space and the selection of the most important keywords for which the site will be optimized. The next section Web Page Optimization describes...

  4. Web Search Studies: Multidisciplinary Perspectives on Web Search Engines

    Science.gov (United States)

    Zimmer, Michael

    Perhaps the most significant tool of our internet age is the web search engine, providing a powerful interface for accessing the vast amount of information available on the world wide web and beyond. While still in its infancy compared to the knowledge tools that precede it - such as the dictionary or encyclopedia - the impact of web search engines on society and culture has already received considerable attention from a variety of academic disciplines and perspectives. This article aims to organize a meta-discipline of “web search studies,” centered around a nucleus of major research on web search engines from five key perspectives: technical foundations and evaluations; transaction log analyses; user studies; political, ethical, and cultural critiques; and legal and policy analyses.

  5. Web Search Engines: Search Syntax and Features.

    Science.gov (United States)

    Ojala, Marydee

    2002-01-01

    Presents a chart that explains the search syntax, features, and commands used by the 12 most widely used general Web search engines. Discusses Web standardization, expanded types of content searched, size of databases, and search engines that include both simple and advanced versions. (LRW)

  6. Web Search Engines: Search Syntax and Features.

    Science.gov (United States)

    Ojala, Marydee

    2002-01-01

    Presents a chart that explains the search syntax, features, and commands used by the 12 most widely used general Web search engines. Discusses Web standardization, expanded types of content searched, size of databases, and search engines that include both simple and advanced versions. (LRW)

  7. A Survey on Semantic Web Search Engine

    Directory of Open Access Journals (Sweden)

    G.Sudeepthi

    2012-03-01

    Full Text Available The tremendous growth in the volume of data and with the terrific growth of number of web pages, traditional search engines now a days are not appropriate and not suitable anymore. Search engine is the most important tool to discover any information in World Wide Web. Semantic Search Engine is born of traditional search engine to overcome the above problem. The Semantic Web is an extension of the current web in which information is given well-defined meaning. Semantic web technologies are playing a crucial role in enhancing traditional web search, as it is working to create machine readable data. but it will not replace traditional search engine. In this paper we made a brief survey on various promising features of some of the best semantic search engines developed so far and we have discussed the various approaches to semantic search. We have summarized the techniques, advantages of some important semantic web search engines that are developed so far.The most prominent part is that how the semantic search engines differ from the traditional searches and their results are shown by giving a sample query as input

  8. The Anatomy of Mitos Web Search Engine

    CERN Document Server

    Papadakos, Panagiotis; Theoharis, Yannis; Armenatzoglou, Nikos; Kopidaki, Stella; Marketakis, Yannis; Daskalakis, Manos; Karamaroudis, Kostas; Linardakis, Giorgos; Makrydakis, Giannis; Papathanasiou, Vangelis; Sardis, Lefteris; Tsialiamanis, Petros; Troullinou, Georgia; Vandikas, Kostas; Velegrakis, Dimitris; Tzitzikas, Yannis

    2008-01-01

    Engineering a Web search engine offering effective and efficient information retrieval is a challenging task. This document presents our experiences from designing and developing a Web search engine offering a wide spectrum of functionalities and we report some interesting experimental results. A rather peculiar design choice of the engine is that its index is based on a DBMS, while some of the distinctive functionalities that are offered include advanced Greek language stemming, real time result clustering, and advanced link analysis techniques (also for spam page detection).

  9. Sexual information seeking on web search engines.

    Science.gov (United States)

    Spink, Amanda; Koricich, Andrew; Jansen, B J; Cole, Charles

    2004-02-01

    Sexual information seeking is an important element within human information behavior. Seeking sexually related information on the Internet takes many forms and channels, including chat rooms discussions, accessing Websites or searching Web search engines for sexual materials. The study of sexual Web queries provides insight into sexually-related information-seeking behavior, of value to Web users and providers alike. We qualitatively analyzed queries from logs of 1,025,910 Alta Vista and AlltheWeb.com Web user queries from 2001. We compared the differences in sexually-related Web searching between Alta Vista and AlltheWeb.com users. Differences were found in session duration, query outcomes, and search term choices. Implications of the findings for sexual information seeking are discussed.

  10. Date restricted queries in web search engines

    OpenAIRE

    Lewandowski, Dirk

    2004-01-01

    Search engines usually offer a date restricted search on their advanced search pages. But determining the actual update of a web page is not without problems. We conduct a study testing date restricted queries on the search engines Google, Teoma and Yahoo!. We find that these searches fail to work properly in the examined engines. We discuss implications of this for further research and search engine development.

  11. A Feedback-Based Web Search Engine

    Institute of Scientific and Technical Information of China (English)

    ZHANG Wei-feng; XU Bao-wen; ZHOU Xiao-yu

    2004-01-01

    Web search engines are very useful information service tools in the Internet.The current web search engines produce search results relating to the search terms and the actual information collected by them.Since the selections of the search results cannot affect the future ones, they may not cover most people's interests.In this paper, feedback information produced by the users' accessing lists will be represented by the rough set and can reconstruct the query string and influence the search results.And thus the search engines can provide self-adaptability.

  12. The Use of Web Search Engines in Information Science Research.

    Science.gov (United States)

    Bar-Ilan, Judit

    2004-01-01

    Reviews the literature on the use of Web search engines in information science research, including: ways users interact with Web search engines; social aspects of searching; structure and dynamic nature of the Web; link analysis; other bibliometric applications; characterizing information on the Web; search engine evaluation and improvement; and…

  13. Graph visualization techniques for web clustering engines.

    Science.gov (United States)

    Di Giacomo, Emilio; Didimo, Walter; Grilli, Luca; Liotta, Giuseppe

    2007-01-01

    One of the most challenging issues in mining information from the World Wide Web is the design of systems that present the data to the end user by clustering them into meaningful semantic categories. We show that the analysis of the results of a clustering engine can significantly take advantage of enhanced graph drawing and visualization techniques. We propose a graph-based user interface for Web clustering engines that makes it possible for the user to explore and visualize the different semantic categories and their relationships at the desired level of detail.

  14. Adding a visualization feature to web search engines: it's time.

    Science.gov (United States)

    Wong, Pak Chung

    2008-01-01

    It's widely recognized that all Web search engines today are almost identical in presentation layout and behavior. In fact, the same presentation approach has been applied to depicting search engine results pages (SERPs) since the first Web search engine launched in 1993. In this Visualization Viewpoints article, I propose to add a visualization feature to Web search engines and suggest that the new addition can improve search engines' performance and capabilities, which in turn lead to better Web search technology.

  15. Weighting Relations Using Web Search Engine

    Science.gov (United States)

    Oka, Mizuki; Matsuo, Yutaka

    Measuring the weight of the relation between a pair of entities is necessary to use social networks for various purposes. Intuitively, a pair of entities has a stronger relation than another. It should therefore be weighted higher. We propose a method, using a Web search engine, to compute the weight of the relation existing between a pair of entities. Our method receives a pair of entities and various relations that exist between entities as input. It then outputs the weighted value for the pair of entities. The method explores how search engine results can be used as evidence for how strongly the two entities pertain to the relation.

  16. A Cooperative Schema between Web Sever and Search Engine for Improving Freshness of Web Repository

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    Because the web is huge and web pages are updated frequently, the index maintained by a search engine has to refresh web pages periodically. This is extremely resource consuming because the search engine needs to crawl the web and download web pages to refresh its index. Based on present technologies of web refreshing, we present a cooperative schema between web server and search engine for maintaining freshness of web repository. The web server provides meta-data defined through XML standard to describe web sites. Before updating the web page the crawler visits the meta-data files. If the meta-data indicates that the page is not modified, then the crawler will not update it. So this schema can save bandwidth resource. A primitive model based on the schema is implemented. The cost and efficiency of the schema are analyzed.

  17. Engineering Compensations in Web Service Environment

    DEFF Research Database (Denmark)

    Schäfer, Micahel; Dolog, Peter; Nejdl, Wolfgang

    2007-01-01

    compensations based on forward recovery principles. We extend the existing Web service transaction coordination architecture and infrastructure in order to support flexible compensation operations. A contract-based approach is being used, which allows the specification of permitted compensations at runtime. We......Business to business integration has recently been performed by employing Web service environments. Moreover, such environments are being provided by major players on the technology markets. Those environments are based on open specifications for transaction coordination. When a failure...... in such an environment occurs, a compensation can be initiated to recover from the failure. However, current environments have only limited capabilities for compensations, and are usually based on backward recovery. In this paper, we introduce an engineering approach and an environment to deal with advanced...

  18. An Introduction to Search Engines and Web Navigation

    CERN Document Server

    Levene, Mark

    2010-01-01

    This book is a second edition, updated and expanded to explain the technologies that help us find information on the web.  Search engines and web navigation tools have become ubiquitous in our day to day use of the web as an information source, a tool for commercial transactions and a social computing tool. Moreover, through the mobile web we have access to the web's services when we are on the move.  This book demystifies the tools that we use when interacting with the web, and gives the reader a detailed overview of where we are and where we are going in terms of search engine

  19. Intelligent Semantic Web Search Engines: A Brief Survey

    CERN Document Server

    Madhu, G; Rajinikanth, Dr T V

    2011-01-01

    The World Wide Web (WWW) allows the people to share the information (data) from the large database repositories globally. The amount of information grows billions of databases. We need to search the information will specialize tools known generically search engine. There are many of search engines available today, retrieving meaningful information is difficult. However to overcome this problem in search engines to retrieve meaningful information intelligently, semantic web technologies are playing a major role. In this paper we present survey on the search engine generations and the role of search engines in intelligent web and semantic search technologies.

  20. Design and Implementation of a Simple Web Search Engine

    CERN Document Server

    Mirzal, Andri

    2011-01-01

    We present a simple web search engine for indexing and searching html documents using python programming language. Because python is well known for its simple syntax and strong support for main operating systems, we hope it will be beneficial for learning information retrieval techniques, especially web search engine technology.

  1. Experience Report: Developing the Servo Web Browser Engine using Rust

    OpenAIRE

    Anderson, Brian; Bergstrom, Lars; Herman,David; Matthews, Josh; McAllister, Keegan; Goregaokar, Manish; Moffitt, Jack; Sapin, Simon

    2015-01-01

    All modern web browsers - Internet Explorer, Firefox, Chrome, Opera, and Safari - have a core rendering engine written in C++. This language choice was made because it affords the systems programmer complete control of the underlying hardware features and memory in use, and it provides a transparent compilation model. Servo is a project started at Mozilla Research to build a new web browser engine that preserves the capabilities of these other browser engines but also both takes advantage of ...

  2. Bioprocess-Engineering Education with Web Technology

    NARCIS (Netherlands)

    Sessink, O.

    2006-01-01

    Development of learning material that is distributed through and accessible via the World Wide Web. Various options from web technology are exploited to improve the quality and efficiency of learning material.

  3. A Case Study of Search Engine on World Wide Web for Chemical Fiber Engineering

    Institute of Scientific and Technical Information of China (English)

    张利; 邵世煌; 曾献辉; 尹美华

    2001-01-01

    Search engine is an effective approach to promote the service quality of the World Wide Web. On terms of the analysis of search engines at home and abroad, the developing principle of search engines is given according to the requirement of Web information for chemical fiber engineering. The implementation method for the communication and dynamic refreshment of information on home page of the search engines are elaborated by using programming technology of Active Server Page 3.0 (ASP3.0). The query of chemical fiber information and automatic linking of chemical fiber Web sites can be easily realized by the developed search engine under Internet environment according to users' requirement.

  4. Competence Centered Specialization in Web Engineering Topics in a Software Engineering Masters Degree Programme

    DEFF Research Database (Denmark)

    Dolog, Peter; Thomsen, Lone Leth; Thomsen, Bent

    2010-01-01

    Web applications and Web-based systems are becoming increasingly complex as a result of either customer requests or technology evolution which has eased other aspects of software engineering. Therefore, there is an increasing demand for highly skilled software engineers able to build and also...... as on general competence based web engineering proles oered also for those who specialize in other areas of software engineering. We describe the current state of the art and our experience with a web engineering curriculum within the software engineering masters degree programme. We also discuss an evolution...... advance the systems on the one hand as well as professionals who are able to evaluate their eectiveness on the other hand. With this idea in mind, the computer science department at Aalborg University is continuously working on improvements in its specialization in web engineering topics as well...

  5. Competence Centered Specialization in Web Engineering Topics in a Software Engineering Masters Degree Programme

    DEFF Research Database (Denmark)

    Dolog, Peter; Thomsen, Lone Leth; Thomsen, Bent

    2010-01-01

    Web applications and Web-based systems are becoming increasingly complex as a result of either customer requests or technology evolution which has eased other aspects of software engineering. Therefore, there is an increasing demand for highly skilled software engineers able to build and also...... as on general competence based web engineering proles oered also for those who specialize in other areas of software engineering. We describe the current state of the art and our experience with a web engineering curriculum within the software engineering masters degree programme. We also discuss an evolution...... advance the systems on the one hand as well as professionals who are able to evaluate their eectiveness on the other hand. With this idea in mind, the computer science department at Aalborg University is continuously working on improvements in its specialization in web engineering topics as well...

  6. Web Service Architecture for a Meta Search Engine

    Directory of Open Access Journals (Sweden)

    K.Srinivas

    2011-10-01

    Full Text Available With the rapid advancements in Information Technology, Information Retrieval on Internet is gaining its importance day by day. Nowadays there are millions of Websites and billions of homepages available on the Internet. Search Engines are the essential tools for the purpose of retrieving the required information from the Web. But the existing search engines have many problems such as not having wide scope, imbalance in accessing the sites etc. So, the effectiveness of a search engine plays a vital role. Meta search engines are such systems that can provide effective information by accessing multiple existing search engines such as Dog Pile, Meta Crawler etc, but most of them cannot successfully operate on heterogeneous and fully dynamic web environment. In this paper we propose a Web Service Architecture for Meta Search Engine to cater the need of heterogeneous and dynamic web environment. The objective of our proposal is to exploit most of the features offered by Web Services through the implementation of a Web Service Meta Search Engine.

  7. Using Advanced Search Operators on Web Search Engines.

    Science.gov (United States)

    Jansen, Bernard J.

    Studies show that the majority of Web searchers enter extremely simple queries, so a reasonable system design approach would be to build search engines to compensate for this user characteristic. One hundred representative queries were selected from the transaction log of a major Web search service. These 100 queries were then modified using the…

  8. Using Advanced Search Operators on Web Search Engines.

    Science.gov (United States)

    Jansen, Bernard J.

    Studies show that the majority of Web searchers enter extremely simple queries, so a reasonable system design approach would be to build search engines to compensate for this user characteristic. One hundred representative queries were selected from the transaction log of a major Web search service. These 100 queries were then modified using the…

  9. A CLIR Interface to a Web search engine.

    Science.gov (United States)

    Daumke, Philipp; Schulz, Stefan; Markó, Kornél

    2005-01-01

    Medical document retrieval presents a unique combination of challenges for the design and implementation of retrieval engines. We introduce a method to meet these challenges by implementing a multilingual retrieval interface for biomedical content in the World Wide Web. To this end we developed an automated method for interlingual query construction by which a standard Web search engine is enabled to process non-English queries from the biomedical domain in order to retrieve English documents.

  10. F-OWL: An Inference Engine for Semantic Web

    Science.gov (United States)

    Zou, Youyong; Finin, Tim; Chen, Harry

    2004-01-01

    Understanding and using the data and knowledge encoded in semantic web documents requires an inference engine. F-OWL is an inference engine for the semantic web language OWL language based on F-logic, an approach to defining frame-based systems in logic. F-OWL is implemented using XSB and Flora-2 and takes full advantage of their features. We describe how F-OWL computes ontology entailment and compare it with other description logic based approaches. We also describe TAGA, a trading agent environment that we have used as a test bed for F-OWL and to explore how multiagent systems can use semantic web concepts and technology.

  11. A Domain Specific Ontology Based Semantic Web Search Engine

    CERN Document Server

    Mukhopadhyay, Debajyoti; Mukherjee, Sreemoyee; Bhattacharya, Jhilik; Kim, Young-Chon

    2011-01-01

    Since its emergence in the 1990s the World Wide Web (WWW) has rapidly evolved into a huge mine of global information and it is growing in size everyday. The presence of huge amount of resources on the Web thus poses a serious problem of accurate search. This is mainly because today's Web is a human-readable Web where information cannot be easily processed by machine. Highly sophisticated, efficient keyword based search engines that have evolved today have not been able to bridge this gap. So comes up the concept of the Semantic Web which is envisioned by Tim Berners-Lee as the Web of machine interpretable information to make a machine processable form for expressing information. Based on the semantic Web technologies we present in this paper the design methodology and development of a semantic Web search engine which provides exact search results for a domain specific search. This search engine is developed for an agricultural Website which hosts agricultural information about the state of West Bengal.

  12. Text Retrieval Online: Historical Perspective on Web Search Engines.

    Science.gov (United States)

    Hahn, Trudi Bellardo

    1998-01-01

    Provides an overview of online systems and search engines, highlighting search (relationships between terms and interpretation of words), browse, and Web search engine capabilities, iterative searches, canned or stored queries, vocabulary browsing, delivery of full source documents, simple and advanced user interfaces, and global access. Notes…

  13. Social Dimension of Web 2.0 in Engineering Education

    Science.gov (United States)

    Ahrens, Andreas; Zascerinska, Jelena

    2010-01-01

    Contemporary engineers need to become more cognizant and more responsive to the emerging needs of the market for engineering and technology services. Social dimension of Web 2.0 which penetrates our society more thoroughly with the availability of broadband services has the potential to contribute decisively to the sustainable development of…

  14. Engineering Compensations in Web Service Environment

    DEFF Research Database (Denmark)

    Schäfer, Micahel; Dolog, Peter; Nejdl, Wolfgang

    2007-01-01

    Business to business integration has recently been performed by employing Web service environments. Moreover, such environments are being provided by major players on the technology markets. Those environments are based on open specifications for transaction coordination. When a failure...... compensations based on forward recovery principles. We extend the existing Web service transaction coordination architecture and infrastructure in order to support flexible compensation operations. A contract-based approach is being used, which allows the specification of permitted compensations at runtime. We...

  15. GoWeb: a semantic search engine for the life science web.

    Science.gov (United States)

    Dietze, Heiko; Schroeder, Michael

    2009-10-01

    Current search engines are keyword-based. Semantic technologies promise a next generation of semantic search engines, which will be able to answer questions. Current approaches either apply natural language processing to unstructured text or they assume the existence of structured statements over which they can reason. Here, we introduce a third approach, GoWeb, which combines classical keyword-based Web search with text-mining and ontologies to navigate large results sets and facilitate question answering. We evaluate GoWeb on three benchmarks of questions on genes and functions, on symptoms and diseases, and on proteins and diseases. The first benchmark is based on the BioCreAtivE 1 Task 2 and links 457 gene names with 1352 functions. GoWeb finds 58% of the functional GeneOntology annotations. The second benchmark is based on 26 case reports and links symptoms with diseases. GoWeb achieves 77% success rate improving an existing approach by nearly 20%. The third benchmark is based on 28 questions in the TREC genomics challenge and links proteins to diseases. GoWeb achieves a success rate of 79%. GoWeb's combination of classical Web search with text-mining and ontologies is a first step towards answering questions in the biomedical domain. GoWeb is online at: http://www.gopubmed.org/goweb.

  16. Semantic Web and Model-Driven Engineering

    CERN Document Server

    Parreiras, Fernando S

    2012-01-01

    The next enterprise computing era will rely on the synergy between both technologies: semantic web and model-driven software development (MDSD). The semantic web organizes system knowledge in conceptual domains according to its meaning. It addresses various enterprise computing needs by identifying, abstracting and rationalizing commonalities, and checking for inconsistencies across system specifications. On the other side, model-driven software development is closing the gap among business requirements, designs and executables by using domain-specific languages with custom-built syntax and se

  17. Measuring Security of Web Services in Requirement Engineering Phase

    Directory of Open Access Journals (Sweden)

    Davoud Mougouei

    2015-05-01

    Full Text Available Addressing security in early stages of web service development has always been a major engineering trend. However, to assure security of web services it is required to perform security evaluation in a rigorous and tangible manner. The results of such an evaluation if performed in early stages of the development process can be used to improve the quality of the target web service. On the other hand, it is impossible to remove all of the security faults during the security analysis of web services. As a result, absolute security is never possible to achieve and a security failure may occur during the execution of web service. To avoid security failures, a measurable level of fault tolerance is required to be achieved through partial satisfaction of security goals. Thus any proposed measurement technique must care for this partiality. Even though there are some approaches toward assessing the security of web services but still there is no precise model for evaluation of security goal satisfaction specifically during the requirement engineering phase. This paper introduces a Security Measurement Model (SMM for evaluating the Degree of Security (DS in security requirements of web services by taking into consideration partial satisfaction of security goals. The proposed model evaluates overall security of the target service through measuring the security in Security Requirement Model (SRM of the service. The proposed SMM also takes into account cost, technical ability, impact and flexibility as the key features of security evaluation.

  18. Getting Off the Beaten Track: Specialized Web Search Engines.

    Science.gov (United States)

    Sullivan, Danny

    1998-01-01

    Describes specialty or vertical Web search engines that may provide more relevant results for information retrieval. Highlights include regional services, including filtering by domain and custom crawling; language searching; family-safe listings, including the pros and cons of filtering; news searches; and subject-oriented searching. (LRW)

  19. Computing Semantic Similarity Measure Between Words Using Web Search Engine

    Directory of Open Access Journals (Sweden)

    Pushpa C N

    2013-05-01

    Full Text Available Semantic Similarity measures between words plays an important role in information retrieval, natural language processing and in various tasks on the web. In this paper, we have proposed a Modified Pattern Extraction Algorithm to compute th e supervised semantic similarity measure between the words by combining both page count meth od and web snippets method. Four association measures are used to find semantic simi larity between words in page count method using web search engines. We use a Sequential Minim al Optimization (SMO support vector machines (SVM to find the optimal combination of p age counts-based similarity scores and top-ranking patterns from the web snippets method. The SVM is trained to classify synonymous word-pairs and non-synonymous word-pairs. The propo sed Modified Pattern Extraction Algorithm outperforms by 89.8 percent of correlatio n value.

  20. Analysis of the Temporal Behaviour of Search Engine Crawlers at Web Sites

    Directory of Open Access Journals (Sweden)

    Jeeva Jose

    2013-06-01

    Full Text Available Web log mining is the extraction of web logs to analyze user behaviour at web sites. In addition to user information, web logs provide immense information about search engine traffic and behaviour. Search engine crawlers are highly automated programs that periodically visit the web site to collect information. The behaviour of search engines could be used in analyzing server load, quality of search engines, dynamics of search engine crawlers, ethics of search engines etc. The time spent by various crawlers is significant in identifying the server load as major proportion of the server load is constituted by search engine crawlers. A temporal analysis of the search engine crawlers were done to identify their behaviour. It was found that there is a significant difference in the total time spent by various crawlers. The presence of search engine crawlers at web sites on hourly basis was also done to identify the dynamics of search engine crawlers at web sites.

  1. Classifying web genres in context: a case study documenting the web genres used by a software engineer

    NARCIS (Netherlands)

    Montesi, M.; Navarrete, T.

    2008-01-01

    This case study analyzes the Internet-based resources that a software engineer uses in his daily work. Methodologically, we studied the web browser history of the participant, classifying all the web pages he had seen over a period of 12 days into web genres. We interviewed him before and after the

  2. Folksonomies, the Web and Search Engines

    Directory of Open Access Journals (Sweden)

    Louise Spiteri

    2008-09-01

    Full Text Available The aim of this special issue of Webology is to explore developments in the design of folksonomies, knowledge organization systems, and search engines to reflect end user preferences for describing items of interest. Particular emphasis is placed on folksonomies, an area of study that has grown exponentially since the term was first coined by Thomas Vander Wal in 2004: "Folksonomy is the result of personal free tagging of information and objects (anything with a URL for one's own retrieval. The tagging is done in a social environment (usually shared and open to others. Folksonomy is created from the act of tagging by the person consuming the information" (Vander Wal, 2007. Since 2004, social software applications and their use of tagging have continued to increase in popularity; in its site dedicated to such applications, Wikipedia (2008 lists no less that 11 extant media sharing sites and 26 social bookmarking sites. This list does not take into account the approximate 20 media cataloguing sites, not to mention the innumerable blogging sites that employ tagging.

  3. Web Engineering as a Specialization of Software Engineering: Differences in Project Management Education

    Directory of Open Access Journals (Sweden)

    Herwig Mayr

    2005-10-01

    Full Text Available We present the motivation and our concept of introducing "Web Engineering" as a specialization of our "Software Engineering" curriculum. Our main focus lies on the differences in project management education for both areas as well as the necessary process models and tools. First we discuss the principal differences of software project management and web project management, focusing on the main difficulties of teaching such management skills to primarily technophile students. Then we analyze the composition of modern software development teams and changes within such teams implied by the development of web applications. We illustrate this transition showing how a merely document-driven process - as can be found in many traditional software development projects - is turned into a highly tool-supported, agile development process, which is characteristic for web development projects.

  4. Relevant Pages in semantic Web Search Engines using Ontology

    Directory of Open Access Journals (Sweden)

    Jemimah Simon

    2012-03-01

    Full Text Available In general, search engines are the most popular means of searching any kind of information from the Internet. Generally, keywords are given to the search engine and the Web database returns the documents containing specified keywords. In many situations, irrelevant results are given as results to the user query since different keywords are used in different forms in various documents. The development of the next generation Web, Semantic Web, will change this situation. This paper proposes a prototype of relation-based search engine which ranks the page according to the user query and on annotated results. Page sub graph is computed for each annotated page in the result set by generating all possible combinations for the relation in the sub graph. A relevance score is computed for each annotated page using a probability measure. A relation based ranking model is used which displays the pages in the final result set according to their relevance score. This ranking is provided by considering keyword-concept associations. Thus, the final result set contains pages in the order of their constrained relevant scores.

  5. A development process meta-model for Web based expert systems: The Web engineering point of view

    DEFF Research Database (Denmark)

    Dokas, I.M.; Alapetite, Alexandre

    2006-01-01

    Similar to many legacy computer systems, expert systems can be accessed via the Web, forming a set of Web applications known as Web based expert systems. The tough Web competition, the way people and organizations rely on Web applications and theincreasing user requirements for better services have...... raised their complexity. Unfortunately, there is so far no clear answer to the question: How may the methods and experience of Web engineering and expert systems be combined and applied in order todevelop effective and successful Web based expert systems? In an attempt to answer this question......, a development process meta-model for Web based expert systems will be presented. Based on this meta-model, a publicly available Web based expert systemcalled Landfill Operation Management Advisor (LOMA) was developed. In addition, the results of an accessibility evaluation on LOMA – the first ever reported...

  6. Situational Requirements Engineering for the Development of Content Management System-based Web Applications

    OpenAIRE

    2005-01-01

    Web applications are evolving towards strong content-centered Web applications. The development processes and implementation of these applications are unlike the development and implementation of traditional information systems. In this paper we propose WebEngineering Method; a method for developing content management system (CMS) based Web applications. Critical to a successful development of CMS-based Web applications, is the adaptation to the dynamic business. We first define CMS-based Web...

  7. Web Spam, Social Propaganda and the Evolution of Search Engine Rankings

    Science.gov (United States)

    Metaxas, Panagiotis Takis

    Search Engines have greatly influenced the way we experience the web. Since the early days of the web, users have been relying on them to get informed and make decisions. When the web was relatively small, web directories were built and maintained using human experts to screen and categorize pages according to their characteristics. By the mid 1990's, however, it was apparent that the human expert model of categorizing web pages does not scale. The first search engines appeared and they have been evolving ever since, taking over the role that web directories used to play.

  8. The invisible Web uncovering information sources search engines can't see

    CERN Document Server

    Sherman, Chris

    2001-01-01

    Enormous expanses of the Internet are unreachable with standard web search engines. This book provides the key to finding these hidden resources by identifying how to uncover and use invisible web resources. Mapping the invisible Web, when and how to use it, assessing the validity of the information, and the future of Web searching are topics covered in detail. Only 16 percent of Net-based information can be located using a general search engine. The other 84 percent is what is referred to as the invisible Web-made up of information stored in databases. Unlike pages on the visible Web, informa

  9. EuroGOV: Engineering a Multilingual Web Corpus

    NARCIS (Netherlands)

    Sigurbjörnsson, B.; Kamps, J.; de Rijke, M.

    2005-01-01

    EuroGOV is a multilingual web corpus that was created to serve as the document collection for WebCLEF, the CLEF 2005 web retrieval task. EuroGOV is a collection of web pages crawled from the European Union portal, European Union member state governmental web sites, and Russian government web sites.

  10. EuroGOV: Engineering a Multilingual Web Corpus

    NARCIS (Netherlands)

    Sigurbjörnsson, B.; Kamps, J.; de Rijke, M.

    2005-01-01

    EuroGOV is a multilingual web corpus that was created to serve as the document collection for WebCLEF, the CLEF 2005 web retrieval task. EuroGOV is a collection of web pages crawled from the European Union portal, European Union member state governmental web sites, and Russian government web sites.

  11. Engineering Semantic Web Applications by Using Object-Oriented Paradigm

    CERN Document Server

    Farooq, Amjad; Shah, Abad

    2010-01-01

    The web information resources are growing explosively in number and volume. Now to retrieve relevant data from web has become very difficult and time-consuming. Semantic Web envisions that these web resources should be developed in machine-processable way in order to handle irrelevancy and manual processing problems. Whereas, the Semantic Web is an extension of current web, in which web resources are equipped with formal semantics about their interpretation through machines. These web resources are usually contained in web applications and systems, and their formal semantics are normally represented in the form of web-ontologies. In this research paper, an object-oriented design methodology (OODM) is upgraded for developing semantic web applications. OODM has been developed for designing of web applications for the current web. This methodology is good enough to develop web applications. It also provides a systematic approach for the web applications development but it is not helpful in generating machine-poc...

  12. Web Feet Guide to Search Engines: Finding It on the Net.

    Science.gov (United States)

    Web Feet, 2001

    2001-01-01

    This guide to search engines for the World Wide Web discusses selecting the right search engine; interpreting search results; major search engines; online tutorials and guides; search engines for kids; specialized search tools for various subjects; and other specialized engines and gateways. (LRW)

  13. Journey of Web Search Engines: Milestones, Challenges & Innovations

    Directory of Open Access Journals (Sweden)

    Mamta Kathuria

    2016-12-01

    Full Text Available Past few decades have witnessed an information big bang in the form of World Wide Web leading to gigantic repository of heterogeneous data. A humble journey that started with the network connection between few computers at ARPANET project has reached to a level wherein almost all the computers and other communication devices of the world have joined together to form a huge global information network that makes available most of the information related to every possible heterogeneous domain. Not only the managing and indexing of this repository is a big concern but to provide a quick answer to the user‘s query is also of critical importance. Amazingly, rather miraculously, the task is being done quite efficiently by the current web search engines. This miracle has been possible due to a series of mathematical and technological innovations continuously being carried out in the area of search techniques. This paper takes an overview of search engine evolution from primitive to the present.

  14. REPTREE CLASSIFIER FOR IDENTIFYING LINK SPAM IN WEB SEARCH ENGINES

    Directory of Open Access Journals (Sweden)

    S.K. Jayanthi

    2013-01-01

    Full Text Available Search Engines are used for retrieving the information from the web. Most of the times, the importance is laid on top 10 results sometimes it may shrink as top 5, because of the time constraint and reliability on the search engines. Users believe that top 10 or 5 of total results are more relevant. Here comes the problem of spamdexing. It is a method to deceive the search result quality. Falsified metrics such as inserting enormous amount of keywords or links in website may take that website to the top 10 or 5 positions. This paper proposes a classifier based on the Reptree (Regression tree representative. As an initial step Link-based features such as neighbors, pagerank, truncated pagerank, trustrank and assortativity related attributes are inferred. Based on this features, tree is constructed. The tree uses the feature inference to differentiate spam sites from legitimate sites. WEBSPAM-UK-2007 dataset is taken as a base. It is preprocessed and converted into five datasets FEATA, FEATB, FEATC, FEATD and FEATE. Only link based features are taken for experiments. This paper focus on link spam alone. Finally a representative tree is created which will more precisely classify the web spam entries. Results are given. Regression tree classification seems to perform well as shown through experiments.

  15. Enhancing food engineering education with interactive web-based simulations

    Directory of Open Access Journals (Sweden)

    Alexandros Koulouris

    2015-04-01

    Full Text Available In the traditional deductive approach in teaching any engineering topic, teachers would first expose students to the derivation of the equations that govern the behavior of a physical system and then demonstrate the use of equations through a limited number of textbook examples. This methodology, however, is rarely adequate to unmask the cause-effect and quantitative relationships between the system variables that the equations embody. Web-based simulation, which is the integration of simulation and internet technologies, has the potential to enhance the learning experience by offering an interactive and easily accessible platform for quick and effortless experimentation with physical phenomena.This paper presents the design and development of a web-based platform for teaching basic food engineering phenomena to food technology students. The platform contains a variety of modules (“virtual experiments” covering the topics of mass and energy balances, fluid mechanics and heat transfer. In this paper, the design and development of three modules for mass balances and heat transfer is presented. Each webpage representing an educational module has the following features: visualization of the studied phenomenon through graphs, charts or videos, computation through a mathematical model and experimentation.  The student is allowed to edit key parameters of the phenomenon and observe the effect of these changes on the outputs. Experimentation can be done in a free or guided fashion with a set of prefabricated examples that students can run and self-test their knowledge by answering multiple-choice questions.

  16. WebVR——Web Virtual Reality Engine Based on P2P network

    Directory of Open Access Journals (Sweden)

    zhihan LV

    2011-07-01

    Full Text Available WebVR, a multi-user online virtual reality engine, is introduced. The main contributions are mapping the geographical space and virtual space to the P2P overlay network space, and dividing the three spaces by quad-tree method. The geocoding is identified with Hash value, which is used to index the user list, terrain data, and the model object data. Sharing of data through improved Kademlia network model is designed and implemented. In this model, XOR algorithm is used to calculate the distance of the virtual space. The model greatly improves the hit rate of 3D geographic data search under P2P overlay network. Some data preprocessing methods have been adopted to accelerate the data transfer. 3D Global data is used for testing the engine. The test result indicates that, without considering the client bandwidth limit, the more users, the faster loading.

  17. INTELLIGENT SEARCH ENGINE-BASED UNIVERSAL DESCRIPTION, DISCOVERY AND INTEGRATION FOR WEB SERVICE DISCOVERY

    Directory of Open Access Journals (Sweden)

    Tamilarasi Karuppiah

    2014-01-01

    Full Text Available Web Services standard has been broadly acknowledged by industries and academic researches along with the progress of web technology and e-business. Increasing number of web applications have been bundled as web services that can be published, positioned and invoked across the web. The importance of the issues regarding their publication and innovation attains a maximum as web services multiply and become more advanced and mutually dependent. With the intension of determining the web services through effiective manner with in the minimum time period in this study proposes an UDDI with intelligent serach engine. In order to publishing and discovering web services initially, the web services are published in the UDDI registry subsequently the published web services are indexed. To improve the efficiency of discovery of web services, the indexed web services are saved as index database. The search query is compared with the index database for discovering of web services and the discovered web services are given to the service customer. The way of accessing the web services is stored in a log file, which is then utilized to provide personalized web services to the user. The finding of web service is enhanced significantly by means of an efficient exploring capability provided by the proposed system and it is accomplished of providing the maximum appropriate web service. Universal Description, Discovery and Integration (UDDI.

  18. Situational Requirements Engineering for the Development of Content Management System-based Web Applications

    NARCIS (Netherlands)

    Souer, J.; van de Weerd, I.; Versendaal, J.M.; Brinkkemper, S.

    2005-01-01

    Web applications are evolving towards strong content-centered Web applications. The development processes and implementation of these applications are unlike the development and implementation of traditional information systems. In this paper we propose WebEngineering Method; a method for developing

  19. Adding to the Students' Toolbox: Using Directories, Search Engines, and the Hidden Web in Search Processes.

    Science.gov (United States)

    Mardis, Marcia A.

    2002-01-01

    Discussion of searching for information on the Web focuses on resources that are not always found by traditional Web searches. Describes sources on the hidden Web, including full-text databases, clearinghouses, digital libraries, and learning objects; explains how search engines operate; and suggests that traditional print sources are still…

  20. Situational Requirements Engineering for the Development of Content Management System-based Web Applications

    NARCIS (Netherlands)

    Souer, J.; van de Weerd, I.; Versendaal, J.M.; Brinkkemper, S.

    2005-01-01

    Web applications are evolving towards strong content-centered Web applications. The development processes and implementation of these applications are unlike the development and implementation of traditional information systems. In this paper we propose WebEngineering Method; a method for developing

  1. Semantic similarity measure in biomedical domain leverage web search engine.

    Science.gov (United States)

    Chen, Chi-Huang; Hsieh, Sheau-Ling; Weng, Yung-Ching; Chang, Wen-Yung; Lai, Feipei

    2010-01-01

    Semantic similarity measure plays an essential role in Information Retrieval and Natural Language Processing. In this paper we propose a page-count-based semantic similarity measure and apply it in biomedical domains. Previous researches in semantic web related applications have deployed various semantic similarity measures. Despite the usefulness of the measurements in those applications, measuring semantic similarity between two terms remains a challenge task. The proposed method exploits page counts returned by the Web Search Engine. We define various similarity scores for two given terms P and Q, using the page counts for querying P, Q and P AND Q. Moreover, we propose a novel approach to compute semantic similarity using lexico-syntactic patterns with page counts. These different similarity scores are integrated adapting support vector machines, to leverage the robustness of semantic similarity measures. Experimental results on two datasets achieve correlation coefficients of 0.798 on the dataset provided by A. Hliaoutakis, 0.705 on the dataset provide by T. Pedersen with physician scores and 0.496 on the dataset provided by T. Pedersen et al. with expert scores.

  2. DBLC_SPAMCLUST: SPAMDEXING DETECTION BY CLUSTERING CLIQUE-ATTACKS IN WEB SEARCH ENGIN

    OpenAIRE

    Dr.S.K.JAYANTHI,; Ms.S.Sasikala

    2011-01-01

    Search engines are playing a more and more important role in discovering information on the web now a day. Spam web pages, however, are employing various tricks to bamboozle search engines, therefore achieving undeserved ranks. In this paper an algorithm DBLCSPAMCLUST is proposed for spam detection based on content and link attributes details, which is an extension of DBSpamClust [1]. As showing through experiments such a method can filter out web spam effectively.

  3. DBLC_SPAMCLUST: SPAMDEXING DETECTION BY CLUSTERING CLIQUE-ATTACKS IN WEB SEARCH ENGIN

    Directory of Open Access Journals (Sweden)

    Dr.S.K.JAYANTHI,

    2011-06-01

    Full Text Available Search engines are playing a more and more important role in discovering information on the web now a day. Spam web pages, however, are employing various tricks to bamboozle search engines, therefore achieving undeserved ranks. In this paper an algorithm DBLCSPAMCLUST is proposed for spam detection based on content and link attributes details, which is an extension of DBSpamClust [1]. As showing through experiments such a method can filter out web spam effectively.

  4. White Hat Search Engine Optimization (SEO: Structured Web Data for Libraries

    Directory of Open Access Journals (Sweden)

    Dan Scott

    2015-06-01

    Full Text Available “White hat” search engine optimization refers to the practice of publishing web pages that are useful to humans, while enabling search engines and web applications to better understand the structure and content of your website. This article teaches you to add structured data to your website so that search engines can more easily connect patrons to your library locations, hours, and contact information. A web page for a branch of the Greater Sudbury Public Library retrieved in January 2015 is used as the basis for examples that progressively enhance the page with structured data. Finally, some of the advantages structured data enables beyond search engine optimization are explored

  5. A Taxonomic Search Engine: Federating taxonomic databases using web services

    Directory of Open Access Journals (Sweden)

    Page Roderic DM

    2005-03-01

    Full Text Available Abstract Background The taxonomic name of an organism is a key link between different databases that store information on that organism. However, in the absence of a single, comprehensive database of organism names, individual databases lack an easy means of checking the correctness of a name. Furthermore, the same organism may have more than one name, and the same name may apply to more than one organism. Results The Taxonomic Search Engine (TSE is a web application written in PHP that queries multiple taxonomic databases (ITIS, Index Fungorum, IPNI, NCBI, and uBIO and summarises the results in a consistent format. It supports "drill-down" queries to retrieve a specific record. The TSE can optionally suggest alternative spellings the user can try. It also acts as a Life Science Identifier (LSID authority for the source taxonomic databases, providing globally unique identifiers (and associated metadata for each name. Conclusion The Taxonomic Search Engine is available at http://darwin.zoology.gla.ac.uk/~rpage/portal/ and provides a simple demonstration of the potential of the federated approach to providing access to taxonomic names.

  6. Architecture of A Scalable Dynamic Parallel WebCrawler with High Speed Downloadable Capability for a Web Search Engine

    CERN Document Server

    Mukhopadhyay, Debajyoti; Ghosh, Soumya; Kar, Saheli; Kim, Young-Chon

    2011-01-01

    Today World Wide Web (WWW) has become a huge ocean of information and it is growing in size everyday. Downloading even a fraction of this mammoth data is like sailing through a huge ocean and it is a challenging task indeed. In order to download a large portion of data from WWW, it has become absolutely essential to make the crawling process parallel. In this paper we offer the architecture of a dynamic parallel Web crawler, christened as "WEB-SAILOR," which presents a scalable approach based on Client-Server model to speed up the download process on behalf of a Web Search Engine in a distributed Domain-set specific environment. WEB-SAILOR removes the possibility of overlapping of downloaded documents by multiple crawlers without even incurring the cost of communication overhead among several parallel "client" crawling processes.

  7. A study of medical and health queries to web search engines.

    Science.gov (United States)

    Spink, Amanda; Yang, Yin; Jansen, Jim; Nykanen, Pirrko; Lorence, Daniel P; Ozmutlu, Seda; Ozmutlu, H Cenk

    2004-03-01

    This paper reports findings from an analysis of medical or health queries to different web search engines. We report results: (i). comparing samples of 10000 web queries taken randomly from 1.2 million query logs from the AlltheWeb.com and Excite.com commercial web search engines in 2001 for medical or health queries, (ii). comparing the 2001 findings from Excite and AlltheWeb.com users with results from a previous analysis of medical and health related queries from the Excite Web search engine for 1997 and 1999, and (iii). medical or health advice-seeking queries beginning with the word 'should'. Findings suggest: (i). a small percentage of web queries are medical or health related, (ii). the top five categories of medical or health queries were: general health, weight issues, reproductive health and puberty, pregnancy/obstetrics, and human relationships, and (iii). over time, the medical and health queries may have declined as a proportion of all web queries, as the use of specialized medical/health websites and e-commerce-related queries has increased. Findings provide insights into medical and health-related web querying and suggests some implications for the use of the general web search engines when seeking medical/health information.

  8. Comparison of Physics Frameworks for WebGL-Based Game Engine

    Directory of Open Access Journals (Sweden)

    Yogya Resa

    2014-03-01

    Full Text Available Recently, a new technology called WebGL shows a lot of potentials for developing games. However since this technology is still new, there are still many potentials in the game development area that are not explored yet. This paper tries to uncover the potential of integrating physics frameworks with WebGL technology in a game engine for developing 2D or 3D games. Specifically we integrated three open source physics frameworks: Bullet, Cannon, and JigLib into a WebGL-based game engine. Using experiment, we assessed these frameworks in terms of their correctness or accuracy, performance, completeness and compatibility. The results show that it is possible to integrate open source physics frameworks into a WebGLbased game engine, and Bullet is the best physics framework to be integrated into the WebGL-based game engine.

  9. Overview of the Web Search Engine%Web搜索引擎综述

    Institute of Scientific and Technical Information of China (English)

    张卫丰; 徐宝文; 周晓宇; 许蕾; 李东

    2001-01-01

    With the explosive increase of the network information,people can find information more and more difficultly. The occurrence of the Web search engine overcomes this problem in some degree. This paper tells about the history of the search engine ,the current state of the search engine. Some guidelines about the search engine are analysed and the related checking methods are also given. In this basis, we introduce the trend of the search engine.

  10. A reverse engineering approach for automatic annotation of Web pages

    NARCIS (Netherlands)

    R. de Virgilio (Roberto); F. Frasincar (Flavius); W. Hop (Wim); S. Lachner (Stephan)

    2013-01-01

    textabstractThe Semantic Web is gaining increasing interest to fulfill the need of sharing, retrieving, and reusing information. Since Web pages are designed to be read by people, not machines, searching and reusing information on the Web is a difficult task without human participation. To this aim

  11. Soil food web changes during spontaneous succession at post mining sites: a possible ecosystem engineering effect on food web organization?

    Science.gov (United States)

    Frouz, Jan; Thébault, Elisa; Pižl, Václav; Adl, Sina; Cajthaml, Tomáš; Baldrián, Petr; Háněl, Ladislav; Starý, Josef; Tajovský, Karel; Materna, Jan; Nováková, Alena; de Ruiter, Peter C

    2013-01-01

    Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi) and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1- 41-year-old post-mining sites near the town of Sokolov (Czech Republic). These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel.

  12. The Effectiveness of Web Search Engines to Index New Sites from Different Countries

    Science.gov (United States)

    Pirkola, Ari

    2009-01-01

    Introduction: Investigates how effectively Web search engines index new sites from different countries. The primary interest is whether new sites are indexed equally or whether search engines are biased towards certain countries. If major search engines show biased coverage it can be considered a significant economic and political problem because…

  13. Review of Metadata Elements within the Web Pages Resulting from Searching in General Search Engines

    Directory of Open Access Journals (Sweden)

    Sima Shafi’ie Alavijeh

    2009-12-01

    Full Text Available The present investigation was aimed to study the scope of presence of Dublin Core metadata elements and HTML meta tags in web pages. Ninety web pages were chosen by searching general search engines (Google, Yahoo and MSN. The scope of metadata elements (Dublin Core and HTML Meta tags present in these pages as well as existence of a significant correlation between presence of meta elements and type of search engines were investigated. Findings indicated very low presence of both Dublin Core metadata elements and HTML meta tags in the pages retrieved which in turn illustrates the very low usage of meta data elements in web pages. Furthermore, findings indicated that there are no significant correlation between the type of search engine used and presence of metadata elements. From the standpoint of including metadata in retrieval of web sources, search engines do not significantly differ from one another.

  14. Dispersion engineered cob-web photonic crystal fibers for efficient supercontinuum generation

    OpenAIRE

    Sørensen, Niels Thorkild; Nikolov, N. I.; Bang, Ole; Bjarklev, Anders Overgaard; Hougaard, Kristian G.; Hansen, Kim Per

    2004-01-01

    Highly nonlinear cob-web photonic crystal fibers are engineered to have dispersion profiles for efficient direct degenerate four-wave mixing and optimized supercontinuum generation with low-power picosecond pulses. This process is robust to fiber irregularities.

  15. Can Interactive Web-Based CAD Tools Improve the Learning of Engineering Drawing? A Case Study

    Science.gov (United States)

    Pando Cerra, Pablo; Suárez González, Jesús M.; Busto Parra, Bernardo; Rodríguez Ortiz, Diana; Álvarez Peñín, Pedro I.

    2014-01-01

    Many current Web-based learning environments facilitate the theoretical teaching of a subject but this may not be sufficient for those disciplines that require a significant use of graphic mechanisms to resolve problems. This research study looks at the use of an environment that can help students learn engineering drawing with Web-based CAD…

  16. Uncovering the Hidden Web, Part I: Finding What the Search Engines Don't. ERIC Digest.

    Science.gov (United States)

    Mardis, Marcia

    Currently, the World Wide Web contains an estimated 7.4 million sites (OCLC, 2001). Yet even the most experienced searcher, using the most robust search engines, can access only about 16% of these pages (Dahn, 2001). The other 84% of the publicly available information on the Web is referred to as the "hidden,""invisible," or…

  17. Curating the Web: Building a Google Custom Search Engine for the Arts

    Science.gov (United States)

    Hennesy, Cody; Bowman, John

    2008-01-01

    Google's first foray onto the web made search simple and results relevant. With its Co-op platform, Google has taken another step toward dramatically increasing the relevancy of search results, further adapting the World Wide Web to local needs. Google Custom Search Engine, a tool on the Co-op platform, puts one in control of his or her own search…

  18. Can Interactive Web-Based CAD Tools Improve the Learning of Engineering Drawing? A Case Study

    Science.gov (United States)

    Pando Cerra, Pablo; Suárez González, Jesús M.; Busto Parra, Bernardo; Rodríguez Ortiz, Diana; Álvarez Peñín, Pedro I.

    2014-01-01

    Many current Web-based learning environments facilitate the theoretical teaching of a subject but this may not be sufficient for those disciplines that require a significant use of graphic mechanisms to resolve problems. This research study looks at the use of an environment that can help students learn engineering drawing with Web-based CAD…

  19. Optimal Threshold Control by the Robots of Web Search Engines with Obsolescence of Documents

    CERN Document Server

    Avrachenkov, Konstantin; Klimenok, Valentina; Nain, Philippe; Semenova, Olga; 10.1016/j.comnet.2011.01.013

    2012-01-01

    A typical web search engine consists of three principal parts: crawling engine, indexing engine, and searching engine. The present work aims to optimize the performance of the crawling engine. The crawling engine finds new web pages and updates web pages existing in the database of the web search engine. The crawling engine has several robots collecting information from the Internet. We first calculate various performance measures of the system (e.g., probability of arbitrary page loss due to the buffer overflow, probability of starvation of the system, the average time waiting in the buffer). Intuitively, we would like to avoid system starvation and at the same time to minimize the information loss. We formulate the problem as a multi-criteria optimization problem and attributing a weight to each criterion. We solve it in the class of threshold policies. We consider a very general web page arrival process modeled by Batch Marked Markov Arrival Process and a very general service time modeled by Phase-type dis...

  20. Dispersion engineered cob-web photonic crystal fibers for efficient supercontinuum generation

    DEFF Research Database (Denmark)

    Sørensen, Niels Thorkild; Nikolov, N.I.; Bang, Ole;

    2004-01-01

    Highly nonlinear cob-web photonic crystal fibers are engineered to have dispersion profiles for efficient direct degenerate four-wave mixing and optimized supercontinuum generation with low-power picosecond pulses. This process is robust to fiber irregularities.......Highly nonlinear cob-web photonic crystal fibers are engineered to have dispersion profiles for efficient direct degenerate four-wave mixing and optimized supercontinuum generation with low-power picosecond pulses. This process is robust to fiber irregularities....

  1. Web Image Retrieval Search Engine based on Semantically Shared Annotation

    Directory of Open Access Journals (Sweden)

    Alaa Riad

    2012-03-01

    Full Text Available This paper presents a new majority voting technique that combines the two basic modalities of Web images textual and visual features of image in a re-annotation and search based framework. The proposed framework considers each web page as a voter to vote the relatedness of keyword to the web image, the proposed approach is not only pure combination between image low level feature and textual feature but it take into consideration the semantic meaning of each keyword that expected to enhance the retrieval accuracy. The proposed approach is not used only to enhance the retrieval accuracy of web images; but also able to annotated the unlabeled images.

  2. Philosophical engineering toward a philosophy of the web

    CERN Document Server

    Halpin, Harry

    2013-01-01

    This is the first interdisciplinary exploration of the philosophical foundations of the Web, a new area of inquiry that has important implications across a range of domains. Contains twelve essays that bridge the fields of philosophy, cognitive science, and phenomenologyTackles questions such as the impact of Google on intelligence and epistemology, the philosophical status of digital objects, ethics on the Web, semantic and ontological changes caused by the Web, and the potential of the Web to serve as a genuine cognitive extensionBrings together insightful new scholarship from well-known an

  3. Feature-based engineering of compensations in web service environment

    DEFF Research Database (Denmark)

    Schaefer, Michael; Dolog, Peter

    2009-01-01

    In this paper, we introduce a product line approach for developing Web services with extended compensation capabilities. We adopt a feature modelling approach in order to describe variable and common compensation properties of Web service variants, as well as service consumer application...

  4. Case and Relation (CARE based Page Rank Algorithm for Semantic Web Search Engines

    Directory of Open Access Journals (Sweden)

    N. Preethi

    2012-05-01

    Full Text Available Web information retrieval deals with a technique of finding relevant web pages for any given query from a collection of documents. Search engines have become the most helpful tool for obtaining useful information from the Internet. The next-generation Web architecture, represented by the Semantic Web, provides the layered architecture possibly allowing data to be reused across application. The proposed architecture use a hybrid methodology named Case and Relation (CARE based Page Rank algorithm which uses past problem solving experience maintained in the case base to form a best matching relations and then use them for generating graphs and spanning forests to assign a relevant score to the pages.

  5. Soil food web changes during spontaneous succession at post mining sites: a possible ecosystem engineering effect on food web organization?

    Directory of Open Access Journals (Sweden)

    Jan Frouz

    Full Text Available Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1- 41-year-old post-mining sites near the town of Sokolov (Czech Republic. These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel.

  6. The Indexed Web: Engineering Tools for Cataloging, Storing, and Delivering Web-Based Documents.

    Science.gov (United States)

    Nanfito, Michael

    1999-01-01

    Describes the development of Java-based software to use controlled vocabularies and subject headings in the cataloging of Web-based HTML articles on a subject-specific legal resources Web site. Topics include requirements for cataloging and for searching; data structure; software required; and metadata with Dublin Core requirements. (LRW)

  7. Engineers and the Web: an analysis of real life gaps in information usage

    NARCIS (Netherlands)

    Kraaijenbrink, Jeroen

    2007-01-01

    Engineers face a wide range of gaps when trying to identify, acquire, and utilize information from the Web. To be able to avoid creating such gaps, it is essential to understand them in detail. This paper reports the results of a study of the real life gaps in information usage processes of 17 engin

  8. Index Compression and Efficient Query Processing in Large Web Search Engines

    Science.gov (United States)

    Ding, Shuai

    2013-01-01

    The inverted index is the main data structure used by all the major search engines. Search engines build an inverted index on their collection to speed up query processing. As the size of the web grows, the length of the inverted list structures, which can easily grow to hundreds of MBs or even GBs for common terms (roughly linear in the size of…

  9. Effects of Web-Based Interactive Modules on Engineering Students' Learning Motivations

    Science.gov (United States)

    Bai, Haiyan; Aman, Amjad; Xu, Yunjun; Orlovskaya, Nina; Zhou, Mingming

    2016-01-01

    The purpose of this study is to assess the impact of a newly developed modules, Interactive Web-Based Visualization Tools for Gluing Undergraduate Fuel Cell Systems Courses system (IGLU), on learning motivations of engineering students using two samples (n[subscript 1] = 144 and n[subscript 2] = 135) from senior engineering classes. The…

  10. Social Dimension of Web 2.0 in Engineering Education: Students' View

    Science.gov (United States)

    Zascerinska, Jelena; Bassus, Olaf; Ahrens, Andreas

    2010-01-01

    Contemporary engineers need to become more cognizant and more responsive to the emerging needs of the market for engineering and technology services. Social dimension of Web 2.0 which penetrates our society more thoroughly with the availability of broadband services has the potential to contribute decisively to the sustainable development of…

  11. Search Engines and Resource Discovery on the Web: Is Dublin Core an Impact Factor?

    Directory of Open Access Journals (Sweden)

    Mehdi Safari

    2005-08-01

    Full Text Available This study evaluates the effectiveness of the Dublin Core metadata elements on the retrieval of web pages in a suite of six search engines, AlltheWeb, AltaVista, Google, Excite, Lycos, and WebCrawler. The effectiveness of four elements, including title, creator, subject and contributor, that concentrate on resource discovery was experimentally evaluated. Searches were made of the keywords extracted from web pages of the Iranian International Journal of Science, before and after metadata implementation. In each search, the ranking of the first specific reference to the exact web page was recorded. The comparison of results and statistical analysis did not reveal a significant difference between control and experimental groups in the retrieval ranks of the web pages.

  12. Cost estimation in software engineering projects with web components development

    Directory of Open Access Journals (Sweden)

    Javier de Andrés

    2015-01-01

    Full Text Available Existen multitud de modelos propuestos para la predicción de co stes en proyectos de software, al gunos orientados específicamen te para proyectos Web. Este trabajo analiza si los modelos específicos para proyectos Web están justifi cados, examinando el comportami ento diferencial de los costes entre proyectos de desarrollo softwar e Web y no Web. Se analizan dos aspectos del cálculo de costes: las deseconomías de escala, y el im pacto de algunas características de estos proyectos que son utilizadas como cost drivers. Se en uncian dos hipótesis: (a en estos proyect os las deseconomías de escala so n mayores y (b el incremento de coste que provocan los cost dr ivers es menor para los proyectos Web. Se contrastaron estas hipótesis a nalizando un conjunto de proyectos reales. Los resultados sugie ren que ambas hipótesis se cumplen. Por lo tanto, la principal contribu ción a la literatura de esta inv estigación es que el desarrollo de modelos específicos para los proyectos Web está justificado.

  13. A UDDI Search Engine for SVG Federated Medical Imaging Web Services

    Directory of Open Access Journals (Sweden)

    Sabah Mohammed

    2006-01-01

    Full Text Available With more and more medical web services appearing on the web, web service’s discovery mechanism becomes essential. UDDI is an online registry standard to facilitate the discovery of business partners and services. However, most medical imaging applications exist within their own protected domain and were never designed to participate and operate with other applications across the web. However, private UDDI registries in federated organizations should be able to share the service descriptions as well as to access them if they are authorized. The new initiatives on Federated Web Services Identity Management can resolve a range of both technical and political barriers to enable wide-scale participation and interoperation of separate domains into a singular, robust user experience. However, there is no widely acceptable standard for federated web services and most of the available venders frameworks concentrate only on the security issue of the federation leaving the issue of searching and discovering web services largely primitive. Federated web services security and web services searching are uniquely intertwined, mutually reliant on each other and are poised to finally solve a long-running problem in both IT and systems security. Traditional keyword search is insufficient for web services search as the very small text fragments in web services are unsuitable for keyword search and the underlying structure and semantics of the web service are not exploited. Engineering solutions that address the security and accessibility concerns of web services, however, is a challenging task. This article introduces an extension to the traditional UDDI that enables sophisticated types of searching based on a lightweight web services federated security infrastructure.

  14. Key word placing in Web page body text to increase visibility to search engines

    Directory of Open Access Journals (Sweden)

    W. T. Kritzinger

    2007-11-01

    Full Text Available The growth of the World Wide Web has spawned a wide variety of new information sources, which has also left users with the daunting task of determining which sources are valid. Many users rely on the Web as an information source because of the low cost of information retrieval. It is also claimed that the Web has evolved into a powerful business tool. Examples include highly popular business services such as Amazon.com and Kalahari.net. It is estimated that around 80% of users utilize search engines to locate information on the Internet. This, by implication, places emphasis on the underlying importance of Web pages being listed on search engines indices. Empirical evidence that the placement of key words in certain areas of the body text will have an influence on the Web sites' visibility to search engines could not be found in the literature. The result of two experiments indicated that key words should be concentrated towards the top, and diluted towards the bottom of a Web page to increase visibility. However, care should be taken in terms of key word density, to prevent search engine algorithms from raising the spam alarm.

  15. 25 Years of Model-Driven Web Engineering: What we achieved, What is missing

    Directory of Open Access Journals (Sweden)

    Gustavo Rossi

    2016-12-01

    Full Text Available Model-Driven Web Engineering (MDWE approaches aim to improve the Web applications development process by focusing on modeling instead of coding, and deriving the running application by transformations from conceptual models to code. The emergence of the Interaction Flow Modeling Language (IFML has been an important milestone in the evolution of Web modeling languages, indicating not only the maturity of the field but also a final convergence of languages. In this paper we explain the evolution of modeling and design approaches since the early years (in the 90’s detailing the forces which drove that evolution and discussing the strengths and weaknesses of some of those approaches. A brief presentation of the IFML is accompanied with a thorough analysis of the most important achievements of the MDWE community as well as the problems and obstacles that hinder the dissemination of model-driven techniques in the Web engineering field.

  16. An assessment of the visibility of MeSH-indexed medical web catalogs through search engines.

    Science.gov (United States)

    Zweigenbaum, P; Darmoni, S J; Grabar, N; Douyère, M; Benichou, J

    2002-01-01

    Manually indexed Internet health catalogs such as CliniWeb or CISMeF provide resources for retrieving high-quality health information. Users of these quality-controlled subject gateways are most often referred to them by general search engines such as Google, AltaVista, etc. This raises several questions, among which the following: what is the relative visibility of medical Internet catalogs through search engines? This study addresses this issue by measuring and comparing the visibility of six major, MeSH-indexed health catalogs through four different search engines (AltaVista, Google, Lycos, Northern Light) in two languages (English and French). Over half a million queries were sent to the search engines; for most of these search engines, according to our measures at the time the queries were sent, the most visible catalog for English MeSH terms was CliniWeb and the most visible one for French MeSH terms was CISMeF.

  17. A World Wide Web Region-Based Image Search Engine

    DEFF Research Database (Denmark)

    Kompatsiaris, Ioannis; Triantafyllou, Evangelia; Strintzis, Michael G.

    2001-01-01

    information. These features along with additional information such as the URL location and the date of index procedure are stored in a database. The user can access and search this indexed content through the Web with an advanced and user friendly interface. The output of the system is a set of links...

  18. World Wide Web Search Engines: AltaVista and Yahoo.

    Science.gov (United States)

    Machovec, George S., Ed.

    1996-01-01

    Examines the history, structure, and search capabilities of Internet search tools AltaVista and Yahoo. AltaVista provides relevance-ranked feedback on full-text searches. Yahoo indexes Web "citations" only but does organize information hierarchically into predefined categories. Yahoo has recently become a publicly held company and…

  19. SOLE: Applying Semantics and Social Web to Support Technology Enhanced Learning in Software Engineering

    Science.gov (United States)

    Colomo-Palacios, Ricardo; Jiménez-López, Diego; García-Crespo, Ángel; Blanco-Iglesias, Borja

    eLearning educative processes are a challenge for educative institutions and education professionals. In an environment in which learning resources are being produced, catalogued and stored using innovative ways, SOLE provides a platform in which exam questions can be produced supported by Web 2.0 tools, catalogued and labeled via semantic web and stored and distributed using eLearning standards. This paper presents, SOLE, a social network of exam questions sharing particularized for Software Engineering domain, based on semantics and built using semantic web and eLearning standards, such as IMS Question and Test Interoperability specification 2.1.

  20. Using the open Web as an information resource and scholarly Web search engines as retrieval tools for academic and research purposes

    Directory of Open Access Journals (Sweden)

    Filistea Naude

    2010-08-01

    Full Text Available This study provided insight into the significance of the open Web as an information resource and Web search engines as research tools amongst academics. The academic staff establishment of the University of South Africa (Unisa was invited to participate in a questionnaire survey and included 1188 staff members from five colleges. This study culminated in a PhD dissertation in 2008. One hundred and eighty seven respondents participated in the survey which gave a response rate of 15.7%. The results of this study show that academics have indeed accepted the open Web as a useful information resource and Web search engines as retrieval tools when seeking information for academic and research work. The majority of respondents used the open Web and Web search engines on a daily or weekly basis to source academic and research information. The main obstacles presented by using the open Web and Web search engines included lack of time to search and browse the Web, information overload, poor network speed and the slow downloading speed of webpages.

  1. Using social annotation and web log to enhance search engine

    CERN Document Server

    Nguyen, Vu Thanh

    2009-01-01

    Search services have been developed rapidly in social Internet. It can help web users easily to find their documents. So that, finding a best method search is always an imagine. This paper would like introduce hybrid method of LPageRank algorithm and Social Sim Rank algorithm. LPageRank is the method using link structure to rank priority of page. It doesn't care content of page and content of query. Therefore, we want to use benefit of social annotations to create the latent semantic association between queries and annotations. This model, we use algorithm SocialPageRank and LPageRank to enhance accuracy of search system. To experiment and evaluate the proposed of the new model, we have used this model for Music Machine Website with their web logs.

  2. Exploiting Semantic Web Technologies to Develop OWL-Based Clinical Practice Guideline Execution Engines.

    Science.gov (United States)

    Jafarpour, Borna; Abidi, Samina Raza; Abidi, Syed Sibte Raza

    2016-01-01

    Computerizing paper-based CPG and then executing them can provide evidence-informed decision support to physicians at the point of care. Semantic web technologies especially web ontology language (OWL) ontologies have been profusely used to represent computerized CPG. Using semantic web reasoning capabilities to execute OWL-based computerized CPG unties them from a specific custom-built CPG execution engine and increases their shareability as any OWL reasoner and triple store can be utilized for CPG execution. However, existing semantic web reasoning-based CPG execution engines suffer from lack of ability to execute CPG with high levels of expressivity, high cognitive load of computerization of paper-based CPG and updating their computerized versions. In order to address these limitations, we have developed three CPG execution engines based on OWL 1 DL, OWL 2 DL and OWL 2 DL + semantic web rule language (SWRL). OWL 1 DL serves as the base execution engine capable of executing a wide range of CPG constructs, however for executing highly complex CPG the OWL 2 DL and OWL 2 DL + SWRL offer additional executional capabilities. We evaluated the technical performance and medical correctness of our execution engines using a range of CPG. Technical evaluations show the efficiency of our CPG execution engines in terms of CPU time and validity of the generated recommendation in comparison to existing CPG execution engines. Medical evaluations by domain experts show the validity of the CPG-mediated therapy plans in terms of relevance, safety, and ordering for a wide range of patient scenarios.

  3. Comparing the Scale of Web Subject Directories Precision in Technical-Engineering Information Retrieval

    Directory of Open Access Journals (Sweden)

    Mehrdokht Wazirpour Keshmiri

    2012-07-01

    Full Text Available The main purpose of this research was to compare the scale of web subject directories precision in information retrieval of technical-engineering science. Information gathering was documentary and webometric. Keywords of technical-engineering science were chosen at twenty different subjects from IEEE (Institute of Electrical and Electronics Engineers and engineering magazines that situated in sciencedirect site. These keywords are used at five subject directories Yahoo, Google, Infomine, Intute, Dmoz, that were web directories high-utilization. Usually first results in searching tools are connected to searching keywords. Because, first ten results was evaluated in every search. These assessments to consist of scale of precision, scale of error, scale retrieval items in technical-engineering categories to retrieval items entirely. The used criteria for determining the scale of precision that was according to high-utilization standards in different documents, to consist of presence of the keywords in title, appearance of keywords at the part of web retrieved pages, keywords adjacency, URL of page, page description and subject categories. Information analysis was according to Kruskal-Wallis Test and L.S.D fisher. Results revealed that there was meaningful difference about precision of web subject directories in information retrieval of technical-engineering science, Therefore this theory was confirmed.web subject directories ranked from point of precision as follows. Google, Yahoo, Intute, Dmoz, and Infomine. The scale of observed error at the first results was another criterion that was used for comparing web subject directories. In this research, Yahoo had minimum scale of error and Infomine had most of error. This research also compared the scale of retrieval items in all of categories web subject directories entirely to retrieval items in technical-engineering categories, results revealed that there was meaningful difference between them. And

  4. Adding a Visualization Feature to Web Search Engines: It’s Time

    Energy Technology Data Exchange (ETDEWEB)

    Wong, Pak C.

    2008-11-11

    Since the first world wide web (WWW) search engine quietly entered our lives in 1994, the “information need” behind web searching has rapidly grown into a multi-billion dollar business that dominates the internet landscape, drives e-commerce traffic, propels global economy, and affects the lives of the whole human race. Today’s search engines are faster, smarter, and more powerful than those released just a few years ago. With the vast investment pouring into research and development by leading web technology providers and the intense emotion behind corporate slogans such as “win the web” or “take back the web,” I can’t help but ask why are we still using the very same “text-only” interface that was used 13 years ago to browse our search engine results pages (SERPs)? Why has the SERP interface technology lagged so far behind in the web evolution when the corresponding search technology has advanced so rapidly? In this article I explore some current SERP interface issues, suggest a simple but practical visual-based interface design approach, and argue why a visual approach can be a strong candidate for tomorrow’s SERP interface.

  5. Categorization of web pages - Performance enhancement to search engine

    Digital Repository Service at National Institute of Oceanography (India)

    Lakshminarayana, S.

    are the major areas of research in IR and strive to improve the effectiveness of interactive IR and can be used as performance evaluation tool. The classification studies at early stages were with strong human interaction than machine learning. The term... and the location of the link. In the absence such works, the spider/worm either moves to the next page available at the least time or by network selection. This classification serves in judgment of traversal of web spider/worm and minimization. Such processes...

  6. Using Exclusive Web Crawlers to Store Better Results in Search Engines' Database

    Directory of Open Access Journals (Sweden)

    Ali Tourani

    2013-05-01

    Full Text Available Crawler-based search engines are the mostly used search engines among web and Internet users , involveweb crawling, storing in database, ranking, indexing and displaying to the user. But it is noteworthy thatbecause of increasing changes in web sites search engines suffer high time and transfers costs which areconsumed to investigate the existence of each page in database while crawling, updating database andeven investigating its existence in any crawling operations."Exclusive Web Crawler" proposes guidelines for crawling features, links, media and other elements and tostore crawling results in a certain table in its database on the web. With doing this, search engines storeeach site's tables in their databases and implement their ranking results on them. Thus, accuracy of data inevery table (and its being up-to-date is ensured and no 404 result is shown in search results since, in fact,this data crawler crawls data entered by webmaster and the database stores whatever he wants to display.

  7. Introduction to Chemical Engineering Reactor Analysis: A Web-Based Reactor Design Game

    Science.gov (United States)

    Orbey, Nese; Clay, Molly; Russell, T.W. Fraser

    2014-01-01

    An approach to explain chemical engineering through a Web-based interactive game design was developed and used with college freshman and junior/senior high school students. The goal of this approach was to demonstrate how to model a lab-scale experiment, and use the results to design and operate a chemical reactor. The game incorporates both…

  8. First 20 Precision among World Wide Web Search Services (Search Engines).

    Science.gov (United States)

    Leighton, H. Vernon; Srivastava, Jaideep

    1999-01-01

    Compares five World Wide Web search engines for precision on the first 20 results returned for 15 queries, adding weight for ranking effectiveness. Discusses methods to lessen evaluator bias, evaluation criteria, definition of relevance, experimental design, the structure of queries, and future work. (Author/LRW)

  9. Web-Based Simulation Games for the Integration of Engineering and Business Fundamentals

    Science.gov (United States)

    Calfa, Bruno; Banholzer, William; Alger, Monty; Doherty, Michael

    2017-01-01

    This paper describes a web-based suite of simulation games that have the purpose to enhance the chemical engineering curriculum with business-oriented decisions. Two simulation cases are discussed whose teaching topics include closing material and energy balances, importance of recycle streams, price-volume relationship in a dynamic market, impact…

  10. Introduction to Chemical Engineering Reactor Analysis: A Web-Based Reactor Design Game

    Science.gov (United States)

    Orbey, Nese; Clay, Molly; Russell, T.W. Fraser

    2014-01-01

    An approach to explain chemical engineering through a Web-based interactive game design was developed and used with college freshman and junior/senior high school students. The goal of this approach was to demonstrate how to model a lab-scale experiment, and use the results to design and operate a chemical reactor. The game incorporates both…

  11. WEB SEARCH ENGINE BASED SEMANTIC SIMILARITY MEASURE BETWEEN WORDS USING PATTERN RETRIEVAL ALGORITHM

    Directory of Open Access Journals (Sweden)

    Pushpa C N

    2013-02-01

    Full Text Available Semantic Similarity measures plays an important role in information retrieval, natural language processing and various tasks on web such as relation extraction, community mining, document clustering, and automatic meta-data extraction. In this paper, we have proposed a Pattern Retrieval Algorithm [PRA] to compute the semantic similarity measure between the words by combining both page count method and web snippets method. Four association measures are used to find semantic similarity between words in page count method using web search engines. We use a Sequential Minimal Optimization (SMO support vector machines (SVM to find the optimal combination of page counts-based similarity scores and top-ranking patterns from the web snippets method. The SVM is trained to classify synonymous word-pairs and nonsynonymous word-pairs. The proposed approach aims to improve the Correlation values, Precision, Recall, and F-measures, compared to the existing methods. The proposed algorithm outperforms by 89.8 % of correlation value.

  12. Applying Web-Based Tools for Research, Engineering, and Operations

    Science.gov (United States)

    Ivancic, William D.

    2011-01-01

    Personnel in the NASA Glenn Research Center Network and Architectures branch have performed a variety of research related to space-based sensor webs, network centric operations, security and delay tolerant networking (DTN). Quality documentation and communications, real-time monitoring and information dissemination are critical in order to perform quality research while maintaining low cost and utilizing multiple remote systems. This has been accomplished using a variety of Internet technologies often operating simultaneously. This paper describes important features of various technologies and provides a number of real-world examples of how combining Internet technologies can enable a virtual team to act efficiently as one unit to perform advanced research in operational systems. Finally, real and potential abuses of power and manipulation of information and information access is addressed.

  13. Reverse Engineering and Software Products Reuse to Teach Collaborative Web Portals: A Case Study with Final-Year Computer Science Students

    Science.gov (United States)

    Medina-Dominguez, Fuensanta; Sanchez-Segura, Maria-Isabel; Mora-Soto, Arturo; Amescua, Antonio

    2010-01-01

    The development of collaborative Web applications does not follow a software engineering methodology. This is because when university students study Web applications in general, and collaborative Web portals in particular, they are not being trained in the use of software engineering techniques to develop collaborative Web portals. This paper…

  14. Reverse Engineering and Software Products Reuse to Teach Collaborative Web Portals: A Case Study with Final-Year Computer Science Students

    Science.gov (United States)

    Medina-Dominguez, Fuensanta; Sanchez-Segura, Maria-Isabel; Mora-Soto, Arturo; Amescua, Antonio

    2010-01-01

    The development of collaborative Web applications does not follow a software engineering methodology. This is because when university students study Web applications in general, and collaborative Web portals in particular, they are not being trained in the use of software engineering techniques to develop collaborative Web portals. This paper…

  15. Delivering Collaborative Web Labs as a Service for Engineering Education

    OpenAIRE

    Bochicchio, Mario A.; Antonella Longo

    2012-01-01

    As Internet speed grows up and academic networks reach more users, engineering schools take interest in online laboratories as a mean to increase the spectrum of offered services and to reduce costs by sharing expensive lab equipments. In this perspective, online labs must comply both with the scientific and pedagogic requirements coming from the lab users (students, researchers, …) and with the requirements coming from the administrative and technical staff in charge to manage and deliver th...

  16. Googling social interactions: web search engine based social network construction.

    Science.gov (United States)

    Lee, Sang Hoon; Kim, Pan-Jun; Ahn, Yong-Yeol; Jeong, Hawoong

    2010-07-21

    Social network analysis has long been an untiring topic of sociology. However, until the era of information technology, the availability of data, mainly collected by the traditional method of personal survey, was highly limited and prevented large-scale analysis. Recently, the exploding amount of automatically generated data has completely changed the pattern of research. For instance, the enormous amount of data from so-called high-throughput biological experiments has introduced a systematic or network viewpoint to traditional biology. Then, is "high-throughput" sociological data generation possible? Google, which has become one of the most influential symbols of the new Internet paradigm within the last ten years, might provide torrents of data sources for such study in this (now and forthcoming) digital era. We investigate social networks between people by extracting information on the Web and introduce new tools of analysis of such networks in the context of statistical physics of complex systems or socio-physics. As a concrete and illustrative example, the members of the 109th United States Senate are analyzed and it is demonstrated that the methods of construction and analysis are applicable to various other weighted networks.

  17. Using Web 2.0 Techniques in NASA's Ares Engineering Operations Network (AEON) Environment - First Impressions

    Science.gov (United States)

    Scott, David W.

    2010-01-01

    The Mission Operations Laboratory (MOL) at Marshall Space Flight Center (MSFC) is responsible for Engineering Support capability for NASA s Ares rocket development and operations. In pursuit of this, MOL is building the Ares Engineering and Operations Network (AEON), a web-based portal to support and simplify two critical activities: Access and analyze Ares manufacturing, test, and flight performance data, with access to Shuttle data for comparison Establish and maintain collaborative communities within the Ares teams/subteams and with other projects, e.g., Space Shuttle, International Space Station (ISS). AEON seeks to provide a seamless interface to a) locally developed engineering applications and b) a Commercial-Off-The-Shelf (COTS) collaborative environment that includes Web 2.0 capabilities, e.g., blogging, wikis, and social networking. This paper discusses how Web 2.0 might be applied to the typically conservative engineering support arena, based on feedback from Integration, Verification, and Validation (IV&V) testing and on searching for their use in similar environments.

  18. Web-based Image Search Engines%因特网上的图像搜索引擎

    Institute of Scientific and Technical Information of China (English)

    陈立娜

    2001-01-01

    The operating principle of Web-based image search engines is briefly described. A detailed evaluation of some of image search engines is made. Finally, the paper points out the deficiencies of the present image search engines and their development trend.

  19. Web search engine:characteristics of user behaviors and their implication

    Institute of Scientific and Technical Information of China (English)

    王建勇; 单松巍; 雷鸣; 谢正茂; 李晓明

    2001-01-01

    In this paper, first studied are the distribution characteristics of user behaviors based on log data from a massive web search engine. Analysis shows that stochastic distribution of user queries accords with the characteristics of power-law function and exhibits strong similarity, and the user's queries and clicked URLs present dramatic locality, which implies that query cache and ‘hot click'cache can be employed to improve system performance. Then three typical cache replacement policies are compared, including LRU, FIFO, and LFU with attenuation. In addition, the distribution characteristics of web information are also analyzed, which demonstrates that the link popularity and replica popularity of a URL have positive influence on its importance. Finally, variance between the link popularity and user popularity, and variance between replica popularity and user popularity are analyzed, which give us some important insight that helps us improve the ranking algorithms in a search engine.

  20. Can Interactive Web-based CAD Tools Improve the Learning of Engineering Drawing? A Case Study

    Science.gov (United States)

    Pando Cerra, Pablo; Suárez González, Jesús M.; Busto Parra, Bernardo; Rodríguez Ortiz, Diana; Álvarez Peñín, Pedro I.

    2014-06-01

    Many current Web-based learning environments facilitate the theoretical teaching of a subject but this may not be sufficient for those disciplines that require a significant use of graphic mechanisms to resolve problems. This research study looks at the use of an environment that can help students learn engineering drawing with Web-based CAD tools, including a self-correction component. A comparative study of 121 students was carried out. The students were divided into two experimental groups using Web-based interactive CAD tools and into two control groups using traditional learning tools. A statistical analysis of all the samples was carried out in order to study student behavior during the research and the effectiveness of these self-study tools in the learning process. The results showed that a greater number of students in the experimental groups passed the test and improved their test scores. Therefore, the use Web-based graphic interactive tools to learn engineering drawing can be considered a significant improvement in the teaching of this kind of academic discipline.

  1. AN EFFICIENT APPROACH FOR KEYWORD SELECTION; IMPROVING ACCESSIBILITY OF WEB CONTENTS BY GENERAL SEARCH ENGINES

    Directory of Open Access Journals (Sweden)

    H. H. Kian

    2011-11-01

    Full Text Available General search engines often provide low precise results even for detailed queries. So there is a vital needto elicit useful information like keywords for search engines to provide acceptable results for user’s searchqueries. Although many methods have been proposed to show how to extract keywords automatically, allattempt to get a better recall, precision and other criteria which describe how the method has done its jobas an author. This paper presents a new automatic keyword extraction method which improves accessibilityof web content by search engines. The proposed method defines some coefficients determining featuresefficiency and tries to optimize them by using a genetic algorithm. Furthermore, it evaluates candidatekeywords by a function that utilizes the result of search engines. When comparing to the other methods,experiments demonstrate that by using the proposed method, a higher score is achieved from searchengines without losing noticeable recall or precision.

  2. Reconsidering the Rhizome: A Textual Analysis of Web Search Engines as Gatekeepers of the Internet

    Science.gov (United States)

    Hess, A.

    Critical theorists have often drawn from Deleuze and Guattari's notion of the rhizome when discussing the potential of the Internet. While the Internet may structurally appear as a rhizome, its day-to-day usage by millions via search engines precludes experiencing the random interconnectedness and potential democratizing function. Through a textual analysis of four search engines, I argue that Web searching has grown hierarchies, or "trees," that organize data in tracts of knowledge and place users in marketing niches rather than assist in the development of new knowledge.

  3. Personal learning environments in a global higher engineering education Web 2.0 realm

    OpenAIRE

    Gillet D.; Law E.L.-C.; Chatterjee A.

    2010-01-01

    This paper presents investigations on formal and informal requirements for personal learning environments taking into account students’ personal and social learning practices. The potential of global Web 2.0 educational service bundles and informal learning communities, as well as their recommendation by educators are addressed. A scenario showing how these new paradigms can be integrated in engineering education as a way to bring together personal and social learning practices is drawn.

  4. Search Engine Optimization for Flash Best Practices for Using Flash on the Web

    CERN Document Server

    Perkins, Todd

    2009-01-01

    Search Engine Optimization for Flash dispels the myth that Flash-based websites won't show up in a web search by demonstrating exactly what you can do to make your site fully searchable -- no matter how much Flash it contains. You'll learn best practices for using HTML, CSS and JavaScript, as well as SWFObject, for building sites with Flash that will stand tall in search rankings.

  5. A unified architecture for biomedical search engines based on semantic web technologies.

    Science.gov (United States)

    Jalali, Vahid; Matash Borujerdi, Mohammad Reza

    2011-04-01

    There is a huge growth in the volume of published biomedical research in recent years. Many medical search engines are designed and developed to address the over growing information needs of biomedical experts and curators. Significant progress has been made in utilizing the knowledge embedded in medical ontologies and controlled vocabularies to assist these engines. However, the lack of common architecture for utilized ontologies and overall retrieval process, hampers evaluating different search engines and interoperability between them under unified conditions. In this paper, a unified architecture for medical search engines is introduced. Proposed model contains standard schemas declared in semantic web languages for ontologies and documents used by search engines. Unified models for annotation and retrieval processes are other parts of introduced architecture. A sample search engine is also designed and implemented based on the proposed architecture in this paper. The search engine is evaluated using two test collections and results are reported in terms of precision vs. recall and mean average precision for different approaches used by this search engine.

  6. How to Boost Engineering Support Via Web 2.0 - Seeds for the Ares Project...and/or Yours?

    Science.gov (United States)

    Scott, David W.

    2010-01-01

    The Mission Operations Laboratory (MOL) at Marshall Space Flight Center (MSFC) is responsible for Engineering Support capability for NASA s Ares launch system development. In pursuit of this, MOL is building the Ares Engineering and Operations Network (AEON), a web-based portal intended to provide a seamless interface to support and simplify two critical activities: a) Access and analyze Ares manufacturing, test, and flight performance data, with access to Shuttle data for comparison. b) Provide archive storage for engineering instrumentation data to support engineering design, development, and test. A mix of NASA-written and COTS software provides engineering analysis tools. A by-product of using a data portal to access and display data is access to collaborative tools inherent in a Web 2.0 environment. This paper discusses how Web 2.0 techniques, particularly social media, might be applied to the traditionally conservative and formal engineering support arena. A related paper by the author [1] considers use

  7. Using the open Web as an information resource and scholarly Web search engines as retrieval tools for academic and research purposes

    OpenAIRE

    Filistea Naude; Chris Rensleigh; Adeline S.A. du Toit

    2010-01-01

    This study provided insight into the significance of the open Web as an information resource and Web search engines as research tools amongst academics. The academic staff establishment of the University of South Africa (Unisa) was invited to participate in a questionnaire survey and included 1188 staff members from five colleges. This study culminated in a PhD dissertation in 2008. One hundred and eighty seven respondents participated in the survey which gave a response rate of 15.7%. The re...

  8. An Evidence-Based Review of Academic Web Search Engines, 2014-2016: Implications for Librarians' Practice and Research Agenda

    National Research Council Canada - National Science Library

    Jody Condit Fagan

    2017-01-01

    Academic web search engines have become central to scholarly research. While the fitness of Google Scholar for research purposes has been examined repeatedly, Microsoft Academic and Google Books have not received much attention...

  9. Web-based Peer Assessment: A Case Study with Civil Engineering Students

    Directory of Open Access Journals (Sweden)

    Gonçalo Cruz

    2013-02-01

    Full Text Available Peer-assessed online collaborative writing is of growing importance for higher education. Based on the quantitative and qualitative analysis of collected data, this work presents the results of a case study in a graduate civil engineering course at the University of Trás-os-Montes e Alto Douro, Vila Real, Portugal. After an overall presentation of the learning activity, it examines students’ attitudes towards web-based peer assessment and identifies which factors are related to students’ use and appreciation of feedback. In addition, an analysis of the type of feedback is presented in an attempt to evaluate students’  written communication skills. The results of this study provide indications on how to improve the design and implementation of future web-based peer assessment tasks and raise questions on the effectiveness of peer/teacher grading in the use of feedback.

  10. Enhanced Trustworthy and High-Quality Information Retrieval System for Web Search Engines

    CERN Document Server

    Ramachandran, S; Joseph, S; Ramaraj, V

    2009-01-01

    The WWW is the most important source of information. But, there is no guarantee for information correctness and lots of conflicting information is retrieved by the search engines and the quality of provided information also varies from low quality to high quality. We provide enhanced trustworthiness in both specific (entity) and broad (content) queries in web searching. The filtering of trustworthiness is based on 5 factors: Provenance, Authority, Age, Popularity, and Related Links. The trustworthiness is calculated based on these 5 factors and it is stored thereby increasing the performance in retrieving trustworthy websites. The calculated trustworthiness is stored only for static websites. Quality is provided based on policies selected by the user. Quality based ranking of retrieved trusted information is provided using WIQA (Web Information Quality Assessment) Framework.

  11. Developing as new search engine and browser for libraries to search and organize the World Wide Web library resources

    OpenAIRE

    Sreenivasulu, V.

    2000-01-01

    Internet Granthalaya urges world wide advocates and targets at the task of creating a new search engine and dedicated browseer. Internet Granthalaya may be the ultimate search engine exclusively dedicated for every library use to search and organize the world wide web libary resources

  12. Developing as new search engine and browser for libraries to search and organize the World Wide Web library resources

    OpenAIRE

    SREENIVASULU, V.

    2000-01-01

    Internet Granthalaya urges world wide advocates and targets at the task of creating a new search engine and dedicated browseer. Internet Granthalaya may be the ultimate search engine exclusively dedicated for every library use to search and organize the world wide web libary resources

  13. A Web portal for the Engineering and Equipment Data Management System at CERN

    Science.gov (United States)

    Tsyganov, A.; Petit, S.; Martel, P.; Milenkovic, S.; Suwalska, A.; Delamare, C.; Widegren, D.; Mallón Amérigo, S.; Pettersson, T.

    2010-04-01

    CERN, the European Laboratory for Particle Physics, located in Geneva - Switzerland, has recently started the Large Hadron Collider (LHC), a 27 km particle accelerator. The CERN Engineering and Equipment Data Management Service (EDMS) provides support for managing engineering and equipment information throughout the entire lifecycle of a project. Based on several both in-house developed and commercial data management systems, this service supports management and follow-up of different kinds of information throughout the lifecycle of the LHC project: design, manufacturing, installation, commissioning data, maintenance and more. The data collection phase, carried out by specialists, is now being replaced by a phase during which data will be consulted on an extensive basis by non-experts users. In order to address this change, a Web portal for the EDMS has been developed. It brings together in one space all the aspects covered by the EDMS: project and document management, asset tracking and safety follow-up. This paper presents the EDMS Web portal, its dynamic content management and its "one click" information search engine.

  14. A Web portal for the Engineering and Equipment Data Management System at CERN

    Energy Technology Data Exchange (ETDEWEB)

    Tsyganov, A; Petit, S; Martel, P; Milenkovic, S; Suwalska, A; Delamare, C; Widegren, D; Amerigo, S Mallon; Pettersson, T [CERN, GS Department, CH-1211 Geneva 23 (Switzerland)

    2010-04-01

    CERN, the European Laboratory for Particle Physics, located in Geneva - Switzerland, has recently started the Large Hadron Collider (LHC), a 27 km particle accelerator. The CERN Engineering and Equipment Data Management Service (EDMS) provides support for managing engineering and equipment information throughout the entire lifecycle of a project. Based on several both in-house developed and commercial data management systems, this service supports management and follow-up of different kinds of information throughout the lifecycle of the LHC project: design, manufacturing, installation, commissioning data, maintenance and more. The data collection phase, carried out by specialists, is now being replaced by a phase during which data will be consulted on an extensive basis by non-experts users. In order to address this change, a Web portal for the EDMS has been developed. It brings together in one space all the aspects covered by the EDMS: project and document management, asset tracking and safety follow-up. This paper presents the EDMS Web portal, its dynamic content management and its 'one click' information search engine.

  15. ThermoData Engine (TDE): software implementation of the dynamic data evaluation concept. 6. Dynamic web-based data dissemination through the NIST Web Thermo Tables.

    Science.gov (United States)

    Kroenlein, Kenneth; Muzny, Chris D; Diky, Vladimir; Kazakov, Andrei F; Chirico, Robert D; Magee, Joseph W; Abdulagatov, Ilmutdin; Frenkel, Michael

    2011-06-27

    ThermoData Engine (TDE) is the first full-scale software implementation of the dynamic data evaluation concept, as reported recently in this journal. In the present paper, we describe the development of a World Wide Web-based interface to TDE evaluations of pure compound properties, including critical properties, phase boundary equilibria (vapor pressures, sublimation pressures, and crystal-liquid boundary pressures), densities, energetic properties, and transport properties. This includes development of a system for caching evaluation results to maintain high availability and an advanced window-in-window interface that leverages modern Web-browser technologies. Challenges associated with bringing the principal advantages of the TDE technology to the Web are described, as are compromises to maintain general access and speed of interaction while remaining true to the tenets of dynamic data evaluation. Future extensions of the interface and associated Web-services are outlined.

  16. MODEST: a web-based design tool for oligonucleotide-mediated genome engineering and recombineering

    DEFF Research Database (Denmark)

    Bonde, Mads; Klausen, Michael Schantz; Anderson, Mads Valdemar

    2014-01-01

    Recombineering and multiplex automated genome engineering (MAGE) offer the possibility to rapidly modify multiple genomic or plasmid sites at high efficiencies. This enables efficient creation of genetic variants including both single mutants with specifically targeted modifications as well......, which confers the corresponding genetic change, is performed manually. To address these challenges, we have developed the MAGE Oligo Design Tool (MODEST). This web-based tool allows designing of MAGE oligos for (i) tuning translation rates by modifying the ribosomal binding site, (ii) generating...... efficiency recombineering and MAGE. MODEST is available for free and is open to all users at http://modest.biosustain.dtu.dk....

  17. GeNemo: a search engine for web-based functional genomic data

    OpenAIRE

    Zhang, Yongqing; Cao, Xiaoyi; Zhong, Sheng

    2016-01-01

    A set of new data types emerged from functional genomic assays, including ChIP-seq, DNase-seq, FAIRE-seq and others. The results are typically stored as genome-wide intensities (WIG/bigWig files) or functional genomic regions (peak/BED files). These data types present new challenges to big data science. Here, we present GeNemo, a web-based search engine for functional genomic data. GeNemo searches user-input data against online functional genomic datasets, including the entire collection of E...

  18. Prototyping of Remote Experiment and Exercise Systems for an Engineering Education based on World Wide Web

    Science.gov (United States)

    Iwatsuki, Masami; Kato, Yoriyuki; Yonekawa, Akira

    State-of-the-art Internet technologies allow us to provide advanced and interactive distance education services. However, we could not help but gather students for experiments and exercises in an education for engineering because large-scale equipments and expensive software are required. On the other hand, teleoperation systems with robot manipulator or vehicle via Internet have been developed in the field of robotics. By fusing these two techniques, we can realize remote experiment and exercise systems for the engineering education based on World Wide Web. This paper presents how to construct the remote environment that allows students to take courses on experiment and exercise independently of their locations. By using the proposed system, users can exercise and practice remotely about control of a manipulator and a robot vehicle and programming of image processing.

  19. A Web portal for the Engineering and Equipment Data Management System at CERN

    CERN Document Server

    Tsyganov, A; Martel, P; Milenkovic, S; Suwalska, A; Delamare, Christophe; Widegren, David; Mallon Amerigo, S; Pettersson, Thomas Sven

    2010-01-01

    CERN, the European Laboratory for Particle Physics, located in Geneva – Switzerland, has recently started the Large Hadron Collider (LHC), a 27 km particle accelerator. The CERN Engineering and Equipment Data Management Service (EDMS) provides support for managing engineering and equipment information throughout the entire lifecycle of a project. Based on several both in-house developed and commercial data management systems, this service supports management and follow-up of different kinds of information throughout the lifecycle of the LHC project: design, manufacturing, installation, commissioning data, maintenance and more. The data collection phase, carried out by specialists, is now being replaced by a phase during which data will be consulted on an extensive basis by non-experts users. In order to address this change, a Web portal for the EDMS has been developed. It brings together in one space all the aspects covered by the EDMS: project and document management, asset tracking and safety follow-up. T...

  20. Fast segmentation of satellite images using SLIC, WebGL and Google Earth Engine

    Science.gov (United States)

    Donchyts, Gennadii; Baart, Fedor; Gorelick, Noel; Eisemann, Elmar; van de Giesen, Nick

    2017-04-01

    Google Earth Engine (GEE) is a parallel geospatial processing platform, which harmonizes access to petabytes of freely available satellite images. It provides a very rich API, allowing development of dedicated algorithms to extract useful geospatial information from these images. At the same time, modern GPUs provide thousands of computing cores, which are mostly not utilized in this context. In the last years, WebGL became a popular and well-supported API, allowing fast image processing directly in web browsers. In this work, we will evaluate the applicability of WebGL to enable fast segmentation of satellite images. A new implementation of a Simple Linear Iterative Clustering (SLIC) algorithm using GPU shaders will be presented. SLIC is a simple and efficient method to decompose an image in visually homogeneous regions. It adapts a k-means clustering approach to generate superpixels efficiently. While this approach will be hard to scale, due to a significant amount of data to be transferred to the client, it should significantly improve exploratory possibilities and simplify development of dedicated algorithms for geoscience applications. Our prototype implementation will be used to improve surface water detection of the reservoirs using multispectral satellite imagery.

  1. Integrating Ecosystem Engineering and Food Web Ecology: Testing the Effect of Biogenic Reefs on the Food Web of a Soft-Bottom Intertidal Area.

    Science.gov (United States)

    De Smet, Bart; Fournier, Jérôme; De Troch, Marleen; Vincx, Magda; Vanaverbeke, Jan

    2015-01-01

    The potential of ecosystem engineers to modify the structure and dynamics of food webs has recently been hypothesised from a conceptual point of view. Empirical data on the integration of ecosystem engineers and food webs is however largely lacking. This paper investigates the hypothesised link based on a field sampling approach of intertidal biogenic aggregations created by the ecosystem engineer Lanice conchilega (Polychaeta, Terebellidae). The aggregations are known to have a considerable impact on the physical and biogeochemical characteristics of their environment and subsequently on the abundance and biomass of primary food sources and the macrofaunal (i.e. the macro-, hyper- and epibenthos) community. Therefore, we hypothesise that L. conchilega aggregations affect the structure, stability and isotopic niche of the consumer assemblage of a soft-bottom intertidal food web. Primary food sources and the bentho-pelagic consumer assemblage of a L. conchilega aggregation and a control area were sampled on two soft-bottom intertidal areas along the French coast and analysed for their stable isotopes. Despite the structural impacts of the ecosystem engineer on the associated macrofaunal community, the presence of L. conchilega aggregations only has a minor effect on the food web structure of soft-bottom intertidal areas. The isotopic niche width of the consumer communities of the L. conchilega aggregations and control areas are highly similar, implying that consumer taxa do not shift their diet when feeding in a L. conchilega aggregation. Besides, species packing and hence trophic redundancy were not affected, pointing to an unaltered stability of the food web in the presence of L. conchilega.

  2. Characterizing interdisciplinarity of researchers and research topics using web search engines.

    Directory of Open Access Journals (Sweden)

    Hiroki Sayama

    Full Text Available Researchers' networks have been subject to active modeling and analysis. Earlier literature mostly focused on citation or co-authorship networks reconstructed from annotated scientific publication databases, which have several limitations. Recently, general-purpose web search engines have also been utilized to collect information about social networks. Here we reconstructed, using web search engines, a network representing the relatedness of researchers to their peers as well as to various research topics. Relatedness between researchers and research topics was characterized by visibility boost-increase of a researcher's visibility by focusing on a particular topic. It was observed that researchers who had high visibility boosts by the same research topic tended to be close to each other in their network. We calculated correlations between visibility boosts by research topics and researchers' interdisciplinarity at the individual level (diversity of topics related to the researcher and at the social level (his/her centrality in the researchers' network. We found that visibility boosts by certain research topics were positively correlated with researchers' individual-level interdisciplinarity despite their negative correlations with the general popularity of researchers. It was also found that visibility boosts by network-related topics had positive correlations with researchers' social-level interdisciplinarity. Research topics' correlations with researchers' individual- and social-level interdisciplinarities were found to be nearly independent from each other. These findings suggest that the notion of "interdisciplinarity" of a researcher should be understood as a multi-dimensional concept that should be evaluated using multiple assessment means.

  3. Characterizing interdisciplinarity of researchers and research topics using web search engines.

    Science.gov (United States)

    Sayama, Hiroki; Akaishi, Jin

    2012-01-01

    Researchers' networks have been subject to active modeling and analysis. Earlier literature mostly focused on citation or co-authorship networks reconstructed from annotated scientific publication databases, which have several limitations. Recently, general-purpose web search engines have also been utilized to collect information about social networks. Here we reconstructed, using web search engines, a network representing the relatedness of researchers to their peers as well as to various research topics. Relatedness between researchers and research topics was characterized by visibility boost-increase of a researcher's visibility by focusing on a particular topic. It was observed that researchers who had high visibility boosts by the same research topic tended to be close to each other in their network. We calculated correlations between visibility boosts by research topics and researchers' interdisciplinarity at the individual level (diversity of topics related to the researcher) and at the social level (his/her centrality in the researchers' network). We found that visibility boosts by certain research topics were positively correlated with researchers' individual-level interdisciplinarity despite their negative correlations with the general popularity of researchers. It was also found that visibility boosts by network-related topics had positive correlations with researchers' social-level interdisciplinarity. Research topics' correlations with researchers' individual- and social-level interdisciplinarities were found to be nearly independent from each other. These findings suggest that the notion of "interdisciplinarity" of a researcher should be understood as a multi-dimensional concept that should be evaluated using multiple assessment means.

  4. A Web Centric Architecture for Deploying Multi-Disciplinary Engineering Design Processes

    Science.gov (United States)

    Woyak, Scott; Kim, Hongman; Mullins, James; Sobieszczanski-Sobieski, Jaroslaw

    2004-01-01

    There are continuous needs for engineering organizations to improve their design process. Current state of the art techniques use computational simulations to predict design performance, and optimize it through advanced design methods. These tools have been used mostly by individual engineers. This paper presents an architecture for achieving results at an organization level beyond individual level. The next set of gains in process improvement will come from improving the effective use of computers and software within a whole organization, not just for an individual. The architecture takes advantage of state of the art capabilities to produce a Web based system to carry engineering design into the future. To illustrate deployment of the architecture, a case study for implementing advanced multidisciplinary design optimization processes such as Bi-Level Integrated System Synthesis is discussed. Another example for rolling-out a design process for Design for Six Sigma is also described. Each example explains how an organization can effectively infuse engineering practice with new design methods and retain the knowledge over time.

  5. Semantic similarity measures in the biomedical domain by leveraging a web search engine.

    Science.gov (United States)

    Hsieh, Sheau-Ling; Chang, Wen-Yung; Chen, Chi-Huang; Weng, Yung-Ching

    2013-07-01

    Various researches in web related semantic similarity measures have been deployed. However, measuring semantic similarity between two terms remains a challenging task. The traditional ontology-based methodologies have a limitation that both concepts must be resided in the same ontology tree(s). Unfortunately, in practice, the assumption is not always applicable. On the other hand, if the corpus is sufficiently adequate, the corpus-based methodologies can overcome the limitation. Now, the web is a continuous and enormous growth corpus. Therefore, a method of estimating semantic similarity is proposed via exploiting the page counts of two biomedical concepts returned by Google AJAX web search engine. The features are extracted as the co-occurrence patterns of two given terms P and Q, by querying P, Q, as well as P AND Q, and the web search hit counts of the defined lexico-syntactic patterns. These similarity scores of different patterns are evaluated, by adapting support vector machines for classification, to leverage the robustness of semantic similarity measures. Experimental results validating against two datasets: dataset 1 provided by A. Hliaoutakis; dataset 2 provided by T. Pedersen, are presented and discussed. In dataset 1, the proposed approach achieves the best correlation coefficient (0.802) under SNOMED-CT. In dataset 2, the proposed method obtains the best correlation coefficient (SNOMED-CT: 0.705; MeSH: 0.723) with physician scores comparing with measures of other methods. However, the correlation coefficients (SNOMED-CT: 0.496; MeSH: 0.539) with coder scores received opposite outcomes. In conclusion, the semantic similarity findings of the proposed method are close to those of physicians' ratings. Furthermore, the study provides a cornerstone investigation for extracting fully relevant information from digitizing, free-text medical records in the National Taiwan University Hospital database.

  6. A Webometric Analysis of ISI Medical Journals Using Yahoo, AltaVista, and All the Web Search Engines

    Directory of Open Access Journals (Sweden)

    Zohreh Zahedi

    2010-12-01

    Full Text Available The World Wide Web is an important information source for scholarly communications. Examining the inlinks via webometrics studies has attracted particular interests among information researchers. In this study, the number of inlinks to 69 ISI medical journals retrieved by Yahoo, AltaVista, and All The web Search Engines were examined via a comparative and Webometrics study. For data analysis, SPSS software was employed. Findings revealed that British Medical Journal website attracted the most links of all in the three search engines. There is a significant correlation between the number of External links and the ISI impact factor. The most significant correlation in the three search engines exists between external links of Yahoo and AltaVista (100% and the least correlation is found between external links of All The web & the number of pages of AltaVista (0.51. There is no significant difference between the internal links & the number of pages found by the three search engines. But in case of impact factors, significant differences are found between these three search engines. So, the study shows that journals with higher impact factor attract more links to their websites. It also indicates that the three search engines are significantly different in terms of total links, outlinks and web impact factors

  7. Electronic Grey Literature in Accelerator Science and Its Allied Subjects : Selected Web Resources for Scientists and Engineers

    CERN Document Server

    Rajendiran, P

    2006-01-01

    Grey literature Web resources in the field of accelerator science and its allied subjects are collected for the scientists and engineers of RRCAT (Raja Ramanna Centre for Advanced Technology). For definition purposes the different types of grey literature are described. The Web resources collected and compiled in this article (with an overview and link for each) specifically focus on technical reports, preprints or e-prints, which meet the main information needs of RRCAT users.

  8. A web-based online collaboration platform for formulating engineering design projects

    Science.gov (United States)

    Varikuti, Sainath

    Effective communication and collaboration among students, faculty and industrial sponsors play a vital role while formulating and solving engineering design projects. With the advent in the web technology, online platforms and systems have been proposed to facilitate interactions and collaboration among different stakeholders in the context of senior design projects. However, there are noticeable gaps in the literature with respect to understanding the effects of online collaboration platforms for formulating engineering design projects. Most of the existing literature is focused on exploring the utility of online platforms on activities after the problem is defined and teams are formed. Also, there is a lack of mechanisms and tools to guide the project formation phase in senior design projects, which makes it challenging for students and faculty to collaboratively develop and refine project ideas and to establish appropriate teams. In this thesis a web-based online collaboration platform is designed and implemented to share, discuss and obtain feedback on project ideas and to facilitate collaboration among students and faculty prior to the start of the semester. The goal of this thesis is to understand the impact of an online collaboration platform for formulating engineering design projects, and how a web-based online collaboration platform affects the amount of interactions among stakeholders during the early phases of design process. A survey measuring the amount of interactions among students and faculty is administered. Initial findings show a marked improvement in the students' ability to share project ideas and form teams with other students and faculty. Students found the online platform simple to use. The suggestions for improving the tool generally included features that were not necessarily design specific, indicating that the underlying concept of this collaborative platform provides a strong basis and can be extended for future online platforms

  9. A web-based search engine for triplex-forming oligonucleotide target sequences.

    Science.gov (United States)

    Gaddis, Sara S; Wu, Qi; Thames, Howard D; DiGiovanni, John; Walborg, Earl F; MacLeod, Michael C; Vasquez, Karen M

    2006-01-01

    Triplex technology offers a useful approach for site-specific modification of gene structure and function both in vitro and in vivo. Triplex-forming oligonucleotides (TFOs) bind to their target sites in duplex DNA, thereby forming triple-helical DNA structures via Hoogsteen hydrogen bonding. TFO binding has been demonstrated to site-specifically inhibit gene expression, enhance homologous recombination, induce mutation, inhibit protein binding, and direct DNA damage, thus providing a tool for gene-specific manipulation of DNA. We have developed a flexible web-based search engine to find and annotate TFO target sequences within the human and mouse genomes. Descriptive information about each site, including sequence context and gene region (intron, exon, or promoter), is provided. The engine assists the user in finding highly specific TFO target sequences by eliminating or flagging known repeat sequences and flagging overlapping genes. A convenient way to check for the uniqueness of a potential TFO binding site is provided via NCBI BLAST. The search engine may be accessed at spi.mdanderson.org/tfo.

  10. Development of Web-Based Learning Environment Model to Enhance Cognitive Skills for Undergraduate Students in the Field of Electrical Engineering

    Science.gov (United States)

    Lakonpol, Thongmee; Ruangsuwan, Chaiyot; Terdtoon, Pradit

    2015-01-01

    This research aimed to develop a web-based learning environment model for enhancing cognitive skills of undergraduate students in the field of electrical engineering. The research is divided into 4 phases: 1) investigating the current status and requirements of web-based learning environment models. 2) developing a web-based learning environment…

  11. Collaborative Science Using Web Services and the SciFlo Grid Dataflow Engine

    Science.gov (United States)

    Wilson, B. D.; Manipon, G.; Xing, Z.; Yunck, T.

    2006-12-01

    The General Earth Science Investigation Suite (GENESIS) project is a NASA-sponsored partnership between the Jet Propulsion Laboratory, academia, and NASA data centers to develop a new suite of Web Services tools to facilitate multi-sensor investigations in Earth System Science. The goal of GENESIS is to enable large-scale, multi-instrument atmospheric science using combined datasets from the AIRS, MODIS, MISR, and GPS sensors. Investigations include cross-comparison of spaceborne climate sensors, cloud spectral analysis, study of upper troposphere-stratosphere water transport, study of the aerosol indirect cloud effect, and global climate model validation. The challenges are to bring together very large datasets, reformat and understand the individual instrument retrievals, co-register or re-grid the retrieved physical parameters, perform computationally-intensive data fusion and data mining operations, and accumulate complex statistics over months to years of data. To meet these challenges, we have developed a Grid computing and dataflow framework, named SciFlo, in which we are deploying a set of versatile and reusable operators for data access, subsetting, registration, mining, fusion, compression, and advanced statistical analysis. SciFlo leverages remote Web Services, called via Simple Object Access Protocol (SOAP) or REST (one-line) URLs, and the Grid Computing standards (WS-* &Globus Alliance toolkits), and enables scientists to do multi-instrument Earth Science by assembling reusable Web Services and native executables into a distributed computing flow (tree of operators). The SciFlo client &server engines optimize the execution of such distributed data flows and allow the user to transparently find and use datasets and operators without worrying about the actual location of the Grid resources. In particular, SciFlo exploits the wealth of datasets accessible by OpenGIS Consortium (OGC) Web Mapping Servers & Web Coverage Servers (WMS/WCS), and by Open Data

  12. Disulfide by Design 2.0: a web-based tool for disulfide engineering in proteins.

    Science.gov (United States)

    Craig, Douglas B; Dombkowski, Alan A

    2013-12-01

    Disulfide engineering is an important biotechnological tool that has advanced a wide range of research. The introduction of novel disulfide bonds into proteins has been used extensively to improve protein stability, modify functional characteristics, and to assist in the study of protein dynamics. Successful use of this technology is greatly enhanced by software that can predict pairs of residues that will likely form a disulfide bond if mutated to cysteines. We had previously developed and distributed software for this purpose: Disulfide by Design (DbD). The original DbD program has been widely used; however, it has a number of limitations including a Windows platform dependency. Here, we introduce Disulfide by Design 2.0 (DbD2), a web-based, platform-independent application that significantly extends functionality, visualization, and analysis capabilities beyond the original program. Among the enhancements to the software is the ability to analyze the B-factor of protein regions involved in predicted disulfide bonds. Importantly, this feature facilitates the identification of potential disulfides that are not only likely to form but are also expected to provide improved thermal stability to the protein. DbD2 provides platform-independent access and significantly extends the original functionality of DbD. A web server hosting DbD2 is provided at http://cptweb.cpt.wayne.edu/DbD2/.

  13. GeNemo: a search engine for web-based functional genomic data.

    Science.gov (United States)

    Zhang, Yongqing; Cao, Xiaoyi; Zhong, Sheng

    2016-07-08

    A set of new data types emerged from functional genomic assays, including ChIP-seq, DNase-seq, FAIRE-seq and others. The results are typically stored as genome-wide intensities (WIG/bigWig files) or functional genomic regions (peak/BED files). These data types present new challenges to big data science. Here, we present GeNemo, a web-based search engine for functional genomic data. GeNemo searches user-input data against online functional genomic datasets, including the entire collection of ENCODE and mouse ENCODE datasets. Unlike text-based search engines, GeNemo's searches are based on pattern matching of functional genomic regions. This distinguishes GeNemo from text or DNA sequence searches. The user can input any complete or partial functional genomic dataset, for example, a binding intensity file (bigWig) or a peak file. GeNemo reports any genomic regions, ranging from hundred bases to hundred thousand bases, from any of the online ENCODE datasets that share similar functional (binding, modification, accessibility) patterns. This is enabled by a Markov Chain Monte Carlo-based maximization process, executed on up to 24 parallel computing threads. By clicking on a search result, the user can visually compare her/his data with the found datasets and navigate the identified genomic regions. GeNemo is available at www.genemo.org.

  14. Knowledge-based personalized search engine for the Web-based Human Musculoskeletal System Resources (HMSR) in biomechanics.

    Science.gov (United States)

    Dao, Tien Tuan; Hoang, Tuan Nha; Ta, Xuan Hien; Tho, Marie Christine Ho Ba

    2013-02-01

    Human musculoskeletal system resources of the human body are valuable for the learning and medical purposes. Internet-based information from conventional search engines such as Google or Yahoo cannot response to the need of useful, accurate, reliable and good-quality human musculoskeletal resources related to medical processes, pathological knowledge and practical expertise. In this present work, an advanced knowledge-based personalized search engine was developed. Our search engine was based on a client-server multi-layer multi-agent architecture and the principle of semantic web services to acquire dynamically accurate and reliable HMSR information by a semantic processing and visualization approach. A security-enhanced mechanism was applied to protect the medical information. A multi-agent crawler was implemented to develop a content-based database of HMSR information. A new semantic-based PageRank score with related mathematical formulas were also defined and implemented. As the results, semantic web service descriptions were presented in OWL, WSDL and OWL-S formats. Operational scenarios with related web-based interfaces for personal computers and mobile devices were presented and analyzed. Functional comparison between our knowledge-based search engine, a conventional search engine and a semantic search engine showed the originality and the robustness of our knowledge-based personalized search engine. In fact, our knowledge-based personalized search engine allows different users such as orthopedic patient and experts or healthcare system managers or medical students to access remotely into useful, accurate, reliable and good-quality HMSR information for their learning and medical purposes.

  15. Assessment and Comparison of Search capabilities of Web-based Meta-Search Engines: A Checklist Approach

    Directory of Open Access Journals (Sweden)

    Alireza Isfandiyari Moghadam

    2010-03-01

    Full Text Available   The present investigation concerns evaluation, comparison and analysis of search options existing within web-based meta-search engines. 64 meta-search engines were identified. 19 meta-search engines that were free, accessible and compatible with the objectives of the present study were selected. An author’s constructed check list was used for data collection. Findings indicated that all meta-search engines studied used the AND operator, phrase search, number of results displayed setting, previous search query storage and help tutorials. Nevertheless, none of them demonstrated any search options for hypertext searching and displaying the size of the pages searched. 94.7% support features such as truncation, keywords in title and URL search and text summary display. The checklist used in the study could serve as a model for investigating search options in search engines, digital libraries and other internet search tools.

  16. Personal Search Engine Based on Web Services%基于Web Services的个人搜索引擎

    Institute of Scientific and Technical Information of China (English)

    曹龙; 杜亚军; 刘倬; 李战胜

    2005-01-01

    随着Internet的出现,要从浩如烟海的网络信息中找出自己所想要的信息是非常困难的,Google是非常著名的搜索引擎,其搜索机制是面向大众的.其提供的Web services功能利用标准化的Web协议在任何一种网络环境下可以使用这种服务,为定制个性化的搜索引擎提供了方便.文章探讨了Google搜索引擎提供的Web Service及程序实现方式,并使用Delphi提供的Web Services组件与Google服务接口,实现了一个个性化的搜索引擎,实验表明其搜索结果与Google本身的搜索功能有了大的改进.

  17. Sagace: A web-based search engine for biomedical databases in Japan

    Directory of Open Access Journals (Sweden)

    Morita Mizuki

    2012-10-01

    Full Text Available Abstract Background In the big data era, biomedical research continues to generate a large amount of data, and the generated information is often stored in a database and made publicly available. Although combining data from multiple databases should accelerate further studies, the current number of life sciences databases is too large to grasp features and contents of each database. Findings We have developed Sagace, a web-based search engine that enables users to retrieve information from a range of biological databases (such as gene expression profiles and proteomics data and biological resource banks (such as mouse models of disease and cell lines. With Sagace, users can search more than 300 databases in Japan. Sagace offers features tailored to biomedical research, including manually tuned ranking, a faceted navigation to refine search results, and rich snippets constructed with retrieved metadata for each database entry. Conclusions Sagace will be valuable for experts who are involved in biomedical research and drug development in both academia and industry. Sagace is freely available at http://sagace.nibio.go.jp/en/.

  18. E-TALEN: a web tool to design TALENs for genome engineering.

    Science.gov (United States)

    Heigwer, Florian; Kerr, Grainne; Walther, Nike; Glaeser, Kathrin; Pelz, Oliver; Breinig, Marco; Boutros, Michael

    2013-11-01

    Use of transcription activator-like effector nucleases (TALENs) is a promising new technique in the field of targeted genome engineering, editing and reverse genetics. Its applications span from introducing knockout mutations to endogenous tagging of proteins and targeted excision repair. Owing to this wide range of possible applications, there is a need for fast and user-friendly TALEN design tools. We developed E-TALEN (http://www.e-talen.org), a web-based tool to design TALENs for experiments of varying scale. E-TALEN enables the design of TALENs against a single target or a large number of target genes. We significantly extended previously published design concepts to consider genomic context and different applications. E-TALEN guides the user through an end-to-end design process of de novo TALEN pairs, which are specific to a certain sequence or genomic locus. Furthermore, E-TALEN offers a functionality to predict targeting and specificity for existing TALENs. Owing to the computational complexity of many of the steps in the design of TALENs, particular emphasis has been put on the implementation of fast yet accurate algorithms. We implemented a user-friendly interface, from the input parameters to the presentation of results. An additional feature of E-TALEN is the in-built sequence and annotation database available for many organisms, including human, mouse, zebrafish, Drosophila and Arabidopsis, which can be extended in the future.

  19. THE EFFECTIVENESS OF WEB-BASED INTERACTIVE BLENDED LEARNING MODEL IN ELECTRICAL ENGINEERING COURSES

    Directory of Open Access Journals (Sweden)

    Hansi Effendi

    2015-12-01

    Full Text Available The study was to test the effectiveness of the Web-Based Interactive Blended Learning Model (BLIBW for subjects in the Department of Electrical Engineering, Padang State University. The design that the researcher employed was a quasi-experimental design with one group pretest-posttest, which was conducted on a group of students consisting of 30 people and the test was conducted for two times. The effectiveness of BLIBW Model was tested by comparing the average pretest scores and the average posttest scores both in the first trial and the second trial. The average prestest and posttest scores in the first trial were 14.13 and 33.80. The increase in the average score was significant at alpha 0.05. Then, the average pretest and posttest scores in the second trial were 18.67 and 47.03. The result was also significant at alpha 0.05. The effectiveness of BLIBW Model in the second trial was higher than in the first test. Those result were not entirely satisfactory and it might be caused several weaknesses in both tests such as: the number of sessions were limited, there was only one subject, and the number of students who were subjected too limited. However, the researcher would like to conclude that the BLIBW Model might be implemented as a replacement alternative for the face-to-face instruction.

  20. Search Engines and Search Technologies for Web-based Text Data%网络文本数据搜索引擎与搜索技术

    Institute of Scientific and Technical Information of China (English)

    李勇

    2001-01-01

    This paper describes the functions, characteristics and operating principles of search engines based on Web text, and the searching and data mining technologies for Web-based text information. Methods of computer-aided text clustering and abstacting are also given. Finally, it gives some guidelines for the assessment of searching quality.

  1. The Effects of Web-Based and Face-to-Face Discussion on Computer Engineering Majors' Performance on the Karnaugh Map

    Science.gov (United States)

    Hung, Yen-Chu

    2011-01-01

    This study investigates the different effects of web-based and face-to-face discussion on computer engineering majors' performance using the Karnaugh map in digital logic design. Pretest and posttest scores for two treatment groups (web-based discussion and face-to-face discussion) and a control group were compared and subjected to covariance…

  2. The Invisible Web: Uncovering Information Sources Search Engines Can't See.

    Science.gov (United States)

    Sherman, Chris; Price, Gary

    This book takes a detailed look at the nature and extent of the Invisible Web, and offers pathfinders for accessing the valuable information it contains. It is designed to fit the needs of both novice and advanced Web searchers. Chapter One traces the development of the Internet and many of the early tools used to locate and share information via…

  3. The Invisible Web: Uncovering Information Sources Search Engines Can't See.

    Science.gov (United States)

    Sherman, Chris; Price, Gary

    This book takes a detailed look at the nature and extent of the Invisible Web, and offers pathfinders for accessing the valuable information it contains. It is designed to fit the needs of both novice and advanced Web searchers. Chapter One traces the development of the Internet and many of the early tools used to locate and share information via…

  4. A fuzzy method for improving the functionality of search engines based on user's web interactions

    Directory of Open Access Journals (Sweden)

    Farzaneh Kabirbeyk

    2015-04-01

    Full Text Available Web mining has been widely used to discover knowledge from various sources in the web. One of the important tools in web mining is mining of web user’s behavior that is considered as a way to discover the potential knowledge of web user’s interaction. Nowadays, Website personalization is regarded as a popular phenomenon among web users and it plays an important role in facilitating user access and provides information of users’ requirements based on their own interests. Extracting important features about web user behavior plays a significant role in web usage mining. Such features are page visit frequency in each session, visit duration, and dates of visiting a certain pages. This paper presents a method to predict user’s interest and to propose a list of pages based on their interests by identifying user’s behavior based on fuzzy techniques called fuzzy clustering method. Due to the user’s different interests and use of one or more interest at a time, user’s interest may belong to several clusters and fuzzy clustering provide a possible overlap. Using the resulted cluster helps extract fuzzy rules. This helps detecting user’s movement pattern and using neural network a list of suggested pages to the users is provided.

  5. Engineering

    National Research Council Canada - National Science Library

    Includes papers in the following fields: Aerospace Engineering, Agricultural Engineering, Chemical Engineering, Civil Engineering, Electrical Engineering, Environmental Engineering, Industrial Engineering, Materials Engineering, Mechanical...

  6. DESIGN AND IMPLEMENTATION OF WEB SERVICES SEARCH ENGINE%Web服务搜索引擎的设计与实现

    Institute of Scientific and Technical Information of China (English)

    贺财平; 覃事刚; 刘建勋

    2011-01-01

    With the gradually increasing number of open Web services, it is crucial to obtain in full and effectively such open Web services scattered on the internet as well as manage them.In this paper,we designed and implemented a WSSE (Web Services Search Engine) in order to solve this issue.A robot was developed in WSSE to crawl along the existing Web sites incessantly for searching the Web services, and the found Web services would then be stored in a centralized management system.Finally we used the open-source Lucene to index the found Web services for enhancing the efficiency of Web services retrieval.%随着开放的Web服务数量的逐渐递增,全面而有效地获取散落于Internet上的这类Web服务并进行管理就成为一个十分重要的问题.针对该问题,设计与实现了一个Web服务搜索引擎WSSE(Web Services Search Engine).它不断爬行现有Web站点以搜索Web服务,并对搜集到的Web服务进行集中式管理,最后采用开源的Lucene对搜索到的Web服务建立索引,提高了Web服务的检索效率.

  7. Comparative Study of Web-based Virtual Library and Web Search Engine on Internet%Web资源虚拟图书馆与搜索引擎的比较研究

    Institute of Scientific and Technical Information of China (English)

    贺亚锋

    2000-01-01

    Particularly based on the comparative study of Web-based virtual library and Web search engine on the Internet, this paper discusses the similarities and differences in retrieval theory, retrieval performance and effect,in order to put forward suggestions for the development of Web-based virtual library and the promotion of Web search engine.%本文对Web上的主要信息检索工具-图书馆制作的Web资源虚拟图书馆和ICP研制的搜索引擎作分析比较,目的在于探讨两种检索工具之间的检索理论、检索性能和检索效果的异同,以期对Web资源虚拟图书馆的发展和搜索引擎的改进提供借鉴。

  8. The efficacy of using search engines in procuring information about orthopaedic foot and ankle problems from the World Wide Web.

    Science.gov (United States)

    Nogler, M; Wimmer, C; Mayr, E; Ofner, D

    1999-05-01

    This study has attempted to demonstrate the feasibility of obtaining information specific to foot and ankle orthopaedics from the World Wide Web (WWW). Six search engines (Lycos, AltaVista, Infoseek, Excite, Webcrawler, and HotBot) were used in scanning the Web for the following key words: "cavus foot," "diabetic foot," "hallux valgus,"and "pes equinovarus." Matches were classified by language, provider, type, and relevance to medical professionals or to patients. Sixty percent (407 sites) of the visited websites contained information intended for use by physicians and other medical professionals; 30% (206 sites) were related to patient information; 10% of the sites were not easily classifiable. Forty-one percent (169 sites) of the websites were commercially oriented homepages that included advertisements.

  9. Design of web platform for science and engineering in the model of open market

    Science.gov (United States)

    Demichev, A. P.; Kryukov, A. P.

    2016-09-01

    This paper presents a design and operation algorithms of a web-platform for convenient, secure and effective remote interaction on the principles of the open market of users and providers of scientific application software and databases.

  10. Research on Search Engine Based on Semantic Web%基于语义Web的搜索引擎研究

    Institute of Scientific and Technical Information of China (English)

    吴根斌; 丁振凡

    2012-01-01

    传统搜索引擎是基于关键字的检索,然而文档的关键字未必和文档有关,而相关的文档也未必显式地包含此关键字.基于语义Web的搜索引擎利用本体技术,可以很好地对关键字进行语义描述.当收到用户提交的搜索请求时,先在已经建立好的本体库的基础上对该请求进行概念推理,然后将推理结果提交给传统的搜索引擎,最终将搜索结果返回给用户.相对于传统的搜索引擎,基于语义Web的搜索引擎有效地提高了搜索的查全率和查准率.%Traditional search engines are keyword-based retrieval. However, the keywords of a document may not be related with the document, and the related documents may also not explicitly contain this keyword. Using ontology technology, search engine based on semantic Web could semantically describe the keywords. When receiving a user' s request of query, system executes logical reasoning on the ontology already created and sends the result of logical reasoning to the traditional engine. The system re-turns the result of search to user at last. Comparing with the traditional search engine, the search engine based on semantic Web can effectively improve the coverage rate and precision rate.

  11. Start Your Search Engines. Part One: Taming Google--and Other Tips to Master Web Searches

    Science.gov (United States)

    Adam, Anna; Mowers, Helen

    2008-01-01

    There are a lot of useful tools on the Web, all those social applications, and the like. Still most people go online for one thing--to perform a basic search. For most fact-finding missions, the Web is there. But--as media specialists well know--the sheer wealth of online information can hamper efforts to focus on a few reliable references.…

  12. An open-source highly scalable web service architecture for the Apertium machine translation engine

    OpenAIRE

    Sánchez-Cartagena, Víctor M.; Pérez-Ortiz, Juan Antonio

    2009-01-01

    Some machine translation services like Google Ajax Language API have become very popular as they make the collaboratively created contents of the web 2.0 available to speakers of many languages. One of the keys of its success is its clear and easy-to-use application programming interface (API) and a scalable and reliable service. This paper describes a highly scalable implementation of an Apertium-based translation web service, that aims to make contents available to speakers of lesser resour...

  13. Web Similarity

    NARCIS (Netherlands)

    Cohen, A.R.; Vitányi, P.M.B.

    2015-01-01

    Normalized web distance (NWD) is a similarity or normalized semantic distance based on the World Wide Web or any other large electronic database, for instance Wikipedia, and a search engine that returns reliable aggregate page counts. For sets of search terms the NWD gives a similarity on a scale fr

  14. Cooperative Project-Based Learning in a Web-Based Software Engineering Course

    Science.gov (United States)

    Piccinini, Nicola; Scollo, Giuseppe

    2006-01-01

    Even in self-organized project-based learning, the instructors' role re-mains critical, especially in the initial orientation provided to the students in order to grasp the educational goals and the various roles they may undertake to achieve them. In this paper we survey a few questions proposed to that purpose in a web-based software engineering…

  15. MODEST: a web-based design tool for oligonucleotide-mediated genome engineering and recombineering

    DEFF Research Database (Denmark)

    Bonde, Mads; Klausen, Michael Schantz; Anderson, Mads Valdemar

    2014-01-01

    , which confers the corresponding genetic change, is performed manually. To address these challenges, we have developed the MAGE Oligo Design Tool (MODEST). This web-based tool allows designing of MAGE oligos for (i) tuning translation rates by modifying the ribosomal binding site, (ii) generating...

  16. An efficient and flexible web services-based multidisciplinary design optimisation framework for complex engineering systems

    Science.gov (United States)

    Li, Liansheng; Liu, Jihong

    2012-08-01

    Multidisciplinary design optimisation (MDO) involves multiple disciplines, multiple coupled relationships and multiple processes, which is implemented by different specialists dispersed geographically on heterogeneous platforms with different analysis and optimisation tools. The product design data integration and data sharing among the participants hampers the development and applications of MDO in enterprises seriously. Therefore, a multi-hierarchical integrated product design data model (MH-iPDM) supporting the MDO in the web environment and a web services-based multidisciplinary design optimisation (Web-MDO) framework are proposed in this article. Based on the enabling technologies including web services, ontology, workflow, agent, XML and evidence theory, the proposed framework enables the designers geographically dispersed to work collaboratively in the MDO environment. The ontology-based workflow enables the logical reasoning of MDO to be processed dynamically. The evidence theory-based uncertainty reasoning and analysis supports the quantification, aggregation and analysis of the conflicting epistemic uncertainty from multiple sources, which improves the quality of product. Finally, a proof-of-concept prototype system is developed using J2EE and an example of supersonic business jet is demonstrated to verify the autonomous execution of MDO strategies and the effectiveness of the proposed approach.

  17. Cooperative Project-Based Learning in a Web-Based Software Engineering Course

    Science.gov (United States)

    Piccinini, Nicola; Scollo, Giuseppe

    2006-01-01

    Even in self-organized project-based learning, the instructors' role re-mains critical, especially in the initial orientation provided to the students in order to grasp the educational goals and the various roles they may undertake to achieve them. In this paper we survey a few questions proposed to that purpose in a web-based software engineering…

  18. A geospatial search engine for discovering multi-format geospatial data across the web

    Science.gov (United States)

    Christopher Bone; Alan Ager; Ken Bunzel; Lauren Tierney

    2014-01-01

    The volume of publically available geospatial data on the web is rapidly increasing due to advances in server-based technologies and the ease at which data can now be created. However, challenges remain with connecting individuals searching for geospatial data with servers and websites where such data exist. The objective of this paper is to present a publically...

  19. Engineering web maps with gradual content zoom based on streaming vector data

    Science.gov (United States)

    Huang, Lina; Meijers, Martijn; Šuba, Radan; van Oosterom, Peter

    2016-04-01

    Vario-scale data structures have been designed to support gradual content zoom and the progressive transfer of vector data, for use with arbitrary map scales. The focus to date has been on the server side, especially on how to convert geographic data into the proposed vario-scale structures by means of automated generalisation. This paper contributes to the ongoing vario-scale research by focusing on the client side and communication, particularly on how this works in a web-services setting. It is claimed that these functionalities are urgently needed, as many web-based applications, both desktop and mobile, require gradual content zoom, progressive transfer and a high performance level. The web-client prototypes developed in this paper make it possible to assess the behaviour of vario-scale data and to determine how users will actually see the interactions. Several different options of web-services communication architectures are possible in a vario-scale setting. These options are analysed and tested with various web-client prototypes, with respect to functionality, ease of implementation and performance (amount of transmitted data and response times). We show that the vario-scale data structure can fit in with current web-based architectures and efforts to standardise map distribution on the internet. However, to maximise the benefits of vario-scale data, a client needs to be aware of this structure. When a client needs a map to be refined (by means of a gradual content zoom operation), only the 'missing' data will be requested. This data will be sent incrementally to the client from a server. In this way, the amount of data transferred at one time is reduced, shortening the transmission time. In addition to these conceptual architecture aspects, there are many implementation and tooling design decisions at play. These will also be elaborated on in this paper. Based on the experiments conducted, we conclude that the vario-scale approach indeed supports gradual

  20. The Einstein Suite: A Web-Based Tool for Rapid and Collaborative Engineering Design and Analysis

    Science.gov (United States)

    Palmer, Richard S.

    1997-01-01

    Taken together the components of the Einstein Suite provide two revolutionary capabilities - they have the potential to change the way engineering and financial engineering are performed by: (1) providing currently unavailable functionality, and (2) providing a 10-100 times improvement over currently available but impractical or costly functionality.

  1. A Strategic Analysis of Search Engine Advertising in Web based-commerce

    Directory of Open Access Journals (Sweden)

    Ela Kumar

    2007-08-01

    Full Text Available Endeavor of this paper is to explore the role play of Search Engine in Online Business Industry. This paper discusses the Search Engine advertising programs and provides an insight about the revenue generated online via Search Engine. It explores the growth of Online Business Industry in India and emphasis on the role of Search Engine as the major advertising vehicle. A case study on re volution of Indian Advertising Industry has been conducted and its impact on online revenu e evaluated. Search Engine advertising strategies have been discussed in detail and the impact of Search Engine on Indian Advertising Industry has been analyzed. It also provides an analytical and competitive study of online advertising strategies with traditional advertising tools to evaluate their efficiencies against important advertising parameters. The paper concludes with a brief discussion on the malpractices that have adversarial impact on the efficiency of the Search Engine advertising model and highlight key hurdle Search Engine Industry is facing in Indian Business Scenario

  2. New Architectures for Presenting Search Results Based on Web Search Engines Users Experience

    Science.gov (United States)

    Martinez, F. J.; Pastor, J. A.; Rodriguez, J. V.; Lopez, Rosana; Rodriguez, J. V., Jr.

    2011-01-01

    Introduction: The Internet is a dynamic environment which is continuously being updated. Search engines have been, currently are and in all probability will continue to be the most popular systems in this information cosmos. Method: In this work, special attention has been paid to the series of changes made to search engines up to this point,…

  3. The Algorithm and Implementation of Picking 3D Figure by Web3D Engine%Web3D引擎中三维图形对象拾取的算法与实现

    Institute of Scientific and Technical Information of China (English)

    陈煜; 林玮

    2011-01-01

    应用Web3D引擎开发的计算机仿真系统和虚拟现实系统均需在Web浏览器上运行,需要其能快速下载和运行,因而要求尽可能提高Web3D引擎包括拾取功能在内的计算效率.在算法上提出了射线的分层次求交,先采用包围盒算法来实现对选择对象的快速拾取;为优化拾取的精度,再用三角形算法来实现准确拾取.在提出算法的基础上,用JAVA语言实现了在开源Web3D引擎上对所选图形对象拾取功能的开发.%Since computer simulation systems and virtual reality systems developed by the application of Web3D engine needs to run in a Web browser, downloading and running these systems shall be fast. It requires that the calculation speed of Web3D engine including picking function shall be improved as far as possible. A hierarchical algorithm of rays intersection is presented. The first layer is a rapid object selection by AABB algorithm, the second layer is the improvement to the picking precision by triangle algorithm. Based on the proposed algorithm, the picking function of selecting figure object is implemented on an open source Web3D engine by JAVA language.

  4. Representing Human Expertise by the OWL Web Ontology Language to Support Knowledge Engineering in Decision Support Systems.

    Science.gov (United States)

    Ramzan, Asia; Wang, Hai; Buckingham, Christopher

    2014-01-01

    Clinical decision support systems (CDSSs) often base their knowledge and advice on human expertise. Knowledge representation needs to be in a format that can be easily understood by human users as well as supporting ongoing knowledge engineering, including evolution and consistency of knowledge. This paper reports on the development of an ontology specification for managing knowledge engineering in a CDSS for assessing and managing risks associated with mental-health problems. The Galatean Risk and Safety Tool, GRiST, represents mental-health expertise in the form of a psychological model of classification. The hierarchical structure was directly represented in the machine using an XML document. Functionality of the model and knowledge management were controlled using attributes in the XML nodes, with an accompanying paper manual for specifying how end-user tools should behave when interfacing with the XML. This paper explains the advantages of using the web-ontology language, OWL, as the specification, details some of the issues and problems encountered in translating the psychological model to OWL, and shows how OWL benefits knowledge engineering. The conclusions are that OWL can have an important role in managing complex knowledge domains for systems based on human expertise without impeding the end-users' understanding of the knowledge base. The generic classification model underpinning GRiST makes it applicable to many decision domains and the accompanying OWL specification facilitates its implementation.

  5. GLIDERS - A web-based search engine for genome-wide linkage disequilibrium between HapMap SNPs

    Directory of Open Access Journals (Sweden)

    Broxholme John

    2009-10-01

    Full Text Available Abstract Background A number of tools for the examination of linkage disequilibrium (LD patterns between nearby alleles exist, but none are available for quickly and easily investigating LD at longer ranges (>500 kb. We have developed a web-based query tool (GLIDERS: Genome-wide LInkage DisEquilibrium Repository and Search engine that enables the retrieval of pairwise associations with r2 ≥ 0.3 across the human genome for any SNP genotyped within HapMap phase 2 and 3, regardless of distance between the markers. Description GLIDERS is an easy to use web tool that only requires the user to enter rs numbers of SNPs they want to retrieve genome-wide LD for (both nearby and long-range. The intuitive web interface handles both manual entry of SNP IDs as well as allowing users to upload files of SNP IDs. The user can limit the resulting inter SNP associations with easy to use menu options. These include MAF limit (5-45%, distance limits between SNPs (minimum and maximum, r2 (0.3 to 1, HapMap population sample (CEU, YRI and JPT+CHB combined and HapMap build/release. All resulting genome-wide inter-SNP associations are displayed on a single output page, which has a link to a downloadable tab delimited text file. Conclusion GLIDERS is a quick and easy way to retrieve genome-wide inter-SNP associations and to explore LD patterns for any number of SNPs of interest. GLIDERS can be useful in identifying SNPs with long-range LD. This can highlight mis-mapping or other potential association signal localisation problems.

  6. 在搜索引擎中增加新网站的方法%The Method to Increase New Web Site in the Search Engine

    Institute of Scientific and Technical Information of China (English)

    方兰平; 杨晓梅

    2001-01-01

    Several search engines in common use at present in this text are the example,and the method that increases the new web site in the search engine has been briefly introduced.%以目前常用的几个搜索引擎为例,对怎样在搜索引擎中增加新网站的方法作了一些简介。

  7. 一种面向网络发布的海洋数据库引擎%A Marine Remote Sensing Spatial Database Engine for Web Publishing

    Institute of Scientific and Technical Information of China (English)

    陈志荣; 徐财江

    2008-01-01

    To meet the requirements of efficient management and web publishing for marine remote sensing data,a spatial database engine,named MRSSDE,is designed independently.The logical model,physical model,and optimization method of MRSSDE are discussed in detail.Compared to the ArcSDE,which is the leading product of Spatial Database Engine,the MRSSDE proved to be more effective.

  8. Problem-Based Learning in Web Environments: The Case of ``Virtual eBMS'' for Business Engineering Education

    Science.gov (United States)

    Elia, Gianluca; Secundo, Giustina; Taurino, Cesare

    This chapter presents a case study where Problem Based Learning (PBL) approach is applied to a Web-based environment. It first describes the main features behind the PBL for creating Business Engineers able to face the grand technological challenges of the 2020. Then it introduces a Web Based system supporting the PBL strategy, called the “Virtual eBMS”. This system has been designed and implemented at the e-Business Management Section of the Scuola Superiore ISUFI - University of Salento (Italy), in the framework of a research project carried out in collaboration with IBM. Besides the logical and technological description of Virtual eBMS, the chapter presents two applications of the platform in two different contexts: an academic context (international master) and an entrepreneurial context (awareness workshop with companies and entrepreneurs). The system is illustrated starting from the description of an operational framework for designing curricula PBL based from the author perspective and, then, illustrating a typical scenario of a learner accessing to the curricula. In the description, it is highlighted both the “structured” way and the “unstructured” way to create and follow an entire learning path.

  9. A Powerful, Cost Effective, Web Based Engineering Solution Supporting Conjunction Detection and Visual Analysis

    Science.gov (United States)

    Novak, Daniel M.; Biamonti, Davide; Gross, Jeremy; Milnes, Martin

    2013-08-01

    An innovative and visually appealing tool is presented for efficient all-vs-all conjunction analysis on a large catalogue of objects. The conjunction detection uses a nearest neighbour search algorithm, based on spatial binning and identification of pairs of objects in adjacent bins. This results in the fastest all vs all filtering the authors are aware of. The tool is constructed on a server-client architecture, where the server broadcasts to the client the conjunction data and ephemerides, while the client supports the user interface through a modern browser, without plug-in. In order to make the tool flexible and maintainable, Java software technologies were used on the server side, including Spring, Camel, ActiveMQ and CometD. The user interface and visualisation are based on the latest web technologies: HTML5, WebGL, THREE.js. Importance has been given on the ergonomics and visual appeal of the software. In fact certain design concepts have been borrowed from the gaming industry.

  10. Finding Business Information on the "Invisible Web": Search Utilities vs. Conventional Search Engines.

    Science.gov (United States)

    Darrah, Brenda

    Researchers for small businesses, which may have no access to expensive databases or market research reports, must often rely on information found on the Internet, which can be difficult to find. Although current conventional Internet search engines are now able to index over on billion documents, there are many more documents existing in…

  11. Chemical compound navigator: a web-based chem-BLAST, chemical taxonomy-based search engine for browsing compounds.

    Science.gov (United States)

    Prasanna, M D; Vondrasek, Jiri; Wlodawer, Alexander; Rodriguez, H; Bhat, T N

    2006-06-01

    A novel technique to annotate, query, and analyze chemical compounds has been developed and is illustrated by using the inhibitor data on HIV protease-inhibitor complexes. In this method, all chemical compounds are annotated in terms of standard chemical structural fragments. These standard fragments are defined by using criteria, such as chemical classification; structural, chemical, or functional groups; and commercial, scientific or common names or synonyms. These fragments are then organized into a data tree based on their chemical substructures. Search engines have been developed to use this data tree to enable query on inhibitors of HIV protease (http://xpdb.nist.gov/hivsdb/hivsdb.html). These search engines use a new novel technique, Chemical Block Layered Alignment of Substructure Technique (Chem-BLAST) to search on the fragments of an inhibitor to look for its chemical structural neighbors. This novel technique to annotate and query compounds lays the foundation for the use of the Semantic Web concept on chemical compounds to allow end users to group, sort, and search structural neighbors accurately and efficiently. During annotation, it enables the attachment of "meaning" (i.e., semantics) to data in a manner that far exceeds the current practice of associating "metadata" with data by creating a knowledge base (or ontology) associated with compounds. Intended users of the technique are the research community and pharmaceutical industry, for which it will provide a new tool to better identify novel chemical structural neighbors to aid drug discovery.

  12. Database with web interface and search engine as a diagnostics tool for electromagnetic calorimeter

    CERN Document Server

    Paluoja, Priit

    2017-01-01

    During 2016 data collection, the Compact Muon Solenoid Data Acquisition (CMS DAQ) system has shown a very good reliability. Nevertheless, the high complexity of the hardware and the software involved is, by its nature, prone to some occasional problems. As CMS subdetector, electromagnetic calorimeter (ECAL) is affected in the same way. Some of the issues are not predictable and can appear during the year more than once such as components getting noisy, power shortcuts or failing communication between machines. The chain detection-diagnosis-intervention must be as fast as possible to minimise the downtime of the detector. The aim of this project was to create a diagnostic software for ECAL crew, which consists of database and its web interface that allows to search, add and edit the contents of the database.

  13. Top-d Rank Aggregation in Web Meta-search Engine

    Science.gov (United States)

    Fang, Qizhi; Xiao, Han; Zhu, Shanfeng

    In this paper, we consider the rank aggregation problem for information retrieval over Web making use of a kind of metric, the coherence, which considers both the normalized Kendall-τ distance and the size of overlap between two partial rankings. In general, the top-d coherence aggregation problem is defined as: given collection of partial rankings Π = {τ 1,τ 2, ⋯ , τ K }, how to find a final ranking π with specific length d, which maximizes the total coherence Φ(π,Pi)=sum_{i=1}^K Φ(π,tau_i). The corresponding complexity and algorithmic issues are discussed in this paper. Our main technical contribution is a polynomial time approximation scheme (PTAS) for a restricted top-d coherence aggregation problem.

  14. Web Project Management

    OpenAIRE

    Suralkar, Sunita; Joshi, Nilambari; Meshram, B B

    2013-01-01

    This paper describes about the need for Web project management, fundamentals of project management for web projects: what it is, why projects go wrong, and what's different about web projects. We also discuss Cost Estimation Techniques based on Size Metrics. Though Web project development is similar to traditional software development applications, the special characteristics of Web Application development requires adaption of many software engineering approaches or even development of comple...

  15. Web Project Management

    OpenAIRE

    2013-01-01

    This paper describes about the need for Web project management, fundamentals of project management for web projects: what it is, why projects go wrong, and what's different about web projects. We also discuss Cost Estimation Techniques based on Size Metrics. Though Web project development is similar to traditional software development applications, the special characteristics of Web Application development requires adaption of many software engineering approaches or even development of comple...

  16. Web Personalization Using Web Mining

    Directory of Open Access Journals (Sweden)

    Ms.Kavita D.Satokar,

    2010-03-01

    Full Text Available The information on the web is growing dramatically. The users has to spend lots of time on the web finding the information they are interested in. Today, he traditional search engines do not give users enough personalized help but provide the user with lots of irrelevant information. In this paper, we present a personalize Web searchsystem, which can helps users to get the relevant web pages based on their selection from the domain list. Thus, users can obtain a set of interested domains and the web pages from the system. The system is based on features extracted from hyperlinks, such as anchor terms or URL tokens. Our methodology uses an innovative weighted URL Rank algorithm based on user interested domains and user query.

  17. A Study about Web 3.0 based E-Learning Tools in Electrical and Electronics Engineering

    Directory of Open Access Journals (Sweden)

    ALBU Razvan-Daniel

    2014-05-01

    Full Text Available Initially designed to be a simple method of sharing documents with the whole world. Web evolved rapidly into one of the most complex nowadays technologies. At WorldWideWebsize.com we can fiind the actual size of the Web, in real-time. In my opinion, the future of Web will be strongly focused on Cloud technologies, a Web “delivered as a service” and the internet of the future will be the Internet of people, things and services. Web 1.0 is the name of the Web we had in the beginning of the 90’s. That was the period in which Web became commercial. In Web 1.0 we can observe an exponentially increase of the amount of information stored on Web sites. Nowadays, Web 3.0 seems to be the predominant technology. There is a great interest of research regarding Web based learning. The purpose of this research work was to assess the efficiency of learning using Web 3.0 technologies. This paper present also a study performed on students at different domains, from University of Oradea and explores some methodological challenges in designing a collaborative Web 3.0 learning environment. The purpose is helping instructors and trainers design an efficient collaborative Web 3.0 e-learning system that takes into account the differences of various study domains.

  18. 基于元搜索的网页去重算法%An algorithm of duplicated web pages detection based on meta-search engine

    Institute of Scientific and Technical Information of China (English)

    张玉连; 王莎莎; 宋桂江

    2011-01-01

    针对元搜索的重复网页问题,提出基于元搜索的网页去重算法,并通过实验对算法进行有效性验证.该算法首先对各成员搜索引擎返回来的结果网页的URL进行比较,然后对各结果网页的标题进行有关处理,提取出网页的主题信息,再对摘要进行分词,计算摘要的相似度,三者结合能很好的检测出重复网页,实现网页去重.该算法有效,并且比以往算法有明显的优势,更接近人工统计结果.%According to the duplicated web pages returning from meta-search engine, an algorithm of deletion of duplicated web pages based on meta-search engine is proposed.The effectiveness of the algorithm is verified through experiments.Firstly, the URL ofresult web pages is compared, which is retum by single search engines.Secondly, the titles of result web pages are processed,and thematic information of pages is extracted.Finally, the word segmentation on the summary is canied out, and the similarity of the summary is calculated.By combining these, the algorithm is able to test the duplicated web pages, realize the goal of deletion of duplicated web pages.Compared with the previous algorithms, the algorithm has obvious advantages and is closer to artificial results.

  19. Facilities Management and Digital Application of Web Engineering: Implications for Business Informatics Systems

    Directory of Open Access Journals (Sweden)

    Ezendu Ariwa

    2010-12-01

    Full Text Available A challenge that is facing businesses every day is that of how to promote themselves on the still vastly growing electronic stage. A company‟s web presence connotes a great deal to its customers, including how professional and reliable the organisation is. Companies are usually quite successful at making their websites look the part and creating a successful corporate image for themselves through their website. There are a vast number of website designers out there who can create a fantastic and aesthetically pleasing website to promote a company‟s image. But that is not all there is to a successful website. When a website is to be used as a marketing tool and additional line of income things are not so simple. The website provides the gateway to products and services as well as information required by the customer. The true knack to a successful website is to promote these as effectively as possible, and ensuring repeat visits goes hand in hand with customer loyalty.

  20. Expanding the docosahexaenoic acid food web for sustainable production: engineering lower plant pathways into higher plants.

    Science.gov (United States)

    Petrie, James R; Singh, Surinder P

    2011-01-01

    Algae are becoming an increasingly important component of land plant metabolic engineering projects. Land plants and algae have similar enough genetics to allow relatively straightforward gene transfer and they also share enough metabolic similarities that algal enzymes often function in a plant cell environment. Understanding metabolic systems in algae can provide insights into homologous systems in land plants. As examples, algal models are currently being used by several groups to better understand starch and lipid metabolism and catabolism, fields which have relevance in land plants. Importantly, land plants and algae also have enough metabolic divergence that algal genes can often provide new metabolic traits to plants. Furthermore, many algal genomes have now been sequenced, with many more in progress, and this easy access to genome-wide information has revealed that algal genomes are often relatively simple when compared with plants. One example of the importance of algal, and in particular microalgal, resources to land plant research is the metabolic engineering of long-chain polyunsaturated fatty acids into oilseed crops which typically uses microalgal genes to extend existing natural plant biosynthetic pathways. This review describes both recent progress and remaining challenges in this field.

  1. Web搜索引擎技术研究%Research on Web Search Engine Technology

    Institute of Scientific and Technical Information of China (English)

    申健; 柴艳娜

    2016-01-01

    Information in Internet is exponential growth with the development of science and technology. There should be a tool to help users to manage the big data effectively and get the useful information what they want,and locate and index information quickly and prop-erly,which is the target of search engine,and why search engine has been an essential tool in daily life. The search engine technologies are researched and their internal principle and mechanism are discussed,and their technical architecture and the information retrieval are ana-lyzed. In the working principle,the relative algorithm and strategy is studied. At the same time,the core technology and algorithm adopted by Google’ s search engine are studied and compared with the traditional technology,analyzing their superiority. In addition,the indexes and SEO the search engine working process involves are discussed respectively. It is pointed out that the information retrieval tools are im-portant for huge amounts of information processing and advanced in information retrieval,the development of which will drive the pro-gress of information science.%科技的进步导致了互联网中的信息以指数级速度增长。如何有效地管理和组织信息,帮助用户在海量的信息里获取有用的信息,并快速定位和索引,既是搜索引擎的目标,也是搜索引擎能够成为网络用户不可或缺的基础工具的原因。对搜索引擎技术进行了研究,讨论其内在原理和运行机制,分析其技术架构和信息抓取方法,并从工作原理上对其采用的算法和策略进行了分析。同时,对实际中Google搜索引擎所采用的核心技术和算法进行研究并与传统技术进行了对比,分析其所具备的先进性。另外,对搜索引擎工作流程涉及到的索引问题、SEO等都分别进行了探讨。指出信息检索工具对于海量信息数据处理的重要性,以及在信息检索方面搜索引擎体现的优越性

  2. Discussion on Web testing technology based on academic search engine websites%基于学术搜索引擎网站浅谈Web测试

    Institute of Scientific and Technical Information of China (English)

    杨小萍; 李德录; 王昱

    2011-01-01

    Web测试是软件测试的一部分,Web测试方法日新月异,基于对学术搜索引擎网站进行测试,文中主要介绍Web测试的相关知识,从功能测试、性能测试、用户界面测试3个方面进行了阐述,对Web测试常用工具分析对比,系统的探讨了Web测试的过程和常用方法.%Web testing methods, as a part of software testing, have been changing rapidly. Based on testing academic search engine websites, this article mainly introduces relevant knowledge on Web testing , including function testing, performance testing, and user interface testing. Also, it analyses and compares among general testing tools and discusses the process of Web testing and commonly used testing methods.

  3. SEMANTIC WEB-BASED SOFTWARE ENGINEERING BY AUTOMATED REQUIREMENTS ONTOLOGY GENERATION IN SOA

    Directory of Open Access Journals (Sweden)

    Vahid Rastgoo

    2014-04-01

    Full Text Available This paper presents an approach for automated generation of requirements ontology using UML diagrams in service-oriented architecture (SOA. The goal of this paper is to convenience progress of software engineering processes like software design, software reuse, service discovering and etc. The proposed method is based on a four conceptual layers. The first layer includes requirements achieved by stakeholders, the second one designs service-oriented diagrams from the data in first layer and extracts XMI codes of them. The third layer includes requirement ontology and protocol ontology to describe behavior of services and relationships between them semantically. Finally the forth layer makes standard the concepts exists in ontologies of previous layer. The generated ontology exceeds absolute domain ontology because it considers the behavior of services moreover the hierarchical relationship of them. Experimental results conducted on a set of UML4Soa diagrams in different scopes demonstrate the improvement of the proposed approach from different points of view such as: completeness of requirements ontology, automatic generation and considering SOA.

  4. 基于Web 2.0的综合搜索引擎%SYNTHESIZED SEARCH ENGINE BASED ON WEB 2.0

    Institute of Scientific and Technical Information of China (English)

    程陈; 齐开悦; 陈剑波

    2010-01-01

    Web 2.0的出现使网络中的信息量呈井喷局势,给搜索引擎带来了新的挑战,目前的搜索引擎已经不能满足大多数用户的需求.针对这种情况,首先分析了当前搜索引擎的现状和优缺点,其次针对新的情况下的用户需求作出分析,参考和利用了当前Web 2.0的一些先进技术,提出一个基于Web 2.0社区的综合搜索引擎.

  5. 用Delphi和Google Web API开发自己的搜索引擎%Creating Your Own Search Engine with Delphi and Google Web API

    Institute of Scientific and Technical Information of China (English)

    任树怀; 孙桂春

    2004-01-01

    简述Delphi是Google Web API提供的一种Web服务,允许开发人员选择自己喜爱的编程语言开发应用程序,通过基于XML的SOAP信息交换协议与远程的Google服务器连接来调用Google的Web服务;通过实例详细介绍用Delphi和Google Web API开发搜索引擎的方法和步骤.

  6. Web of Science, Scopus, and Google Scholar citation rates: a case study of medical physics and biomedical engineering: what gets cited and what doesn't?

    Science.gov (United States)

    Trapp, Jamie

    2016-12-01

    There are often differences in a publication's citation count, depending on the database accessed. Here, aspects of citation counts for medical physics and biomedical engineering papers are studied using papers published in the journal Australasian physical and engineering sciences in medicine. Comparison is made between the Web of Science, Scopus, and Google Scholar. Papers are categorised into subject matter, and citation trends are examined. It is shown that review papers as a group tend to receive more citations on average; however the highest cited individual papers are more likely to be research papers.

  7. Web服务搜索引擎的WSRank方法研究%Research on the WSRank Method for the Web Service Search Engine

    Institute of Scientific and Technical Information of China (English)

    胡蓉; 刘建勋

    2011-01-01

    Web service retrieval difficulties hampers the speed of its application and development.After realizing a web service search engine named WSSE, how to rank the services becomes the focus.The novel WSRank algorithm is advanced through analyzing the web services distribution structure and mutual relationship based on the crawling features, and learning from the famous PageRank Algorithm and its improved research achievements.The rank values are iteratively calculated, and then the web services are sorted by the values in a non-increasing order.The experiments show that the algorithm can improve the accuracy of Web service search.%Web服务检索的困难阻碍了其应用和发展的速度.在实现了一个Web服务搜索引擎WSSE后,服务的排序成为需要解决的问题.通过Web服务爬虫的爬行特点分析Web服务的分布结构和相互关系,借鉴著名的网页排序算法PageRank及其改进算法的研究成果,创新地提出WSRank算法.迭代计算各服务的排序值,按值进行非递增排序.实验表明,本算法能提高Web服务检索的准确性.

  8. W eb垂直搜索引擎实现过程的研究%Research on implementation process of Web vertical search engine

    Institute of Scientific and Technical Information of China (English)

    张弘弦; 田玉玲

    2016-01-01

    Web垂直搜索引擎是一个复杂的信息系统,目前大多数研究都集中在解决搜索引擎中出现的某一个方面的问题,仍缺乏对Web垂直搜索引擎完整实现过程的相关研究。针对这个问题,提出一种三层架构的Web垂直搜索引擎的实现过程,整个过程包含数据准备、查询处理和界面交互。使用Java语言和相关的开源工具,对实现过程描述的具体任务进行实际操作,实现了一个查询手机信息的Web垂直搜索引擎。该三层架构和实现过程有效地为构建面向主题的完整Web垂直搜索引擎提供了理论依据和实践指导。%The Web search engine is a complex information system. However,most researches are concentrated on one de⁃tailed problem appearing in a aspect of the search engine,but they lack of the correlational research on the complete implemen⁃tation process of Web vertical search engines. Aiming at this problem,the implementation process of a Web vertical search en⁃gine with a three⁃layer architecture is proposed,in which data preparation,query processing and interface interaction are con⁃tained. An actual operation of a certain task describing the implementation process was performed with Java platform and relative open⁃source tools. And by this operation,the Web vertical search engine that could query mobile information was realized. The three⁃layer architecture and implementation process provide a theoretical basis and practical guidance for building a complete subject⁃oriented Web vertical search engine.

  9. THE FLAG: A Web Resource of Innovative Assessment Tools for Faculty in College Science, Mathematics, Engineering, and Technology

    Science.gov (United States)

    Zeilik, M.; Mathieu, R. D.; National InstituteScience Education; College Level-One Team

    2000-12-01

    Even the most dedicated college faculty often discover that their students fail to learn what was taught in their courses and that much of what students do learn is quickly forgotten after the final exam. To help college faculty improve student learning in college Science, Mathematics, Engineering and Technology (SMET), the College Level - One Team of the National Institute for Science Education has created the "FLAG" a Field-tested Learning Assessment Guide for SMET faculty. Developed with funding from the National Science Foundation, the FLAG presents in guidebook format a diverse and robust collection of field-tested classroom assessment techniques (CATs), with supporting information on how to apply them in the classroom. Faculty can download the tools and techniques from the website, which also provides a goals clarifier, an assessment primer, a searchable database, and links to additional resources. The CATs and tools have been reviewed by an expert editorial board and the NISE team. These assessment strategies can help faculty improve the learning environments in their SMET courses especially the crucial introductory courses that most strongly shape students' college learning experiences. In addition, the FLAG includes the web-based Student Assessment of Learning Gains. The SALG offers a convenient way to evaluate the impact of your courses on students. It is based on findings that students' estimates of what they gained are more reliable and informative than their observations of what they liked about the course or teacher. It offers accurate feedback on how well the different aspects of teaching helped the students to learn. Students complete the SALG online after a generic template has been modified to fit the learning objectives and activities of your course. The results are presented to the teacher as summary statistics automatically. The FLAG can be found at the NISE "Innovations in SMET Education" website at www.wcer.wisc.edu/nise/cl1

  10. The Evolution of Web Searching.

    Science.gov (United States)

    Green, David

    2000-01-01

    Explores the interrelation between Web publishing and information retrieval technologies and lists new approaches to Web indexing and searching. Highlights include Web directories; search engines; portalisation; Internet service providers; browser providers; meta search engines; popularity based analysis; natural language searching; links-based…

  11. 语义Web环境下文献搜索引擎功能的研究%Research on the literature search engine function of semantic Web environment

    Institute of Scientific and Technical Information of China (English)

    袁辉; 李延香

    2013-01-01

    Based on solving the problem of when Handle user submitted the key word the traditional information retrieval technology Will contain key pages returned to the user, and contain key pages returned to users, we use demand irrelevant web page also return together with the purpose of this problem, the semantic web environment design in literature search engine search method, through the semantic web information retrieval system design and implementation of the test, it is concluded that the semantic web can provide users of the key word to logical understanding, achieve more accurate search function of the conclusion.%基于解决传统信息检索技术的搜索引擎在处理用户提交的关键字时,将含有关键字的页面返还给用户,与用户需求无关的网页也一并返还这一问题的目的,采用了在语义网环境下设计文献搜索引擎进行检索的方法,通过对语义网文献检索系统的设计与实现的试验,得出语义网可将用户提供的关键字进行逻辑理解,实现更加精准的搜索功能的结论.

  12. 基于IA和Web Service的FTP搜索引擎设计%Design of FTP Search Engine Based on IA and Web Service Technology

    Institute of Scientific and Technical Information of China (English)

    龚达

    2003-01-01

    设计了一个基于IA技术和Web Service技术的FTP搜索引擎.它在搜索服务器间实现了某种程度的peer to peer模式的分布,并可在有防火墙的环境中进行数据库的更新和任务委托;通过Spider技术实现信息的自动收集与分类整理,同时利用分词字典降低用户请求的检索粒度.

  13. El factor de impacto Web de los archivos universitarios latinoamericanos: la fiabilidad de los motores de búsqueda para calcularlo Web Impact Factor of Latin American University Archives: reliability of calculations using search engines

    Directory of Open Access Journals (Sweden)

    Isabel María Sanz Caballero

    2010-04-01

    Full Text Available El objetivo principal de la presente investigación consiste en analizar el Factor de Impacto Web (FIW de los Archivos Universitarios Latinoamericanos durante un periodo temporal de seis meses utilizando dos herramientas distintas para calcularlo: Google y Altavista. Los resultados demuestran que no existe correlación entre ambos motores a la hora de calcular el FIW de los Arhivos en cuestión, lo que permite concluir que el uso de distintas herramientas o motores utilizados para calcular el FIW interfiere en los resultados obtenidos, por lo que hay que tener cuidado al momento de establecer un ranking de calidad de los AUL, según su FIW.The main objective of this research is to analyze the Web Impact Factor (WIF of the Latin American University Archives during a six month period using two different analytical tools: Google and Altavista. The results show no correlation between the two engines when calculating the WIF of these Archives, allowing us to conclude that the use of different tools or engines to calculate the WIF affects the results, so that we must be careful when setting up a quality ranking of the LUA, according to their WIF.

  14. Moving Real Exergaming Engines on the Web: The webFitForAll Case Study in an Active and Healthy Ageing Living Lab Environment.

    Science.gov (United States)

    Konstantinidis, Evdokimos I; Bamparopoulos, Giorgos; Bamidis, Panagiotis D

    2017-05-01

    Exergames have been the subject of research and technology innovations for a number of years. Different devices and technologies have been utilized to train the body and the mind of senior people or different patient groups. In the past, we presented FitForAll, the protocol efficacy of which was proven through widely taken (controlled) pilots with more than 116 seniors for a period of two months. The current piece of work expands this and presents the first truly web exergaming platform, which is solely based on HTML5 and JavaScript without any browser plugin requirements. The adopted architecture (controller application communication framework) combines a unified solution for input devices such as MS Kinect and Wii Balance Βoard which may seamlessly be exploited through standard physical exercise protocols (American College of Sports Medicine guidelines) and accommodate high detail logging; this allows for proper pilot testing and usability evaluations in ecologically valid Living Lab environments. The latter type of setups is also used herein for evaluating the web application with more than a dozen of real elderly users following quantitative approaches.

  15. Web application to access U.S. Army Corps of Engineers Civil Works and Restoration Projects information for the Rio Grande Basin, southern Colorado, New Mexico, and Texas

    Science.gov (United States)

    Archuleta, Christy-Ann M.; Eames, Deanna R.

    2009-01-01

    The Rio Grande Civil Works and Restoration Projects Web Application, developed by the U.S. Geological Survey in cooperation with the U.S. Army Corps of Engineers (USACE) Albuquerque District, is designed to provide publicly available information through the Internet about civil works and restoration projects in the Rio Grande Basin. Since 1942, USACE Albuquerque District responsibilities have included building facilities for the U.S. Army and U.S. Air Force, providing flood protection, supplying water for power and public recreation, participating in fire remediation, protecting and restoring wetlands and other natural resources, and supporting other government agencies with engineering, contracting, and project management services. In the process of conducting this vast array of engineering work, the need arose for easily tracking the locations of and providing information about projects to stakeholders and the public. This fact sheet introduces a Web application developed to enable users to visualize locations and search for information about USACE (and some other Federal, State, and local) projects in the Rio Grande Basin in southern Colorado, New Mexico, and Texas.

  16. Exploring the academic invisible web

    OpenAIRE

    Lewandowski, Dirk; Mayr, Philipp

    2006-01-01

    Purpose: To provide a critical review of Bergman’s 2001 study on the Deep Web. In addition, we bring a new concept into the discussion, the Academic Invisible Web (AIW). We define the Academic Invisible Web as consisting of all databases and collections relevant to academia but not searchable by the general-purpose internet search engines. Indexing this part of the Invisible Web is central to scientific search engines. We provide an overview of approaches followed thus far. Design/methodol...

  17. B2C垂直搜索引擎的网络爬虫设计%Design of Web Crawlers for B2C Vertical Search Engines

    Institute of Scientific and Technical Information of China (English)

    杨亮; 刘利伟; 胡华莲

    2013-01-01

    A vertical search engine web crawler system based on Beautiful-Soup information extraction technique is developed. Experiment results indicate that the effective rate of information extraction is up to 95%and can meet commercial requirements.%  开发了基于 Beautiful-Soup 信息提取的 B2C 类垂直搜索引擎爬虫系统。测试结果表明:该爬虫的抓取有效率实际达到95%以上,满足商业应用的要求。

  18. Web services foundations

    CERN Document Server

    Bouguettaya, Athman; Daniel, Florian

    2013-01-01

    Web services and Service-Oriented Computing (SOC) have become thriving areas of academic research, joint university/industry research projects, and novel IT products on the market. SOC is the computing paradigm that uses Web services as building blocks for the engineering of composite, distributed applications out of the reusable application logic encapsulated by Web services. Web services could be considered the best-known and most standardized technology in use today for distributed computing over the Internet.Web Services Foundations is the first installment of a two-book collection coverin

  19. Excavando la web

    OpenAIRE

    Ricardo, Baeza-Yates

    2004-01-01

    The web is the internet's most important phenomenon, as demonstrated by its exponential growth and diversity. Hence, due to the volume and wealth of its data, search engines have become among the web's main tools. They are useful when we know what we are looking for. However, certainly the web holds answers to questions never imagined. The process of finding relations or interesting patterns within a data set is called "data mining" and in the case of the web, "web mining". In this article...

  20. Penerapan teknik web scraping pada mesin pencari artikel ilmiah

    OpenAIRE

    Josi, Ahmad; Abdillah, Leon Andretti; Suryayusra

    2014-01-01

    Search engines are a combination of hardware and computer software supplied by a particular company through the website which has been determined. Search engines collect information from the web through bots or web crawlers that crawls the web periodically. The process of retrieval of information from existing websites is called "web scraping." Web scraping is a technique of extracting information from websites. Web scraping is closely related to Web indexing, as for how to develop a web scra...

  1. Evaluation Method of Web Site Based on Web Structure Mining

    Institute of Scientific and Technical Information of China (English)

    LiJun-e; ZhouDong-ru

    2003-01-01

    The structure of Web site became more complex than before. During the design period of a Web site, the lack of model and method results in improper Web structure,which depend on the designer's experience. From the point of view of software engineering, every period in the software life must be evaluated before starting the next period's work. It is very important and essential to search relevant methods for evaluating Web structure before the site is completed. In this work, after studying the related work about the Web structure mining and analyzing the major structure mining methods (Page-rank and Hub/Authority), a method based on the Page-rank for Web structure evaluation in design stage is proposed. A Web structure modeling language WSML is designed, and the implement strategies for evaluating system of the Web site structure are given out. Web structure mining has being used mainly in search engines before. It is the first time to employ the Web structure mining technology to evaluate a Web structure in the design period of a Web site. It contributes to the formalization of the design documents for Web site and the improving of software engineering for large scale Web site, and the evaluating system is a practical tool for Web site construction.

  2. Research and Implementation of Web Search Engine Based on Lucene%基于Lucene的Web搜索引擎的研究和实现

    Institute of Scientific and Technical Information of China (English)

    周凤丽; 林晓丽

    2012-01-01

    Search engine has made a constant development with the development of the internet, but its gradual shifting to commercial operation makes the technical details of search engine more and more hidden. Based on research and analysis of the system structure,model and indexer of Lucene,it implements a search engine system,this system uses a non-recursive mode to take responsibility for Web crawling in the Web and distributing ,handling of URL links in the process of crawling.it manages multiple crawling threads by multi-threading technology,implements concurrently Web pages crawling and improves the system operating efficiency. And then,use JSP technology to design a simple news search engine clients. The system can run stable in line which achieves the search engine's principles and has certain significance.%互联网的快速发展也使搜索引擎不断的发展着,而搜索引擎逐渐转向商业化运行,使得搜索引擎的技术细节越来越隐蔽.文章研究和分析了搜索引擎工具Lucene的原理、模型和索引器,设计了一个搜索引擎系统.该系统采用了非递 归的方式负责Web站点的网页爬取以及爬取过程中URL链接的存储、处理等,并通过多线程技术管理多个抓取线程,实现了并发抓取网页,提高了系统的运行效率.最后采用JSP技术设计了一个简易的新闻搜索引擎客户端,系统可以稳定运行,基本符合搜索引擎原理的探索,具有一定的现实意义.

  3. Embedded Web Technology in Engineering Monitoring%嵌入式Web技术在工程安全监测领域的应用

    Institute of Scientific and Technical Information of China (English)

    孙骏; 徐骏; 邓检华

    2011-01-01

    A method for Web technology in the engineering monitoring is introduced. In the method, all kinds of functions in the engineering monitoring are designed as application programs of standard common gateway interface (CGI), accordingly each field monitoring device becomes a Web server. All the operations to the sever (the device) are carried out in the form of homepages so as to realize the mode of IE browser accessing to the server, that is, the client/server (C/S) mode. By means of this mode, the users can monitor the field situation of engineering safety through IE browser and the engineers can complete the field debugging without extra software setup.%介绍了一种Web技术在工程监测领域的应用方法,该方法将工程监测领域的监控装置的各类功能设计为标准的通用网关接口(CGI)的应用程序,使得现场监控装置成为一台Web服务器.对web服务器的所有操作都是以网页形式执行,从而实现了浏览器到服务器的访问模式,即客户/服务器(C/S)模式.通过这种访问模式,客户使用IE浏览器便能对工程安全状况进行监控,现场工程人员不需要安装专业软件便能进行工程现场的调试工作.

  4. 一种Web地图服务引擎的设计与实践%Design and Implementation of a Web Map Service Engine

    Institute of Scientific and Technical Information of China (English)

    李晓斌

    2011-01-01

    In order to overcome the shortage of slow response speed of the WebGIS currently applied, this paper designs and develops a map service engine based on tile map, and discuss the system architecture, positioning theory,grid map and business data creation, implementation of client ends Maps API etc. key technique. That engine has already been applied to the portal website and obtains good application effect.%为了解决目前WebGIS应用中存在的响应速度慢等不足,设计和开发了一种基于瓦片的Web地图服务引擎,并探讨了系统架构设计、定位原理、栅格地图制作、业务数据制作、客户端Maps API实现等关键技术.该地图服务引擎已应用于门户网站中,取得了良好的应用效果.

  5. Comparative Study of CAD Software, Web3D Technologies and Existing Solutions to Support Distance-Learning Students of Engineering Profile

    Directory of Open Access Journals (Sweden)

    Zona Kostic

    2012-07-01

    Full Text Available Interactive simulations, virtual environments and 3D interfaces significantly contribute to the process of learning through collab-oration and interaction with distant participants, and through simulation of real environments and real life situations. Depend-ing on the needs and case studies, different technologies, soft-ware packages and their combinations are used. In this paper, we present a comparative study based on the use of different CAD software in conjunction with Web3D technologies for teaching students at the Department of Engineering Management. With respect to the specificity of the profile of students undergoing training, the selection of CAD packages is reduced to the five most commonly used in this field: SolidWorks, Inventor, CATIA, Pro/ENGINEER and AutoCAD. During research, we found that Inventor and AutoCAD do not export VRML/X3D file types; hence, this paper provides only the basic functions of Inventor without integration of Web3D technologies. Combining with AutoCAD is described due to the existence of a number of plug-ins, as well as exporters and working groups working on Auto-CAD-X3D development. The main criteria for evaluating and ranking CAD packages used here are learning curve, export and import from CAD to X3D and vice versa, file types and sizes, types of nodes and material properties. We also review the tech-nology used to display 3D content on the Internet, and the rea-sons why we decided to use Web3D technologies in combination with these CAD packages. This paper presents specific conclu-sions, the advantages and disadvantages of software and technol-ogy, as well as predictions regarding further development of existing platforms and environments.

  6. RL_Spider: AN INDEPENDENT VERTICAL SEARCH ENGINE WEB CRAWLER%RL_Spider:一种自主垂直搜索引擎网络爬虫

    Institute of Scientific and Technical Information of China (English)

    黄蔚; 刘忠; 刘全

    2011-01-01

    在分析相关spider技术的基础上,提出了将强化学习技术应用到垂直搜索引擎的可控网络爬虫方法.该方法通过强化学习技术得到一些控制“经验信息”,根据这些信息来预测较远的回报,按照某一主题进行搜索,以使累积返回的回报值最大.将得到的网页存储、索引,用户通过搜索引擎的搜索接口,就可以得到最佳的搜索结果.对多个网站进行主题爬虫搜索,实验结果表明,该方法对于网络的查全率和查准率都具有较大的提高.%Based on the analysis of related spider techniques, the approach for applying reinforcement learning technology to controllable web crawler of vertical search engine is proposed in the paper. It predicts the future reward based on some control " experience information" obtained through reinforcement learning, focuses on specific topic search to maximise the accumulated returned reward value. By storing and indexing the searched web pages, users can search through search interface provided by search engine to gain the optimal search results. The topic crawler searches have been executed on various websites, experimental results show the obvious enhancement in the recall and precision of the web.

  7. Síntesis y crítica de las evaluaciones de la efectividad de los motores de búsqueda en la Web. (Synthesis and critical review of evaluations of the effectiveness of Web search engines

    Directory of Open Access Journals (Sweden)

    Francisco Javier Martínez Méndez

    2003-01-01

    Full Text Available A considerable number of proposals for measuring the effectiveness of information retrieval systems have been made since the early days of such systems. The consolidation of the World Wide Web as the paradigmatic method for developing the Information Society, and the continuous multiplication of the number of documents published in this environment, has led to the implementation of the most advanced, and extensive information retrieval systems, in the shape of web search engines. Nevertheless, there is an underlying concern about the effectiveness of these systems, especially when they usually present, in response to a question, many documents with little relevance to the users' information needs. The evaluation of these systems has been, up to now, dispersed and various. The scattering is due to the lack of uniformity in the criteria used in evaluation, and this disparity derives from their a periodicity and variable coverage. In this review, we identify three groups of studies: explicit evaluations, experimental evaluations and, more recently, several proposals for the establishment of a global framework to evaluate these systems.

  8. The Process, Dialogues, and Attitudes of Vocational Engineering High School Students in a Web Problem-Based Learning (WPBL) System

    Science.gov (United States)

    Tseng, Kuo-Hung; Chang, Chi-Cheng; Lou, Shi-Jer

    2012-01-01

    This study aims to explore how high school students collaboratively solve problems in a web problem-based learning (WPBL) system in an 8-week digital logic course using discourse analysis. Employing in-depth interviews, this study also investigated the students' attitudes toward the WPBL system. The number of teaching assistants' responses had a…

  9. Attitudes, Perceptions, and Behavioral Intentions of Engineering Workers toward Web 2.0 Tools in the Workplace

    Science.gov (United States)

    Krause, Jaclyn A.

    2010-01-01

    As Web 2.0 tools and technologies increase in popularity in consumer markets, enterprises are seeking ways to take advantage of the rich social knowledge exchanges that these tools offer. The problem this study addresses is that it remains unknown whether employees perceive that these tools offer value to the organization and therefore will be…

  10. A Situational Implementation Method for Web-based Content Management System-applications: Method Engineering and Validation in Practice

    NARCIS (Netherlands)

    Weerd, I. van de; Brinkkemper, S.; Souer, J.; Versendaal, J.M.

    2006-01-01

    The usage of data-intensive web applications raises problems concerning consistency, navigation, and data duplication. Content management systems (CMSs) can overcome these problems. In this research, we focus on special types ofweb contentmanagement systems – webbased CMS applications. Currently, no

  11. Spider webs: Damage control

    Science.gov (United States)

    Omenetto, Fiorenzo G.; Kaplan, David L.

    2012-04-01

    A study reveals that spider orb webs fail in a nonlinear fashion, owing to the hierarchical organization of the silk proteins. The discovery may serve as inspiration for engineers for the design of aerial, light-weight, robust architectures.

  12. Web数据挖掘在校园网搜索引擎系统中的应用研究%Applied Research of Web Data Mining in Search Engine System of Campus Network

    Institute of Scientific and Technical Information of China (English)

    牛凯

    2014-01-01

    在阐述了Web数据挖掘的分类、Web数据挖掘的方法和Web数据挖掘具体过程的基础上,设计了校园网搜索引擎系统的整体架构,论述了搜索引擎系统主要功能模块设计,提出了Web数据挖掘技术在校园网搜索引擎系统中的应用。%Based on the elaboration of the classification of Web Data Mining and the Method of Web Data Mining,the specific process is discussed in the paper,And the architecture of campus network search engine system,is designed the main module of Search engine system is discussed,Finally the application of Web data mining technology in campus network search engine.

  13. Study of Search Engine Transaction Logs Shows Little Change in How Users use Search Engines. A review of: Jansen, Bernard J., and Amanda Spink. “How Are We Searching the World Wide Web? A Comparison of Nine Search Engine Transaction Logs.” Information Processing & Management 42.1 (2006: 248‐263.

    Directory of Open Access Journals (Sweden)

    David Hook

    2006-09-01

    Full Text Available Objective – To examine the interactions between users and search engines, and how they have changed over time. Design – Comparative analysis of search engine transaction logs. Setting – Nine major analyses of search engine transaction logs. Subjects – Nine web search engine studies (4 European, 5 American over a seven‐year period, covering the search engines Excite, Fireball, AltaVista, BWIE and AllTheWeb. Methods – The results from individual studies are compared by year of study for percentages of single query sessions, one term queries, operator (and, or, not, etc. usage and single result page viewing. As well, the authors group the search queries into eleven different topical categories and compare how the breakdown has changed over time. Main Results – Based on the percentage of single query sessions, it does not appear that the complexity of interactions has changed significantly for either the U.S.‐based or the European‐based search engines. As well, there was little change observed in the percentage of one‐term queries over the years of study for either the U.S.‐based or the European‐based search engines. Few users (generally less than 20% use Boolean or other operators in their queries, and these percentages have remained relatively stable. One area of noticeable change is in the percentage of users viewing only one results page, which has increased over the years of study. Based on the studies of the U.S.‐based search engines, the topical categories of ‘People, Place or Things’ and ‘Commerce, Travel, Employment or Economy’ are becoming more popular, while the categories of ‘Sex and Pornography’ and ‘Entertainment or Recreation’ are declining. Conclusions – The percentage of users viewing only one results page increased during the years of the study, while the percentages of single query sessions, oneterm sessions and operator usage remained stable. The increase in single result page viewing

  14. A Research Framework of Web Search Engine Usage Mining%Web搜索引擎日志挖掘研究框架

    Institute of Scientific and Technical Information of China (English)

    王继民; 李雷明子; 孟涛

    2011-01-01

    Log files of search engines record the interactive procedure between users and the system completely. Mining the logs can help us to discover the characteristics of user behaviors and to improve the performance of search systems. This paper gives a framework on Web search engine usage mining, which includes the choice of data collections, the methods of data preprocessing, and an analysis and comparison of search behaviors from different countries. We also explore its applications on improving the effectiveness and efficiency of search engines.%搜索引擎日志记录了用户与系统交互的整个过程.对日志文件进行挖掘,可以发现用户进行Web搜索的行为特征与规律,有效改善搜索引擎系统的性能.在对国内外相关研究进行系统梳理和总结的基础上,文章提出了一个Web搜索引擎日志挖掘的研究框架,主要包括日志挖掘的研究内容、数据集的选择方法、数据预处理的方法、不同地域用户行为的特征与比较、如何应用于系统性能的改善等内容.

  15. Web Search Engines and Indexing and Ranking the Content Object Including Metadata Elements Available at the Dynamic Information Environments

    Directory of Open Access Journals (Sweden)

    Faezeh sadat Tabatabai Amiri

    2012-10-01

    Full Text Available The purpose of this research was to make exam the indexing and ranking of XML content objects containing Dublin Core and MARC 21 metadata elements in dynamic online information environments by general search engines and comparing them together in a comparative-analytical approach. 100 XML content objects in two groups were analyzed: those with DCXML elements and those with MARCXML elements were published in website http://www.marcdcmi.ir. from late Mordad 1388 till Khordad 1389. Then the website was introduced to Google and Yahoo search engines. Google search engine was able to retrieve fully all the content objects during the study period through their Dublin Core and MARC 21 metadata elements; Yahoo search engine, however, did not respond at all. The indexing of metadata elements embedded in content objects in dynamic online information environments and different between indexing and ranking of them were examined. Findings showed all Dublin Core and MARC 21 metadata elements by Google search engine were indexed. And there was not observed difference between indexing and ranking DCXML and MARCXML metadata elements in dynamic online information environments by Google search engine.

  16. Advanced web services

    CERN Document Server

    Bouguettaya, Athman; Daniel, Florian

    2013-01-01

    Web services and Service-Oriented Computing (SOC) have become thriving areas of academic research, joint university/industry research projects, and novel IT products on the market. SOC is the computing paradigm that uses Web services as building blocks for the engineering of composite, distributed applications out of the reusable application logic encapsulated by Web services. Web services could be considered the best-known and most standardized technology in use today for distributed computing over the Internet. This book is the second installment of a two-book collection covering the state-o

  17. Digging Deeper: The Deep Web.

    Science.gov (United States)

    Turner, Laura

    2001-01-01

    Focuses on the Deep Web, defined as Web content in searchable databases of the type that can be found only by direct query. Discusses the problems of indexing; inability to find information not indexed in the search engine's database; and metasearch engines. Describes 10 sites created to access online databases or directly search them. Lists ways…

  18. Digging Deeper: The Deep Web.

    Science.gov (United States)

    Turner, Laura

    2001-01-01

    Focuses on the Deep Web, defined as Web content in searchable databases of the type that can be found only by direct query. Discusses the problems of indexing; inability to find information not indexed in the search engine's database; and metasearch engines. Describes 10 sites created to access online databases or directly search them. Lists ways…

  19. Correct software in web applications and web services

    CERN Document Server

    Thalheim, Bernhard; Prinz, Andreas; Buchberger, Bruno

    2015-01-01

    The papers in this volume aim at obtaining a common understanding of the challenging research questions in web applications comprising web information systems, web services, and web interoperability; obtaining a common understanding of verification needs in web applications; achieving a common understanding of the available rigorous approaches to system development, and the cases in which they have succeeded; identifying how rigorous software engineering methods can be exploited to develop suitable web applications; and at developing a European-scale research agenda combining theory, methods a

  20. 面向道路工程Web3D档案管理的三维可视化方法%Three-dimensional visualization for road engineering Web3D archives management

    Institute of Scientific and Technical Information of China (English)

    程丹; 李伟; 彭永; 涂星宇

    2014-01-01

    Road engineering Web3D archives management is important to improve decision-making ability. The key tech-nologies including scene organization and management, integrated building and simplification method for road 3D model are studied. An organization and management method based on directed acyclic graph is proposed. Child nodes are derived gradually in depth by subdivision. Brother nodes are divided by mileage-section or sense type in breadth. Standard model nodes are shared by related father nodes. Based on constrained Delaunay triangulation, the road 3D integrated model is created. An error metric for 3D model simplification considering road constrained edges is put forward, in which, the geometric error for every edge and an additional error for constrained edge are included. The constrained edges can be refined preferentially. Benefit from above methods a road engineering Web3D archives management system is developed out, in which, road 3D model quickly view-dependent reconstruction is realized and constrained edges are retained on the client. This system has been applied to many highways construction and operating management.%道路工程Web3D档案管理技术对辅助提高决策能力具有重要意义。对其中的海量数据组织管理、三维整体建模及模型简化方法进行了研究。提出了基于有向无环图的三维场景组织管理方法,深度方向从总体到细节的逐步细分,广度方向按里程分段或模型类型进行划分,标准模型节点为多个父节点所共享;基于约束Delaunay三角网建立了道路三维整体模型;提出了顾及约束的三维模型简化误差度量准则,考虑了对任意边进行简化时的几何误差,并为约束边建立了一个附加误差项,保证约束边尽早细化。基于上述方法开发了道路工程Web3D档案管理系统,远程客户端可快速重构出三维模型并较好地保留了约束边,目前已应用在多条高速公路的建设、运营管理中。

  1. Searching for information on the World Wide Web with a search engine: a pilot study on cognitive flexibility in younger and older users.

    Science.gov (United States)

    Dommes, Aurelie; Chevalier, Aline; Rossetti, Marilyne

    2010-04-01

    This pilot study investigated the age-related differences in searching for information on the World Wide Web with a search engine. 11 older adults (6 men, 5 women; M age=59 yr., SD=2.76, range=55-65 yr.) and 12 younger adults (2 men, 10 women; M=23.7 yr., SD=1.07, range=22-25 yr.) had to conduct six searches differing in complexity, and for which a search method was or was not induced. The results showed that the younger and older participants provided with an induced search method were less flexible than the others and produced fewer new keywords. Moreover, older participants took longer than the younger adults, especially in the complex searches. The younger participants were flexible in the first request and spontaneously produced new keywords (spontaneous flexibility), whereas the older participants only produced new keywords when confronted by impasses (reactive flexibility). Aging may influence web searches, especially the nature of keywords used.

  2. 一种基于知识工程的DeepWeb信息抽取方法%A DeepWeb Information Extraction Method Based on Knowledge Engineering

    Institute of Scientific and Technical Information of China (English)

    乌尔柯西; 杨抒; 王业; 游香薷

    2016-01-01

    The information contained in DeepWeb is more and more huge with great value. But due to the factors that DeepWeb informa-tion is highly heterogeneous, autonomous, dynamic and incomplete, and the design style, page structure, display contents of DeepWeb theme website are different,and the use of JavaScript technology is widespread,the traditional extraction technology can’ t be effectively automated integration of high quality information contained in DeepWeb resources. Presents a DeepWeb extraction method based on knowledge engineering. The page mode,HTML structure and visual features of DeepWeb are analyzed and integrated. Applies HTML DOM tree parsing algorithm to match the template accorded with page mode,HTML structure and object information source by automatic or semi-automatic way,locating the information in DeepWeb to obtain the free text,structured and semi-structured data. Using a large number of site data with nested structure as data source,the effectiveness of extraction method is verified.%DeepWeb中蕴含的信息越发庞大并且价值可观。但是由于DeepWeb信息的高度异构性、自主性、动态性以及不完整性,DeepWeb主题性网站的设计风格、页面结构、显示内容的不同,JavaScript技术的广泛使用等因素,使传统的抽取技术无法有效自动化集成蕴含在DeepWeb中的高质量信息资源。提出一种基于KBE(知识工程)的DeepWeb信息抽取方法。通过对目标DeepWeb的页面模式、页面HTML结构、页面视觉信息等进行分析、整合,利用HTML DOM ( Document Object Model)树解析算法,自动或半自动方式匹配出符合页面模式、页面HTML结构以及目标信息源等元素的模板,来对Deep-Web中的信息进行定位,从而得到页面中的自由文本,结构化和半结构化数据。实验以大量嵌套结构的网站数据作为数据来源,验证了抽取方法的有效性。

  3. RESTful web services with Dropwizard

    CERN Document Server

    Dallas, Alexandros

    2014-01-01

    A hands-on focused step-by-step tutorial to help you create Web Service applications using Dropwizard. If you are a software engineer or a web developer and want to learn more about building your own Web Service application, then this is the book for you. Basic knowledge of Java and RESTful Web Service concepts is assumed and familiarity with SQL/MySQL and command-line scripting would be helpful.

  4. Construction of web-based nutrition education contents and searching engine for usage of healthy menu of children.

    Science.gov (United States)

    Hong, Soon-Myung; Lee, Tae-Kyong; Chung, Hea-Jung; Park, Hye-Kyung; Lee, Eun-Ju; Nam, Hye-Seon; Jung, Soon-Im; Cho, Jee-Ye; Lee, Jin-Hee; Kim, Gon; Kim, Min-Chan

    2008-01-01

    A diet habit, which is developed in childhood, lasts for a life time. In this sense, nutrition education and early exposure to healthy menus in childhood is important. Children these days have easy access to the internet. Thus, a web-based nutrition education program for children is an effective tool for nutrition education of children. This site provides the material of the nutrition education for children with characters which are personified nutrients. The 151 menus are stored in the site together with video script of the cooking process. The menus are classified by the criteria based on age, menu type and the ethnic origin of the menu. The site provides a search function. There are three kinds of search conditions which are key words, menu type and "between" expression of nutrients such as calorie and other nutrients. The site is developed with the operating system Windows 2003 Server, the web server ZEUS 5, development language JSP, and database management system Oracle 10 g.

  5. Indexing and Retrieval for the Web.

    Science.gov (United States)

    Rasmussen, Edie M.

    2003-01-01

    Explores current research on indexing and ranking as retrieval functions of search engines on the Web. Highlights include measuring search engine stability; evaluation of Web indexing and retrieval; Web crawlers; hyperlinks for indexing and ranking; ranking for metasearch; document structure; citation indexing; relevance; query evaluation;…

  6. Hidden Page WebCrawler Model for Secure Web Pages

    Directory of Open Access Journals (Sweden)

    K. F. Bharati

    2013-03-01

    Full Text Available The traditional search engines available over the internet are dynamic in searching the relevant content over the web. The search engine has got some constraints like getting the data asked from a varied source, where the data relevancy is exceptional. The web crawlers are designed only to more towards a specific path of the web and are restricted in moving towards a different path as they are secured or at times restricted due to the apprehension of threats. It is possible to design a web crawler that will have the capability of penetrating through the paths of the web, not reachable by the traditional web crawlers, in order to get a better solution in terms of data, time and relevancy for the given search query. The paper makes use of a newer parser and indexer for coming out with a novel idea of web crawler and a framework to support it. The proposed web crawler is designed to attend Hyper Text Transfer Protocol Secure (HTTPS based websites and web pages that needs authentication to view and index. User has to fill a search form and his/her creditionals will be used by the web crawler to attend secure web server for authentication. Once it is indexed the secure web server will be inside the web crawler’s accessible zone

  7. Extracting Macroscopic Information from Web Links.

    Science.gov (United States)

    Thelwall, Mike

    2001-01-01

    Discussion of Web-based link analysis focuses on an evaluation of Ingversen's proposed external Web Impact Factor for the original use of the Web, namely the interlinking of academic research. Studies relationships between academic hyperlinks and research activities for British universities and discusses the use of search engines for Web link…

  8. Research and design of vertical search engine for DCI based on web%基于Web的DCI垂直搜索引擎的研究与设计

    Institute of Scientific and Technical Information of China (English)

    吴洁明; 冀单单; 韩云辉

    2013-01-01

    为了解决用户能够快速、准确的搜索互联网上数字作品信息的问题,分析设计了一个对数字作品版权唯一标识符(Digital Copyright Identifier简称DCI)数字作品的垂直搜索引擎.首先基于Heritrix网络爬虫技术,对互联网上的数字作品进行数据采集和正文信息抽取,并将抽取的数据保存到本地;然后基于Lucene的全文检索工具包,对本地数据进行分词、倒排索引、索引检索和改进的相关度排序等处理,最终设计实现了一个通用可扩展的DCI垂直搜索引擎.实验结果表明,该搜索引擎在很大程度上提高了网页信息抽取的准确度和数据的检索效率.%In order to solve the users' problem for searching digital works information quickly and correctly, a vertical search engine about digital work's Copyright Identifier is analyzed and designed. In the first place, based on the Heritrix web crawler, the network digital work's data acquisition and text information extraction are presented and the extracted data is saved to the local; In the second place, on the basis of the Lucene's full-text retrieval toolkit, segmentation, inverted index, index retrieval and improved sorting algorithm technology are taken to handle the collected data. a general and extensible DCI vertical search engine is designed and achieved. The experimenal results show that this search engine does enhance web page information extraction accuracy and data indexing efficiency in great degree.

  9. Semantic Oriented Agent based Approach towards Engineering Data Management, Web Information Retrieval and User System Communication Problems

    CERN Document Server

    Ahmed, Zeeshan

    2010-01-01

    The four intensive problems to the software rose by the software industry .i.e., User System Communication / Human Machine Interface, Meta Data extraction, Information processing & management and Data representation are discussed in this research paper. To contribute in the field we have proposed and described an intelligent semantic oriented agent based search engine including the concepts of intelligent graphical user interface, natural language based information processing, data management and data reconstruction for the final user end information representation.

  10. Adaptable Web Modules to Stimulate Active Learning in Engineering Hydrology using Data and Model Simulations of Three Regional Hydrologic Systems

    Science.gov (United States)

    Habib, E. H.; Tarboton, D. G.; Lall, U.; Bodin, M.; Rahill-Marier, B.; Chimmula, S.; Meselhe, E. A.; Ali, A.; Williams, D.; Ma, Y.

    2013-12-01

    The hydrologic community has long recognized the need for broad reform in hydrologic education. A paradigm shift is critically sought in undergraduate hydrology and water resource education by adopting context-rich, student-centered, and active learning strategies. Hydrologists currently deal with intricate issues rooted in complex natural ecosystems containing a multitude of interconnected processes. Advances in the multi-disciplinary field include observational settings such as Critical Zone and Water, Sustainability and Climate Observatories, Hydrologic Information Systems, instrumentation and modeling methods. These research advances theory and practices call for similar efforts and improvements in hydrologic education. The typical, text-book based approach in hydrologic education has focused on specific applications and/or unit processes associated with the hydrologic cycle with idealizations, rather than the contextual relations in the physical processes and the spatial and temporal dynamics connecting climate and ecosystems. An appreciation of the natural variability of these processes will lead to graduates with the ability to develop independent learning skills and understanding. This appreciation cannot be gained in curricula where field components such as observational and experimental data are deficient. These types of data are also critical when using simulation models to create environments that support this type of learning. Additional sources of observations in conjunction with models and field data are key to students understanding of the challenges associated with using models to represent such complex systems. Recent advances in scientific visualization and web-based technologies provide new opportunities for the development of active learning techniques utilizing ongoing research. The overall goal of the current study is to develop visual, case-based, data and simulation driven learning experiences to instructors and students through a web

  11. APLIKASI WEB CRAWLER UNTUK WEB CONTENT PADA MOBILE PHONE

    Directory of Open Access Journals (Sweden)

    Sarwosri Sarwosri

    2009-01-01

    Full Text Available Crawling is the process behind a search engine, which served through the World Wide Web in a structured and with certain ethics. Applications that run the crawling process is called Web Crawler, also called web spider or web robot. The growth of mobile search services provider, followed by growth of a web crawler that can browse web pages in mobile content type. Crawler Web applications can be accessed by mobile devices and only web pages that type Mobile Content to be explored is the Web Crawler. Web Crawler duty is to collect a number of Mobile Content. A mobile application functions as a search application that will use the results from the Web Crawler. Crawler Web server consists of the Servlet, Mobile Content Filter and datastore. Servlet is a gateway connection between the client with the server. Datastore is the storage media crawling results. Mobile Content Filter selects a web page, only the appropriate web pages for mobile devices or with mobile content that will be forwarded.

  12. Characteristics of scientific web publications

    DEFF Research Database (Denmark)

    Thorlund Jepsen, Erik; Seiden, Piet; Ingwersen, Peter Emil Rerup

    2004-01-01

    Because of the increasing presence of scientific publications on the Web, combined with the existing difficulties in easily verifying and retrieving these publications, research on techniques and methods for retrieval of scientific Web publications is called for. In this article, we report on the......Vista and AllTheWeb retrieved a higher degree of accessible scientific content than Google. Because of the search engine cutoffs of accessible URLs, the feasibility of using search engine output for Web content analysis is also discussed....

  13. 基于DC元数据的Web网络搜索引擎系统设计%Design about WEB Search Engine System Based on DC Metadata

    Institute of Scientific and Technical Information of China (English)

    伯琼; 胡飞; 钟国祥

    2011-01-01

    传统搜索引擎通常靠抓取全文关键词进行分析,由此带来三大缺陷:缺乏语义描述导致查准率低;检索结果冗余模糊导致检索效率低;检索途径不足.基于DC元数据描述网络资源的优越性,课题组设计了一个基于DC元数据的网络搜索引擎系统DCSE,力图克服传统搜索引擎的上述缺陷.DCSE系统自动抓取含DC描述的Web网页,把DC描述信息存入到数据库,排序索引后提供用户检索.检索界面设计成以15个DC元素为检索项的多项逻辑组合检索,检索结果以各DC元素的描述内容来显示,如标题、创建者、描述、日期等.用户通过多项组合检索提高查准率,并通过清晰的结果显示对所需信息做出快速判断、选择,从而达到提高检索效率的目的.%The traditional search engines analyze key words that are snatched in the full text, leading to three flaws:lowly precision ratio; bad search efficiency, lacking search channel. Based on advantage of describing Web resource by DC metadata, the seminar designed a DCSE system in order to overcome flaws of traditional search engine. DCSE grasped automatically WEB page and storaged DC describe information to database that supplied for user following index. Search interface was designed as multiple logical combinations with various retrieval based on 15 DC elements, the retrieval results are displayed with discribled content of DC element, such as title, creator, description, date , etc. The user can raise the precision ratio by multiterm assemble search and can select rapidly needed information by display clearly search result,thereby raise retrieval efficiency.

  14. Applying Web Usage Mining for Personalizing Hyperlinks in Web-Based Adaptive Educational Systems

    Science.gov (United States)

    Romero, Cristobal; Ventura, Sebastian; Zafra, Amelia; de Bra, Paul

    2009-01-01

    Nowadays, the application of Web mining techniques in e-learning and Web-based adaptive educational systems is increasing exponentially. In this paper, we propose an advanced architecture for a personalization system to facilitate Web mining. A specific Web mining tool is developed and a recommender engine is integrated into the AHA! system in…

  15. OneWeb: web content adaptation platform based on W3C Mobile Web Initiative guidelines

    Directory of Open Access Journals (Sweden)

    Francisco O. Martínez P.

    2011-01-01

    Full Text Available  Restrictions regardingnavigability and user-friendliness are the main challenges the Mobile Web faces to be accepted worldwide. W3C has recently developed the Mobile Web Initiative (MWI, a set of directives for the suitable design and presentation of mobile Web interfaces. This article presents the main features and functional modules of OneWeb, an MWI-based Web content adaptation platform developed by Mobile Devices Applications Development Interest Group’s  (W@PColombia research activities, forming part of the Universidad de Cauca’s Telematics Engineering Group.Some performance measurementresults and comparison with other Web content adaptation platforms are presented. Tests have shown suitable response times for Mobile Web environments; MWI guidelines were applied to over twenty Web pages selected for testing purposes.  

  16. An Efficient Web Page Ranking for Semantic Web

    Science.gov (United States)

    Chahal, P.; Singh, M.; Kumar, S.

    2014-01-01

    With the enormous amount of information presented on the web, the retrieval of relevant information has become a serious problem and is also the topic of research for last few years. The most common tools to retrieve information from web are search engines like Google. The Search engines are usually based on keyword searching and indexing of web pages. This approach is not very efficient as the result-set of web pages obtained include large irrelevant pages. Sometimes even the entire result-set may contain lot of irrelevant pages for the user. The next generation of search engines must address this problem. Recently, many semantic web search engines have been developed like Ontolook, Swoogle, which help in searching meaningful documents presented on semantic web. In this process the ranking of the retrieved web pages is very crucial. Some attempts have been made in ranking of semantic web pages but still the ranking of these semantic web documents is neither satisfactory and nor up to the user's expectations. In this paper we have proposed a semantic web based document ranking scheme that relies not only on the keywords but also on the conceptual instances present between the keywords. As a result only the relevant page will be on the top of the result-set of searched web pages. We explore all relevant relations between the keywords exploring the user's intention and then calculate the fraction of these relations on each web page to determine their relevance. We have found that this ranking technique gives better results than those by the prevailing methods.

  17. Exploring the academic invisible web

    OpenAIRE

    Lewandowski, Dirk

    2006-01-01

    The Invisible Web is often discussed in the academic context, where its contents (mainly in the form of databases) are of great importance. But this discussion is mainly based on some seminal research done by Sherman and Price (2001) and Bergman (2001), respectively. We focus on the types of Invisible Web content relevant for academics and the improvements made by search engines to deal with these content types. In addition, we question the volume of the Invisible Web as stated by Bergman. Ou...

  18. Software engineering

    CERN Document Server

    Sommerville, Ian

    2010-01-01

    The ninth edition of Software Engineering presents a broad perspective of software engineering, focusing on the processes and techniques fundamental to the creation of reliable, software systems. Increased coverage of agile methods and software reuse, along with coverage of 'traditional' plan-driven software engineering, gives readers the most up-to-date view of the field currently available. Practical case studies, a full set of easy-to-access supplements, and extensive web resources make teaching the course easier than ever.

  19. Distributed Deep Web Search

    NARCIS (Netherlands)

    Tjin-Kam-Jet, Kien

    2013-01-01

    The World Wide Web contains billions of documents (and counting); hence, it is likely that some document will contain the answer or content you are searching for. While major search engines like Bing and Google often manage to return relevant results to your query, there are plenty of situations in

  20. Yahoo! Cataloging the Web.

    Science.gov (United States)

    Callery, Anne

    The Internet has the potential to be the ultimate information resource, but it needs to be organized in order to be useful. This paper discusses how the subject guide, "Yahoo!" is different from most web search engines, and how best to search for information on Yahoo! The strength in Yahoo! lies in the subject hierarchy. Advantages to…

  1. Distributed deep web search

    NARCIS (Netherlands)

    Tjin-Kam-Jet, Kien-Tsoi Theodorus Egbert

    2013-01-01

    The World Wide Web contains billions of documents (and counting); hence, it is likely that some document will contain the answer or content you are searching for. While major search engines like Bing and Google often manage to return relevant results to your query, there are plenty of situations in

  2. CRISPy-web

    DEFF Research Database (Denmark)

    Blin, Kai; Pedersen, Lasse Ebdrup; Weber, Tilmann

    2016-01-01

    CRISPR/Cas9-based genome editing has been one of the major achievements of molecular biology, allowing the targeted engineering of a wide range of genomes. The system originally evolved in prokaryotes as an adaptive immune system against bacteriophage infections. It now sees widespread application...... designing sgRNAs for non-model organisms exist. Here, we present CRISPy-web (http://crispy.secondarymetabolites.org/), an easy to use web tool based on CRISPy to design sgRNAs for any user-provided microbial genome. CRISPy-web allows researchers to interactively select a region of their genome of interest...... to scan for possible sgRNAs. After checks for potential off-target matches, the resulting sgRNA sequences are displayed graphically and can be exported to text files. All steps and information are accessible from a web browser without the requirement to install and use command line scripts....

  3. Survey of Techniques for Deep Web Source Selection and Surfacing the Hidden Web Content

    OpenAIRE

    Khushboo Khurana; M.B. Chandak

    2016-01-01

    Large and continuously growing dynamic web content has created new opportunities for large-scale data analysis in the recent years. There is huge amount of information that the traditional web crawlers cannot access, since they use link analysis technique by which only the surface web can be accessed. Traditional search engine crawlers require the web pages to be linked to other pages via hyperlinks causing large amount of web data to be hidden from the crawlers. Enormous data is available in...

  4. Survey of Techniques for Deep Web Source Selection and Surfacing the Hidden Web Content

    OpenAIRE

    Khushboo Khurana; M B Chandak

    2016-01-01

    Large and continuously growing dynamic web content has created new opportunities for large-scale data analysis in the recent years. There is huge amount of information that the traditional web crawlers cannot access, since they use link analysis technique by which only the surface web can be accessed. Traditional search engine crawlers require the web pages to be linked to other pages via hyperlinks causing large amount of web data to be hidden from the crawlers. Enormous data is available in...

  5. Collection Selection for Distributed Web Search

    NARCIS (Netherlands)

    Bockting, S.

    2009-01-01

    Current popular web search engines, such as Google, Live Search and Yahoo!, rely on crawling to build an index of the World Wide Web. Crawling is a continuous process to keep the index fresh and generates an enormous amount of data traffic. By far the largest part of the web remains unindexed, becau

  6. Personalized Spiders for Web Search and Analysis.

    Science.gov (United States)

    Chau, Michael; Zeng, Daniel; Chen, Hsinchun

    Searching for useful information on the World Wide Web has become increasingly difficult. While Internet search engines have been helping people to search on the Web, low recall rate and outdated indexes have become more and more problematic as the Web grows. In addition, search tools usually present to the user only a list of search results,…

  7. Evaluation Method of Web Site Structure Based on Web Structure Mining

    Institute of Scientific and Technical Information of China (English)

    Li Jun-e; Zhou Dong-ru

    2003-01-01

    The structure of Web site hecarne more complex titan before. During the design period of a Web site, the lack of model and method results in improper Web structure,which depend on the designer's experience. From the point of view of software engineering, every period in the software life must be evaluated before starting the next period's work. It is very important and essential to search relevant methods for evaluating Web structure before the site is completed. In this work, after studying the related work about the Web struc lure mining and analyzing the major structure mining methods (Page-rank and Hub/Authority), a method based on the Page-rank for Web structure evaluation in design stage is proposecL A Web structure modeling language WSML is designed, and the implement strategies for evaluating system of the Web site structure are given out. Web structure mining has being used mainly in search engines before. It is the first time to employ the Web structure mining technology to evaluate a Web structure in the design period of a Web site. It contributes to the formalization of the design documents for Web site and the improving of software engineering for large scale Web site, and the evaluating system is a practical tool for Web site construction.

  8. Modeling of Web-based engineering camouflage feature information resource platform%基于Web的工程伪装特征信息资源平台建模

    Institute of Scientific and Technical Information of China (English)

    王志东; 林伟; 叶立超; 刘志

    2013-01-01

    Such characteristics of multifarious content, various format and scattered distribution make it difficult for engineering camouflage feature data to be unified description, centralized stored and quickly searched. To solve these problems, the paper constructed a web-based engineering camouflage feature information resource platform. Relying on the platform, such resources can be overall administrated and fully utilized. With formalized method,the platform model was described, and multi-view structures were constructed by means of software engineering. The platform software prototype of the mass of the engineering camouflage data resource management was realized by the M-L algorithm and Web-Tree scheme. The feasibility of the model and high performance of the algorithm technique were validated by the enormous test data.%为了解决工程伪装特征信息资源内容庞杂、形式多样、分布零散,给统一描述、集中存储和快速检索带来的困难,构建了基于Web的工程伪装特征信息资源管理平台,实现了此类资源的集中统一管理和高效利用.运用形式化方法对平台模型进行了描述,运用软件工程化方法对平台结构进行了建模.在此基础上,基于M-L算法和Web-Tree组件技术实现了平台海量伪装特征信息资源管理的实验原型.通过大量测试数据验证了平台模型的可行性、算法和技术方案的高效性.

  9. DESIGNING OF SIMULATION FOR ENGINE ROOM KM. SINABUNG WITH CONTROL MONITORING WEB SERVER BASED BY WIRELESS NETWORK AND POWER LINE COMMUNICATION

    Directory of Open Access Journals (Sweden)

    Eko Sasmito Hadi

    2013-06-01

    Full Text Available Engine room monitoring control system is monitoring and controlling main engine and auxiliary engine from long distance by powerline communication network and wireless network to ease the operator in operating the ship and save operational cost. To prevent error in programming the main engine and auxiliary engine, a simulation using instrument software is needed to know the machine characteristic. After simulation result fulfills the requirement which is approached the value of test record, it can be applied to the real machine. In this study, some  steps were done such as: getting know type of main engine and auxiliary engine which performance will be simulated and programmed, getting test record of main engine and auxiliary engine, getting know how the main engine and auxiliary engine work, making simulation of main engine and auxiliary engine work system, doing monitoring control simulation by powerline   communication and wireless network, comparing the results between simulation and test record of main engine and auxiliary engine. Engine programming can be set after simulation result meets the requirement. Result of simulation fulfills the requirement, the difference values between machine simulation using software instrument and test record of main engine and auxiliary engine are around 1%-2%. If engine room monitoring control system by wireless and powerline communication is applied in the ship, the ship owner will get advantages because it will prolong ship durability and can monitor operasional of main engine or Auxiliary engine from long distance ; while the operator will be easier in operating the ship. The disadvantage is only the higher cost

  10. 主题搜索引擎中网络蜘蛛搜索策略的研究%Study of Search Strategy in Topic -oriented Web Spider for Topic- driven Search Engine

    Institute of Scientific and Technical Information of China (English)

    王明国; 胡敬仓

    2011-01-01

    主题网络蜘蛛的搜索策略是主题搜索引擎的核心部分,是近年来主题搜索引擎研究中的热点问题之一.深入研究了主题网络蜘蛛的关键技术,阐述了多线程网络蜘蛛的实现过程,并对传统的VSM算法和PageRank算法进行了改进,提高了主题网络蜘蛛采集信息的有效性和准确性.%The search strategy in topic -oriented web spider is the key component of topic -driven search engine, and is really hot in research in recent years. Based on in - depth research in the search strategy in topic - driven search engine, this article expounds the development processes of the multi -thread web spider, at the same time, improves the traditional VSM algorithm and the PageRank algorithm, and enhances the validation and accuracy of information collection for topic - oriented web spider.

  11. Introduction to Webometrics Quantitative Web Research for the Social Sciences

    CERN Document Server

    Thelwall, Michael

    2009-01-01

    Webometrics is concerned with measuring aspects of the web: web sites, web pages, parts of web pages, words in web pages, hyperlinks, web search engine results. The importance of the web itself as a communication medium and for hosting an increasingly wide array of documents, from journal articles to holiday brochures, needs no introduction. Given this huge and easily accessible source of information, there are limitless possibilities for measuring or counting on a huge scale (e.g., the number of web sites, the number of web pages, the number of blogs) or on a smaller scale (e.g., the number o

  12. Dark Web

    CERN Document Server

    Chen, Hsinchun

    2012-01-01

    The University of Arizona Artificial Intelligence Lab (AI Lab) Dark Web project is a long-term scientific research program that aims to study and understand the international terrorism (Jihadist) phenomena via a computational, data-centric approach. We aim to collect "ALL" web content generated by international terrorist groups, including web sites, forums, chat rooms, blogs, social networking sites, videos, virtual world, etc. We have developed various multilingual data mining, text mining, and web mining techniques to perform link analysis, content analysis, web metrics (technical

  13. Using Google App Engine

    CERN Document Server

    Severance, Charles

    2009-01-01

    Build exciting, scalable web applications quickly and confidently using Google App Engine and this book, even if you have little or no experience in programming or web development. App Engine is perhaps the most appealing web technology to appear in the last year, providing an easy-to-use application framework with basic web tools. While Google's own tutorial assumes significant experience, Using Google App Engine will help anyone get started with this platform. By the end of this book, you'll know how to build complete, interactive applications and deploy them to the cloud using the same s

  14. Modeling and Analyze the Deep Web: Surfacing Hidden Value

    OpenAIRE

    Suneet Kumar; Anuj Kumar Yadav; Rakesh Bharati; Rani Choudhary

    2011-01-01

    Focused web crawlers have recently emerged as an alternative to the well-established web search engines. While the well-known focused crawlers retrieve relevant web-pages, there are various applications which target whole websites instead of single web-pages. For example, companies are represented by websites, not by individual web-pages. To answer queries targeted at Websites, web directories are an established solution. In this paper, we introduce a novel focused website crawler to employ t...

  15. Accurate And Efficient Crawling The Deep Web: Surfacing Hidden Value

    OpenAIRE

    Suneet Kumar; Anuj Kumar Yadav; Rakesh Bharti; Rani Choudhary

    2011-01-01

    Searching Focused web crawlers have recently emerged as an alternative to the well-established web search engines. While the well-known focused crawlers retrieve relevant web-pages, there are various applications which target whole websites instead of single web-pages. For example, companies are represented by websites, not by individual web-pages. To answer queries targeted at websites, web directories are an established solution. In this paper, we introduce a novel focused website crawler t...

  16. Factsheets Web Application

    Energy Technology Data Exchange (ETDEWEB)

    VIGIL,FRANK; REEDER,ROXANA G.

    2000-10-30

    The Factsheets web application was conceived out of the requirement to create, update, publish, and maintain a web site with dynamic research and development (R and D) content. Before creating the site, a requirements discovery process was done in order to accurately capture the purpose and functionality of the site. One of the high priority requirements for the site would be that no specialized training in web page authoring would be necessary. All functions of uploading, creation, and editing of factsheets needed to be accomplished by entering data directly into web form screens generated by the application. Another important requirement of the site was to allow for access to the factsheet web pages and data via the internal Sandia Restricted Network and Sandia Open Network based on the status of the input data. Important to the owners of the web site would be to allow the published factsheets to be accessible to all personnel within the department whether or not the sheets had completed the formal Review and Approval (R and A) process. Once the factsheets had gone through the formal review and approval process, they could then be published both internally and externally based on their individual publication status. An extended requirement and feature of the site would be to provide a keyword search capability to search through the factsheets. Also, since the site currently resides on both the internal and external networks, it would need to be registered with the Sandia search engines in order to allow access to the content of the site by the search engines. To date, all of the above requirements and features have been created and implemented in the Factsheet web application. These have been accomplished by the use of flat text databases, which are discussed in greater detail later in this paper.

  17. Location-based Web Search

    Science.gov (United States)

    Ahlers, Dirk; Boll, Susanne

    In recent years, the relation of Web information to a physical location has gained much attention. However, Web content today often carries only an implicit relation to a location. In this chapter, we present a novel location-based search engine that automatically derives spatial context from unstructured Web resources and allows for location-based search: our focused crawler applies heuristics to crawl and analyze Web pages that have a high probability of carrying a spatial relation to a certain region or place; the location extractor identifies the actual location information from the pages; our indexer assigns a geo-context to the pages and makes them available for a later spatial Web search. We illustrate the usage of our spatial Web search for location-based applications that provide information not only right-in-time but also right-on-the-spot.

  18. Keeping Dublin Core Simple: Cross-Domain Discovery or Resource Description?; First Steps in an Information Commerce Economy: Digital Rights Management in the Emerging E-Book Environment; Interoperability: Digital Rights Management and the Emerging EBook Environment; Searching the Deep Web: Direct Query Engine Applications at the Department of Energy.

    Science.gov (United States)

    Lagoze, Carl; Neylon, Eamonn; Mooney, Stephen; Warnick, Walter L.; Scott, R. L.; Spence, Karen J.; Johnson, Lorrie A.; Allen, Valerie S.; Lederman, Abe

    2001-01-01

    Includes four articles that discuss Dublin Core metadata, digital rights management and electronic books, including interoperability; and directed query engines, a type of search engine designed to access resources on the deep Web that is being used at the Department of Energy. (LRW)

  19. Keeping Dublin Core Simple: Cross-Domain Discovery or Resource Description?; First Steps in an Information Commerce Economy: Digital Rights Management in the Emerging E-Book Environment; Interoperability: Digital Rights Management and the Emerging EBook Environment; Searching the Deep Web: Direct Query Engine Applications at the Department of Energy.

    Science.gov (United States)

    Lagoze, Carl; Neylon, Eamonn; Mooney, Stephen; Warnick, Walter L.; Scott, R. L.; Spence, Karen J.; Johnson, Lorrie A.; Allen, Valerie S.; Lederman, Abe

    2001-01-01

    Includes four articles that discuss Dublin Core metadata, digital rights management and electronic books, including interoperability; and directed query engines, a type of search engine designed to access resources on the deep Web that is being used at the Department of Energy. (LRW)

  20. Establishment of Satisfaction Model and Evaluation Criteria System For Web Search Engine%Web搜索引擎满意度模型与评价指标体系构建

    Institute of Scientific and Technical Information of China (English)

    叶凤云; 汪传雷

    2011-01-01

    Based on the ACSI, this paper not only establishes Web search engine satisfaction model, but also establishes evaluation criteria system with the past search engine evaluation criteria system and the upper model. It can provide references for evaluating Web search engine satisfaction and creating the basement to verify the model fully.%在ACSI(美国客户满意度指数)模型基础上,构建Web搜索引擎满意度(简称WSES)模型。同时,依据已有的搜索引擎评价指标体系,结合所构建的WSES模型,建立相应的测量指标体系,为进一步进行结构方程模型的验证分析建立基础,并为评价Web搜索引擎满意度提供参考。

  1. Study on Web usage mining in search engine of university library%高校图书馆搜索引擎中Web使用记录挖掘研究

    Institute of Scientific and Technical Information of China (English)

    赵静

    2013-01-01

    针对高校的信息资源检索的命中率低问题提出了运用Web使用记录挖掘的高校图书馆搜索引擎.通过采用Web使用记录挖掘技术和Clementine对高校图书馆网站的Web访问日志记录进行挖掘.在Web使用记录挖掘流程中,提出一个基于用户IP、登陆时间、网站的拓扑图、引用网页和Agent采识别出单个用户的新算法,获得有效提高识别用户的实验结果.最后用路径分析来挖掘模式,优化网站结构,从而提高高校图书馆搜索引擎的命中率.%Because the hit rate of university information resource retrieval is low, the university library search engine applying Web usage mining is put forward. Through Web usage mining technology and Clementine, Web access log record of uni-versity library website was excavated. In the process of Web usage mining, a new algorithm that identifies individual users is pro-posed based on user IP, log time, site topological graph, cited webpage and Agent, so as to improve the effect of user recognition. The path analysis is used to excavate the pattern and optimize the website structure, so that the hit rate of university library search engine is raised up.

  2. Rendimiento de los sistemas de recuperación de información en la web: evalución de servicios de búsqueda (search engines.

    Directory of Open Access Journals (Sweden)

    Olvera Lobo, María Dolores

    2000-09-01

    Full Text Available Ten search engines, Altavista, Excite, Hotbot, Infoseek, Lycos. Magellan, OpenText, WebCrawler, WWWWorm, Yahoo, were evaluated, by means of a questionnaire with 20 items (adding up to a total of 200 questions. The 20 first results for each question were analysed in terms of relevance, and values of precision and recall were computed for the resulting 4000 references. The results are also analyzed in terms of the type of question (boolean or natural language and topic (specialized vs. general interest. The results showed that Excite, Infoseek and AltaVista performed generally better. The conclusion of this methodological trial was that the method used allows the evaluation of the performance of Information Retrieval Systems in the Web. As for the results, web search engines are not very precise but extremely exhaustive.

    Se han evaluado diez servicios de búsqueda: Altavista, Excite, Hotbot, Infoseek, Lycos, Magellan, OpenText, WebCrawler, WWWWorm, Yahoo. Se formularon 20 preguntas a cada uno de los 10 sistemas evaluados por lo que se realizaron 200 consultas. Además, se examinó la relevancia de los primeros 20 resultados de cada consulta lo que significa que, en total, se revisaron aproximadamente 4.000 referencias, para cada una de las cuales se calcularon los valores de precisión y exhaustividad. Los análisis muestran que Excite, Infoseek y Altavista son los tres servicios que, de forma genérica, muestran mejor rendimiento. Se analizan también los resultados en función del tipo de pregunta (booleanas o de frase y del tema (ocio o especializada. Se concluye que el método empleado permite analizar el rendimiento de los SRI de la W3 y que los resultados ponen de manifiesto que los buscadores no son sistemas de recuperación de información muy precisos aunque sí muy exhaustivos.

  3. Search Engine Optimization

    CERN Document Server

    Davis, Harold

    2006-01-01

    SEO--short for Search Engine Optimization--is the art, craft, and science of driving web traffic to web sites. Web traffic is food, drink, and oxygen--in short, life itself--to any web-based business. Whether your web site depends on broad, general traffic, or high-quality, targeted traffic, this PDF has the tools and information you need to draw more traffic to your site. You'll learn how to effectively use PageRank (and Google itself); how to get listed, get links, and get syndicated; and much more. The field of SEO is expanding into all the possible ways of promoting web traffic. This

  4. Python for Google app engine

    CERN Document Server

    Pippi, Massimiliano

    2015-01-01

    If you are a Python developer, whether you have experience in web applications development or not, and want to rapidly deploy a scalable backend service or a modern web application on Google App Engine, then this book is for you.

  5. 基于超链接引导和链接图分析的主题搜索引擎%Research on Focused Search Engine Based on Hyperlink Induced and Web Structure

    Institute of Scientific and Technical Information of China (English)

    唐苏; 刘循

    2011-01-01

    Focused search engine is a tool designed to query information on a particular subject or theme information. Considering the advantages and disadvantages of current focused search engine technologies, put forward the IPageRank - IND algorithm that combining the hyperlink - induced technology based on text-inspired with the PageRank algorithm based on web structure analysis to improve the accuracy of relativity judgment and the coverage of focused resources research, and classifying the web page by sub-topic in order to retrieve efficiently. Then, experiment with a search engine to build,to compare the algorithm with several other algorithms,see the advantage of IPageRank-IND algorithm is obvious.%主题搜索引擎足专为查询某一学科或主题信息而出现的查询工具.针对目前各种主题搜索引擎在主题搜索上的优缺点,提出将基于文字内容启发的超链接引导技术与基于 Web 链接图的 PageRank 算法相结合的 IPageRank-IND 算法,以提高链接相关度判断的准确性和主题资源搜索的覆盖率,并将网页按照 VSM 算法进行内容相关度判断和自动分类,从而提高检索效率.最后构建一个搜索引擎进行实验,通过比较该算法与其他几种算法的实验结果,能够看到 IPageRank-IND 算法的优势是明显的.

  6. 维、哈、柯多文种搜索引擎中web文本分类的研究%Study on Web Document Classification of Uyghur, Kazak, Kirgiz Multi-lingual Search Engine

    Institute of Scientific and Technical Information of China (English)

    海丽且木·艾沙; 维尼拉·木沙江

    2011-01-01

    研究维、哈、柯多文种搜索引擎中web文本分类问题.根据维、哈、柯Web文本具有结构信息的特点,提出分类系统框架,采用基于改进的KNN的Web文本分类方法,并结合具体实验在对数据进行预处理的基础上实现了改进的KNN分类算法.实验表明,改进的KNN方法在维吾尔文Web文本分类中能够获得较好的分类效果.%This paper studies the problems of Uyghur, Kazak, Kirgiz Multi-lingual search engine web document classification. According to the structure of Uyghur, Kazak, Kirgiz web text, a system frame is provided, a Web document classification algorithm based on improved KNN method presented , based on data preprocess, improved KNN method is implemented. The result of research indicates that the impact of the new method is better.

  7. Development of Content Management System-based Web Applications

    NARCIS (Netherlands)

    Souer, J.

    2012-01-01

    Web engineering is the application of systematic and quantifiable approaches (concepts, methods, techniques, tools) to cost-effective requirements analysis, design, implementation, testing, operation, and maintenance of high quality web applications. Over the past years, Content Management Systems (

  8. Language engineering for the Semantic Web: a digital library for endangered languages. Endangered languages, Ontology, Digital library, Multimedia, EMELD, Intelligent querying and retrieval, ImageSpace

    Directory of Open Access Journals (Sweden)

    Lu Shiyong

    2004-01-01

    Full Text Available In this paper, we describe the effort undertaken at Wayne State University to preserve endangered languages using the state-of-the-art information technologies. In particular, we discuss the issues involved in such an effort, and present the architecture of a distributed digital library for endangered languages which will contain various data of endangered languages in the forms of text, image, video, audio and include advanced tools for intelligent cataloguing, indexing, searching and browsing information on languages and language analysis. We use various Semantic Web technologies such as XML, OLAC, ontologies so that our digital library becomes a useful linguistic resource on the Semantic Web.

  9. A Novel Method for Bilingual Web Page Mining Via Search Engines%基于搜索引擎的双语混合网页识别新方法

    Institute of Scientific and Technical Information of China (English)

    冯艳卉; 洪宇; 颜振祥; 姚建民; 朱巧明

    2011-01-01

    A new approach has been developed for acquiring bilingual web pages from the result pages of search engines, which is composed of two challenging tasks. The first task is to detect web records embedded in the result pages automatically via a clustering method of a sample page. Identifying these useful records through the clustering method allows the generation of highly effective features for the next task which is high-quality bilingual web page acquisition. The task of high-quality bilingual web page acquisition is assumed as a classification problem. One advantage of our approach is that it is independent of the search engine and the domain. The test is based on 2 516 records extracted from six search engines automatically and annotated manually, which gets a high precision of 81.3% and a recall of 94.93%. The experimental results indicate that our approach is very effective.%该文提出了一种从搜索引擎返回的结果网页中获取双语网页的新方法,该方法分为两个任务.第一个任务是自动地检测并收集搜索引擎返回的结果网页中的数据记录.该步骤通过聚类的方法识别出有用的记录摘要并且为下一个任务即高质量双语混合网页的验证及其荻取提供有效特征.该文中把双语混合网页的验证看作是有效的分类问题,该方法不依赖于特定领域和搜索引擎.基于从搜索引擎收集并经过人工标注的2516务检索结果记录,该文提出的方法取得了81.3%的精确率和94.93%的召回率.

  10. 基于Web数据挖掘的个性化搜索引擎的应用和发展趋势%Application and Development Trend of Personalized Search Engine Based on Web Data Mining

    Institute of Scientific and Technical Information of China (English)

    王丽; 曹家琏

    2009-01-01

    Web数据挖掘是将数据挖掘技术和理论应用于对www资源进行挖掘的一个新兴的研究领域.论述Web数据挖掘的发展现状、发展趋势以及将来可能的研究方向.并简单介绍个性化搜索引擎的一些情况,论述Web数据挖掘在个性化搜索引擎中的应用.%Web data mining is a new developing research field in which data mining technology and application of the theory apply to be-ing excavated to www resources. Describe current situation of the development, development trend and possible research direction in thefuture that Web data excavate, and introduce some situations of the individualized search engine briefly, Elaborates the application of Webdata mining in the individualized search engine.

  11. Borderless Geospatial Web (bolegweb)

    Science.gov (United States)

    Cetl, V.; Kliment, T.; Kliment, M.

    2016-06-01

    The effective access and use of geospatial information (GI) resources acquires a critical value of importance in modern knowledge based society. Standard web services defined by Open Geospatial Consortium (OGC) are frequently used within the implementations of spatial data infrastructures (SDIs) to facilitate discovery and use of geospatial data. This data is stored in databases located in a layer, called the invisible web, thus are ignored by search engines. SDI uses a catalogue (discovery) service for the web as a gateway to the GI world through the metadata defined by ISO standards, which are structurally diverse to OGC metadata. Therefore, a crosswalk needs to be implemented to bridge the OGC resources discovered on mainstream web with those documented by metadata in an SDI to enrich its information extent. A public global wide and user friendly portal of OGC resources available on the web ensures and enhances the use of GI within a multidisciplinary context and bridges the geospatial web from the end-user perspective, thus opens its borders to everybody. Project "Crosswalking the layers of geospatial information resources to enable a borderless geospatial web" with the acronym BOLEGWEB is ongoing as a postdoctoral research project at the Faculty of Geodesy, University of Zagreb in Croatia (http://bolegweb.geof.unizg.hr/). The research leading to the results of the project has received funding from the European Union Seventh Framework Programme (FP7 2007-2013) under Marie Curie FP7-PEOPLE-2011-COFUND. The project started in the November 2014 and is planned to be finished by the end of 2016. This paper provides an overview of the project, research questions and methodology, so far achieved results and future steps.

  12. Resource Selection for Federated Search on the Web

    OpenAIRE

    Nguyen, Dong Van; Demeester, Thomas; Trieschnigg, Dolf; Hiemstra, Djoerd

    2016-01-01

    A publicly available dataset for federated search reflecting a real web environment has long been absent, making it difficult for researchers to test the validity of their federated search algorithms for the web setting. We present several experiments and analyses on resource selection on the web using a recently released test collection containing the results from more than a hundred real search engines, ranging from large general web search engines such as Google, Bing and Yahoo to small do...

  13. Reviews Equipment: Data logger Book: Imagined Worlds Equipment: Mini data loggers Equipment: PICAXE-18M2 data logger Books: Engineering: A Very Short Introduction and To Engineer Is Human Book: Soap, Science, & Flat-Screen TVs Equipment: uLog and SensorLab Web Watch

    Science.gov (United States)

    2012-07-01

    WE RECOMMEND Data logger Fourier NOVA LINK: data logging and analysis To Engineer is Human Engineering: essays and insights Soap, Science, & Flat-Screen TVs People, politics, business and science overlap uLog sensors and sensor adapter A new addition to the LogIT range offers simplicity and ease of use WORTH A LOOK Imagined Worlds Socio-scientific predictions for the future Mini light data logger and mini temperature data logger Small-scale equipment for schools SensorLab Plus LogIT's supporting software, with extra features HANDLE WITH CARE CAXE110P PICAXE-18M2 data logger Data logger 'on view' but disappoints Engineering: A Very Short Introduction A broad-brush treatment fails to satisfy WEB WATCH Two very different websites for students: advanced physics questions answered and a more general BBC science resource

  14. Self-Regulated Learning Skills and Online Activities between Higher and Lower Performers on a Web-Intensive Undergraduate Engineering Course

    Science.gov (United States)

    Lawanto, Oenardi; Santoso, Harry B.; Lawanto, Kevin N.; Goodridge, Wade

    2017-01-01

    The objective of this study was to evaluate students' self-regulated learning (SRL) skills used in a Web-intensive learning environment. The research question guiding the study was: How did the use of student SRL skills and student engagement in online activities compare between higher- and lower-performing students participating in a…

  15. Setting Engineering Students up for Success in the 21st Century: Integrating Gamification and Crowdsourcing into a CDIO-Based Web Design Course

    Science.gov (United States)

    Song, Donglei; Tavares, Adriano; Pinto, Sandro; Xu, Hao

    2017-01-01

    Over the past few decades, many researchers have tested course designs that may better engage students in developing countries, accommodate for Millennials' desires to learn and teach at will, and teach students the skills they need for their first jobs. The vision of this paper for a web design course seeks to address these issues for engineering…

  16. Performance Support Engineering: Building Performance-Centered Web-based Systems, Information Systems, and Knowledge Management Systems in the 21st Century.

    Science.gov (United States)

    Raybould, Barry

    2000-01-01

    Examines the growth in developing performance-centered systems in business. Discusses Web-based systems, including the Internet and intranets; knowledge management systems; knowledge acquisition; performance-centered design; performance support; group processes; systems approach; focus on goals; electronic performance support systems;…

  17. Web Classification Using DYN FP Algorithm

    Directory of Open Access Journals (Sweden)

    Bhanu Pratap Singh

    2014-01-01

    Full Text Available Web mining is the application of data mining techniques to extract knowledge from Web. Web mining has been explored to a vast degree and different techniques have been proposed for a variety of applications that includes Web Search, Classification and Personalization etc. The primary goal of the web site is to provide the relevant information to the users. Web mining technique is used to categorize users and pages by analyzing users behavior, the content of pages and order of URLs accessed. In this paper, proposes an auto-classification algorithm of web pages using data mining techniques. The problem of discovering association rules between terms in a set of web pages belonging to a category in a search engine database, and present an auto – classification algorithm for solving this problem that are fundamentally based on FP-growth algorithm

  18. Web Page Recommendation Using Web Mining

    Directory of Open Access Journals (Sweden)

    Modraj Bhavsar

    2014-07-01

    Full Text Available On World Wide Web various kind of content are generated in huge amount, so to give relevant result to user web recommendation become important part of web application. On web different kind of web recommendation are made available to user every day that includes Image, Video, Audio, query suggestion and web page. In this paper we are aiming at providing framework for web page recommendation. 1 First we describe the basics of web mining, types of web mining. 2 Details of each web mining technique.3We propose the architecture for the personalized web page recommendation.

  19. Pro JavaScript for web apps

    CERN Document Server

    Freeman, Adam

    2012-01-01

    JavaScript is the engine behind every web app, and a solid knowledge of it is essential for all modern web developers. Pro JavaScript for Web Apps gives you all of the information that you need to create professional, optimized, and efficient JavaScript applications that will run across all devices. It takes you through all aspects of modern JavaScript application creation, showing you how to combine JavaScript with the new features of HTML5 and CSS3 to make the most of the new web technologies. The focus of the book is on creating professional web applications, ensuring that your app provides

  20. Survey of Technologies for Web Application Development

    CERN Document Server

    Doyle, Barry

    2008-01-01

    Web-based application developers face a dizzying array of platforms, languages, frameworks and technical artifacts to choose from. We survey, classify, and compare technologies supporting Web application development. The classification is based on (1) foundational technologies; (2)integration with other information sources; and (3) dynamic content generation. We further survey and classify software engineering techniques and tools that have been adopted from traditional programming into Web programming. We conclude that, although the infrastructure problems of the Web have largely been solved, the cacophony of technologies for Web-based applications reflects the lack of a solid model tailored for this domain.

  1. Developing a Consensus-Driven, Core Competency Model to Shape Future Audio Engineering Technology Curriculum: A Web-Based Modified Delphi Study

    Science.gov (United States)

    Tough, David T.

    2009-01-01

    The purpose of this online study was to create a ranking of essential core competencies and technologies required by AET (audio engineering technology) programs 10 years in the future. The study was designed to facilitate curriculum development and improvement in the rapidly expanding number of small to medium sized audio engineering technology…

  2. Capturing, Sharing, and Discovering Product Data at a Semantic Level--Moving Forward to the Semantic Web for Advancing the Engineering Product Design Process

    Science.gov (United States)

    Zhu, Lijuan

    2011-01-01

    Along with the greater productivity that CAD automation provides nowadays, the product data of engineering applications needs to be shared and managed efficiently to gain a competitive edge for the engineering product design. However, exchanging and sharing the heterogeneous product data is still challenging. This dissertation first presents a…

  3. Capturing, Sharing, and Discovering Product Data at a Semantic Level--Moving Forward to the Semantic Web for Advancing the Engineering Product Design Process

    Science.gov (United States)

    Zhu, Lijuan

    2011-01-01

    Along with the greater productivity that CAD automation provides nowadays, the product data of engineering applications needs to be shared and managed efficiently to gain a competitive edge for the engineering product design. However, exchanging and sharing the heterogeneous product data is still challenging. This dissertation first presents a…

  4. Semantic web services for web databases

    CERN Document Server

    Ouzzani, Mourad

    2011-01-01

    Semantic Web Services for Web Databases introduces an end-to-end framework for querying Web databases using novel Web service querying techniques. This includes a detailed framework for the query infrastructure for Web databases and services. Case studies are covered in the last section of this book. Semantic Web Services For Web Databases is designed for practitioners and researchers focused on service-oriented computing and Web databases.

  5. Improved Relevance Ranking in WebGather

    Institute of Scientific and Technical Information of China (English)

    LEI Ming; WANG Jianyong; CHEN Baojue; LI Xiaoming

    2001-01-01

    The amount of information on the web is growing rapidly, and search engines that rely on keyword matching usually return too many low quality matches. To improve search results, a challenging task for search engines is how to effectively calculate a relevance ranking for each web page. This paper discusses in what order a search engine should return the URLs it has produced in response to a user's query, so as to show more relevant pages first.Emphasis is given on the ranking functions adopted by WebGather that take link structure and user popularity factors into account. Experimental results are also presented to evaluate the proposed strategy.

  6. Sensor web

    Science.gov (United States)

    Delin, Kevin A. (Inventor); Jackson, Shannon P. (Inventor)

    2011-01-01

    A Sensor Web formed of a number of different sensor pods. Each of the sensor pods include a clock which is synchronized with a master clock so that all of the sensor pods in the Web have a synchronized clock. The synchronization is carried out by first using a coarse synchronization which takes less power, and subsequently carrying out a fine synchronization to make a fine sync of all the pods on the Web. After the synchronization, the pods ping their neighbors to determine which pods are listening and responded, and then only listen during time slots corresponding to those pods which respond.

  7. Advanced Metasearch Engine Technology

    CERN Document Server

    Meng, Weiyi

    2010-01-01

    Among the search tools currently on the Web, search engines are the most well known thanks to the popularity of major search engines such as Google and Yahoo!. While extremely successful, these major search engines do have serious limitations. This book introduces large-scale metasearch engine technology, which has the potential to overcome the limitations of the major search engines. Essentially, a metasearch engine is a search system that supports unified access to multiple existing search engines by passing the queries it receives to its component search engines and aggregating the returned

  8. Search Engine Design and Implementation of Web Data Mining Based on Concept Clustering%基于概念聚类的Web数据挖掘搜索引擎的设计与实现

    Institute of Scientific and Technical Information of China (English)

    刘典型; 刘完芳; 钟钢

    2015-01-01

    For the search of the web data mining,its accuracy depends to the numbers of keywords that user has inputted very much,as well as the agreement of user’s intent and the semantic analysis by search engine,the analysis of the search engine including clustering method based on link and based on concept.In this paper,to overcome the defect of the clustering method based on link,through using the method of clustering based on concept,starting from the concept and the storage method of bipartite graph,designed and implemented a personalized search engine of web data mining,its superiority is verified.%针对Web数据挖掘的搜索过程,其准确度很大程度取决于用户输入的关键词的数量,以及搜索引擎对关键词的语义的解析与用户原意的吻合度,而搜索引擎对关键词的解析,包括基于链接的聚类方法和基于概念的聚类方法。本文克服基于链接的聚类方法的缺陷,采用基于概念聚类的方法,从二分图的概念和存储方法入手,设计和实现了个性化的Web数据挖掘搜索引擎,并验证了其优越性。

  9. Curriculum Optimization of Web Design and Making Basing on Search Engine Optimization%基于搜索引擎优化的《网页设计与制作》课程改革

    Institute of Scientific and Technical Information of China (English)

    关晓惠; 周志敏

    2013-01-01

      Now more and more companies begin to pay attention to economic benefits of their website .They will demand more talented persons engaged web design and making .The curriculum of Web Design and Making is a core professional course of computer specialty.It is used to train students’ methods and techniques of designing and making websites .In or-der to improve the ranking of website in search engines of Google or Baidu , this course should be reformed on the idea of search engine optimization.In this paper, existing problems of the course are analyze and approaches on how to optimize processing of implementation in the light of search engine optimization are also propose .Finally teaching mode of JAVA curriculum is also explored.%  《网页设计与制作》作为计算机专业的核心专业课程,不但要培养学生的网页设计与制作的方法和技巧,还要依据搜索引擎优化的理念来设计和制作网页,以提高网站或网页在各类搜索引擎中的排名。通过分析《网页设计与制作》课程教学中存在的问题,并根据搜索引擎优化的技术优化该课程内容,采用三阶段递进的教学模式,以培养出符合社会需求的网页设计与制作专业人才。

  10. Automating Information Discovery Within the Invisible Web

    Science.gov (United States)

    Sweeney, Edwina; Curran, Kevin; Xie, Ermai

    A Web crawler or spider crawls through the Web looking for pages to index, and when it locates a new page it passes the page on to an indexer. The indexer identifies links, keywords, and other content and stores these within its database. This database is searched by entering keywords through an interface and suitable Web pages are returned in a results page in the form of hyperlinks accompanied by short descriptions. The Web, however, is increasingly moving away from being a collection of documents to a multidimensional repository for sounds, images, audio, and other formats. This is leading to a situation where certain parts of the Web are invisible or hidden. The term known as the "Deep Web" has emerged to refer to the mass of information that can be accessed via the Web but cannot be indexed by conventional search engines. The concept of the Deep Web makes searches quite complex for search engines. Google states that the claim that conventional search engines cannot find such documents as PDFs, Word, PowerPoint, Excel, or any non-HTML page is not fully accurate and steps have been taken to address this problem by implementing procedures to search items such as academic publications, news, blogs, videos, books, and real-time information. However, Google still only provides access to a fraction of the Deep Web. This chapter explores the Deep Web and the current tools available in accessing it.

  11. The Semantic Web: opportunities and challenges for next-generation Web applications

    Directory of Open Access Journals (Sweden)

    2002-01-01

    Full Text Available Recently there has been a growing interest in the investigation and development of the next generation web - the Semantic Web. While most of the current forms of web content are designed to be presented to humans, but are barely understandable by computers, the content of the Semantic Web is structured in a semantic way so that it is meaningful to computers as well as to humans. In this paper, we report a survey of recent research on the Semantic Web. In particular, we present the opportunities that this revolution will bring to us: web-services, agent-based distributed computing, semantics-based web search engines, and semantics-based digital libraries. We also discuss the technical and cultural challenges of realizing the Semantic Web: the development of ontologies, formal semantics of Semantic Web languages, and trust and proof models. We hope that this will shed some light on the direction of future work on this field.

  12. Engineering Adaptive Applications

    DEFF Research Database (Denmark)

    Dolog, Peter

    . The different requirements might be satisfied by different variants of features maintained and provided by Web applications. An adaptive Web application can be seen as a family of Web applications where application instances are those generated for particular user based on his characteristics relevant...... for a domain.In this book, we propose a new domain engineering framework which extends a development process of Web applications with techniques required when designing such adaptive customizable Web applications. The framework is provided with design abstractions which deal separately with information served...

  13. Database selection and result merging in P2P web search

    NARCIS (Netherlands)

    Chernov, S.; Serdyukov, P.; Bender, M.; Michel, S.; Weikum, G.; Zimmer, C.

    2005-01-01

    Intelligent Web search engines are extremely popular now. Currently, only the commercial centralized search engines like Google can process terabytes of Web data. Alternative search engines fulfilling collaborative Web search on a voluntary basis are usually based on a blooming Peer-to-Peer (P2P) te

  14. A Survey on Web Crawling%Web信息采集研究进展

    Institute of Scientific and Technical Information of China (English)

    李盛韬; 余智华; 程学旗; 白硕

    2003-01-01

    As a basic component of search engine and a series of other services on Web,Web crawler is playing an important role. Roughly,a Web crawler is a program which automatically traverses the Web by downloading documents and following links from page to page. This article detailedly explains the principles and difficulties on the Web crawler,comprehensively argues several hot directions of Web crawler,and at last views the new direction of Web crawler.

  15. BORDERLESS GEOSPATIAL WEB (BOLEGWEB

    Directory of Open Access Journals (Sweden)

    V. Cetl

    2016-06-01

    Full Text Available The effective access and use of geospatial information (GI resources acquires a critical value of importance in modern knowledge based society. Standard web services defined by Open Geospatial Consortium (OGC are frequently used within the implementations of spatial data infrastructures (SDIs to facilitate discovery and use of geospatial data. This data is stored in databases located in a layer, called the invisible web, thus are ignored by search engines. SDI uses a catalogue (discovery service for the web as a gateway to the GI world through the metadata defined by ISO standards, which are structurally diverse to OGC metadata. Therefore, a crosswalk needs to be implemented to bridge the OGC resources discovered on mainstream web with those documented by metadata in an SDI to enrich its information extent. A public global wide and user friendly portal of OGC resources available on the web ensures and enhances the use of GI within a multidisciplinary context and bridges the geospatial web from the end-user perspective, thus opens its borders to everybody. Project “Crosswalking the layers of geospatial information resources to enable a borderless geospatial web” with the acronym BOLEGWEB is ongoing as a postdoctoral research project at the Faculty of Geodesy, University of Zagreb in Croatia (http://bolegweb.geof.unizg.hr/. The research leading to the results of the project has received funding from the European Union Seventh Framework Programme (FP7 2007-2013 under Marie Curie FP7-PEOPLE-2011-COFUND. The project started in the November 2014 and is planned to be finished by the end of 2016. This paper provides an overview of the project, research questions and methodology, so far achieved results and future steps.

  16. Automatic generation of Web mining environments

    Science.gov (United States)

    Cibelli, Maurizio; Costagliola, Gennaro

    1999-02-01

    The main problem related to the retrieval of information from the world wide web is the enormous number of unstructured documents and resources, i.e., the difficulty of locating and tracking appropriate sources. This paper presents a web mining environment (WME), which is capable of finding, extracting and structuring information related to a particular domain from web documents, using general purpose indices. The WME architecture includes a web engine filter (WEF), to sort and reduce the answer set returned by a web engine, a data source pre-processor (DSP), which processes html layout cues in order to collect and qualify page segments, and a heuristic-based information extraction system (HIES), to finally retrieve the required data. Furthermore, we present a web mining environment generator, WMEG, that allows naive users to generate a WME specific to a given domain by providing a set of specifications.

  17. Multilabel Learning for Automatic Web Services Tagging

    OpenAIRE

    Mustapha AZNAG; Mohamed QUAFAFOU; Jarir, Zahi

    2014-01-01

    Recently, some web services portals and search engines as Biocatalogue and Seekda!, have allowed users to manually annotate Web services using tags. User Tags provide meaningful descriptions of services and allow users to index and organize their contents. Tagging technique is widely used to annotate objects in Web 2.0 applications. In this paper we propose a novel probabilistic topic model (which extends the CorrLDA model - Correspondence Latent Dirichlet Allocation-) to automatically tag we...

  18. Bridging data mining and semantic web

    OpenAIRE

    Aman, Edris

    2016-01-01

    Nowadays Semantic Web is widely adopted standard of knowledge representation. Hence, knowledge engineers are applying sophisticated methods to capture, discover and represent knowledge in Semantic Web form. Studies show that, to represent knowledge in Semantic Web standard, data mining techniques such as Decision Trees, Association Rules, etc., play an important role. These techniques are implemented in publicly available Data Mining tools. These tools represent knowledge discovered in human ...

  19. WEB BASED TRANSLATION OF CHINESE ORGANIZATION NAME

    Institute of Scientific and Technical Information of China (English)

    Yang Muyun; Liu Daxin; Zhao Tiejun; Qi Haoliang; Lin Kaiming

    2009-01-01

    A web-based translation method for Chinese organization name is proposed. After analyzing the structure of Chinese organization name, the methods of bilingual query formulation and maximum entropy based translation re-ranking are suggested to retrieve the English translation from the web via public search engine. The experiments on Chinese university names demonstrate the validness of this approach.

  20. A neural click model for web search

    NARCIS (Netherlands)

    Borisov, A.; Markov, I.; de Rijke, M.; Serdyukov, P.

    2016-01-01

    Understanding user browsing behavior in web search is key to improving web search effectiveness. Many click models have been proposed to explain or predict user clicks on search engine results. They are based on the probabilistic graphical model (PGM) framework, in which user behavior is represented

  1. Getting To Know the "Invisible Web."

    Science.gov (United States)

    Smith, C. Brian

    2001-01-01

    Discusses the portions of the World Wide Web that cannot be accessed via directories or search engines, explains why they can't be accessed, and offers suggestions for reference librarians to find these sites. Lists helpful resources and gives examples of invisible Web sites which are often databases. (LRW)

  2. Social Networking on the Semantic Web

    Science.gov (United States)

    Finin, Tim; Ding, Li; Zhou, Lina; Joshi, Anupam

    2005-01-01

    Purpose: Aims to investigate the way that the semantic web is being used to represent and process social network information. Design/methodology/approach: The Swoogle semantic web search engine was used to construct several large data sets of Resource Description Framework (RDF) documents with social network information that were encoded using the…

  3. Faster and Efficient Web Crawling with Parallel Migrating Web Crawler

    Directory of Open Access Journals (Sweden)

    Akansha Singh

    2010-05-01

    Full Text Available A Web crawler is a module of a search engine that fetches data from various servers. Web crawlers are an essential component to search engines; running a web crawler is a challenging task. It is a time-taking process to gather data from various sources around the world. Such a single process faces limitations on the processing power of a single machine and one network connection. This module demands much processing power and network consumption. This paper aims at designing and implementing such a parallel migrating crawler in which the work of a crawler is divided amongst a number of independent and parallel crawlers which migrate to different machines to improve network efficiency and speed up the downloading. The migration and parallel working of the proposed design was experimented and the results were recorded.

  4. Computational origami environment on the web

    Institute of Scientific and Technical Information of China (English)

    Asem KASEM; Tetsuo IDA

    2008-01-01

    We present a computing environment for ori-gami on the web. The environment consists of the compu-tational origami engine Eos for origami construction, visualization, and geometrical reasoning, WEвEOS for pro-viding web interface to the functionalities of Eos, and web service system SCORUM for symbolic computing web ser-vices. WEBEOS is developed using Web2.0 technologies, and provides a graphical interactive web interface for ori-gami construction and proving. In SCORUM, we are prepar-ing web services for a wide range of symbolic computing systems, and are using these services in our origami envir-onment. We explain the functionalities of this environment, and discuss its architectural and technological features.

  5. A Reverse Engineering Method of Web Software System Based on Data Stream Technology%一种基于数据流的网络软件系统逆向工程方法

    Institute of Scientific and Technical Information of China (English)

    伍智平

    2011-01-01

    Due to analytical data of web software system quickly generate, one of the biggest challenges facing the outlier detection is how to get results of reverse engineering by processing the analytical data rapidly. Based on analysis of existing research results, the analytical data by using the view of data stream are designed and a algorithm of Web software system reverse engineering based on priority and quadruples is provided solutions. The theo retical analysis and experiments show that algorithms proposed have higher precision and efficiency.%由于网络软件系统中分析数据具有快速产生的特征,网络软件系统逆向工程的最大挑战在于如何及时高效地处理这些数据,从而得到逆向工程的结果.在分析现有研究成果的基础上,对分析数据进行基于数据流的建模,从而能够快速高效地处理大中型网络软件系统中的数据,同时设计了基于优先级与数据流四元组的网络软件系统逆向工程算法,通过不同类型网站的实验,验证了算法的性能与匹配度和其他同类算法相比都有较大的提高.

  6. Engineering Adaptive Applications

    DEFF Research Database (Denmark)

    Dolog, Peter

    for a domain.In this book, we propose a new domain engineering framework which extends a development process of Web applications with techniques required when designing such adaptive customizable Web applications. The framework is provided with design abstractions which deal separately with information served...

  7. Ada & the Analytical Engine.

    Science.gov (United States)

    Freeman, Elisabeth

    1996-01-01

    Presents a brief history of Ada Byron King, Countess of Lovelace, focusing on her primary role in the development of the Analytical Engine--the world's first computer. Describes the Ada Project (TAP), a centralized World Wide Web site that serves as a clearinghouse for information related to women in computing, and provides a Web address for…

  8. Developing an Internet Oriented Platform for Earthquake Engineering Application and Web-based Virtual Reality Simulation System for Seismic hazards: Towards Disaster Mitigation in Metropolises

    Directory of Open Access Journals (Sweden)

    Ali Alaghehbandian

    2003-04-01

    Full Text Available This paper reviews the state of the art on risk communication to the public, with an emphasis on simulation of seismic hazards using VRML. Rapid growth computer technologies, especially the Internet provide human beings new measures to deal with engineering and social problems which were hard to solve in traditional ways. This paper presents a prototype of an application platform based on the Internet using VR (Virtual Reality for civil engineering considering building an information system of risk communication for seismic hazards and at the moment in the case of bridge structure.

  9. A combined strategy of "in silico" transcriptome analysis and web search engine optimization allows an agile identification of reference genes suitable for normalization in gene expression studies.

    Science.gov (United States)

    Faccioli, Primetta; Ciceri, Gian Paolo; Provero, Paolo; Stanca, Antonio Michele; Morcia, Caterina; Terzi, Valeria

    2007-03-01

    Traditionally housekeeping genes have been employed as endogenous reference (internal control) genes for normalization in gene expression studies. Since the utilization of single housekeepers cannot assure an unbiased result, new normalization methods involving multiple housekeeping genes and normalizing using their mean expression have been recently proposed. Moreover, since a gold standard gene suitable for every experimental condition does not exist, it is also necessary to validate the expression stability of every putative control gene on the specific requirements of the planned experiment. As a consequence, finding a good set of reference genes is for sure a non-trivial problem requiring quite a lot of lab-based experimental testing. In this work we identified novel candidate barley reference genes suitable for normalization in gene expression studies. An advanced web search approach aimed to collect, from publicly available web resources, the most interesting information regarding the expression profiling of candidate housekeepers on a specific experimental basis has been set up and applied, as an example, on stress conditions. A complementary lab-based analysis has been carried out to verify the expression profile of the selected genes in different tissues and during heat shock response. This combined dry/wet approach can be applied to any species and physiological condition of interest and can be considered very helpful to identify putative reference genes to be shortlisted every time a new experimental design has to be set up.

  10. 网络数据库管理技术在CIMS工程中的应用%MANAGEMENT TECHNOLOGY OF WEB DATABASE FOR APPLICATIONS IN CIMS ENGINEERING

    Institute of Scientific and Technical Information of China (English)

    孟飙; 贺向新

    2001-01-01

    CIMS技术在制造业的应用是制造业和信息技术融合必然趋势.作为CIMS应用的支撑平台也随着Internet的迅猛发展不断完善提高,因而Web技术和企业Internet数据的集成一直是人们关注的热点.本文在研究参考了多种Web和CIMS技术之后,通过一定程度的验证,提出了适合中小企业应用的信息方案.%Accompanying by the development of IT since 20 century,all kinds of enterprise have had a great change on its management,production,and so on. It is an inevitable tendency that CIMS and IT technologies are fusing in the applications of CIMS technologies in manufacture. The supporting platform for CIMS applications are becoming flawless unceasingly following the rapidly developing of Internet,hence ,the integrated technology of Web and data from enterprises intranet is a hotspot for a long time. In this article,after researching and referencing manifold Web and CIMS technologies,with a validating in some degree,brings forward an information management project that is suitable to medium or small enterprises.

  11. Advanced Techniques in Web Intelligence-2 Web User Browsing Behaviour and Preference Analysis

    CERN Document Server

    Palade, Vasile; Jain, Lakhmi

    2013-01-01

    This research volume focuses on analyzing the web user browsing behaviour and preferences in traditional web-based environments, social  networks and web 2.0 applications,  by using advanced  techniques in data acquisition, data processing, pattern extraction and  cognitive science for modeling the human actions.  The book is directed to  graduate students, researchers/scientists and engineers  interested in updating their knowledge with the recent trends in web user analysis, for developing the next generation of web-based systems and applications.

  12. Usability Testing and Redesign of Library Web Pages at Lund University, Faculty of Engineering: A Case Study Applying a Two-Phase, Systematic Quality Approach

    Science.gov (United States)

    Persson, Ann-Christin; Langh, Maria; Nilsson, Jessica

    2010-01-01

    Introduction: The Lund University Faculty of Engineering's LibQual+[R] survey 2007 showed that students and faculty had difficulties finding the information they needed at the libraries' Websites. To be able to improve the Websites, we needed to find out how the users navigated the Websites, as well as what content they needed. Method: Twenty-four…

  13. How Google Web Search copes with very similar documents

    NARCIS (Netherlands)

    Mettrop, W.; Nieuwenhuysen, P.; Smulders, H.

    2006-01-01

    A significant portion of the computer files that carry documents, multimedia, programs etc. on the Web are identical or very similar to other files on the Web. How do search engines cope with this? Do they perform some kind of “deduplication”? How should users take into account that web search resul

  14. Two-Dimensional Animation Techniques in Web Courseware

    Institute of Scientific and Technical Information of China (English)

    ZuoZongyi; FengKaiping

    2002-01-01

    To avoid the shortcomings of many Web education systems that look like Web books,a true Web multimedia Engineering Graphics courseware is developed to explore the method and techniques of 2D animation with text,buttons,and graphics merged together,For speeding up the transmission and download,data compression and interactive approach are introduced.

  15. Overview of the TREC 2013 Federated Web Search Track

    NARCIS (Netherlands)

    Demeester, Thomas; Trieschnigg, Dolf; Nguyen, Dong; Hiemstra, Djoerd

    2014-01-01

    The TREC Federated Web Search track is intended to promote research related to federated search in a realistic web setting, and hereto provides a large data collection gathered from a series of online search engines. This overview paper discusses the results of the first edition of the track, FedWeb

  16. Overview of the TREC 2014 Federated Web Search Track

    NARCIS (Netherlands)

    Demeester, Thomas; Trieschnigg, Dolf; Nguyen, Dong-Phuong; Zhou, Ke; Hiemstra, Djoerd

    2014-01-01

    The TREC Federated Web Search track facilitates research in topics related to federated web search, by providing a large realistic data collection sampled from a multitude of online search engines. The FedWeb 2013 challenges of Resource Selection and Results Merging challenges are again included in

  17. Overview of the TREC 2013 Federated Web Search Track

    NARCIS (Netherlands)

    Demeester, Thomas; Trieschnigg, Rudolf Berend; Nguyen, Dong-Phuong; Hiemstra, Djoerd

    The TREC Federated Web Search track is intended to promote research related to federated search in a realistic web setting, and hereto provides a large data collection gathered from a series of online search engines. This overview paper discusses the results of the first edition of the track, FedWeb

  18. Exposing the Hidden-Web Induced by Ajax

    NARCIS (Netherlands)

    Mesbah, A.; Van Deursen, A.

    2008-01-01

    AJAX is a very promising approach for improving rich interactivity and responsiveness of web applications. At the same time, AJAX techniques increase the totality of the hidden web by shattering the metaphor of a web ‘page’ upon which general search engines are based. This paper describes a techniqu

  19. Exposing the Hidden-Web Induced by Ajax

    NARCIS (Netherlands)

    Mesbah, A.; Van Deursen, A.

    2008-01-01

    AJAX is a very promising approach for improving rich interactivity and responsiveness of web applications. At the same time, AJAX techniques increase the totality of the hidden web by shattering the metaphor of a web ‘page’ upon which general search engines are based. This paper describes a

  20. A Personalized Search Engine Based on User Context and Semantic Web Technology%基于用户情境及语义网技术的个性化搜索引擎

    Institute of Scientific and Technical Information of China (English)

    吴芳

    2011-01-01

    With the increasing of web resources, users have strong demands to the network search engine. The existing search engine must be improved in order to satisfy the users' objective, accurate and prompt personalization information retrieval requests and to%随着网络资源的不断增加,用户对网络搜索引擎的要求越来越高。为了满足用户客观、准确、及时、深入的个性化信息检索的需求,必须改进现有搜索引擎,以便用户能更多地参与搜索流程。提出了在搜索引擎检索过程中融入用户的情境信息,即建立一个用户情境模型结构,并将其与搜索引擎集成在一起,在所集成的结构中应用语义网技术,增加搜索引擎对所检索信息的语义理解,从而实现个性化的信息检索。

  1. 语义Web环境下的搜索引擎功能分析%Analyses of Search Engine Functionality on the Semantic Web

    Institute of Scientific and Technical Information of China (English)

    严亚兰

    2010-01-01

    探讨当前搜索引擎存在的问题以及搜索引擎的语义功能需求,然后基于Web搜索引擎和语义Web,提出语义Web环境下的搜索引擎功能流图,并针对crawler、本体与知识库、语义注释、筛选与推理、语义索引、语义检索等对搜索引擎的功能进行分析.语义Web环境下的搜索引擎将促进信息、知识需求得到更好、更精确的语义表述和满足,推动高效的信息和知识管理.

  2. Gestor de contenidos web

    OpenAIRE

    García Populin, Iván

    2014-01-01

    Trabajo final de carrera desarrollado en .NET. Presenta un gestor de contenidos web para generar una web publicitaria. Treball final de carrera desenvolupat en .NET. Presenta un gestor de continguts web per generar una web publicitària.

  3. A New Methodology for Web-Knowledge- Based System Using Systematic Thinking, KM Process and Data & Knowledge Engineering Technology:FBR-GAs-CBRC5.0- CART

    Directory of Open Access Journals (Sweden)

    Patcharaporn Paokanta

    2013-10-01

    Full Text Available In Knowledge Management perspective, Organization Learning and the selection of Knowledge Management tool affects the Knowledge Management strategy planning. Among the various KM theorem such as Learning method, organization knowledge creation, Cognitive theory, Intangible assets and knowledge capital, Measuring knowledge theory etc., Systematic Thinking plays an important role in Knowledge Management activities especially, the creation of Knowledge Management strategy, KM process and Knowledge Management system. DKET is one of several approaches for implementing the Knowledge Management tools based on the KM strategies. They are not only implemented in forms of standalone system but the web-online system also. Generally, DKET namely Ensemble Learning is well known as the technique of using different training data sets or learning algorithms. Currently, a popular learning algorithm is Fuzzy-Based Reasoning (FBR which the concept of this theory is “each item is notmatched to a given cluster but it has a degree of belonging to a certain cluster”. According to these reasons, in this paper, a new methodology for Web-Knowledge-Based System by using SystematicThinking, Knowledge Process and DKET (FBR-GAs-C5.0-CART is proposed in terms of KM perspective. The algorithm performance comparisons of Fuzzy C-Means-CBR-GAs-C5.0-CART in several data sets are presented. The satisfied clustering results of Fuzzy-C Means-GAs-CBR-C5.0-CARTattain RMSE at 5.10 for the case that full data set, on the other hand the best result of using Fuzzy-C Means-CBR-C5.0-CART attain RMSE at 12.03 in the case that unrecoded variables and CBR-C5.0- CART without symptoms variables. In the future, the other KM theories and DKET will be applied to improve the performance of this system.

  4. Research on the Relationships between Chinese Journal Impact Factors and External Web Link Counts and Web Impact Factors

    Science.gov (United States)

    An, Lu; Qiu, Junping

    2004-01-01

    Journal impact factors (JIFs) as determined by the Institute for Scientific and Technological Information of China (ISTIC) of forty-two Chinese engineering journals were compared with external Web link counts, obtained from Lycos, and Web Impact Factors (WIFs) of corresponding journal Web sites to determine if any significant correlation existed…

  5. A New Hidden Web Crawling Approach

    OpenAIRE

    L.Saoudi; A.Boukerram; S.Mhamedi

    2015-01-01

    Traditional search engines deal with the Surface Web which is a set of Web pages directly accessible through hyperlinks and ignores a large part of the Web called hidden Web which is a great amount of valuable information of online database which is “hidden” behind the query forms. To access to those information the crawler have to fill the forms with a valid data, for this reason we propose a new approach which use SQLI technique in order to find the most promising keywords of a specific dom...

  6. Search engines that learn from their users

    NARCIS (Netherlands)

    Schuth, A.G.

    2016-01-01

    More than half the world’s population uses web search engines, resulting in over half a billion search queries every single day. For many people web search engines are among the first resources they go to when a question arises. Moreover, search engines have for many become the most trusted route to

  7. WEB MINING BASED FRAMEWORK FOR ONTOLOGY LEARNING

    Directory of Open Access Journals (Sweden)

    C.Ramesh

    2015-07-01

    Full Text Available Today, the notion of Semantic Web has emerged as a prominent solution to the problem of organizing the immense information provided by World Wide Web, and its focus on supporting a better co-operation between humans and machines is noteworthy. Ontology forms the major component of Semantic Web in its realization. However, manual method of ontology construction is time-consuming, costly, error-prone and inflexible to change and in addition, it requires a complete participation of knowledge engineer or domain expert. To address this issue, researchers hoped that a semi-automatic or automatic process would result in faster and better ontology construction and enrichment. Ontology learning has become recently a major area of research, whose goal is to facilitate construction of ontologies, which reduces the effort in developing ontology for a new domain. However, there are few research studies that attempt to construct ontology from semi-structured Web pages. In this paper, we present a complete framework for ontology learning that facilitates the semi-automation of constructing and enriching web site ontology from semi structured Web pages. The proposed framework employs Web Content Mining and Web Usage mining in extracting conceptual relationship from Web. The main idea behind this concept was to incorporate the web author's ideas as well as web users’ intentions in the ontology development and its evolution.

  8. Research on Key Technologies for the Semantic Web Semantic Search Engine%面向语义网的语义搜索引擎关键技术研究

    Institute of Scientific and Technical Information of China (English)

    韦丽红

    2014-01-01

    随着计算机应用技术的不断发展和网络信息技术的进步,人们的生活和工作以及商品社会的流通和交换都进入了快速的网络信息化时代。人们只需要简单的轻触鼠标就可以了解到世界各地的信息,同世界范围的人进行交流、沟通和分享信息。之所以能够如此简单的实现信息的查找和定位,都要归功于网络搜素引擎技术。该文就面向语义网的语义搜素引擎关键技术进行详细的讨论和研究,希望为网络搜素引擎技术的提高和发展提供一些借鉴和参考。%With the advancement of computer technology continues to develop applications and network information technolo-gy, people live and work and social circulation and exchange of goods have entered the era of rapid information network. People only need a simple touch of the mouse you can learn information from around the world to communicate, communicate and share information with people worldwide. Has been able to achieve such a simple find and locate information, thanks to a net-work search engines technology. This paper conducted a semantic search engines for the key technologies of the Semantic Web detailed discussion and research, hoping to provide some reference for the search engines to improve network technology and de-velopment.

  9. A grammar checker based on web searching

    Directory of Open Access Journals (Sweden)

    Joaquim Moré

    2006-05-01

    Full Text Available This paper presents an English grammar and style checker for non-native English speakers. The main characteristic of this checker is the use of an Internet search engine. As the number of web pages written in English is immense, the system hypothesises that a piece of text not found on the Web is probably badly written. The system also hypothesises that the Web will provide examples of how the content of the text segment can be expressed in a grammatically correct and idiomatic way. Thus, when the checker warns the user about the odd nature of a text segment, the Internet engine searches for contexts that can help the user decide whether he/she should correct the segment or not. By means of a search engine, the checker also suggests use of other expressions that appear on the Web more often than the expression he/she actually wrote.

  10. A Framework for Deep Web Crawler Using Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    K.F.Bharati

    2013-03-01

    Full Text Available The Web has become one of the largest and most readily accessible repositories of human knowledge. The traditional search engines index only surface Web whose pages are easily found. The focus has now been moved to invisible Web or hidden Web, which consists of a large warehouse of useful data such as images, sounds, presentations and many other types of media. To use such data, there is a need for specialized technique to locate those sites as we do with search engines. This paper focuses on an effective design of a Hidden Web Crawler that can automatically discover pages from the Hidden Web by employing multi- agent Web mining system. A framework for deep web with genetic algorithm is used to discover the resource discovery problem and the results show the improvement in the crawling strategy and harvest rate.

  11. Web sites that work secrets from winning web sites

    CERN Document Server

    Smith, Jon

    2012-01-01

    Leading web site entrepreneur Jon Smith has condensed the secrets of his success into 52 inspiring ideas that even the most hopeless technophobe can implement. The brilliant tips and practical advice in Web sites that work will uplift and transform any website, from the simplest to the most complicated. It deals with everything from fundamentals such as how to assess the effectiveness of a website and how to get a site listed on the most popular search engines to more sophisticated challenges like creating a community and dealing with legal requirements. Straight-talking, practical and humorou

  12. Unified Approach to Vulnerability Analysis of Web Applications

    Science.gov (United States)

    Le, H. T.; Loh, P. K. K.

    2008-11-01

    Web vulnerabilities in web-based applications may be detected, classified and documented. Several Web scanners exist for vulnerabilities in Web applications implemented via different technologies. However, none of them provides the technology-independent, generic coverage of possible vulnerabilities. In this project that is funded by Mindef Singapore, we propose a new approach for Web application security and vulnerability analysis. The design addresses the categorization of scanner results with a generic data model and the design of a language-independent rule-based engine that detects, analyses and reports suspected vulnerabilities in web-based applications.

  13. A Deep Web Data Integration System for Job Search

    Institute of Scientific and Technical Information of China (English)

    LIU Wei; LI Xian; LING Yanyan; ZHANG Xiaoyu; MENG Xiaofeng

    2006-01-01

    With the rapid development of Web, there are more and more Web databases available for users to access. At the same time, job searchers often have difficulties in first finding the right sources and then querying over them, providing such an integrated job search system over Web databases has become a Web application in high demand. Based on such consideration, we build a deep Web data integration system that supports unified access for users to multiple job Web sites as a job meta-search engine. In this paper, the architecture of the system is given first, and the key components in the system are introduced.

  14. Automated Security Testing of Web Widget Interactions

    NARCIS (Netherlands)

    Bezemer, C.-P.; Mesbah, A.; Van Deursen, A.

    2009-01-01

    This paper is a pre-print of: Cor-Paul Bezemer, Ali Mesbah, and Arie van Deursen. Automated Security Testing of Web Widget Interactions. In Proceedings of the 7th joint meeting of the European Software Engineering Conference and the ACM SIGSOFT Symposium on the Foundations of Software Engineering (E

  15. Automated Security Testing of Web Widget Interactions

    NARCIS (Netherlands)

    Bezemer, C.-P.; Mesbah, A.; Van Deursen, A.

    2009-01-01

    This paper is a pre-print of: Cor-Paul Bezemer, Ali Mesbah, and Arie van Deursen. Automated Security Testing of Web Widget Interactions. In Proceedings of the 7th joint meeting of the European Software Engineering Conference and the ACM SIGSOFT Symposium on the Foundations of Software Engineering

  16. Automated Security Testing of Web Widget Interactions

    NARCIS (Netherlands)

    Bezemer, C.-P.; Mesbah, A.; Van Deursen, A.

    2009-01-01

    This paper is a pre-print of: Cor-Paul Bezemer, Ali Mesbah, and Arie van Deursen. Automated Security Testing of Web Widget Interactions. In Proceedings of the 7th joint meeting of the European Software Engineering Conference and the ACM SIGSOFT Symposium on the Foundations of Software Engineering (E

  17. How much data resides in a web collection: how to estimate size of a web collection

    NARCIS (Netherlands)

    Khelghati, Mohammadreza; Hiemstra, Djoerd; Keulen, van Maurice

    2013-01-01

    With increasing amount of data in deep web sources (hidden from general search engines behind web forms), accessing this data has gained more attention. In the algorithms applied for this purpose, it is the knowledge of a data source size that enables the algorithms to make accurate decisions in sto

  18. Web Indexing with Meta Fields: A Survey of Web Objects in Polymer Chemistry.

    Science.gov (United States)

    Qin, Jian; Wesley, Kathryn

    1998-01-01

    This survey of 1037 World Wide Web objects on polymer chemistry collected by four search engines (AltaVista, Lycos, Excite, Webcrawler) examined document information and use of meta fields, images, and chemical names. Found that the use of meta fields was not widespread and that knowledge of meta fields in HTML varied among Web object creators.…

  19. How much data resides in a web collection: how to estimate size of a web collection

    NARCIS (Netherlands)

    Khelghati, Mohammadreza; Hiemstra, Djoerd; van Keulen, Maurice

    2013-01-01

    With increasing amount of data in deep web sources (hidden from general search engines behind web forms), accessing this data has gained more attention. In the algorithms applied for this purpose, it is the knowledge of a data source size that enables the algorithms to make accurate decisions in

  20. What is the invisible web? A crawler perspective

    OpenAIRE

    Arroyo, Natalia

    2004-01-01

    The invisible Web, also known as the deep Web or dark matter, is an important problem for Webometrics due to difficulties of conceptualization and measurement. The invisible Web has been defined to be the part of the Web that cannot be indexed by search engines, including databases and dynamically generated pages. Some authors have recognized that this is a quite subjective concept that depends on the point of view of the observer: what is visible for one observer may be invisible for others....

  1. Overview of the TREC 2014 Federated Web Search Track

    OpenAIRE

    Demeester, Thomas; Trieschnigg, Rudolf Berend; Nguyen, Dong-Phuong; Zhou, Ke; Hiemstra, Djoerd

    2014-01-01

    The TREC Federated Web Search track facilitates research in topics related to federated web search, by providing a large realistic data collection sampled from a multitude of online search engines. The FedWeb 2013 challenges of Resource Selection and Results Merging challenges are again included in FedWeb 2014, and we additionally introduced the task of vertical selection. Other new aspects are the required link between the Resource Selection and Results Merging, and the importance of diversi...

  2. Detection And Classification Of Web Robots With Honeypots

    Science.gov (United States)

    2016-03-01

    programs has been attributed to the explosion in content and user-generated social media on the Internet. The Web search engines like Google require...large numbers of automated bots on the Web to build their indexes. Furthermore, the growth of internet has produced a market for businesses, both...played an important role in its evolution and growth. Conversely, the “bad” Web robots have been and continue to be a significant problem. Bad Web robots

  3. WEB 238 Courses Tutorial / indigohelp

    OpenAIRE

    2015-01-01

    WEB 238 Week 2 JavaScript Events WEB 238 Week 3 Cookies WEB 238 Week 4 Dynamic HTML WEB 238 Week 5 Web Programming Languages WEB 238 Week 1 DQs WEB 238 Week 2DQs WEB 238 Week 3DQs WEB 238 Week 4DQs WEB 238 Week 5DQs  

  4. A web service infrastructure for thermochemical data.

    Science.gov (United States)

    Paolini, Christopher P; Bhattacharjee, Subrata

    2008-07-01

    W3C standardized Web Services are becoming an increasingly popular middleware technology used to facilitate the open exchange of chemical data. While several projects in existence use Web Services to wrap existing commercial and open-source tools that mine chemical structure data, no Web Service infrastructure has yet been developed to compute thermochemical properties of substances. This work presents an infrastructure of Web Services for thermochemical data retrieval. Several examples are presented to demonstrate how our Web Services can be called from Java, through JavaScript using an AJAX methodology, and within commonly used commercial applications such as Microsoft Excel and MATLAB for use in computational work. We illustrate how a JANAF table, widely used by chemists and engineers, can be quickly reproduced through our Web Service infrastructure.

  5. Development of Visual Design Process of Clothing Sales Web Page Based on Perceptual Engineering%基于感性工学的服装销售网页视觉设计流程开发

    Institute of Scientific and Technical Information of China (English)

    侯倩

    2015-01-01

    Network sales have increasingly become an important way of enterprise image display and sales. It becomes a design issue of concern that enhancing the visual appeal of the web page rapidly and accurately. So, by collecting and analyzing existing research results, the article formed the development ideas of the visual design process of clothing sales web page based on the perceptual engineering. From the demands of consumers and sellers, the article set up the principle of the process development, established the constructing points, and ifnally completed the process building, to offer methods and theoretical references for the relevant design practice and further research.%网络销售越来越成为企业展示形象和销售的重要途径,快速准确地增强网页视觉吸引力成为备受关注的设计问题。因此本文整理与分析了已有研究成果,形成了基于感性工学的服装销售网页视觉设计流程的构建思路,并从消费者与销售者双方诉求入手,树立了流程开发的原则,明确了构建要点,最终完成了流程的构建,以期为相关设计实践与深入研究提供方法借鉴和理论支撑。

  6. Deep Web Source Discovery Based on Search Engine%基于搜索引擎的Deep Web数据源发现技术

    Institute of Scientific and Technical Information of China (English)

    李文骏; 崔志明

    2008-01-01

    随着Web数据库的广泛应用,Web正在不断"深化".传统搜索引擎只能检索浅层网络,却不能直接索引到深层网络(Deep Web)的资源.为了有效地利用Deep Web资源,必须要对Deep Web数据进行大规模集成.其中,数据源发现是整合DeepWeb资源的首要工作,能否高效地发现DeepWeb站点是DeepWeb数据获取的关键.提出了一种基于传统搜索引擎的Deep Web数据源发现方法,该方法通过分析返回结果来扩展查询,从而进一步提高了数据源发现的效率.实验证明该方法能得到较好的结果.

  7. Usare WebDewey

    OpenAIRE

    Baldi, Paolo

    2016-01-01

    This presentation shows how to use the WebDewey tool. Features of WebDewey. Italian WebDewey compared with American WebDewey. Querying Italian WebDewey. Italian WebDewey and MARC21. Italian WebDewey and UNIMARC. Numbers, captions, "equivalente verbale": Dewey decimal classification in Italian catalogues. Italian WebDewey and Nuovo soggettario. Italian WebDewey and LCSH. Italian WebDewey compared with printed version of Italian Dewey Classification (22. edition): advantages and disadvantages o...

  8. A Novel Personalized Web Search Model

    Institute of Scientific and Technical Information of China (English)

    ZHU Zhengyu; XU Jingqiu; TIAN Yunyan; REN Xiang

    2007-01-01

    A novel personalized Web search model is proposed.The new system, as a middleware between a user and a Web search engine, is set up on the client machine. It can learn a user's preference implicitly and then generate the user profile automatically. When the user inputs query keywords, the system can automatically generate a few personalized expansion words by computing the term-term associations according to the current user profile, and then these words together with the query keywords are submitted to a popular search engine such as Yahoo or Google.These expansion words help to express accurately the user's search intention. The new Web search model can make a common search engine personalized, that is, the search engine can return different search results to different users who input the same keywords. The experimental results show the feasibility and applicability of the presented work.

  9. Web Services Harvesting Approach Based on Internet%基于Internet的Web Services收集方法

    Institute of Scientific and Technical Information of China (English)

    王立杰; 邹艳珍; 金靖; 赵俊峰; 谢冰

    2011-01-01

    Web Services作为一种部署在Internet上的新型的可复用软件资源,得到广泛的重视和应用.但是,现有通用搜索引擎并不能很好地支持Web Services的发现,这使得软件开发人员需要花费很大的代价寻找合适的Web Services.鉴于此,提出了一种基于Intemet的Web Services获取方法,采用通用搜索引擎和特定Web Services信息发布网站相结合的方法,从Internet上收集Web Services,并从Web Services的相关网页中抽取相关描述信息.利用收集到的Web Services数据,对当前互联网上的Web Services现状进行了统计分析.这些统计结果一定程度上反映了Web Services的发展现状.%As a special kind of reusable software resources on the Internet, Web Services have received great attention. However, general Web search engine cannot well support the discovery of Web Services, which makes developers have to spend a lot of efforts in order to acquire their desired Web Services. For this reason, this paper proposes an approach for harvesting Web Services from the Internet by leveraging general search engine and specific Web services portals. In addition, it also proposes an approach to extract descriptions for Web Services from their related Web pages on the Internet. Based on the harvested data, it conducts a statistics analysis on the status of Web Services on the Internet.

  10. Study on online community user motif using web usage mining

    Science.gov (United States)

    Alphy, Meera; Sharma, Ajay

    2016-04-01

    The Web usage mining is the application of data mining, which is used to extract useful information from the online community. The World Wide Web contains at least 4.73 billion pages according to Indexed Web and it contains at least 228.52 million pages according Dutch Indexed web on 6th august 2015, Thursday. It’s difficult to get needed data from these billions of web pages in World Wide Web. Here is the importance of web usage mining. Personalizing the search engine helps the web user to identify the most used data in an easy way. It reduces the time consumption; automatic site search and automatic restore the useful sites. This study represents the old techniques to latest techniques used in pattern discovery and analysis in web usage mining from 1996 to 2015. Analyzing user motif helps in the improvement of business, e-commerce, personalisation and improvement of websites.

  11. 基于搜索引擎的Deep Web数据源发现%Deep Web Data Source Discovery Based on Search Engine

    Institute of Scientific and Technical Information of China (English)

    王海龙; 胡景芝; 赵朋朋; 崔志明

    2011-01-01

    This paper proposes a method for the data source discovery using the search engine. In order to submit high quality key words to the search engine, the paper introduces the ontology to the initial word construction process, classifies all the words according to their frequency in the current domain, and reclassifies these words in accordance with the element quantity of the returned collection, ensures that the key word contributes greatly to the discovery of the data source query interface. Test results in different domains show that the approach proposed can discover a large amount of query interfaces, and its validty is verified.%提出一种利用搜索引擎发现数据源的方法.为向搜索引擎提交高质量的关键词,将本体作为等级化组织词汇的架构引入到初始词构建过程.对所有词汇按在当前领域中出现频率高低进行分类,并根据搜索引擎返回接口元素数量进行二次分类,确保关键词是对发现数据源查询接口贡献较大的词汇.在不同领域上的测试结果表明,该方法能发现相当数量的查询接口,从而验证其有效性.

  12. Semantic Web

    Directory of Open Access Journals (Sweden)

    Anna Lamandini

    2011-06-01

    Full Text Available The semantic Web is a technology at the service of knowledge which is aimed at accessibility and the sharing of content; facilitating interoperability between different systems and as such is one of the nine key technological pillars of TIC (technologies for information and communication within the third theme, programme specific cooperation of the seventh programme framework for research and development (7°PQRS, 2007-2013. As a system it seeks to overcome overload or excess of irrelevant information in Internet, in order to facilitate specific or pertinent research. It is an extension of the existing Web in which the aim is for cooperation between and the computer and people (the dream of Sir Tim Berners –Lee where machines can give more support to people when integrating and elaborating data in order to obtain inferences and a global sharing of data. It is a technology that is able to favour the development of a “data web” in other words the creation of a space in both sets of interconnected and shared data (Linked Data which allows users to link different types of data coming from different sources. It is a technology that will have great effect on everyday life since it will permit the planning of “intelligent applications” in various sectors such as education and training, research, the business world, public information, tourism, health, and e-government. It is an innovative technology that activates a social transformation (socio-semantic Web on a world level since it redefines the cognitive universe of users and enables the sharing not only of information but of significance (collective and connected intelligence.

  13. DERIVING USER ACCESS PATTERNS AND MINING WEB COMMUNITY WITH WEB-LOG DATA FOR PREDICTING USER SESSIONS WITH PAJEK

    Directory of Open Access Journals (Sweden)

    S. Balaji

    2012-10-01

    Full Text Available Web logs are a young and dynamic media type. Due to the intrinsic relationship among Web objects and the deficiency of a uniform schema of web documents, Web community mining has become significant area for Web data management and analysis. The research of Web communities extents a number of research domains. In this paper an ontological model has been present with some recent studies on this topic, which cover finding relevant Web pages based on linkage information, discovering user access patterns through analyzing Web log files from Web data. A simulation has been created with the academic website crawled data. The simulation is done in JAVA and ORACLE environment. Results show that prediction of user session could give us plenty of vital information for the Business Intelligence. Search Engine Optimization could also use these potential results which are discussed in the paper in detail.

  14. Information Collection and Extraction of Web Pages with Public Opinion Search Engine%舆情搜索引擎中网页信息的采集与抽取研究

    Institute of Scientific and Technical Information of China (English)

    王兰成

    2011-01-01

    It is usually different between Internet public opinion search engine and information search. The first is tobe gathering and extracting data within the page depth to the site and effective. It puts forward many new research content and methods for information area. Web information extraction in the templates and page analysis of two ways, based on natural language processing, Ontology extraction and wrapper induction method based on the analysis have been researched. The wrapper induction based manner and in the rule generation module used an expert model has been designed. It improves the accuracy of public opinion and the quality of search engines.%网络舆情搜索引擎与通常的网络信息搜索不同,其最终结果要深入到站点和页面内部采集与抽取有效数据,给情报界提出了许多新的研究内容和方法.在对网页信息抽取的模板和页面分析两种方式、基于自然语言处理、包装器归纳和Ontology抽取方法的分析基础上,使用基于包装器归纳方式并在规则生成模块中采用专家模式,设计一种基于样本学习的新闻抽取方法,通过人工分析网页源代码制定和修改抽取规则,然后根据抽取规则进行信息自动抽取,以提高舆情搜索引擎的精度和质量.

  15. Focused Crawling of the Deep Web Using Service Class Descriptions

    Energy Technology Data Exchange (ETDEWEB)

    Rocco, D; Liu, L; Critchlow, T

    2004-06-21

    Dynamic Web data sources--sometimes known collectively as the Deep Web--increase the utility of the Web by providing intuitive access to data repositories anywhere that Web access is available. Deep Web services provide access to real-time information, like entertainment event listings, or present a Web interface to large databases or other data repositories. Recent studies suggest that the size and growth rate of the dynamic Web greatly exceed that of the static Web, yet dynamic content is often ignored by existing search engine indexers owing to the technical challenges that arise when attempting to search the Deep Web. To address these challenges, we present DynaBot, a service-centric crawler for discovering and clustering Deep Web sources offering dynamic content. DynaBot has three unique characteristics. First, DynaBot utilizes a service class model of the Web implemented through the construction of service class descriptions (SCDs). Second, DynaBot employs a modular, self-tuning system architecture for focused crawling of the DeepWeb using service class descriptions. Third, DynaBot incorporates methods and algorithms for efficient probing of the Deep Web and for discovering and clustering Deep Web sources and services through SCD-based service matching analysis. Our experimental results demonstrate the effectiveness of the service class discovery, probing, and matching algorithms and suggest techniques for efficiently managing service discovery in the face of the immense scale of the Deep Web.

  16. Evaluative Measures of Search Engines

    Directory of Open Access Journals (Sweden)

    Jitendra Nath Singh

    2012-03-01

    Full Text Available The ability to search and retrieve information from the web efficiently and effectively is great challenge of search engine. Information retrieval on the Web is very different from retrieval in traditional indexed databases because it’s hyper-linked character, the heterogeneity of document types and authoring styles. Thus, since Web retrieval is substantially different from information retrieval, new or revised evaluative measures are required to assess retrieval performance using search engines. In this paper we suggested a number of evaluative measures to evaluate the effectiveness of search engines. The motivation behind each of these measures is presented, along with their descriptions and definitions.

  17. Analyzing web log files of the health on the net HONmedia search engine to define typical image search tasks for image retrieval evaluation.

    Science.gov (United States)

    Müller, Henning; Boyer, Célia; Gaudinat, Arnaud; Hersh, William; Geissbuhler, Antoine

    2007-01-01

    Medical institutions produce ever-increasing amount of diverse information. The digital form makes these data available for the use on more than a single patient. Images are no exception to this. However, less is known about how medical professionals search for visual medical information and how they want to use it outside of the context of a single patient. This article analyzes ten months of usage log files of the Health on the Net (HON) medical media search engine. Key words were extracted from all queries and the most frequent terms and subjects were identified. The dataset required much pre-treatment. Problems included national character sets, spelling errors and the use of terms in several languages. The results show that media search, particularly for images, was frequently used. The most common queries were for general concepts (e.g., heart, lung). To define realistic information needs for the ImageCLEFmed challenge evaluation (Cross Language Evaluation Forum medical image retrieval), we used frequent queries that were still specific enough to at least cover two of the three axes on modality, anatomic region, and pathology. Several research groups evaluated their image retrieval algorithms based on these defined topics.

  18. Distributed and Cooperative Information Retrieval on the World Wide Web

    Institute of Scientific and Technical Information of China (English)

    王继成; 金翔宇; 杨晓江; 张福炎

    2000-01-01

    A mass of heterogeneous, distributed and dynamic information on the World Wide Web (the Web) has resulted in "information overload". It's an important and urgent research issue to provide users with effective information retrieval service on the Web. Web search engines attempt to solve this problem, yet their effect is far from satisfying. In this paper, a distributed and cooperative strategy for information retrieval on the Web is proposed to substitute the centralized mode adopted by the current search engines. Then a new information retrieval system model IRSM is presented, which supports the retrieval of metadata about Web documents and uses Z39.50 standard protocol to unify the heterogeneous interfaces of different systems. Based on that, a distributed and cooperative information retrieval framework, called DCIRF, is designed to help users in fast and effective information retrieval on the Web.

  19. A Survey On Various Web Template Detection And Extraction Methods

    Directory of Open Access Journals (Sweden)

    Neethu Mary Varghese

    2015-03-01

    Full Text Available Abstract In todays digital world reliance on the World Wide Web as a source of information is extensive. Users increasingly rely on web based search engines to provide accurate search results on a wide range of topics that interest them. The search engines in turn parse the vast repository of web pages searching for relevant information. However majority of web portals are designed using web templates which are designed to provide consistent look and feel to end users. The presence of these templates however can influence search results leading to inaccurate results being delivered to the users. Therefore to improve the accuracy and reliability of search results identification and removal of web templates from the actual content is essential. A wide range of approaches are commonly employed to achieve this and this paper focuses on the study of the various approaches of template detection and extraction that can be applied across homogenous as well as heterogeneous web pages.

  20. Multilabel Learning for Automatic Web Services Tagging

    Directory of Open Access Journals (Sweden)

    Mustapha AZNAG

    2014-08-01

    Full Text Available Recently, some web services portals and search engines as Biocatalogue and Seekda!, have allowed users to manually annotate Web services using tags. User Tags provide meaningful descriptions of services and allow users to index and organize their contents. Tagging technique is widely used to annotate objects in Web 2.0 applications. In this paper we propose a novel probabilistic topic model (which extends the CorrLDA model - Correspondence Latent Dirichlet Allocation- to automatically tag web services according to existing manual tags. Our probabilistic topic model is a latent variable model that exploits local correlation labels. Indeed, exploiting label correlations is a challenging and crucial problem especially in multi-label learning context. Moreover, several existing systems can recommend tags for web services based on existing manual tags. In most cases, the manual tags have better quality. We also develop three strategies to automatically recommend the best tags for web services. We also propose, in this paper, WS-Portal; An Enriched Web Services Search Engine which contains 7063 providers, 115 sub-classes of category and 22236 web services crawled from the Internet. In WS-Portal, severals technologies are employed to improve the effectiveness of web service discovery (i.e. web services clustering, tags recommendation, services rating and monitoring. Our experiments are performed out based on real-world web services. The comparisons of Precision@n, Normalised Discounted Cumulative Gain (NDCGn values for our approach indicate that the method presented in this paper outperforms the method based on the CorrLDA in terms of ranking and quality of generated tags.

  1. The Ideal Web IR Service Frame%理想的Web IR服务模式的研究

    Institute of Scientific and Technical Information of China (English)

    刘悦; 冯国臻; 程学旗; 薄立彦

    2003-01-01

    Web IR presents a new challenge due to the heterogeneity,the dynamic characteristic and the size of theWeb. A practical IR system that can satisfy the users' demand is very important ,in this paper we research the charac-teristics of Web IR in detail and give out the ideal Web IR service model:it should include search engine spectrum,search engine hierarchy,search engine cooperative network. We also analyze the key technique of this model,propose asimple way for the Web IR service to deal with the huge-scale of Web resources easily,and test part of the ideas in ourprototype system SAInSE.

  2. Survey of Techniques for Deep Web Source Selection and Surfacing the Hidden Web Content

    Directory of Open Access Journals (Sweden)

    Khushboo Khurana

    2016-05-01

    Full Text Available Large and continuously growing dynamic web content has created new opportunities for large-scale data analysis in the recent years. There is huge amount of information that the traditional web crawlers cannot access, since they use link analysis technique by which only the surface web can be accessed. Traditional search engine crawlers require the web pages to be linked to other pages via hyperlinks causing large amount of web data to be hidden from the crawlers. Enormous data is available in deep web that can be useful to gain new insight for various domains, creating need to access the information from the deep web by developing efficient techniques. As the amount of Web content grows rapidly, the types of data sources are proliferating, which often provide heterogeneous data. So we need to select Deep Web Data sources that can be used by the integration systems. The paper discusses various techniques that can be used to surface the deep web information and techniques for Deep Web Source Selection.

  3. Myanmar Language Search Engine

    Directory of Open Access Journals (Sweden)

    Pann Yu Mon

    2011-03-01

    Full Text Available With the enormous growth of the World Wide Web, search engines play a critical role in retrieving information from the borderless Web. Although many search engines are available for the major languages, but they are not much proficient for the less computerized languages including Myanmar. The main reason is that those search engines are not considering the specific features of those languages. A search engine which capable of searching the Web documents written in those languages is highly needed, especially when more and more Web sites are coming up with localized content in multiple languages. In this study, the design and the architecture of language specific search engine for Myanmar language is proposed. The main feature of the system are, (1 it can search the multiple encodings of the Myanmar Web page, (2 the system is designed to comply with the specific features of the Myanmar language. Finally the experiment has been done to prove whether it meets the design requirements.

  4. Earth Science Mining Web Services

    Science.gov (United States)

    Pham, Long; Lynnes, Christopher; Hegde, Mahabaleshwa; Graves, Sara; Ramachandran, Rahul; Maskey, Manil; Keiser, Ken

    2008-01-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at he GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADam components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestras the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to the infusion is the loosely coupled, Web-Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  5. Earth Science Mining Web Services

    Science.gov (United States)

    Pham, L. B.; Lynnes, C. S.; Hegde, M.; Graves, S.; Ramachandran, R.; Maskey, M.; Keiser, K.

    2008-12-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at the GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADaM components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestrates the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to this infusion is the loosely coupled, Web- Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  6. Endnote web

    OpenAIRE

    Uezu, Denis

    2015-01-01

    Представлено краткое руководство по работе с сетевой сервисной программой EndNote Web на платформе Web of Knowledge издательства Thomson Reuters на русском языке. EndNote Web разработана для предоставления помощи исследователям и студентам в процессе написания научных публикаций. Позволяет создавать свои базы данных с собственными библиографическими списками для цитирования в научных работах....

  7. 传统搜索引擎和语义搜索引擎在Web2.0搜索中的差异性%Differences Between Common Search Engine and Semantic Search Engine in Web2.0 Searching

    Institute of Scientific and Technical Information of China (English)

    赵夷平

    2010-01-01

    文章通过对传统搜索引擎和语义搜索引擎在用户创造内容搜索、社会网络搜索和个性化搜索3个方面的对比,展现了两种不同类型搜索引擎在Web2.0环境下提供搜索服务的差异性,为搜索引擎未来的发展提供参考.

  8. IMPROVING PERSONALIZED WEB SEARCH USING BOOKSHELF DATA STRUCTURE

    Directory of Open Access Journals (Sweden)

    S.K. Jayanthi

    2012-10-01

    Full Text Available Search engines are playing a vital role in retrieving relevant information for the web user. In this research work a user profile based web search is proposed. So the web user from different domain may receive different set of results. The main challenging work is to provide relevant results at the right level of reading difficulty. Estimating user expertise and re-ranking the results are the main aspects of this paper. The retrieved results are arranged in Bookshelf Data Structure for easy access. Better presentation of search results hence increases the usability of web search engines significantly in visual mode.

  9. Towards semantic web mining

    OpenAIRE

    Berendt, Bettina; Hotho, Andreas; Stumme, Gerd

    2002-01-01

    Semantic Web Mining aims at combining the two fast-developing research areas Semantic Web and Web Mining. The idea is to improve, on the one hand, the results of Web Mining by exploiting the new semantic structures in the Web; and to make use of Web Mining, on overview of where the two areas meet today, and sketches ways of how a closer integration could be profitable.

  10. Identifying Aspects for Web-Search Queries

    OpenAIRE

    Wu, Fei; Madhavan, Jayant; Halevy, Alon

    2014-01-01

    Many web-search queries serve as the beginning of an exploration of an unknown space of information, rather than looking for a specific web page. To answer such queries effec- tively, the search engine should attempt to organize the space of relevant information in a way that facilitates exploration. We describe the Aspector system that computes aspects for a given query. Each aspect is a set of search queries that together represent a distinct information need relevant to the original search...

  11. Acquiring Evolving Technologies: Web Services Standards

    Science.gov (United States)

    2016-06-30

    2006 Carnegie Mellon University Acquiring Evolving Technologies: Web Services Standards Harry L. Levinson Software Engineering Institute Carnegie...Acquiring Evolving Technologies: Web Services Standards 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT...NUMBER OF PAGES 22 19a. NAME OF RESPONSIBLE PERSON a. REPORT unclassified b. ABSTRACT unclassified c. THIS PAGE unclassified Standard Form

  12. Web TA Production (WebTA)

    Data.gov (United States)

    US Agency for International Development — WebTA is a web-based time and attendance system that supports USAID payroll administration functions, and is designed to capture hours worked, leave used and...

  13. Semantic Web Technologies for the Adaptive Web

    DEFF Research Database (Denmark)

    Dolog, Peter

    2007-01-01

    Ontologies and reasoning are the key terms brought into focus by the semantic web community. Formal representation of ontologies in a common data model on the web can be taken as a foundation for adaptive web technologies as well. This chapter describes how ontologies shared on the semantic web...... means for deciding which links to show, annotate, hide, generate, and reorder. The semantic web technologies provide means to formalize the domain ontologies and metadata created from them. The formalization enables reasoning for personalization decisions. This chapter describes which components...... are crucial to be formalized by the semantic web ontologies for adaptive web. We use examples from an eLearning domain to illustrate the principles which are broadly applicable to any information domain on the web....

  14. Current Application of Search Engines and Their Developing Trend

    Institute of Scientific and Technical Information of China (English)

    ZHANG Li; SHAO Shi-huang; WU Xiao-qiong; ZENG Xian-hui; FAN Xiao-wen

    2002-01-01

    The basic types of current search engines which can help users to perfume laborious information-gathering tasks on Internet is proposed. Basically, the search engines can be classified into index engine, directory engine and agent engine on WWW information service. The key technologies of web mine, automatic classifying of documents and ordering regulation of feedback information are discussed. Finally, the developing trend of search engines is pointed out by analyzing their practical application on World Wide Web.

  15. Semantic web for dummies

    CERN Document Server

    Pollock, Jeffrey T

    2009-01-01

    Semantic Web technology is already changing how we interact with data on the Web. By connecting random information on the Internet in new ways, Web 3.0, as it is sometimes called, represents an exciting online evolution. Whether you're a consumer doing research online, a business owner who wants to offer your customers the most useful Web site, or an IT manager eager to understand Semantic Web solutions, Semantic Web For Dummies is the place to start! It will help you:Know how the typical Internet user will recognize the effects of the Semantic WebExplore all the benefits the data Web offers t

  16. Generating Best Features for Web Page Classification

    Directory of Open Access Journals (Sweden)

    K. Selvakuberan

    2008-03-01

    Full Text Available As the Internet provides millions of web pages for each and every search term, getting interesting and required results quickly from the Web becomes very difficult. Automatic classification of web pages into relevant categories is the current research topic which helps the search engine to get relevant results. As the web pages contain many irrelevant, infrequent and stop words that reduce the performance of the classifier, extracting or selecting representative features from the web page is an essential pre-processing step. The goal of this paper is to find minimum number of highly qualitative features by integrating feature selection techniques. We conducted experiments with various numbers of features selected by different feature selection algorithms on a well defined initial set of features and show that cfssubset evaluator combined with term frequency method gives minimal qualitative features enough to attain considerable classification accuracy.

  17. Fuzzy knowledge management for the semantic web

    CERN Document Server

    Ma, Zongmin; Yan, Li; Cheng, Jingwei

    2014-01-01

    This book goes to great depth concerning the fast growing topic of technologies and approaches of fuzzy logic in the Semantic Web. The topics of this book include fuzzy description logics and fuzzy ontologies, queries of fuzzy description logics and fuzzy ontology knowledge bases, extraction of fuzzy description logics and ontologies from fuzzy data models, storage of fuzzy ontology knowledge bases in fuzzy databases, fuzzy Semantic Web ontology mapping, and fuzzy rules and their interchange in the Semantic Web. The book aims to provide a single record of current research in the fuzzy knowledge representation and reasoning for the Semantic Web. The objective of the book is to provide the state of the art information to researchers, practitioners and graduate students of the Web intelligence and at the same time serve the knowledge and data engineering professional faced with non-traditional applications that make the application of conventional approaches difficult or impossible.

  18. A New Hidden Web Crawling Approach

    Directory of Open Access Journals (Sweden)

    L.Saoudi

    2015-10-01

    Full Text Available Traditional search engines deal with the Surface Web which is a set of Web pages directly accessible through hyperlinks and ignores a large part of the Web called hidden Web which is a great amount of valuable information of online database which is “hidden” behind the query forms. To access to those information the crawler have to fill the forms with a valid data, for this reason we propose a new approach which use SQLI technique in order to find the most promising keywords of a specific domain for automatic form submission. The effectiveness of proposed framework has been evaluated through experiments using real web sites and encouraging preliminary results were obtained

  19. Being, space and time in the Web

    CERN Document Server

    Vafopoulos, Michalis

    2011-01-01

    The Web emerged as the antidote to rapidly increasing quantity of accumulated knowledge because it successfully enables massive representation and communication with minimum costs. Despite the fact that its gigantic scale and impact make difficult to anticipate the effects in humans, we claim from it to be fast, secure, reliable, all-inclusive and trustworthy. It is time for science to compensate and provide an epistemological "antidote" to these issues. On this campaign, Philosophy should be in the front line by forming the relevant questions. We initiate the dialogue for a theory about being in the Web that will serve as a bridge between philosophical thinking and engineering. We analyze existence and spatiotemporality in the Web, as a closed techno-social system, and how it transforms the traditional conceptions about actuality. Location in the Web space is specified by the Web being's URI and the URI's of incoming and outgoing links. The primer role of visiting durations is best approximated by Bergsonian...

  20. Measuring the Utilization of On-Page Search Engine Optimization in Selected Domain

    National Research Council Canada - National Science Library

    Goran Matošević

    2015-01-01

    Search engine optimization (SEO) techniques involve „on-page“ and „off-page“ actions taken by web developers and SEO specialists with aim to increase the ranking of web pages in search engine results pages (SERP...

  1. QOS Aware Formalized Model for Semantic Web Service Selection

    Directory of Open Access Journals (Sweden)

    Divya Sachan

    2014-10-01

    Full Text Available Selecting the most relevant Web Service according to a client requirement is an onerous task, as innumerous number of functionally same Web Services(WS are listed in UDDI registry. WS are functionally same but their Quality and performance varies as per service providers. A web Service Selection Process involves two major points: Recommending the pertinent Web Service and avoiding unjustifiable web service. The deficiency in keyword based searching is that it doesn’t handle the client request accurately as keyword may have ambiguous meaning on different scenarios. UDDI and search engines all are based on keyword search, which are lagging behind on pertinent Web service selection. So the search mechanism must be incorporated with the Semantic behavior of Web Services. In order to strengthen this approach, the proposed model is incorporated with Quality of Services (QoS based Ranking of semantic web services.

  2. An application of TOPSIS for ranking internet web browsers

    Directory of Open Access Journals (Sweden)

    Shahram Rostampour

    2012-07-01

    Full Text Available Web browser is one of the most important internet facilities for surfing the internet. A good web browser must incorporate literally tens of features such as integrated search engine, automatic updates, etc. Each year, ten web browsers are formally introduced as top best reviewers by some organizations. In this paper, we propose the implementation of TOPSIS technique to rank ten web browsers. The proposed model of this paper uses five criteria including speed, features, security, technical support and supported configurations. In terms of speed, Safari is the best web reviewer followed by Google Chrome and Internet Explorer while Opera is the best web reviewer when we look into 20 different features. We have also ranked these web browsers using all five categories together and the results indicate that Opera, Internet explorer, Firefox and Google Chrome are the best web browsers to be chosen.

  3. Keyword search in the Deep Web

    OpenAIRE

    Calì, Andrea; Martinenghi, D.; Torlone, R.

    2015-01-01

    The Deep Web is constituted by data accessible through Web\\ud pages, but not readily indexable by search engines, as they are returned\\ud in dynamic pages. In this paper we propose a framework for accessing\\ud Deep Web sources, represented as relational tables with so-called ac-\\ud cess limitations, with keyword-based queries. We formalize the notion\\ud of optimal answer and investigate methods for query processing. To our\\ud knowledge, this problem has never been studied in a systematic way.

  4. On the performance of web services

    CERN Document Server

    Tari, Zahir; Jayasinghe, Malith

    2011-01-01

    Performance of Web Services provides innovative techniques to improve the performance of Web Services, as well as QoS (Quality of Service) requirements. This includes Qos performance, reliability and security. The author presents two levels of Web Services: the ""kernel"" (ithe SOAP engine which delivers messages from one point to another through various networks), and the ""server side"" (which processes heavy load / requests). The primary objective of this book is execution of applications delivered in a timely fashion. Case studies and examples are provided throughout this book.

  5. Evaluative Measures of Search Engines

    OpenAIRE

    Jitendra Nath Singh; Dr. S.K. Dwivedi

    2012-01-01

    The ability to search and retrieve information from the web efficiently and effectively is great challenge of search engine. Information retrieval on the Web is very different from retrieval in traditional indexed databases because it’s hyper-linked character, the heterogeneity of document types and authoring styles. Thus, since Web retrieval is substantially different from information retrieval, new or revised evaluative measures are required to assess retrieval performance using search engi...

  6. WEB GIS: IMPLEMENTATION ISSUES

    Institute of Scientific and Technical Information of China (English)

    2000-01-01

    With the rapid expansion and development of Internet and WWW (World Wide Web or Web), Web GIS (Web Geographical Information Systen) is becoming ever more popular and as a result numerous sites have added GIS capability on their Web sites. In this paper, the reasons behind developing a Web GIS instead of a “traditional” GIS are first outlined. Then the current status of Web GIS is reviewed, and their implementation methodologies are explored as well.The underlying technologies for developing Web GIS, such as Web Server, Web browser, CGI (Common Gateway Interface), Java, ActiveX, are discussed, and some typical implementation tools from both commercial and public domain are given as well. Finally, the future development direction of Web GIS is predicted.

  7. Next-Gen Search Engines

    Science.gov (United States)

    Gupta, Amardeep

    2005-01-01

    Current search engines--even the constantly surprising Google--seem unable to leap the next big barrier in search: the trillions of bytes of dynamically generated data created by individual web sites around the world, or what some researchers call the "deep web." The challenge now is not information overload, but information overlook.…

  8. A Search Engine Features Comparison.

    Science.gov (United States)

    Vorndran, Gerald

    Until recently, the World Wide Web (WWW) public access search engines have not included many of the advanced commands, options, and features commonly available with the for-profit online database user interfaces, such as DIALOG. This study evaluates the features and characteristics common to both types of search interfaces, examines the Web search…

  9. Raising Reliability of Web Search Tool Research through Replication and Chaos Theory

    OpenAIRE

    Nicholson, Scott

    1999-01-01

    Because the World Wide Web is a dynamic collection of information, the Web search tools (or "search engines") that index the Web are dynamic. Traditional information retrieval evaluation techniques may not provide reliable results when applied to the Web search tools. This study is the result of ten replications of the classic 1996 Ding and Marchionini Web search tool research. It explores the effects that replication can have on transforming unreliable results from one iteration into replica...

  10. Personalized Web Services for Web Information Extraction

    CERN Document Server

    Jarir, Zahi; Erradi, Mahammed

    2011-01-01

    The field of information extraction from the Web emerged with the growth of the Web and the multiplication of online data sources. This paper is an analysis of information extraction methods. It presents a service oriented approach for web information extraction considering both web data management and extraction services. Then we propose an SOA based architecture to enhance flexibility and on-the-fly modification of web extraction services. An implementation of the proposed architecture is proposed on the middleware level of Java Enterprise Edition (JEE) servers.

  11. A web services choreography scenario for interoperating bioinformatics applications

    Directory of Open Access Journals (Sweden)

    Cheung David W

    2004-03-01

    Full Text Available Abstract Background Very often genome-wide data analysis requires the interoperation of multiple databases and analytic tools. A large number of genome databases and bioinformatics applications are available through the web, but it is difficult to automate interoperation because: 1 the platforms on which the applications run are heterogeneous, 2 their web interface is not machine-friendly, 3 they use a non-standard format for data input and output, 4 they do not exploit standards to define application interface and message exchange, and 5 existing protocols for remote messaging are often not firewall-friendly. To overcome these issues, web services have emerged as a standard XML-based model for message exchange between heterogeneous applications. Web services engines have been developed to manage the configuration and execution of a web services workflow. Results To demonstrate the benefit of using web services over traditional web interfaces, we compare the two implementations of HAPI, a gene expression analysis utility developed by the University of California San Diego (UCSD that allows visual characterization of groups or clusters of genes based on the biomedical literature. This utility takes a set of microarray spot IDs as input and outputs a hierarchy of MeSH Keywords that correlates to the input and is grouped by Medical Subject Heading (MeSH category. While the HTML output is easy for humans to visualize, it is difficult for computer applications to interpret semantically. To facilitate the capability of machine processing, we have created a workflow of three web services that replicates the HAPI functionality. These web services use document-style messages, which means that messages are encoded in an XML-based format. We compared three approaches to the implementation of an XML-based workflow: a hard coded Java application, Collaxa BPEL Server and Taverna Workbench. The Java program functions as a web services engine and interoperates

  12. A web services choreography scenario for interoperating bioinformatics applications

    Science.gov (United States)

    de Knikker, Remko; Guo, Youjun; Li, Jin-long; Kwan, Albert KH; Yip, Kevin Y; Cheung, David W; Cheung, Kei-Hoi

    2004-01-01

    Background Very often genome-wide data analysis requires the interoperation of multiple databases and analytic tools. A large number of genome databases and bioinformatics applications are available through the web, but it is difficult to automate interoperation because: 1) the platforms on which the applications run are heterogeneous, 2) their web interface is not machine-friendly, 3) they use a non-standard format for data input and output, 4) they do not exploit standards to define application interface and message exchange, and 5) existing protocols for remote messaging are often not firewall-friendly. To overcome these issues, web services have emerged as a standard XML-based model for message exchange between heterogeneous applications. Web services engines have been developed to manage the configuration and execution of a web services workflow. Results To demonstrate the benefit of using web services over traditional web interfaces, we compare the two implementations of HAPI, a gene expression analysis utility developed by the University of California San Diego (UCSD) that allows visual characterization of groups or clusters of genes based on the biomedical literature. This utility takes a set of microarray spot IDs as input and outputs a hierarchy of MeSH Keywords that correlates to the input and is grouped by Medical Subject Heading (MeSH) category. While the HTML output is easy for humans to visualize, it is difficult for computer applications to interpret semantically. To facilitate the capability of machine processing, we have created a workflow of three web services that replicates the HAPI functionality. These web services use document-style messages, which means that messages are encoded in an XML-based format. We compared three approaches to the implementation of an XML-based workflow: a hard coded Java application, Collaxa BPEL Server and Taverna Workbench. The Java program functions as a web services engine and interoperates with these web

  13. Web Analytics

    OpenAIRE

    Mužík, Zbyněk

    2006-01-01

    Práce se zabývá problematikou měření ukazatelů souvisejících s provozem webových stránek a aplikací a technologickými prostředky k tomu sloužícími ? Web Analytics (WA). Hlavním cílem práce je otestovat a porovnat vybrané zástupce těchto nástrojů a podrobit je srovnání podle objektivních kriterií, dále také kritické zhodnocení možností WA nástrojů obecně. V první části se práce zaměřuje na popis různých způsobů měření provozu na WWW a definuje související metriky. Poskytuje také přehled dostup...

  14. SELECTION OF ONTOLOGY FOR WEB SERVICE DESCRIPTION LANGUAGE TO ONTOLOGY WEB LANGUAGE CONVERSION

    Directory of Open Access Journals (Sweden)

    J. Mannar Mannan

    2014-01-01

    Full Text Available Semantic web is to extend the current human readable web to encoding some of the semantic of resources in a machine processing form. As a Semantic web component, Semantic Web Services (SWS uses a mark-up that makes the data into detailed and sophisticated machine readable way. One such language is Ontology Web Language (OWL. Existing conventional web service annotation can be changed to semantic web service by mapping Web Service Description Language (WSDL with the semantic annotation of OWL-S. In this conversion of WSDL to OWL process, the ontology plays a vital role. Ontology can be stored and retrieved from local repository and selecting the appropriate ontology is a complicated process and this can be achieved by Ontology Searching and Property Matching (OSPM engine. Ontology is stored in the local repository as ontology document and exact matching of ontology for the requested query can be searched using semantic similarity ranking method. High ranked classes of ontology will undergo property matching; here requested concept will be matched with the resulting property. OSPM engine act as the backbone for selecting an exact ontology and reduce the conflict that occurs while selecting the ontology for annotation purpose.

  15. Het WEB leert begrijpen

    CERN Multimedia

    Stroeykens, Steven

    2004-01-01

    The WEB could be much more useful if the computers understood something of information on the Web pages. That explains the goal of the "semantic Web", a project in which takes part, amongst others, Tim Berners Lee, the inventor of the original WEB

  16. Instant responsive web design

    CERN Document Server

    Simmons, Cory

    2013-01-01

    A step-by-step tutorial approach which will teach the readers what responsive web design is and how it is used in designing a responsive web page.If you are a web-designer looking to expand your skill set by learning the quickly growing industry standard of responsive web design, this book is ideal for you. Knowledge of CSS is assumed.

  17. Handbook of web surveys

    NARCIS (Netherlands)

    Bethlehem, J.; Biffignandi, S.

    2012-01-01

    Best practices to create and implementhighly effective web surveys Exclusively combining design and sampling issues, Handbook of Web Surveys presents a theoretical yet practical approach to creating and conducting web surveys. From the history of web surveys to various modes of data collection to ti

  18. Geospatial semantic web

    CERN Document Server

    Zhang, Chuanrong; Li, Weidong

    2015-01-01

    This book covers key issues related to Geospatial Semantic Web, including geospatial web services for spatial data interoperability; geospatial ontology for semantic interoperability; ontology creation, sharing, and integration; querying knowledge and information from heterogeneous data source; interfaces for Geospatial Semantic Web, VGI (Volunteered Geographic Information) and Geospatial Semantic Web; challenges of Geospatial Semantic Web; and development of Geospatial Semantic Web applications. This book also describes state-of-the-art technologies that attempt to solve these problems such as WFS, WMS, RDF, OWL, and GeoSPARQL, and demonstrates how to use the Geospatial Semantic Web technologies to solve practical real-world problems such as spatial data interoperability.

  19. Web Impact Factors for Iranian Universities

    Directory of Open Access Journals (Sweden)

    Alireza Noruzi

    2005-04-01

    Full Text Available This study investigates the Web Impact Factors (WIFs for Iranian universities and introduces a new system of measurement. Counts of links to the web sites of Iranian universities were calculated from the output of AltaVista search engine. The WIFs for Iranian universities were calculated by dividing link page counts by the number of pages found in AltaVista for each university at a given point in time. These WIFs were then compared, to study the impact, visibility, and influence of Iranian university web sites. Overall, Iranian university web sites have a low inlink WIF. While specific features of sites may affect an institution's Web Impact Factor, there is a significant correlation between the proportion of English-language pages at an institution's site and the institution's backlink counts. This indicates that for linguistic reasons, Iranian (Persian-language web sites may not receive and attract the attention that they deserve from the World Wide Web. This raises the possibility that information may be ignored due to linguistic and geographic barriers, and this should be taken into account in the development of the global Web.

  20. Space Physics Data Facility Web Services

    Science.gov (United States)

    Candey, Robert M.; Harris, Bernard T.; Chimiak, Reine A.

    2005-01-01

    The Space Physics Data Facility (SPDF) Web services provides a distributed programming interface to a portion of the SPDF software. (A general description of Web services is available at http://www.w3.org/ and in many current software-engineering texts and articles focused on distributed programming.) The SPDF Web services distributed programming interface enables additional collaboration and integration of the SPDF software system with other software systems, in furtherance of the SPDF mission to lead collaborative efforts in the collection and utilization of space physics data and mathematical models. This programming interface conforms to all applicable Web services specifications of the World Wide Web Consortium. The interface is specified by a Web Services Description Language (WSDL) file. The SPDF Web services software consists of the following components: 1) A server program for implementation of the Web services; and 2) A software developer s kit that consists of a WSDL file, a less formal description of the interface, a Java class library (which further eases development of Java-based client software), and Java source code for an example client program that illustrates the use of the interface.

  1. Web Search Results Summarization Using Similarity Assessment

    Directory of Open Access Journals (Sweden)

    Sawant V.V.

    2014-06-01

    Full Text Available Now day’s internet has become part of our life, the WWW is most important service of internet because it allows presenting information such as document, imaging etc. The WWW grows rapidly and caters to a diversified levels and categories of users. For user specified results web search results are extracted. Millions of information pouring online, users has no time to surf the contents completely .Moreover the information available is repeated or duplicated in nature. This issue has created the necessity to restructure the search results that could yield results summarized. The proposed approach comprises of different feature extraction of web pages. Web page visual similarity assessment has been employed to address the problems in different fields including phishing, web archiving, web search engine etc. In this approach, initially by enters user query the number of search results get stored. The Earth Mover's Distance is used to assessment of web page visual similarity, in this technique take the web page as a low resolution image, create signature of that web page image with color and co-ordinate features .Calculate the distance between web pages by applying EMD method. Compute the Layout Similarity value by using tag comparison algorithm and template comparison algorithm. Textual similarity is computed by using cosine similarity, and hyperlink analysis is performed to compute outward links. The final similarity value is calculated by fusion of layout, text, hyperlink and EMD value. Once the similarity matrix is found clustering is employed with the help of connected component. Finally group of similar web pages i.e. summarized results get displayed to user. Experiment conducted to demonstrate the effectiveness of four methods to generate summarized result on different web pages and user queries also.

  2. Web Science 2015

    OpenAIRE

    Boucher, Andy; Cameron, David; Gaver, William; Hauenstein, Mark; Jarvis, Nadine; Kerridge, Tobie; Michael, Mike; Ovalle, Liliana; Pennington, Sarah; Wilkie, Alex

    2015-01-01

    Web Science 2015 conference exhibition. Web Science is the emergent study of the people and technologies, applications, processes and practices that shape and are shaped by the World Wide Web. Web Science aims to draw together theories, methods and findings from across academic disciplines, and to collaborate with industry, business, government and civil society, to develop knowledge and understanding of the Web: the largest socio-technical infrastructure in human history.

  3. A WebKit-based web crawler%基于WebKit的网络爬虫

    Institute of Scientific and Technical Information of China (English)

    郭津丞; 冯超; 张磊

    2013-01-01

    以AJAX技术为代表的Web应用新技术的出现,赋予了JavaScript更加丰富的功能。但也导致更多的URL以数据形式存在于JavaScript代码中,给网络爬虫的URL提取带来了新的挑战。为了解决这一问题,在此提出了一种基于WebKit的网络爬虫,以WebKit作为爬虫的前端来解析并执行JavaScript。一是实现JavaScript对网页DOM的修改,从而将存在于此类代码中的URL转换成HTML形式并以静态分析方法来提取;二是定位JavaScript页面导航的代码并且劫持输入导航方法及对象的变量以提取变量中的URL。这充分降低了客户端脚本给爬虫带来的障碍,从而更好地提取网页中的URL。%With the emergence of the new web application technologies symbolized by the AJAX technology,the richer functions has been attached to JavaScript,but this also leads to more URL existing in the form of data in JavaScript codes and brings new challenges to the URL extraction of Web crawler. To solve these problems,a WebKit-based web crawler is proposed in this paper. JavaScript is parsed and executed by taking WebKit engine as the front end of the crawler,with which the DOM modification on the web page of JavaScript is implemented to convert URL existing in those codes into HTML style and extract them by static analysis method,and also the page navigation codes of JavaScript is positioned and the input navigation method is intercepted to extract URL in these variables. These two methods has sufficiently reduced the barriers caused by client side script,so URL in the web page can be extracted more perfectly.

  4. Search Engine Optimization through Spanning Forest Generation Algorithm

    Directory of Open Access Journals (Sweden)

    SATYA PAVAN KUMAR SOMAYAJULA

    2011-09-01

    Full Text Available Search engine technology has had to scale dramatically to keep up with the growth of the web. With the tremendous growth of information available to end users through the Web, search engines come to play ever a more critical role. Determining the user intent of Web searches is a difficult problem due to the sparse data available concerning the searcher. We qualitatively analyze samples of queries from seven transaction logs from three different Web search engines containing more than five million queries. The following are our research objectives: Isolate characteristics of informational, navigational, and transactional for Web searching queries by identifying characteristics of each query type that will lead toreal world classification. Validate the taxonomy by automatically classifying a large set of queries from a Web search engine. This paper we deal with now is semantic web search engines is the layeredarchitecture and we use this with relation based page rank algorithm.

  5. Remote Experiments in Control Engineering Education Laboratory

    Directory of Open Access Journals (Sweden)

    Milica B Naumović

    2008-05-01

    Full Text Available This paper presents Automatic Control Engineering Laboratory (ACEL - WebLab, an under-developed, internet-based remote laboratory for control engineering education at the Faculty of Electronic Engineering in Niš. Up to now, the remote laboratory integrates two physical systems (velocity servo system and magnetic levitation system and enables some levels of measurement and control. To perform experiments in ACEL-WebLab, the "LabVIEW Run Time Engine"and a standard web browser are needed.

  6. Overview of the TREC 2013 Federated Web Search Track

    OpenAIRE

    Demeester, Thomas; Trieschnigg, Rudolf Berend; Nguyen, Dong-Phuong; Hiemstra, Djoerd

    2014-01-01

    The TREC Federated Web Search track is intended to promote research related to federated search in a realistic web setting, and hereto provides a large data collection gathered from a series of online search engines. This overview paper discusses the results of the first edition of the track, FedWeb 2013. The focus was on basic challenges in federated search: (1) resource selection, and (2) results merging. After an overview of the provided data collection and the relevance judgments for the ...

  7. [Development of domain specific search engines].

    Science.gov (United States)

    Takai, T; Tokunaga, M; Maeda, K; Kaminuma, T

    2000-01-01

    As cyber space exploding in a pace that nobody has ever imagined, it becomes very important to search cyber space efficiently and effectively. One solution to this problem is search engines. Already a lot of commercial search engines have been put on the market. However these search engines respond with such cumbersome results that domain specific experts can not tolerate. Using a dedicate hardware and a commercial software called OpenText, we have tried to develop several domain specific search engines. These engines are for our institute's Web contents, drugs, chemical safety, endocrine disruptors, and emergent response for chemical hazard. These engines have been on our Web site for testing.

  8. From the Director: Surfing the Web for Health Information

    Science.gov (United States)

    ... authoritative health information from across the World Wide Web. Searching for Reliable Results Most Internet users first visit a search engine — like Google or Yahoo! — when seeking health information. ...

  9. A study on the personalization methods of the web | Hajighorbani ...

    African Journals Online (AJOL)

    A study on the personalization methods of the web. ... the information they need requires the use of contextual and semantic information analysis techniques ... Keywords: personalization, search engine, user preferences, data mining methods ...

  10. Kansei Engineering and Website Design

    DEFF Research Database (Denmark)

    Song, Zheng; Howard, Thomas J.; Achiche, Sofiane

    2012-01-01

    a methodology based on Kansei Engineering, which has done significant work in product and industrial design but not quite been adopted in the IT field, in order to discover implicit emotional needs of users toward web site and transform them into design details. Survey and interview techniques and statistical...... methods were performed in this paper. A prototype web site was produced based on the Kansei results integrated with technical expertise and practical considerations. The results showed that the Kansei Engineering methodology in this paper played a significant role in web site design in terms of satisfying...

  11. Semantic Web Technologies for the Adaptive Web

    DEFF Research Database (Denmark)

    Dolog, Peter

    2007-01-01

    Ontologies and reasoning are the key terms brought into focus by the semantic web community. Formal representation of ontologies in a common data model on the web can be taken as a foundation for adaptive web technologies as well. This chapter describes how ontologies shared on the semantic web...... provide conceptualization for the links which are a main vehicle to access information on the web. The subject domain ontologies serve as constraints for generating only those links which are relevant for the domain a user is currently interested in. Furthermore, user model ontologies provide additional...... means for deciding which links to show, annotate, hide, generate, and reorder. The semantic web technologies provide means to formalize the domain ontologies and metadata created from them. The formalization enables reasoning for personalization decisions. This chapter describes which components...

  12. A Web Service and Interface for Remote Electronic Device Characterization

    Science.gov (United States)

    Dutta, S.; Prakash, S.; Estrada, D.; Pop, E.

    2011-01-01

    A lightweight Web Service and a Web site interface have been developed, which enable remote measurements of electronic devices as a "virtual laboratory" for undergraduate engineering classes. Using standard browsers without additional plugins (such as Internet Explorer, Firefox, or even Safari on an iPhone), remote users can control a Keithley…

  13. 29 CFR 1926.757 - Open web steel joists.

    Science.gov (United States)

    2010-07-01

    ... approval of the project structural engineer of record. (8) Field-bolted joists. (i) Except for steel joists... 29 Labor 8 2010-07-01 2010-07-01 false Open web steel joists. 1926.757 Section 1926.757 Labor... (CONTINUED) SAFETY AND HEALTH REGULATIONS FOR CONSTRUCTION Steel Erection § 1926.757 Open web steel...

  14. Quality of Web-Based Information on Cannabis Addiction

    Science.gov (United States)

    Khazaal, Yasser; Chatton, Anne; Cochand, Sophie; Zullino, Daniele

    2008-01-01

    This study evaluated the quality of Web-based information on cannabis use and addiction and investigated particular content quality indicators. Three keywords ("cannabis addiction," "cannabis dependence," and "cannabis abuse") were entered into two popular World Wide Web search engines. Websites were assessed with a standardized proforma designed…

  15. How to Weave... the Web Into K-8 Science

    Science.gov (United States)

    Wetzel, David R.

    2005-01-01

    Like a search engine for science teachers, How to... "Weave the Web into K-8 Science" is a custom-made guide to bringing the best of the Internet into the classroom. Author David Wetzel has done the work of locating online materials. The book offers resources for Web-based science teaching and learning plus online technical help for both…

  16. Searcher Responsibility for Quality in the Web World.

    Science.gov (United States)

    Halvorson, T. R.

    1998-01-01

    Examines problems with quality of information found on the World Wide Web. Highlights include legal responsibility for complete and accurate searches for clients; rating scales; scope, coverage, and recall in Web search engines; ranking, placement and precision; currency; content quality; and possible solutions. (LRW)

  17. Guiding Students in Finding Information on the Web.

    Science.gov (United States)

    Quible, Zane K.

    1999-01-01

    Argues that business-communication instructors can aid students in their research by introducing them to the terminology and functions of an efficient Web-search process. Discusses the operation of four search tools: Web directories, search engines, indexes, and spiders or robots. Discusses Boolean logic, and other ways to improve the productivity…

  18. Quality of Web-Based Information on Cannabis Addiction

    Science.gov (United States)

    Khazaal, Yasser; Chatton, Anne; Cochand, Sophie; Zullino, Daniele

    2008-01-01

    This study evaluated the quality of Web-based information on cannabis use and addiction and investigated particular content quality indicators. Three keywords ("cannabis addiction," "cannabis dependence," and "cannabis abuse") were entered into two popular World Wide Web search engines. Websites were assessed with a standardized proforma designed…

  19. Hidden Web Data Extraction Using Dynamic Rule Generation

    Directory of Open Access Journals (Sweden)

    Anuradha

    2011-08-01

    Full Text Available World Wide Web is a global information medium of interlinked hypertext documents accessed via computers connected to the internet. Most of the users rely on traditional search engines to search theinformation on the web. These search engines deal with the Surface Web which is a set of Web pages directly accessible through hyperlinks and ignores a large part of the Web called Hidden Web which is hidden to present-day search engines. It lies behind search forms and this part of the web containing an almost endless amount of sources providing high quality information stored in specialized databases can be found in the depths of the WWW. A large amount of this Hidden web is structured i.e Hidden websites contain the information in the form of lists and tables. However visiting dozens of these sites and analyzing the results is very much time consuming task for user. Hence, it is desirable to build a prototype which will minimize user’s effort and give him high quality information in integrated form. This paper proposes a novel method that extracts the data records from the lists and tables of various hidden web sites of same domain using dynamic rule generation and forms a repository which is used for later searching. By searching the data from this repository, user will find the desired data at one place. It reduces the user’s effort to look at various result pages of different hidden websites.

  20. Efficient Approach for Semantic Web Searching Using Markov Model

    Directory of Open Access Journals (Sweden)

    Pradeep Salve

    2012-09-01

    Full Text Available The semantic search usually the web pages for the required information and filter the pages from semantic web searching unnecessary pages by using advanced algorithms. Web pages are vulnerable in answering intelligent semantic search from the user due to the confidence of their consequences on information obtainable in web pages. To get the trusted results semantic web search engines require searching for pages that maintain such information at some place including domain knowledge. The layered model of Semantic Web provides solution to this problem by providing semantic web search based on HMM for optimization of search engines tasks, specialty focusing on how to construct a new model structure to improve the extraction of web pages. We classify the search results using some search engines and some different search keywords provide a significant improvement in search accuracy. Semantic web is segmented from the elicited information of various websites based on their characteristic of semi-structure in order to improve the accuracy and efficiency of the transition matrix. Also, it optimizes the observation probability distribution and the estimation accuracy of state transition sequence by adopting the “voting strategy” and alter Viterbi algorithm. In this paper, we have presented a hybrid system that includes both hidden Markov models and rich markov model that showed the effectiveness of combining implicit search with rich Markov models for a recommender system.

  1. An intelligent method for geographic Web search

    Science.gov (United States)

    Mei, Kun; Yuan, Ying

    2008-10-01

    While the electronically available information in the World-Wide Web is explosively growing and thus increasing, the difficulty to find relevant information is also increasing for search engine user. In this paper we discuss how to constrain web queries geographically. A number of search queries are associated with geographical locations, either explicitly or implicitly. Accurately and effectively detecting the locations where search queries are truly about has huge potential impact on increasing search relevance, bringing better targeted search results, and improving search user satisfaction. Our approach focus on both in the way geographic information is extracted from the web and, as far as we can tell, in the way it is integrated into query processing. This paper gives an overview of a spatially aware search engine for semantic querying of web document. It also illustrates algorithms for extracting location from web documents and query requests using the location ontologies to encode and reason about formal semantics of geographic web search. Based on a real-world scenario of tourism guide search, the application of our approach shows that the geographic information retrieval can be efficiently supported.

  2. Shifts in Search Engine Development: A Review of Past, Present and Future Trends in Research on Search Engines

    OpenAIRE

    Hamid R. Jamali; Saeid Asadi

    2004-01-01

    The World Wide Web has developed fast and many people use search engines to capture information from the Web. This article reviews past, present and future of search engines. Papers published in four major Web and information management conferences were surveyed to track research interests in the last five years. Web search and information retrieval topics such as ranking, filtering and query formulation are still hot topics among researchers. The most important shifts and issues of the futur...

  3. Sounds of Web Advertising

    DEFF Research Database (Denmark)

    Jessen, Iben Bredahl; Graakjær, Nicolai Jørgensgaard

    2010-01-01

    Sound seems to be a neglected issue in the study of web ads. Web advertising is predominantly regarded as visual phenomena–commercial messages, as for instance banner ads that we watch, read, and eventually click on–but only rarely as something that we listen to. The present chapter presents...... an overview of the auditory dimensions in web advertising: Which kinds of sounds do we hear in web ads? What are the conditions and functions of sound in web ads? Moreover, the chapter proposes a theoretical framework in order to analyse the communicative functions of sound in web advertising. The main...... argument is that an understanding of the auditory dimensions in web advertising must include a reflection on the hypertextual settings of the web ad as well as a perspective on how users engage with web content....

  4. Web Crawler Based on Mobile Agent and Java Aglets

    Directory of Open Access Journals (Sweden)

    Md. Abu Kausar

    2013-09-01

    Full Text Available With the huge growth of the Internet, many web pages are available online. Search engines use web crawlers to collect these web pages from World Wide Web for the purpose of storage and indexing. Basically Web Crawler is a program, which finds information from the World Wide Web in a systematic and automated manner. This network load farther will be reduced by using mobile agents.The proposed approach uses mobile agents to crawl the pages. A mobile agent is not bound to the system in which it starts execution. It has the unique ability to transfer itself from one system in a network to another system. The main advantages of web crawler based on Mobile Agents are that the analysis part of the crawling process is done locally rather than remote side. This drastically reduces network load and traffic which can improve the performance and efficiency of the whole crawling process.

  5. Collective Behaviour Learning :A Concept For Filtering Web Pages

    Directory of Open Access Journals (Sweden)

    G. Mercy Bai

    2014-03-01

    Full Text Available The rapid growth of the WWW poses unprecedented challenges for general purpose crawlers and search engines. The Former technique used to crawl web pages was FOCUS (Forum Crawler Under Supervision.This project presents a collective behavior learning algorithm for web crawling. The collective behavior learning algorithm crawl the web pages based on particular keyword. Discriminative learning extracts only the related URL of the particular keyword based on filtering. The goal of this project is to crawl relevant forum content from the web with minimal overhead. The unwanted URL is removed from the web pages and the web page crawling is reduced by using the collective behavior learning. The web pages must be extracted based on certain learning techniques and can be used to collect the unwanted URL’S.

  6. The Ontological Perspectives of the Semantic Web and the Metadata Harvesting Protocol: Applications of Metadata for Improving Web Search.

    Science.gov (United States)

    Fast, Karl V.; Campbell, D. Grant

    2001-01-01

    Compares the implied ontological frameworks of the Open Archives Initiative Protocol for Metadata Harvesting and the World Wide Web Consortium's Semantic Web. Discusses current search engine technology, semantic markup, indexing principles of special libraries and online databases, and componentization and the distinction between data and…

  7. Web Mining and Social Networking

    DEFF Research Database (Denmark)

    Xu, Guandong; Zhang, Yanchun; Li, Lin

    This book examines the techniques and applications involved in the Web Mining, Web Personalization and Recommendation and Web Community Analysis domains, including a detailed presentation of the principles, developed algorithms, and systems of the research in these areas. The applications of web ...... sense of individuals or communities. The volume will benefit both academic and industry communities interested in the techniques and applications of web search, web data management, web mining and web knowledge discovery, as well as web community and social network analysis....

  8. Using Semantic Web Concepts to Achieve Traceability in Cross-disciplinary Development Projects

    OpenAIRE

    2014-01-01

    This thesis contribute to the application of semantic web concepts to achieve traceability in cross-disciplinary development projects. Specifically it has been focused in two specific fields such as mechanical engineering and software engineering.

  9. Effective performance of information retrieval on web by using web crawling

    CERN Document Server

    AbdulNabi, Sk

    2012-01-01

    World Wide Web consists of more than 50 billion pages online. It is highly dynamic i.e. the web continuously introduces new capabilities and attracts many people. Due to this explosion in size, the effective information retrieval system or search engine can be used to access the information. In this paper we have proposed the EPOW (Effective Performance of WebCrawler) architecture. It is a software agent whose main objective is to minimize the overload of a user locating needed information. We have designed the web crawler by considering the parallelization policy. Since our EPOW crawler has a highly optimized system it can download a large number of pages per second while being robust against crashes. We have also proposed to use the data structure concepts for implementation of scheduler & circular Queue to improve the performance of our web crawler.

  10. Web Video Mining: Metadata Predictive Analysis using Classification Techniques

    Directory of Open Access Journals (Sweden)

    Siddu P. Algur

    2016-02-01

    Full Text Available Now a days, the Data Engineering becoming emerging trend to discover knowledge from web audiovisual data such as- YouTube videos, Yahoo Screen, Face Book videos etc. Different categories of web video are being shared on such social websites and are being used by the billions of users all over the world. The uploaded web videos will have different kind of metadata as attribute information of the video data. The metadata attributes defines the contents and features/characteristics of the web videos conceptually. Hence, accomplishing web video mining by extracting features of web videos in terms of metadata is a challenging task. In this work, effective attempts are made to classify and predict the metadata features of web videos such as length of the web videos, number of comments of the web videos, ratings information and view counts of the web videos using data mining algorithms such as Decision tree J48 and navie Bayesian algorithms as a part of web video mining. The results of Decision tree J48 and navie Bayesian classification models are analyzed and compared as a step in the process of knowledge discovery from web videos.

  11. Shifts in search engine development: A review of past, present and future trends in research on search engines

    Directory of Open Access Journals (Sweden)

    Hamid R. Jamali

    2004-12-01

    Full Text Available The World Wide Web has developed fast and many people use search engines to capture information from the Web. This article reviews past, present and future of search engines. Papers published in four major Web and information management conferences were surveyed to track research interests in the last five years. Web search and information retrieval topics such as ranking, filtering and query formulation are still hot topics among researchers. The most important shifts and issues of the future of search engines are mentioned too. Search engine companies are trying to capture the Deep Web and extract structured data to offer high quality results. Using Web page structure, shared search engines, expert recommendations and different mobile search facilities seem to be features of the next generation of search engines.

  12. Why Evaluating Semantic Web Applications is Difficult

    NARCIS (Netherlands)

    Ossenbruggen, J.R. van; Amin, A.K.; Hildebrand, M.

    2008-01-01

    This position paper discusses our experience in evaluating our cultural search and annotation engine. We identify three aspects that determine the quality of a semantic web application as a whole, namely: the quality of data set, the quality of underlying search and inference software and the qualit

  13. Towards a semantic web layered architecture

    CSIR Research Space (South Africa)

    Gerber, AJ

    2007-02-01

    Full Text Available .R. Layered Architecture(s): Princi- ples and Practice in Concurrent and Distributed Systems. In 1997 Workshop on Engineering of Computer-Based Systems (ECBS ’97) 1997, pp. 312– 320. [46] THURAISINGHAM B. Security Issues for the Se- mantic Web...

  14. Preparing SCORM for the semantic web

    NARCIS (Netherlands)

    Aroyo, Lora; Pokraev, Stanislav; Brussee, Rogier

    2003-01-01

    In this paper we argue that the effort within the context of Semantic Web research, such as RDF and DAML-S. will allow for better knowledge representation and engineering of educational systems and easier integration of e-learning with other business processes. We also argue that existing educationa

  15. Web Adventures in K-12 Science.

    Science.gov (United States)

    Friedman, Edward A.; McGrath, Beth; Baron, Joshua

    1997-01-01

    Describes activities at the Center for Improved Engineering and Science Education at Stevens Institute of Technology (New Jersey) that have explored applications of the Internet in elementary and secondary school science classrooms. Highlights include working with real-time data, teacher training for the Web, and examples of curriculum activities.…

  16. A Plausible Comprehensive Web Intelligent System for Investigation of Web User Behaviour Adaptable to Incremental Mining

    Directory of Open Access Journals (Sweden)

    V.V.R. Maheswara Rao

    2010-08-01

    Full Text Available With the continued increase in the usage of the World Wide Web (WWW Web mining has beenestablished as an important area of research. The WWW is a vast repository of unstructured information,in the form of interrelated files, distributed on numerous web servers over wide geographical regions.Web mining deals with the discovering and analyzing of useful information from the WWW. Web usagemining focuses on investigating the potential knowledge from the browsing patterns of users and to findthe correlation between the pages on analysis. To proceed towards web intelligence, obviating the needfor human interaction, need to incorporate and embed artificial intelligence into web tools. Beforeapplying mining techniques, the data in the web log has to be pre-processed, integrated and transformed.The data pre-processing stage is the most important phase in the process of web mining and is criticaland complex in successful extraction of useful data. The web log is non scalable, impractical anddistributed in nature thus conventional data pre-processing techniques are proved to be not suitable asthey assume that the data is static. Hence intelligent system is required for capable of pre processingweblog efficiently. Due to the incremental nature of the web log, it is necessary for web miners to useincremental mining techniques to extract the usage patterns and study the visiting characteristics of user,hence one can require a comprehensive algorithm which reduces the computing cost significantly.This paper introduces an Intelligent System IPS for pre-processing of web log, in addition a learningalgorithm IFP-tree model is proposed for pattern recognition. The Intelligent Pre-processing System(IPS can differentiate human user and web search engine accesses intelligently in less time, and discardssearch engine accesses. The present system reduces the error rate and improves significant learningperformance of the algorithm. The Incremental Frequent Pattern Tree

  17. Web 2.0

    CERN Document Server

    Han, Sam

    2012-01-01

    Web 2.0 is a highly accessible introductory text examining all the crucial discussions and issues which surround the changing nature of the World Wide Web. It not only contextualises the Web 2.0 within the history of the Web, but also goes on to explore its position within the broader dispositif of emerging media technologies.The book uncovers the connections between diverse media technologies including mobile smart phones, hand-held multimedia players, ""netbooks"" and electronic book readers such as the Amazon Kindle, all of which are made possible only by the Web 2.0. In addition, Web 2.0 m

  18. Handbook of web surveys

    CERN Document Server

    Bethlehem, Jelke

    2011-01-01

    BEST PRACTICES TO CREATE AND IMPLEMENTHIGHLY EFFECTIVE WEB SURVEYS Exclusively combining design and sampling issues, Handbook of Web Surveys presents a theoretical yet practical approach to creating and conducting web surveys. From the history of web surveys to various modes of data collection to tips for detecting error, this book thoroughly introduces readers to the this cutting-edge technique and offers tips for creating successful web surveys. The authors provide a history of web surveys and go on to explore the advantages and disadvantages of this mode of dat

  19. Creating OGC Web Processing Service workflows using a web-based editor

    Science.gov (United States)

    de Jesus, J.; Walker, P.; Grant, M.

    2012-04-01

    The OGC WPS (Web Processing Service) specifies how geospatial algorithms may be accessed in an SOA (Service Oriented Architecture). Service providers can encode both simple and sophisticated algorithms as WPS processes and publish them as web services. These services are not only useful individually but may be built into complex processing chains (workflows) that can solve complex data analysis and/or scientific problems. The NETMAR project has extended the Web Processing Service (WPS) framework to provide transparent integration between it and the commonly used WSDL (Web Service Description Language) that describes the web services and its default SOAP (Simple Object Access Protocol) binding. The extensions allow WPS services to be orchestrated using commonly used tools (in this case Taverna Workbench, but BPEL based systems would also be an option). We have also developed a WebGUI service editor, based on HTML5 and the WireIt! Javascript API, that allows users to create these workflows using only a web browser. The editor is coded entirely in Javascript and performs all XSLT transformations needed to produce a Taverna compatible (T2FLOW) workflow description which can be exported and run on a local Taverna Workbench or uploaded to a web-based orchestration server and run there. Here we present the NETMAR WebGUI service chain editor and discuss the problems associated with the development of a WebGUI for scientific workflow editing; content transformation into the Taverna orchestration language (T2FLOW/SCUFL); final orchestration in the Taverna engine and how to deal with the large volumes of data being transferred between different WPS services (possibly running on different servers) during workflow orchestration. We will also demonstrate using the WebGUI for creating a simple workflow making use of published web processing services, showing how simple services may be chained together to produce outputs that would previously have required a GIS (Geographic

  20. ONTOLOGY BASED WEB PAGE ANNOTATION FOR EFFECTIVE INFORMATION RETRIEVAL

    Directory of Open Access Journals (Sweden)

    S.Kalarani

    2010-11-01

    Full Text Available Today’s World Wide Web has large volume of data – billions of documents. So it is a time consuming process to discover effective knowledge from the input data. With today's keyword approach the amount of time and effort required to find the right information is directly proportional to the amount of information on the web.The web has grown exponentially and people are forced to spend more and more time in search for the information they are looking for. Lack of personalization as well as inability to easily separate commercial from non-commercial searches is among other limitations of today's web search technologies. This paper proposes a prototype relation-based search engine. “OntoLook” which has been designed in a virtual semantic web environment. The architecture has been proposed. The Semantic Web is well recognized as an effective infrastructure to enhance visibility of knowledge on the Web. The core of the Semantic Web is “ontology”, which is used to explicitly represent our conceptualizations. Ontology engineering in the Semantic Web isprimarily supported by languages such as RDF, RDFS and OWL. This paper discusses the requirements of ontology in the context of the Web, compares the above three languages with existing knowledge representation formalisms, and surveys tools for managing and applying ontology. Advantages of using ontology in both knowledge-base-style and database-style applications are demonstrated using one real world applications.

  1. WikiWikiWebs: New Ways to Communicate in a Web Environment

    Directory of Open Access Journals (Sweden)

    Brenda Chawner

    2006-03-01

    Full Text Available This paper introduces WikiWikiWeb software, also known as Wiki, for use in library and information management contexts. Wikis provide an environment for Web-based collaboration and can also be used for Web site content management. The article includes an overview of the history and development of Wiki, as well as discussing basic and advanced Wiki features. It compares three Wiki engines and describes seven case studies of real-world library and library-related Wiki applications. The paper concludes with a discussion of factors that can contribute to a successful Wiki project.

  2. Web document clustering using hyperlink structures

    Energy Technology Data Exchange (ETDEWEB)

    He, Xiaofeng; Zha, Hongyuan; Ding, Chris H.Q; Simon, Horst D.

    2001-05-07

    With the exponential growth of information on the World Wide Web there is great demand for developing efficient and effective methods for organizing and retrieving the information available. Document clustering plays an important role in information retrieval and taxonomy management for the World Wide Web and remains an interesting and challenging problem in the field of web computing. In this paper we consider document clustering methods exploring textual information hyperlink structure and co-citation relations. In particular we apply the normalized cut clustering method developed in computer vision to the task of hyperdocument clustering. We also explore some theoretical connections of the normalized-cut method to K-means method. We then experiment with normalized-cut method in the context of clustering query result sets for web search engines.

  3. Discovery and Selection of Semantic Web Services

    CERN Document Server

    Wang, Xia

    2013-01-01

    For advanced web search engines to be able not only to search for semantically related information dispersed over different web pages, but also for semantic services providing certain functionalities, discovering semantic services is the key issue. Addressing four problems of current solution, this book presents the following contributions. A novel service model independent of semantic service description models is proposed, which clearly defines all elements necessary for service discovery and selection. It takes service selection as its gist and improves efficiency. Corresponding selection algorithms and their implementation as components of the extended Semantically Enabled Service-oriented Architecture in the Web Service Modeling Environment are detailed. Many applications of semantic web services, e.g. discovery, composition and mediation, can benefit from a general approach for building application ontologies. With application ontologies thus built, services are discovered in the same way as with single...

  4. Flow Webs: Mechanism and Architecture for the Implementation of Sensor Webs

    Science.gov (United States)

    Gorlick, M. M.; Peng, G. S.; Gasster, S. D.; McAtee, M. D.

    2006-12-01

    -time demands. Flows are the connective tissue of flow webs—massive computational engines organized as directed graphs whose nodes are semi-autonomous components and whose edges are flows. The individual components of a flow web may themselves be encapsulated flow webs. In other words, a flow web subgraph may be presented to a yet larger flow web as a single, seamless component. Flow webs, at all levels, may be edited and modified while still executing. Within a flow web individual components may be added, removed, started, paused, halted, reparameterized, or inspected. The topology of a flow web may be changed at will. Thus, flow webs exhibit an extraordinary degree of adaptivity and robustness as they are explicitly designed to be modified on the fly, an attribute well suited for dynamic model interactions in sensor webs. We describe our concept for a sensor web, implemented as a flow web, in the context of a wildfire disaster management system for the southern California region. Comprehensive wildfire management requires cooperation among multiple agencies. Flow webs allow agencies to share resources in exactly the manner they choose. We will explain how to employ flow webs and agents to integrate satellite remote sensing data, models, in-situ sensors, UAVs and other resources into a sensor web that interconnects organizations and their disaster management tools in a manner that simultaneously preserves their independence and builds upon the individual strengths of agency-specific models and data sources.

  5. Acquiring geographical data with web harvesting

    Science.gov (United States)

    Dramowicz, K.

    2016-04-01

    Many websites contain very attractive and up to date geographical information. This information can be extracted, stored, analyzed and mapped using web harvesting techniques. Poorly organized data from websites are transformed with web harvesting into a more structured format, which can be stored in a database and analyzed. Almost 25% of web traffic is related to web harvesting, mostly while using search engines. This paper presents how to harvest geographic information from web documents using the free tool called the Beautiful Soup, one of the most commonly used Python libraries for pulling data from HTML and XML files. It is a relatively easy task to process one static HTML table. The more challenging task is to extract and save information from tables located in multiple and poorly organized websites. Legal and ethical aspects of web harvesting are discussed as well. The paper demonstrates two case studies. The first one shows how to extract various types of information about the Good Country Index from the multiple web pages, load it into one attribute table and map the results. The second case study shows how script tools and GIS can be used to extract information from one hundred thirty six websites about Nova Scotia wines. In a little more than three minutes a database containing one hundred and six liquor stores selling these wines is created. Then the availability and spatial distribution of various types of wines (by grape types, by wineries, and by liquor stores) are mapped and analyzed.

  6. With News Search Engines

    Science.gov (United States)

    Gunn, Holly

    2005-01-01

    Although there are many news search engines on the Web, finding the news items one wants can be challenging. Choosing appropriate search terms is one of the biggest challenges. Unless one has seen the article that one is seeking, it is often difficult to select words that were used in the headline or text of the article. The limited archives of…

  7. Assessing Bias in Search Engines.

    Science.gov (United States)

    Mowshowitz, Abbe; Kawaguchi, Akira

    2002-01-01

    Addresses the measurement of bias in search engines on the Web, defining bias as the balance and representation of items in a collection retrieved from a database for a set of queries. Assesses bias by measuring the deviation from the ideal of the distribution produced by a particular search engine. (Author/LRW)

  8. EVOLUTION OF THE WORLD WIDE WEB: FROM WEB 1.0 TO WEB 4.0

    Directory of Open Access Journals (Sweden)

    Sareh Aghaei

    2012-02-01

    Full Text Available The World Wide Web as the largest information construct has had much progress since its advent. Thispaper provides a background of the evolution of the web from web 1.0 to web 4.0. Web 1.0 as a web ofinformation connections, Web 2.0 as a web of people connections, Web 3.0 as a web of knowledgeconnections and web 4.0 as a web of intelligence connections are described as four generations of the webin the paper.

  9. Declarative Access Control for WebDSL: Combining Language Integration and Separation of Concerns

    NARCIS (Netherlands)

    Groenewegen, D.; Visser, E.

    2008-01-01

    Preprint of paper published in: ICWE 2008 - 8th International Conference on Web Engineering, 14-18 July 2008; doi:10.1109/ICWE.2008.15 In this paper, we present the extension of WebDSL, a domain-specific language for web application development, with abstractions for declarative definition of acces

  10. On the visibility of information on the Web: an exploratory experimental approach

    NARCIS (Netherlands)

    Wouters, P.; Reddy, C.; Aguillo, I.

    2006-01-01

    On the Web, information that is not presented by the search engine in response to a specific query is in fact inaccessible. This problem has been defined as the invisible web or the deep web problem. Specific characteristics of the format of information may make it inaccessible to the crawlers that

  11. On the visibility of information on the Web: an exploratory experimental approach

    NARCIS (Netherlands)

    Wouters, P.; Reddy, C.; Aguillo, I.

    2006-01-01

    On the Web, information that is not presented by the search engine in response to a specific query is in fact inaccessible. This problem has been defined as the invisible web or the deep web problem. Specific characteristics of the format of information may make it inaccessible to the crawlers that

  12. Graph Structure in Three National Academic Webs: Power Laws with Anomalies.

    Science.gov (United States)

    Thelwall, Mike; Wilkinson, David

    2003-01-01

    Explains how the Web can be modeled as a mathematical graph and analyzes the graph structures of three national university publicly indexable Web sites from Australia, New Zealand, and the United Kingdom. Topics include commercial search engines and academic Web link research; method-analysis environment and data sets; and power laws. (LRW)

  13. A Preliminary Mapping of Web Queries Using Existing Image Query Schemes.

    Science.gov (United States)

    Jansen, Bernard J.

    End user searching on the Web has become the primary method of locating images for many people. This study investigates the nature of Web image queries by attempting to map them to known image classification schemes. In this study, approximately 100,000 image queries from a major Web search engine were collected in 1997, 1999, and 2001. A…

  14. EPA Web Taxonomy

    Data.gov (United States)

    U.S. Environmental Protection Agency — EPA's Web Taxonomy is a faceted hierarchical vocabulary used to tag web pages with terms from a controlled vocabulary. Tagging enables search and discovery of EPA's...

  15. Chemical Search Web Utility

    Data.gov (United States)

    U.S. Environmental Protection Agency — The Chemical Search Web Utility is an intuitive web application that allows the public to easily find the chemical that they are interested in using, and which...

  16. Practical web development

    CERN Document Server

    Wellens, Paul

    2015-01-01

    This book is perfect for beginners who want to get started and learn the web development basics, but also offers experienced developers a web development roadmap that will help them to extend their capabilities.

  17. Wordpress web application development

    CERN Document Server

    Ratnayake, Rakhitha Nimesh

    2015-01-01

    This book is intended for WordPress developers and designers who want to develop quality web applications within a limited time frame and for maximum profit. Prior knowledge of basic web development and design is assumed.

  18. AN OVERVIEW OF SEARCHING AND DISCOVERING WEB BASED INFORMATION RESOURCES

    Directory of Open Access Journals (Sweden)

    Cezar VASILESCU

    2010-01-01

    Full Text Available The Internet becomes for most of us a daily used instrument, for professional or personal reasons. We even do not remember the times when a computer and a broadband connection were luxury items. More and more people are relying on the complicated web network to find the needed information.This paper presents an overview of Internet search related issues, upon search engines and describes the parties and the basic mechanism that is embedded in a search for web based information resources. Also presents ways to increase the efficiency of web searches, through a better understanding of what search engines ignore at websites content.

  19. Survey of Web Technologies

    OpenAIRE

    Špoljar, Boris

    2011-01-01

    The World Wide Web bas become an important platform for developing and running applications. A vital process while developing web applications is the choice of web technologies, on which the application will be build. The developers face a dizzying array of platforms, languages, frameworks and technical artifacts to choose from. The decison carries consequences on most other decisions in the development process. Thesis contains analisis, classifications and comparison of web technologies s...

  20. Semiautomatic Web service generation

    OpenAIRE

    Fuentes, José María de; Corella, Miguel Ángel; Castells, Pablo; Rico, Mariano

    2005-01-01

    Proceedings of the IADIS International Conference WWW/Internet 2005, held in Lisbon (Portugal). The lack of a critical mass of actually deployed web services, semantic or not, is an important hurdle for the advancement and further innovation in web service technologies. In this paper we introduce Federica, a platform for semi-automatic generation and implementation of semantic web services by exploiting existing web applications published in internet. Federica generates semantical...