WorldWideScience

Sample records for generation web part

  1. Web Tools: The Second Generation

    Science.gov (United States)

    Pascopella, Angela

    2008-01-01

    Web 2.0 tools and technologies, or second generation tools, help districts to save time and money, and eliminate the need to transfer or move files back and forth across computers. Many Web 2.0 tools help students think critically and solve problems, which falls under the 21st-century skills. The second-generation tools are growing in popularity…

  2. Towards Second and Third Generation Web-Based Multimedia

    NARCIS (Netherlands)

    J.R. van Ossenbruggen (Jacco); J.P.T.M. Geurts (Joost); F.J. Cornelissen; L. Rutledge (Lloyd); L. Hardman (Lynda)

    2001-01-01

    textabstractFirst generation Web-content encodes information in handwritten (HTML) Web pages. Second generation Web content generates HTML pages on demand, e.g. by filling in templates with content retrieved dynamically from a database or transformation of structured documents using style sheets

  3. Towards Second and Third Generation Web-Based Multimedia

    OpenAIRE

    Ossenbruggen, Jacco; Geurts, Joost; Cornelissen, F.J.; Rutledge, Lloyd; Hardman, Lynda

    2001-01-01

    textabstractFirst generation Web-content encodes information in handwritten (HTML) Web pages. Second generation Web content generates HTML pages on demand, e.g. by filling in templates with content retrieved dynamically from a database or transformation of structured documents using style sheets (e.g. XSLT). Third generation Web pages will make use of rich markup (e.g. XML) along with metadata (e.g. RDF) schemes to make the content not only machine readable but also machine processable - a ne...

  4. Next generation of weather generators on web service framework

    Science.gov (United States)

    Chinnachodteeranun, R.; Hung, N. D.; Honda, K.; Ines, A. V. M.

    2016-12-01

    Weather generator is a statistical model that synthesizes possible realization of long-term historical weather in future. It generates several tens to hundreds of realizations stochastically based on statistical analysis. Realization is essential information as a crop modeling's input for simulating crop growth and yield. Moreover, they can be contributed to analyzing uncertainty of weather to crop development stage and to decision support system on e.g. water management and fertilizer management. Performing crop modeling requires multidisciplinary skills which limit the usage of weather generator only in a research group who developed it as well as a barrier for newcomers. To improve the procedures of performing weather generators as well as the methodology to acquire the realization in a standard way, we implemented a framework for providing weather generators as web services, which support service interoperability. Legacy weather generator programs were wrapped in the web service framework. The service interfaces were implemented based on an international standard that was Sensor Observation Service (SOS) defined by Open Geospatial Consortium (OGC). Clients can request realizations generated by the model through SOS Web service. Hierarchical data preparation processes required for weather generator are also implemented as web services and seamlessly wired. Analysts and applications can invoke services over a network easily. The services facilitate the development of agricultural applications and also reduce the workload of analysts on iterative data preparation and handle legacy weather generator program. This architectural design and implementation can be a prototype for constructing further services on top of interoperable sensor network system. This framework opens an opportunity for other sectors such as application developers and scientists in other fields to utilize weather generators.

  5. Ontology Enabled Generation of Embedded Web Services

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Zhang, Weishan; Soares, Goncalo Teofilo Afonso Pinheiro

    2008-01-01

    Web services are increasingly adopted as a service provision mechanism in pervasive computing environments. Implementing web services on networked, embedded devices raises a number of challenges, for example efficiency of web services, handling of variability and dependencies of hardware...... and software platforms, and of devices state and context changes. To address these challenges, we developed a Web service compiler, Limbo, in which Web Ontology Language (OWL) ontologies are used to make the Limbo compiler aware of its compilation context, such as targeted hardware and software. At the same...... time, knowledge on device details, platform dependencies, and resource/power consumption is built into the supporting ontologies, which are used to configure Limbo for generating resource efficient web service code. A state machine ontology is used to generate stub code to facilitate handling of state...

  6. Ontology Enabled Generation of Embedded Web Services

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Zhang, Weishan; Soares, Goncalo Teofilo Afonso Pinheiro

    2008-01-01

    and software platforms, and of devices state and context changes. To address these challenges, we developed a Web service compiler, Limbo, in which Web Ontology Language (OWL) ontologies are used to make the Limbo compiler aware of its compilation context, such as targeted hardware and software. At the same...... time, knowledge on device details, platform dependencies, and resource/power consumption is built into the supporting ontologies, which are used to configure Limbo for generating resource efficient web service code. A state machine ontology is used to generate stub code to facilitate handling of state...

  7. Web Page Recommendation Using Web Mining

    OpenAIRE

    Modraj Bhavsar; Mrs. P. M. Chavan

    2014-01-01

    On World Wide Web various kind of content are generated in huge amount, so to give relevant result to user web recommendation become important part of web application. On web different kind of web recommendation are made available to user every day that includes Image, Video, Audio, query suggestion and web page. In this paper we are aiming at providing framework for web page recommendation. 1) First we describe the basics of web mining, types of web mining. 2) Details of each...

  8. The Semantic Web: opportunities and challenges for next-generation Web applications

    Directory of Open Access Journals (Sweden)

    2002-01-01

    Full Text Available Recently there has been a growing interest in the investigation and development of the next generation web - the Semantic Web. While most of the current forms of web content are designed to be presented to humans, but are barely understandable by computers, the content of the Semantic Web is structured in a semantic way so that it is meaningful to computers as well as to humans. In this paper, we report a survey of recent research on the Semantic Web. In particular, we present the opportunities that this revolution will bring to us: web-services, agent-based distributed computing, semantics-based web search engines, and semantics-based digital libraries. We also discuss the technical and cultural challenges of realizing the Semantic Web: the development of ontologies, formal semantics of Semantic Web languages, and trust and proof models. We hope that this will shed some light on the direction of future work on this field.

  9. The Aalborg Survey / Part 1 - Web Based Survey

    DEFF Research Database (Denmark)

    Harder, Henrik; Christensen, Cecilie Breinholm

    Background and purpose The Aalborg Survey consists of four independent parts: a web, GPS and an interview based survey and a literature study, which together form a consistent investigation and research into use of urban space, and specifically into young people’s use of urban space: what young......) and the research focus within the cluster of Mobility and Tracking Technologies (MoTT), AAU. Summary / Part 1 Web Base Survey The 1st part of the research project Diverse Urban Spaces (DUS) has been carried out during the period from December 1st 2007 to February 1st 2008 as a Web Based Survey of the 27.040 gross...... [statistikbanken.dk, a] young people aged 14-23 living in Aalborg Municipality in 2008. The web based questionnaire has been distributed among the group of young people studying at upper secondary schools in Aalborg, i.e. 7.680 young people [statistikbanken.dk, b]. The resulting data from those respondents who...

  10. Web service based system for generating input data sets

    International Nuclear Information System (INIS)

    Kralev, Velin; Kraleva, Radoslava

    2011-01-01

    This article deals with a three-layer architectural model of a distributed information system based on Web services, which will be used for automatic generation of sets of input data. The information system will be constructed of a client layer, a service layer and of a data layer. The web services as a tool of developing distributed software systems will be presented briefly. A web service and the implementation of its web methods will be described. A way to use the developed web methods in real application will be proposed. Keywords: web services

  11. Generating an Automated Test Suite by Variable Strength Combinatorial Testing for Web Services

    Directory of Open Access Journals (Sweden)

    Yin Li

    2016-09-01

    Full Text Available Testing Web Services has become the spotlight of software engineering as an important means to assure the quality of Web application. Due to lacking of graphic interface and source code, Web services need an automated testing method, which is an important part in efficiently designing and generating test suite. However, the existing testing methods may lead to the redundancy of test suite and the decrease of fault-detecting ability since it cannot handle scenarios where the strengths of the different interactions are not uniform. With the purpose of solving this problem, firstly the formal tree model based on WSDL is constructed and the actual interaction relationship of each node is made sufficient consideration into, then the combinatorial testing is proposed to generate variable strength combinatorial test suite based on One-test-at-a-time strategy. At last test cases are minimized according to constraint rules. The results show that compared with conventional random testing, the proposed approach can detect more errors with the same amount of test cases which turning out to be more ideal than existing ones in size.

  12. VennDiagramWeb: a web application for the generation of highly customizable Venn and Euler diagrams.

    Science.gov (United States)

    Lam, Felix; Lalansingh, Christopher M; Babaran, Holly E; Wang, Zhiyuan; Prokopec, Stephenie D; Fox, Natalie S; Boutros, Paul C

    2016-10-03

    Visualization of data generated by high-throughput, high-dimensionality experiments is rapidly becoming a rate-limiting step in computational biology. There is an ongoing need to quickly develop high-quality visualizations that can be easily customized or incorporated into automated pipelines. This often requires an interface for manual plot modification, rapid cycles of tweaking visualization parameters, and the generation of graphics code. To facilitate this process for the generation of highly-customizable, high-resolution Venn and Euler diagrams, we introduce VennDiagramWeb: a web application for the widely used VennDiagram R package. VennDiagramWeb is hosted at http://venndiagram.res.oicr.on.ca/ . VennDiagramWeb allows real-time modification of Venn and Euler diagrams, with parameter setting through a web interface and immediate visualization of results. It allows customization of essentially all aspects of figures, but also supports integration into computational pipelines via download of R code. Users can upload data and download figures in a range of formats, and there is exhaustive support documentation. VennDiagramWeb allows the easy creation of Venn and Euler diagrams for computational biologists, and indeed many other fields. Its ability to support real-time graphics changes that are linked to downloadable code that can be integrated into automated pipelines will greatly facilitate the improved visualization of complex datasets. For application support please contact Paul.Boutros@oicr.on.ca.

  13. MadGraph/MadEvent. The new web generation

    International Nuclear Information System (INIS)

    Alwall, J.

    2007-01-01

    The new web-based version of the automatized process and event generator MadGraph/MadEvent is now available. Recent developments are: New models, notably MSSM, 2HDM and a framework for addition of user-defined models, inclusive sample generation and on-line hadronization and detector simulation. Event generation can be done on-line on any of our clusters. (author)

  14. Facilitating Student-Generated Content Using Web 2.0 Technologies

    Science.gov (United States)

    Lee, Eunbae

    2011-01-01

    Web 2.0 technologies have created a trend of user-generated content by supporting media production, collaboration, communication, and dissemination. User-generated content is translated into student-generated content (SGC) in education. SGC engages learners in an authentic project that fosters students' autonomy, creativity, and real-world…

  15. Clever generation of rich SPARQL queries from annotated relational schema: application to Semantic Web Service creation for biological databases.

    Science.gov (United States)

    Wollbrett, Julien; Larmande, Pierre; de Lamotte, Frédéric; Ruiz, Manuel

    2013-04-15

    In recent years, a large amount of "-omics" data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic.

  16. Designing and Implementing Weather Generators as Web Services

    Directory of Open Access Journals (Sweden)

    Rassarin Chinnachodteeranun

    2016-12-01

    Full Text Available Climate and weather realizations are essential inputs for simulating crop growth and yields to analyze the risks associated with future conditions. To simplify the procedure of generating weather realizations and make them available over the Internet, we implemented novel mechanisms for providing weather generators as web services, as well as a mechanism for sharing identical weather realizations given a climatological information. A web service for preparing long-term climate data was implemented based on an international standard, Sensor Observation Service (SOS. The weather generator services, which are the core components of the framework, analyze climatological data, and can take seasonal climate forecasts as inputs for generating weather realizations. The generated weather realizations are encoded in a standard format, which are ready for use to crop modeling. All outputs are generated in SOS standard, which broadens the extent of data sharing and interoperability with other sectoral applications, e.g., water resources management. These services facilitate the development of other applications requiring input weather realizations, as these can be obtained easily by just calling the service. The workload of analysts related to data preparation and handling of legacy weather generator programs can be reduced. The architectural design and implementation presented here can be used as a prototype for constructing further services on top of an interoperable sensor network system.

  17. WEB-IS2: Next Generation Web Services Using Amira Visualization Package

    Science.gov (United States)

    Yang, X.; Wang, Y.; Bollig, E. F.; Kadlec, B. J.; Garbow, Z. A.; Yuen, D. A.; Erlebacher, G.

    2003-12-01

    Amira (www.amiravis.com) is a powerful 3-D visualization package and has been employed recently by the science and engineering communities to gain insight into their data. We present a new web-based interface to Amira, packaged in a Java applet. We have developed a module called WEB-IS/Amira (WEB-IS2), which provides web-based access to Amira. This tool allows earth scientists to manipulate Amira controls remotely and to analyze, render and view large datasets over the internet, without regard for time or location. This could have important ramifications for GRID computing. The design of our implementation will soon allow multiple users to visually collaborate by manipulating a single dataset through a variety of client devices. These clients will only require a browser capable of displaying Java applets. As the deluge of data continues, innovative solutions that maximize ease of use without sacrificing efficiency or flexibility will continue to gain in importance, particularly in the Earth sciences. Major initiatives, such as Earthscope (http://www.earthscope.org), which will generate at least a terabyte of data daily, stand to profit enormously by a system such as WEB-IS/Amira (WEB-IS2). We discuss our use of SOAP (Livingston, D., Advanced SOAP for Web development, Prentice Hall, 2002), a novel 2-way communication protocol, as a means of providing remote commands, and efficient point-to-point transfer of binary image data. We will present our initial experiences with the use of Naradabrokering (www.naradabrokering.org) as a means to decouple clients and servers. Information is submitted to the system as a published item, while it is retrieved through a subscription mechanisms, via what is known as "topics". These topic headers, their contents, and the list of subscribers are automatically tracked by Naradabrokering. This novel approach promises a high degree of fault tolerance, flexibility with respect to client diversity, and language independence for the

  18. Web 2.0 and the Net Generation - A Critical Perspective

    DEFF Research Database (Denmark)

    Ryberg, Thomas

    2012-01-01

    In the recent years, social media and web 2.0 have been hot topics within educational debates and within the research area of networked learning. The latter is evident from symposia and papers from the last years' networked learning conferences, but also European research projects, special issues......, and books have revolved around social media, web 2.0, personal learning environments, student-centred learning, and student-generated content. Alongside these internet developments we have witnessed debates on what schools and universities can do to cater to the 'net-generation' or the 'digital natives' in...

  19. Clever generation of rich SPARQL queries from annotated relational schema: application to Semantic Web Service creation for biological databases

    Science.gov (United States)

    2013-01-01

    Background In recent years, a large amount of “-omics” data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. Results We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. Conclusions BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic. PMID:23586394

  20. TermGenie - a web-application for pattern-based ontology class generation.

    Science.gov (United States)

    Dietze, Heiko; Berardini, Tanya Z; Foulger, Rebecca E; Hill, David P; Lomax, Jane; Osumi-Sutherland, David; Roncaglia, Paola; Mungall, Christopher J

    2014-01-01

    Biological ontologies are continually growing and improving from requests for new classes (terms) by biocurators. These ontology requests can frequently create bottlenecks in the biocuration process, as ontology developers struggle to keep up, while manually processing these requests and create classes. TermGenie allows biocurators to generate new classes based on formally specified design patterns or templates. The system is web-based and can be accessed by any authorized curator through a web browser. Automated rules and reasoning engines are used to ensure validity, uniqueness and relationship to pre-existing classes. In the last 4 years the Gene Ontology TermGenie generated 4715 new classes, about 51.4% of all new classes created. The immediate generation of permanent identifiers proved not to be an issue with only 70 (1.4%) obsoleted classes. TermGenie is a web-based class-generation system that complements traditional ontology development tools. All classes added through pre-defined templates are guaranteed to have OWL equivalence axioms that are used for automatic classification and in some cases inter-ontology linkage. At the same time, the system is simple and intuitive and can be used by most biocurators without extensive training.

  1. Automatic WSDL-guided Test Case Generation for PropEr Testing of Web Services

    Directory of Open Access Journals (Sweden)

    Konstantinos Sagonas

    2012-10-01

    Full Text Available With web services already being key ingredients of modern web systems, automatic and easy-to-use but at the same time powerful and expressive testing frameworks for web services are increasingly important. Our work aims at fully automatic testing of web services: ideally the user only specifies properties that the web service is expected to satisfy, in the form of input-output relations, and the system handles all the rest. In this paper we present in detail the component which lies at the heart of this system: how the WSDL specification of a web service is used to automatically create test case generators that can be fed to PropEr, a property-based testing tool, to create structurally valid random test cases for its operations and check its responses. Although the process is fully automatic, our tool optionally allows the user to easily modify its output to either add semantic information to the generators or write properties that test for more involved functionality of the web services.

  2. The Future of Web Maps in Next Generation Textbooks

    Science.gov (United States)

    DiBiase, D.; Prasad, S.

    2014-12-01

    The reformation of the "Object Formerly Known as Textbook" (coined by the Chronicle of Higher Education) toward a digital future is underway. Emerging nextgen texts look less like electronic books ("ebooks") and more like online courseware. In addition to text and illustrations, nextgen textbooks for STEM subjects are likely to combine quizzes, grade management tools, support for social learning, and interactive media including web maps. Web maps are interactive, multi-scale, online maps that enable teachers and learners to explore, interrogate, and mash up the wide variety of map layers available in the cloud. This presentation will show how web maps coupled with interactive quizzes enable students' purposeful explorations and interpretations of spatial patterns related to humankind's interactions with the earth. Attendees will also learn about Esri's offer to donate ArcGIS Online web mapping subscriptions to every U.S. school as part of the President Obama's ConnectED initiative.

  3. Statmaster and HEROS - web-based courses first and second generation

    DEFF Research Database (Denmark)

    Larsen, Pia Veldt; Rootzen, Helle

    2008-01-01

    With the increasing focus on life-long learning, and with the convenience and accessibility of the Internet, the market for web-based courses has expanded vastly in recent times–in particular in connection with continuing education. However, teaching web-based courses presents various technical...... as well as pedagogical challenges. Some of these challenges are addressed, and means to dealing with them are suggested. A second generation of web-based courses is comprised of learning objects, which allows for tailoring courses for specialized groups of students, and accommodate individualized learning....... The concept of learning objects and how they are used to form new courses are discussed....

  4. Flexible Generation of Pervasive Web Services using OSGi Declarative Services and OWL Ontologies

    DEFF Research Database (Denmark)

    Hansen, Klaus Marius; Zhang, Weishan; Fernandes, Joao

    2008-01-01

    There is a growing trend to deploy web services in pervasive computing environments. Implementing web services on networked, embedded devices leads to a set of challenges, including productivity of development, efficiency of web services, and handling of variability and dependencies of hardware...... and software platforms. To address these challenges, we developed a web service compiler called Limbo, in which Web Ontology Language (OWL) ontologies are used to make the Limbo compiler aware of its compilation context such as device hardware and software details, platform dependencies, and resource....../power consumption. The ontologies are used to configure Limbo for generating resource-efficient web service code. The architecture of Limbo follows the Blackboard architectural style and Limbo is implemented using the OSGi Declarative Services component model. The component model provides high flexibility...

  5. Implementing the WebSocket Protocol Based on Formal Modelling and Automated Code Generation

    DEFF Research Database (Denmark)

    Simonsen, Kent Inge; Kristensen, Lars Michael

    2014-01-01

    with pragmatic annotations for automated code generation of protocol software. The contribution of this paper is an application of the approach as implemented in the PetriCode tool to obtain protocol software implementing the IETF WebSocket protocol. This demonstrates the scalability of our approach to real...... protocols. Furthermore, we perform formal verification of the CPN model prior to code generation, and test the implementation for interoperability against the Autobahn WebSocket test-suite resulting in 97% and 99% success rate for the client and server implementation, respectively. The tests show...

  6. A Virtual Learning Environment for Part-Time MASW Students: An Evaluation of the WebCT

    Science.gov (United States)

    Chan, Charles C.; Tsui, Ming-sum; Chan, Mandy Y. C.; Hong, Joe H.

    2008-01-01

    This study aims to evaluate the perception of a cohort of social workers studying for a part-time master's program in social work in using the popular Web-based learning platform--World Wide Web Course Tools (WebCT) as a complimentary method of teaching and learning. It was noted that social work profession began incorporating computer technology…

  7. BEAT: A Web-Based Boolean Expression Fault-Based Test Case Generation Tool

    Science.gov (United States)

    Chen, T. Y.; Grant, D. D.; Lau, M. F.; Ng, S. P.; Vasa, V. R.

    2006-01-01

    BEAT is a Web-based system that generates fault-based test cases from Boolean expressions. It is based on the integration of our several fault-based test case selection strategies. The generated test cases are considered to be fault-based, because they are aiming at the detection of particular faults. For example, when the Boolean expression is in…

  8. The Web 2.0 as Marketing Tool: Opportunities for SMEs

    NARCIS (Netherlands)

    Constantinides, Efthymios

    2008-01-01

    The new generation of Internet applications widely known as Social Media or Web 2.0 offers corporations a whole range of opportunities for improving their marketing efficiency and internal operations. Web 2.0 applications have already become part of the daily life of an increasing number of

  9. Generating crop calendars with Web search data

    International Nuclear Information System (INIS)

    Van der Velde, Marijn; See, Linda; Fritz, Steffen; Khabarov, Nikolay; Obersteiner, Michael; Verheijen, Frank G A

    2012-01-01

    This paper demonstrates the potential of using Web search volumes for generating crop specific planting and harvesting dates in the USA integrating climatic, social and technological factors affecting crop calendars. Using Google Insights for Search, clear peaks in volume occur at times of planting and harvest at the national level, which were used to derive corn specific planting and harvesting dates at a weekly resolution. Disaggregated to state level, search volumes for corn planting generally are in agreement with planting dates from a global crop calendar dataset. However, harvest dates were less discriminatory at the state level, indicating that peaks in search volume may be blurred by broader searches on harvest as a time of cultural events. The timing of other agricultural activities such as purchase of seed and response to weed and pest infestation was also investigated. These results highlight the future potential of using Web search data to derive planting dates in countries where the data are sparse or unreliable, once sufficient search volumes are realized, as well as the potential for monitoring in real time the response of farmers to climate change over the coming decades. Other potential applications of search volume data of relevance to agronomy are also discussed. (letter)

  10. DietPal: a Web-based dietary menu-generating and management system.

    Science.gov (United States)

    Noah, Shahrul A; Abdullah, Siti Norulhuda; Shahar, Suzana; Abdul-Hamid, Helmi; Khairudin, Nurkahirizan; Yusoff, Mohamed; Ghazali, Rafidah; Mohd-Yusoff, Nooraini; Shafii, Nik Shanita; Abdul-Manaf, Zaharah

    2004-01-30

    Attempts in current health care practice to make health care more accessible, effective, and efficient through the use of information technology could include implementation of computer-based dietary menu generation. While several of such systems already exist, their focus is mainly to assist healthy individuals calculate their calorie intake and to help monitor the selection of menus based upon a prespecified calorie value. Although these prove to be helpful in some ways, they are not suitable for monitoring, planning, and managing patients' dietary needs and requirements. This paper presents a Web-based application that simulates the process of menu suggestions according to a standard practice employed by dietitians. To model the workflow of dietitians and to develop, based on this workflow, a Web-based system for dietary menu generation and management. The system is aimed to be used by dietitians or by medical professionals of health centers in rural areas where there are no designated qualified dietitians. First, a user-needs study was conducted among dietitians in Malaysia. The first survey of 93 dietitians (with 52 responding) was an assessment of information needed for dietary management and evaluation of compliance towards a dietary regime. The second study consisted of ethnographic observation and semi-structured interviews with 14 dietitians in order to identify the workflow of a menu-suggestion process. We subsequently designed and developed a Web-based dietary menu generation and management system called DietPal. DietPal has the capability of automatically calculating the nutrient and calorie intake of each patient based on the dietary recall as well as generating suitable diet and menu plans according to the calorie and nutrient requirement of the patient, calculated from anthropometric measurements. The system also allows reusing stored or predefined menus for other patients with similar health and nutrient requirements. We modeled the workflow of

  11. Beyond Web 2.0 … and Beyond the Semantic Web

    Science.gov (United States)

    Bénel, Aurélien; Zhou, Chao; Cahier, Jean-Pierre

    Tim O'Reilly, the famous technology book publisher, changed the life of many of us when he coined the name "Web 2.0" (O' Reilly 2005). Our research topics suddenly became subjects for open discussion in various cultural formats such as radio and TV, while at the same time they became part of an inappropriate marketing discourse according to several scientific reviewers. Indeed Tim O'Reilly's initial thoughts were about economic consequence, since it was about the resurrection of the Web after the bursting of the dot-com bubble. Some opponents of the concept do not think the term should be used at all since it is underpinned by no technological revolution. In contrast, we think that there was a paradigm shift when several sites based on user-generated content became some of the most visited Web sites and massive adoption of that kind is worthy of researchers' attention.

  12. Semantic Web Technologies for the Adaptive Web

    DEFF Research Database (Denmark)

    Dolog, Peter; Nejdl, Wolfgang

    2007-01-01

    Ontologies and reasoning are the key terms brought into focus by the semantic web community. Formal representation of ontologies in a common data model on the web can be taken as a foundation for adaptive web technologies as well. This chapter describes how ontologies shared on the semantic web...... provide conceptualization for the links which are a main vehicle to access information on the web. The subject domain ontologies serve as constraints for generating only those links which are relevant for the domain a user is currently interested in. Furthermore, user model ontologies provide additional...... means for deciding which links to show, annotate, hide, generate, and reorder. The semantic web technologies provide means to formalize the domain ontologies and metadata created from them. The formalization enables reasoning for personalization decisions. This chapter describes which components...

  13. The Web 2.0 as Marketing Tool: Opportunities for SMEs

    OpenAIRE

    Constantinides, Efthymios

    2008-01-01

    The new generation of Internet applications widely known as Social Media or Web 2.0 offers corporations a whole range of opportunities for improving their marketing efficiency and internal operations. Web 2.0 applications have already become part of the daily life of an increasing number of consumers who regard them as prime channels of communication, information exchange, sharing of expertise, dissemination of individual creativity and entertainment. Web logs, podcasts, online forums and soc...

  14. Recursive macro generator for the TAS-86 language. First part: the macro generator language. Second part: system internal logics

    International Nuclear Information System (INIS)

    Zraick, Samir

    1970-01-01

    A macro-generator is a translator which is able to interpret and translate a programme written in a macro-language. After a first part presenting the main notions and proposing a brief description of the TAS-86 language, the second part of this research thesis reports the development of the macro-generator language, and notably presents the additional functionalities provided by the macro generator. The development is illustrated by logical flowcharts and programming listings

  15. Sharing human-generated observations by integrating HMI and the Semantic Sensor Web.

    Science.gov (United States)

    Sigüenza, Alvaro; Díaz-Pardo, David; Bernat, Jesús; Vancea, Vasile; Blanco, José Luis; Conejero, David; Gómez, Luis Hernández

    2012-01-01

    Current "Internet of Things" concepts point to a future where connected objects gather meaningful information about their environment and share it with other objects and people. In particular, objects embedding Human Machine Interaction (HMI), such as mobile devices and, increasingly, connected vehicles, home appliances, urban interactive infrastructures, etc., may not only be conceived as sources of sensor information, but, through interaction with their users, they can also produce highly valuable context-aware human-generated observations. We believe that the great promise offered by combining and sharing all of the different sources of information available can be realized through the integration of HMI and Semantic Sensor Web technologies. This paper presents a technological framework that harmonizes two of the most influential HMI and Sensor Web initiatives: the W3C's Multimodal Architecture and Interfaces (MMI) and the Open Geospatial Consortium (OGC) Sensor Web Enablement (SWE) with its semantic extension, respectively. Although the proposed framework is general enough to be applied in a variety of connected objects integrating HMI, a particular development is presented for a connected car scenario where drivers' observations about the traffic or their environment are shared across the Semantic Sensor Web. For implementation and evaluation purposes an on-board OSGi (Open Services Gateway Initiative) architecture was built, integrating several available HMI, Sensor Web and Semantic Web technologies. A technical performance test and a conceptual validation of the scenario with potential users are reported, with results suggesting the approach is sound.

  16. WEB LOG EXPLORER – CONTROL OF MULTIDIMENSIONAL DYNAMICS OF WEB PAGES

    Directory of Open Access Journals (Sweden)

    Mislav Šimunić

    2012-07-01

    Full Text Available Demand markets dictate and pose increasingly more requirements to the supplymarket that are not easily satisfied. The supply market presenting its web pages to thedemand market should find the best and quickest ways to respond promptly to the changesdictated by the demand market. The question is how to do that in the most efficient andquickest way. The data on the usage of web pages on a specific web site are recorded in alog file. The data in a log file are stochastic and unordered and require systematicmonitoring, categorization, analyses, and weighing. From the data processed in this way, itis necessary to single out and sort the data by their importance that would be a basis for acontinuous generation of dynamics/changes to the web site pages in line with the criterionchosen. To perform those tasks successfully, a new software solution is required. For thatpurpose, the authors have developed the first version of the WLE (WebLogExplorersoftware solution, which is actually realization of web page multidimensionality and theweb site as a whole. The WebLogExplorer enables statistical and semantic analysis of a logfile and on the basis thereof, multidimensional control of the web page dynamics. Theexperimental part of the work was done within the web site of HTZ (Croatian NationalTourist Board being the main portal of the global tourist supply in the Republic of Croatia(on average, daily "log" consists of c. 600,000 sets, average size of log file is 127 Mb, andc. 7000-8000 daily visitors on the web site.

  17. Teaching Web 2.0 technologies using Web 2.0 technologies.

    Science.gov (United States)

    Rethlefsen, Melissa L; Piorun, Mary; Prince, J Dale

    2009-10-01

    The research evaluated participant satisfaction with the content and format of the "Web 2.0 101: Introduction to Second Generation Web Tools" course and measured the impact of the course on participants' self-evaluated knowledge of Web 2.0 tools. The "Web 2.0 101" online course was based loosely on the Learning 2.0 model. Content was provided through a course blog and covered a wide range of Web 2.0 tools. All Medical Library Association members were invited to participate. Participants were asked to complete a post-course survey. Respondents who completed the entire course or who completed part of the course self-evaluated their knowledge of nine social software tools and concepts prior to and after the course using a Likert scale. Additional qualitative information about course strengths and weaknesses was also gathered. Respondents' self-ratings showed a significant change in perceived knowledge for each tool, using a matched pair Wilcoxon signed rank analysis (P<0.0001 for each tool/concept). Overall satisfaction with the course appeared high. Hands-on exercises were the most frequently identified strength of the course; the length and time-consuming nature of the course were considered weaknesses by some. Learning 2.0-style courses, though demanding time and self-motivation from participants, can increase knowledge of Web 2.0 tools.

  18. Towards Development of Web-based Assessment System Based on Semantic Web Technology

    Directory of Open Access Journals (Sweden)

    Hosam Farouk El-Sofany

    2011-01-01

    Full Text Available The assessment process in an educational system is an important and primordial part of its success to assure the correct way of knowledge transmission and to ensure that students are working correctly and succeed to acquire the needed knowledge. In this study, we aim to include Semantic Web technologies in the E-learning process, as new components. We use Semantic Web (SW to: 1 support the evaluation of open questions in e-learning courses, 2 support the creation of questions and exams automatically, 3 support the evaluation of exams created by the system. These components should allow for measuring academic performance, providing feedback mechanisms, and improving participative and collaborative ideas. Our goal is to use Semantic Web and Wireless technologies to design and implement the assessment system that allows the students, to take: web-based tutorials, quizzes, free exercises, and exams, to download: course reviews, previous exams and their model answers, to access the system through the Mobile and take quick quizzes and exercises. The system facilitates generation of automatic, balanced, and different exam sheets that contain different types of questions covering the entire curriculum, and display gradually from easiness to difficulty. The system provides the teachers and administrators with several services such as: store different types of questions, generate exams with specific criteria, and upload course assignments, exams, and reviews.

  19. What Is Web 2.0: Design Patterns and Business Models for the Next Generation of Software

    OpenAIRE

    O'Reilly, Tim

    2007-01-01

    This paper was the first initiative to try to define Web2.0 and understand its implications for the next generation of software, looking at both design patterns and business modes. Web 2.0 is the network as platform, spanning all connected devices; Web 2.0 applications are those that make the most of the intrinsic advantages of that platform: delivering software as a continually-updated service that gets better the more people use it, consuming and remixing data from multiple sources, includi...

  20. Using Semantic Web technologies for the generation of domain-specific templates to support clinical study metadata standards.

    Science.gov (United States)

    Jiang, Guoqian; Evans, Julie; Endle, Cory M; Solbrig, Harold R; Chute, Christopher G

    2016-01-01

    The Biomedical Research Integrated Domain Group (BRIDG) model is a formal domain analysis model for protocol-driven biomedical research, and serves as a semantic foundation for application and message development in the standards developing organizations (SDOs). The increasing sophistication and complexity of the BRIDG model requires new approaches to the management and utilization of the underlying semantics to harmonize domain-specific standards. The objective of this study is to develop and evaluate a Semantic Web-based approach that integrates the BRIDG model with ISO 21090 data types to generate domain-specific templates to support clinical study metadata standards development. We developed a template generation and visualization system based on an open source Resource Description Framework (RDF) store backend, a SmartGWT-based web user interface, and a "mind map" based tool for the visualization of generated domain-specific templates. We also developed a RESTful Web Service informed by the Clinical Information Modeling Initiative (CIMI) reference model for access to the generated domain-specific templates. A preliminary usability study is performed and all reviewers (n = 3) had very positive responses for the evaluation questions in terms of the usability and the capability of meeting the system requirements (with the average score of 4.6). Semantic Web technologies provide a scalable infrastructure and have great potential to enable computable semantic interoperability of models in the intersection of health care and clinical research.

  1. Web-based Toolkit for Dynamic Generation of Data Processors

    Science.gov (United States)

    Patel, J.; Dascalu, S.; Harris, F. C.; Benedict, K. K.; Gollberg, G.; Sheneman, L.

    2011-12-01

    All computation-intensive scientific research uses structured datasets, including hydrology and all other types of climate-related research. When it comes to testing their hypotheses, researchers might use the same dataset differently, and modify, transform, or convert it to meet their research needs. Currently, many researchers spend a good amount of time performing data processing and building tools to speed up this process. They might routinely repeat the same process activities for new research projects, spending precious time that otherwise could be dedicated to analyzing and interpreting the data. Numerous tools are available to run tests on prepared datasets and many of them work with datasets in different formats. However, there is still a significant need for applications that can comprehensively handle data transformation and conversion activities and help prepare the various processed datasets required by the researchers. We propose a web-based application (a software toolkit) that dynamically generates data processors capable of performing data conversions, transformations, and customizations based on user-defined mappings and selections. As a first step, the proposed solution allows the users to define various data structures and, in the next step, can select various file formats and data conversions for their datasets of interest. In a simple scenario, the core of the proposed web-based toolkit allows the users to define direct mappings between input and output data structures. The toolkit will also support defining complex mappings involving the use of pre-defined sets of mathematical, statistical, date/time, and text manipulation functions. Furthermore, the users will be allowed to define logical cases for input data filtering and sampling. At the end of the process, the toolkit is designed to generate reusable source code and executable binary files for download and use by the scientists. The application is also designed to store all data

  2. Investigation of next-generation sequencing data of Klebsiella pneumoniae using web-based tools.

    Science.gov (United States)

    Brhelova, Eva; Antonova, Mariya; Pardy, Filip; Kocmanova, Iva; Mayer, Jiri; Racil, Zdenek; Lengerova, Martina

    2017-11-01

    Rapid identification and characterization of multidrug-resistant Klebsiella pneumoniae strains is necessary due to the increasing frequency of severe infections in patients. The decreasing cost of next-generation sequencing enables us to obtain a comprehensive overview of genetic information in one step. The aim of this study is to demonstrate and evaluate the utility and scope of the application of web-based databases to next-generation sequenced (NGS) data. The whole genomes of 11 clinical Klebsiella pneumoniae isolates were sequenced using Illumina MiSeq. Selected web-based tools were used to identify a variety of genetic characteristics, such as acquired antimicrobial resistance genes, multilocus sequence types, plasmid replicons, and identify virulence factors, such as virulence genes, cps clusters, urease-nickel clusters and efflux systems. Using web-based tools hosted by the Center for Genomic Epidemiology, we detected resistance to 8 main antimicrobial groups with at least 11 acquired resistance genes. The isolates were divided into eight sequence types (ST11, 23, 37, 323, 433, 495 and 562, and a new one, ST1646). All of the isolates carried replicons of large plasmids. Capsular types, virulence factors and genes coding AcrAB and OqxAB efflux pumps were detected using BIGSdb-Kp, whereas the selected virulence genes, identified in almost all of the isolates, were detected using CLC Genomic Workbench software. Applying appropriate web-based online tools to NGS data enables the rapid extraction of comprehensive information that can be used for more efficient diagnosis and treatment of patients, while data processing is free of charge, easy and time-efficient.

  3. Technical Evaluation Report 61: The World-Wide Inaccessible Web, Part 2: Internet routes

    Directory of Open Access Journals (Sweden)

    Jim Klaas

    2007-06-01

    Full Text Available In the previous report in this series, Web browser loading times were measured in 12 Asian countries, and were found to be up to four times slower than commonly prescribed as acceptable. Failure of webpages to load at all was frequent. The current follow-up study compares these loading times with the complexity of the Internet routes linking the Web users and the Web servers hosting them. The study was conducted in the same 12 Asian countries, with the assistance of members of the International Development Research Centre’s PANdora distance education research network. The data were generated by network members in Bhutan, Cambodia, India, Indonesia, Laos, Mongolia, the Philippines, Sri Lanka, Pakistan, Singapore, Thailand, and Vietnam. Additional data for the follow-up study were collected in China. Using a ‘traceroute’ routine, the study indicates that webpage loading time is linked to the complexity of the Internet routes between Web users and the host server. It is indicated that distance educators can apply such information in the design of improved online delivery and mirror sites, notably in areas of the developing world which currently lack an effective infrastructure for online education.

  4. DISCRETIZATION APPROACH USING RAY-TESTING MODEL IN PARTING LINE AND PARTING SURFACE GENERATION

    Institute of Scientific and Technical Information of China (English)

    HAN Jianwen; JIAN Bin; YAN Guangrong; LEI Yi

    2007-01-01

    Surface classification, 3D parting line, parting surface generation and demoldability analysis which is helpful to select optimal parting direction and optimal parting line are involved in automatic cavity design based on the ray-testing model. A new ray-testing approach is presented to classify the part surfaces to core/cavity surfaces and undercut surfaces by automatic identifying the visibility of surfaces. A simple, direct and efficient algorithm to identify surface visibility is developed. The algorithm is robust and adapted to rather complicated geometry, so it is valuable in computer-aided mold design systems. To validate the efficiency of the approach, an experimental program is implemented. Case studies show that the approach is practical and valuable in automatic parting line and parting surface generation.

  5. Is there a "net generation" in veterinary medicine? A comparative study on the use of the Internet and Web 2.0 by students and the veterinary profession.

    Science.gov (United States)

    Tenhaven, Christoph; Tipold, Andrea; Fischer, Martin R; Ehlers, Jan P

    2013-01-01

    Informal and formal lifelong learning is essential at university and in the workplace. Apart from classical learning techniques, Web 2.0 tools can be used. It is controversial whether there is a so-called net generation amongst people under 30. To test the hypothesis that a net generation among students and young veterinarians exists. An online survey of students and veterinarians was conducted in the German-speaking countries which was advertised via online media and traditional print media. 1780 people took part in the survey. Students and veterinarians have different usage patterns regarding social networks (91.9% vs. 69%) and IM (55.9% vs. 24.5%). All tools were predominantly used passively and in private, to a lesser extent also professionally and for studying. The use of Web 2.0 tools is useful, however, teaching information and media skills, preparing codes of conduct for the internet and verification of user generated content is essential.

  6. Review of „Brand-urile în era Web 2.0. Conținutul generat de consumatori” [Web 2.0 Brands. User-generated Content] by Rodica Săvulescu, București: Tritonic, 2016, 252 p.

    Directory of Open Access Journals (Sweden)

    Alexandra Vițelar

    2017-01-01

    Full Text Available The web 2.0 era has shifted brand ownership from communication specialists towards consumers. This is the main idea on which Rodica Sãvulescu builds her argumentation in her recently published book, `Web 2.0 Brands. User-generated content` (2016. The emergence of new technologies blurs the lines between content producers and consumers. In this book, the author addresses the topic of democratization of content in relation with brand communication.

  7. Actinide colloid generation in groundwater. Part 2

    International Nuclear Information System (INIS)

    Kim, J.I.

    1991-01-01

    The progress made in the investigation of actinide colloid generation in groundwater is summarized and discussed with particular examples relevant to an understanding of the migration behaviour of actinides in natural aquifer systems. The first part deals with the characterization of colloids: groundwater colloids, actinide real-colloids and actinide pseudocolloids. The second part concentrates on the generation processes and migration behaviour of actinide pseudo colloids, which are discussed with some notable experimental examples. Importance is stressed more on the chemical aspects of the actinide colloid generation in groundwater. This work is a contribution to the CEC Mirage II project, in particular the complexation and colloids research area

  8. Dynamic Web Pages: Performance Impact on Web Servers.

    Science.gov (United States)

    Kothari, Bhupesh; Claypool, Mark

    2001-01-01

    Discussion of Web servers and requests for dynamic pages focuses on experimentally measuring and analyzing the performance of the three dynamic Web page generation technologies: CGI, FastCGI, and Servlets. Develops a multivariate linear regression model and predicts Web server performance under some typical dynamic requests. (Author/LRW)

  9. Hypermedia presentation generation for semantic web information systems

    NARCIS (Netherlands)

    Frasincar, F.

    2005-01-01

    Due to Web popularity many information systems have been made available through the Web, resulting in so-called Web Information Systems (WIS). Due to the complex requirements that WIS need to ful??ll, the design of these systems is not a trivial task. Design methodologies provide guidelines for the

  10. PseKNC: a flexible web server for generating pseudo K-tuple nucleotide composition.

    Science.gov (United States)

    Chen, Wei; Lei, Tian-Yu; Jin, Dian-Chuan; Lin, Hao; Chou, Kuo-Chen

    2014-07-01

    The pseudo oligonucleotide composition, or pseudo K-tuple nucleotide composition (PseKNC), can be used to represent a DNA or RNA sequence with a discrete model or vector yet still keep considerable sequence order information, particularly the global or long-range sequence order information, via the physicochemical properties of its constituent oligonucleotides. Therefore, the PseKNC approach may hold very high potential for enhancing the power in dealing with many problems in computational genomics and genome sequence analysis. However, dealing with different DNA or RNA problems may need different kinds of PseKNC. Here, we present a flexible and user-friendly web server for PseKNC (at http://lin.uestc.edu.cn/pseknc/default.aspx) by which users can easily generate many different modes of PseKNC according to their need by selecting various parameters and physicochemical properties. Furthermore, for the convenience of the vast majority of experimental scientists, a step-by-step guide is provided on how to use the current web server to generate their desired PseKNC without the need to follow the complicated mathematical equations, which are presented in this article just for the integrity of PseKNC formulation and its development. It is anticipated that the PseKNC web server will become a very useful tool in computational genomics and genome sequence analysis. Copyright © 2014 Elsevier Inc. All rights reserved.

  11. A Secure Web Application Providing Public Access to High-Performance Data Intensive Scientific Resources - ScalaBLAST Web Application

    International Nuclear Information System (INIS)

    Curtis, Darren S.; Peterson, Elena S.; Oehmen, Chris S.

    2008-01-01

    This work presents the ScalaBLAST Web Application (SWA), a web based application implemented using the PHP script language, MySQL DBMS, and Apache web server under a GNU/Linux platform. SWA is an application built as part of the Data Intensive Computer for Complex Biological Systems (DICCBS) project at the Pacific Northwest National Laboratory (PNNL). SWA delivers accelerated throughput of bioinformatics analysis via high-performance computing through a convenient, easy-to-use web interface. This approach greatly enhances emerging fields of study in biology such as ontology-based homology, and multiple whole genome comparisons which, in the absence of a tool like SWA, require a heroic effort to overcome the computational bottleneck associated with genome analysis. The current version of SWA includes a user account management system, a web based user interface, and a backend process that generates the files necessary for the Internet scientific community to submit a ScalaBLAST parallel processing job on a dedicated cluster

  12. The World Wide Web and the Television Generation.

    Science.gov (United States)

    Maddux, Cleborne D.

    1996-01-01

    The hypermedia nature of the World Wide Web may represent a true paradigm shift in telecommunications, but barriers exist to the Web having similar impact on education. Some of today's college students compare the Web with "bad TV"--lengthy pauses, links that result in error messages, and animation and sound clips that are too brief.…

  13. DEVELOPING WEB MAPPING APPLICATION USING ARCGIS SERVER WEB APPLICATION DEVELOPMEN FRAMEWORK (ADF FOR GEOSPATIAL DATA GENERATED DURING REHABILITATION AND RECONSTRUCTION PROCESS OF POST-TSUNAMI 2004 DISASTER IN ACEH

    Directory of Open Access Journals (Sweden)

    Nizamuddin Nizamuddin

    2014-04-01

    Full Text Available ESRI ArcGIS Server is equipped with ArcGIS Server Web Application Development Framework (ADF and ArcGIS Web Controls integration for Visual Studio.NET. Both the ArcGIS Server Manager for .NET and ArcGIS Web Controls can be easily utilized for developing the ASP.NET based ESRI Web mapping application. In  this study we implemented both tools for developing the ASP.NET based ESRI Web mapping application for geospatial data generated dring rehabilitation and reconstruction process of post-tsunami 2004 disaster in Aceh province. Rehabilitation and reconstruction process has produced a tremendous amount of geospatial data. This method was chosen in this study because in the process of developing  a web mapping application, one can easily and quickly create Mapping Services of huge geospatial data and also develop Web mapping application without writing any code. However, when utilizing Visual Studio.NET 2008, one needs to have some coding ability.

  14. Lipids: Part of the tangled web

    Energy Technology Data Exchange (ETDEWEB)

    Krauss, R.M.

    1992-08-01

    Analysis of LDL subclasses by non-denaturing gradient gel electrophoresis has led to the identification of a subclass pattern characterized by predominance of small LDL, designated LDL subclass pattern B. The prevalence of pattern B in the general population is approximately 25%, but varies as a function of age and gender, being relatively uncommon in children and in premenopausal women. The remainder of the population has a predominance of larger LDL (pattern A) or an intermediate pattern. Our findings indicate that LDL subclass pattern B is an integral part of the ``tangled web`` of interrelated coronary disease risk factors associated with insulin resistance. It may be that the pathologic features of this lipoprotein profile, including the relative atherogenicity of small, dense LDL and IDL, contribute importantly to the increased risk of cardiovascular disease in subjects with insulin resistance and hypertension. Furthermore, pattern B serves as a marker for a common genetic trait which may underlie a substantial portion of the familial predisposition to coronary artery disease in the general population. Studies of hormonal, dietary, and pharmacologic influences on expression of this atherogenic phenotype should lead to more effective identification and management of high-risk individuals, and improved approaches to disease prevention in high-risk families.

  15. Lipids: Part of the tangled web

    Energy Technology Data Exchange (ETDEWEB)

    Krauss, R.M.

    1992-08-01

    Analysis of LDL subclasses by non-denaturing gradient gel electrophoresis has led to the identification of a subclass pattern characterized by predominance of small LDL, designated LDL subclass pattern B. The prevalence of pattern B in the general population is approximately 25%, but varies as a function of age and gender, being relatively uncommon in children and in premenopausal women. The remainder of the population has a predominance of larger LDL (pattern A) or an intermediate pattern. Our findings indicate that LDL subclass pattern B is an integral part of the tangled web'' of interrelated coronary disease risk factors associated with insulin resistance. It may be that the pathologic features of this lipoprotein profile, including the relative atherogenicity of small, dense LDL and IDL, contribute importantly to the increased risk of cardiovascular disease in subjects with insulin resistance and hypertension. Furthermore, pattern B serves as a marker for a common genetic trait which may underlie a substantial portion of the familial predisposition to coronary artery disease in the general population. Studies of hormonal, dietary, and pharmacologic influences on expression of this atherogenic phenotype should lead to more effective identification and management of high-risk individuals, and improved approaches to disease prevention in high-risk families.

  16. Getting to the Source: a Survey of Quantitative Data Sources Available to the Everyday Librarian: Part 1: Web Server Log Analysis

    Directory of Open Access Journals (Sweden)

    Lisa Goddard

    2007-03-01

    Full Text Available This is the first part of a two‐part article that provides a survey of data sources which are likely to be immediately available to the typical practitioner who wishes to engage instatistical analysis of collections and services within his or her own library. Part I outlines the data elements which can be extracted from web server logs, and discusses web log analysis tools. Part II looks at logs, reports, and data sources from proxy servers, resource vendors, link resolvers, federated search engines, institutional repositories, electronic reference services, and the integrated library system.

  17. Web site development: applying aesthetics to promote breast health education and awareness.

    Science.gov (United States)

    Thomas, Barbara; Goldsmith, Susan B; Forrest, Anne; Marshall, Renée

    2002-01-01

    This article describes the process of establishing a Web site as part of a collaborative project using visual art to promote breast health education. The need for a more "user-friendly" comprehensive breast health Web site that is aesthetically rewarding was identified after an analysis of current Web sites available through the World Wide Web. Two predetermined sets of criteria, accountability and aesthetics, were used to analyze these sites and to generate ideas for creating a breast health education Web site using visual art. Results of the analyses conducted are included as well as the factors to consider for incorporating into a Web site. The process specified is thorough and can be applied to establish a Web site that is aesthetically rewarding and informative for a variety of educational purposes.

  18. A Two-Tiered Model for Analyzing Library Web Site Usage Statistics, Part 1: Web Server Logs.

    Science.gov (United States)

    Cohen, Laura B.

    2003-01-01

    Proposes a two-tiered model for analyzing web site usage statistics for academic libraries: one tier for library administrators that analyzes measures indicating library use, and a second tier for web site managers that analyzes measures aiding in server maintenance and site design. Discusses the technology of web site usage statistics, and…

  19. Het WEB leert begrijpen

    CERN Multimedia

    Stroeykens, Steven

    2004-01-01

    The WEB could be much more useful if the computers understood something of information on the Web pages. That explains the goal of the "semantic Web", a project in which takes part, amongst others, Tim Berners Lee, the inventor of the original WEB

  20. Web-based control application using WebSocket

    International Nuclear Information System (INIS)

    Furukawa, Y.

    2012-01-01

    The WebSocket allows asynchronous full-duplex communication between a Web-based (i.e. Java Script-based) application and a Web-server. WebSocket started as a part of HTML5 standardization but has now been separated from HTML5 and has been developed independently. Using WebSocket, it becomes easy to develop platform independent presentation layer applications for accelerator and beamline control software. In addition, a Web browser is the only application program that needs to be installed on client computer. The WebSocket-based applications communicate with the WebSocket server using simple text-based messages, so WebSocket is applicable message-based control system like MADOCA, which was developed for the SPring-8 control system. A simple WebSocket server for the MADOCA control system and a simple motor control application were successfully made as a first trial of the WebSocket control application. Using Google-Chrome (version 13.0) on Debian/Linux and Windows 7, Opera (version 11.0) on Debian/Linux and Safari (version 5.0.3) on Mac OS X as clients, the motors can be controlled using a WebSocket-based Web-application. Diffractometer control application use in synchrotron radiation diffraction experiment was also developed. (author)

  1. Process-oriented semantic web search

    CERN Document Server

    Tran, DT

    2011-01-01

    The book is composed of two main parts. The first part is a general study of Semantic Web Search. The second part specifically focuses on the use of semantics throughout the search process, compiling a big picture of Process-oriented Semantic Web Search from different pieces of work that target specific aspects of the process.In particular, this book provides a rigorous account of the concepts and technologies proposed for searching resources and semantic data on the Semantic Web. To collate the various approaches and to better understand what the notion of Semantic Web Search entails, this bo

  2. Social web artifacts for boosting recommenders theory and implementation

    CERN Document Server

    Ziegler, Cai-Nicolas

    2013-01-01

    Recommender systems, software programs that learn from human behavior and make predictions of what products we are expected to appreciate and purchase, have become an integral part of our everyday life. They proliferate across electronic commerce around the globe and exist for virtually all sorts of consumable goods, such as books, movies, music, or clothes. At the same time, a new evolution on the Web has started to take shape, commonly known as the “Web 2.0” or the “Social Web”: Consumer-generated media has become rife, social networks have emerged and are pulling significant shares of Web traffic. In line with these developments, novel information and knowledge artifacts have become readily available on the Web, created by the collective effort of millions of people. This textbook presents approaches to exploit the new Social Web fountain of knowledge, zeroing in first and foremost on two of those information artifacts, namely classification taxonomies and trust networks. These two are used to impr...

  3. Web-page Prediction for Domain Specific Web-search using Boolean Bit Mask

    OpenAIRE

    Sinha, Sukanta; Duttagupta, Rana; Mukhopadhyay, Debajyoti

    2012-01-01

    Search Engine is a Web-page retrieval tool. Nowadays Web searchers utilize their time using an efficient search engine. To improve the performance of the search engine, we are introducing a unique mechanism which will give Web searchers more prominent search results. In this paper, we are going to discuss a domain specific Web search prototype which will generate the predicted Web-page list for user given search string using Boolean bit mask.

  4. Comparison of elution efficiency of 99Mo/99mTc generator using theoretical and a free web based software method

    International Nuclear Information System (INIS)

    Kiran Kumar, J.K.; Sharma, S.; Chakraborty, D.; Singh, B.; Bhattacharaya, A.; Mittal, B.R.; Gayana, S.

    2010-01-01

    Full text: Generator is constructed on the principle of decay growth relationship between a long lived parent radionuclide and short lived daughter radionuclide. Difference in chemical properties of daughter and parent radionuclide helps in efficient separation of the two radionuclides. Aim and Objectives: The present study was designed to calculate the elution efficiency of the generator using the traditional formula based method and free web based software method. Materials and Methods: 99 Mo/ 99m Tc MON.TEK (Monrol, Gebze) generator and sterile 0.9% NaCl vial and vacuum vial in the lead shield were used for the elution. A new 99 Mo/ 99m Tc generator (calibrated activity 30GBq) calibrated for thursday was received on monday morning in our department. Generator was placed behind lead bricks in fume hood. The rubber plugs of both vacuum and 0.9% NaCl vial were wiped with 70% isopropyl alcohol swabs. Vacuum vial placed inside the lead shield was inserted in the vacuum position simultaneously 10 ml NaCl vial was inserted in the second slot. After 1-2 min vacuum vial was removed without moving the emptied 0.9%NaCl vial. The vacuum slot was covered with another sterile vial to maintain sterility. The RAC was measured in the calibrated dose calibrator (Capintec, 15 CRC). The elution efficiency was calculated theoretically and using free web based software (Apache Web server (www.apache.org) and PHP (www.php.net). Web site of the Italian Association of Nuclear Medicine and Molecular Imaging (www.aimn.it). Results: The mean elution efficiency calculated by theoretical method was 93.95% +0.61. The mean elution efficiency as calculated by the software was 92.85% + 0.89. There was no statistical difference in both the methods. Conclusion: The free web based software provides precise and reproducible results and thus saves time and mathematical calculation steps. This enables a rational use of available activity and also enabling a selection of the type and number of

  5. IS 37 FORM ON EDH WEB

    CERN Multimedia

    2000-01-01

    To Staff Members in charge of the execution of works The “Issuers” are reminded to fill in - if necessary - the form attached to Safety Instruction 37 when disabling all or part of the system generating a level 3 alarm. Reminder: The request must be completed by the issuer and authorised by the TSO/GLIMOS responsible for the building or area. After completion of the works, the TSO/GLIMOS make sure that the system is recommissioned. Please note that the computerized version of this form is available on the web. The icon can be found on the EDH Web Desktop Homepage. The paper version is still in use. If you have any questions, please contact A. Chouvelon/TIS, tel. 74229.

  6. Mapping Commercial Web 2.0 Worlds: Towards a New Critical Ontogenesis

    Directory of Open Access Journals (Sweden)

    Kenneth Werbin

    2009-01-01

    Full Text Available At the 2007 International Communication Association Conference, Web 2.0 was highlighted as an emergent topic of research with a keynote panel entitled 'What's so Significant about Social Networking? Web 2.0 and its Critical Potentials'. One of the thought-provoking moments during the panel was the juxtaposition of two very different and at first, contradictory theoretical approaches to the relationships between Web 2.0 and user-generated content. While Henry Jenkins focused on the democratic potential of online participatory culture as enabling new modes of knowledge production, Titziana Terranova argued for a post-Marxist perspective on Web 2.0 as a site of cultural colonization and expansion of new forms of capitalization on culture, affect and knowledge. The juxtaposition of these two very different critical approaches did not simply rehash the old divide between cultural theory, particularly active audience theory, and post-Marxist critical theory; rather, this debate over Web 2.0 suggested new possibilities for the synthesis and continued development of both sets of critiques. In other words, the event reinforced our belief that corporate colonization arguments do not provide an entirely adequate model for understanding Web 2.0. After all, commercial Web 2.0 spaces such as Facebook, YouTube and MySpace are important sites of cultural exchange and political discussion, in part because they almost entirely rely on user-generated content to exist.

  7. Web-ADARE: A Web-Aided Data Repairing System

    KAUST Repository

    Gu, Binbin

    2017-03-08

    Data repairing aims at discovering and correcting erroneous data in databases. In this paper, we develop Web-ADARE, an end-to-end web-aided data repairing system, to provide a feasible way to involve the vast data sources on the Web in data repairing. Our main attention in developing Web-ADARE is paid on the interaction problem between web-aided repairing and rule-based repairing, in order to minimize the Web consultation cost while reaching predefined quality requirements. The same interaction problem also exists in crowd-based methods but this is not yet formally defined and addressed. We first prove in theory that the optimal interaction scheme is not feasible to be achieved, and then propose an algorithm to identify a scheme for efficient interaction by investigating the inconsistencies and the dependencies between values in the repairing process. Extensive experiments on three data collections demonstrate the high repairing precision and recall of Web-ADARE, and the efficiency of the generated interaction scheme over several baseline ones.

  8. Web-ADARE: A Web-Aided Data Repairing System

    KAUST Repository

    Gu, Binbin; Li, Zhixu; Yang, Qiang; Xie, Qing; Liu, An; Liu, Guanfeng; Zheng, Kai; Zhang, Xiangliang

    2017-01-01

    Data repairing aims at discovering and correcting erroneous data in databases. In this paper, we develop Web-ADARE, an end-to-end web-aided data repairing system, to provide a feasible way to involve the vast data sources on the Web in data repairing. Our main attention in developing Web-ADARE is paid on the interaction problem between web-aided repairing and rule-based repairing, in order to minimize the Web consultation cost while reaching predefined quality requirements. The same interaction problem also exists in crowd-based methods but this is not yet formally defined and addressed. We first prove in theory that the optimal interaction scheme is not feasible to be achieved, and then propose an algorithm to identify a scheme for efficient interaction by investigating the inconsistencies and the dependencies between values in the repairing process. Extensive experiments on three data collections demonstrate the high repairing precision and recall of Web-ADARE, and the efficiency of the generated interaction scheme over several baseline ones.

  9. Web archives

    DEFF Research Database (Denmark)

    Finnemann, Niels Ole

    2018-01-01

    This article deals with general web archives and the principles for selection of materials to be preserved. It opens with a brief overview of reasons why general web archives are needed. Section two and three present major, long termed web archive initiatives and discuss the purposes and possible...... values of web archives and asks how to meet unknown future needs, demands and concerns. Section four analyses three main principles in contemporary web archiving strategies, topic centric, domain centric and time-centric archiving strategies and section five discuss how to combine these to provide...... a broad and rich archive. Section six is concerned with inherent limitations and why web archives are always flawed. The last sections deal with the question how web archives may fit into the rapidly expanding, but fragmented landscape of digital repositories taking care of various parts...

  10. Designing a responsive web site

    OpenAIRE

    Fejzić , Diana

    2016-01-01

    Due to the increasing prevalence of smartphones and tablet computers design became a crucial part of web design. For a user, responsive web design enables the best user experience, regardless of whether a user is visiting the site via a mobile phone, a tablet or a computer. This thesis covers the process of planning, designing and responsive web site development, for a fictitious company named “Creative Design d.o.o.”, with the help of web technologies. In the initial part of the thesis, w...

  11. Usability Testing for e-Resource Discovery: How Students Find and Choose e-Resources Using Library Web Sites

    Science.gov (United States)

    Fry, Amy; Rich, Linda

    2011-01-01

    In early 2010, library staff at Bowling Green State University (BGSU) in Ohio designed and conducted a usability study of key parts of the library web site, focusing on the web pages generated by the library's electronic resources management system (ERM) that list and describe the library's databases. The goal was to discover how users find and…

  12. Worldwide telemedicine services based on distributed multimedia electronic patient records by using the second generation Web server hyperwave.

    Science.gov (United States)

    Quade, G; Novotny, J; Burde, B; May, F; Beck, L E; Goldschmidt, A

    1999-01-01

    A distributed multimedia electronic patient record (EPR) is a central component of a medicine-telematics application that supports physicians working in rural areas of South America, and offers medical services to scientists in Antarctica. A Hyperwave server is used to maintain the patient record. As opposed to common web servers--and as a second generation web server--Hyperwave provides the capability of holding documents in a distributed web space without the problem of broken links. This enables physicians to browse through a patient's record by using a standard browser even if the patient's record is distributed over several servers. The patient record is basically implemented on the "Good European Health Record" (GEHR) architecture.

  13. Writing for the web composing, coding, and constructing web sites

    CERN Document Server

    Applen, JD

    2013-01-01

    Writing for the Web unites theory, technology, and practice to explore writing and hypertext for website creation. It integrates such key topics as XHTML/CSS coding, writing (prose) for the Web, the rhetorical needs of the audience, theories of hypertext, usability and architecture, and the basics of web site design and technology. Presenting information in digestible parts, this text enables students to write and construct realistic and manageable Web sites with a strong theoretical understanding of how online texts communicate to audiences. Key features of the book

  14. Social network extraction based on Web: 3. the integrated superficial method

    Science.gov (United States)

    Nasution, M. K. M.; Sitompul, O. S.; Noah, S. A.

    2018-03-01

    The Web as a source of information has become part of the social behavior information. Although, by involving only the limitation of information disclosed by search engines in the form of: hit counts, snippets, and URL addresses of web pages, the integrated extraction method produces a social network not only trusted but enriched. Unintegrated extraction methods may produce social networks without explanation, resulting in poor supplemental information, or resulting in a social network of durmise laden, consequently unrepresentative social structures. The integrated superficial method in addition to generating the core social network, also generates an expanded network so as to reach the scope of relation clues, or number of edges computationally almost similar to n(n - 1)/2 for n social actors.

  15. Web buckling behavior under in-plane compression and shear loads for web reinforced composite sandwich core

    Science.gov (United States)

    Toubia, Elias Anis

    Sandwich construction is one of the most functional forms of composite structures developed by the composite industry. Due to the increasing demand of web-reinforced core for composite sandwich construction, a research study is needed to investigate the web plate instability under shear, compression, and combined loading. If the web, which is an integral part of the three dimensional web core sandwich structure, happens to be slender with respect to one or two of its spatial dimensions, then buckling phenomena become an issue in that it must be quantified as part of a comprehensive strength model for a fiber reinforced core. In order to understand the thresholds of thickness, web weight, foam type, and whether buckling will occur before material yielding, a thorough investigation needs to be conducted, and buckling design equations need to be developed. Often in conducting a parametric study, a special purpose analysis is preferred over a general purpose analysis code, such as a finite element code, due to the cost and effort usually involved in generating a large number of results. A suitable methodology based on an energy method is presented to solve the stability of symmetrical and specially orthotropic laminated plates on an elastic foundation. Design buckling equations were developed for the web modeled as a laminated plate resting on elastic foundations. The proposed equations allow for parametric studies without limitation regarding foam stiffness, geometric dimensions, or mechanical properties. General behavioral trends of orthotropic and symmetrical anisotropic plates show pronounced contribution of the elastic foundation and fiber orientations on the buckling resistance of the plate. The effects of flexural anisotropy on the buckling behavior of long rectangular plates when subjected to pure shear loading are well represented in the model. The reliability of the buckling equations as a design tool is confirmed by comparison with experimental results

  16. Development of Web GIS for complex processing and visualization of climate geospatial datasets as an integral part of dedicated Virtual Research Environment

    Science.gov (United States)

    Gordov, Evgeny; Okladnikov, Igor; Titov, Alexander

    2017-04-01

    For comprehensive usage of large geospatial meteorological and climate datasets it is necessary to create a distributed software infrastructure based on the spatial data infrastructure (SDI) approach. Currently, it is generally accepted that the development of client applications as integrated elements of such infrastructure should be based on the usage of modern web and GIS technologies. The paper describes the Web GIS for complex processing and visualization of geospatial (mainly in NetCDF and PostGIS formats) datasets as an integral part of the dedicated Virtual Research Environment for comprehensive study of ongoing and possible future climate change, and analysis of their implications, providing full information and computing support for the study of economic, political and social consequences of global climate change at the global and regional levels. The Web GIS consists of two basic software parts: 1. Server-side part representing PHP applications of the SDI geoportal and realizing the functionality of interaction with computational core backend, WMS/WFS/WPS cartographical services, as well as implementing an open API for browser-based client software. Being the secondary one, this part provides a limited set of procedures accessible via standard HTTP interface. 2. Front-end part representing Web GIS client developed according to a "single page application" technology based on JavaScript libraries OpenLayers (http://openlayers.org/), ExtJS (https://www.sencha.com/products/extjs), GeoExt (http://geoext.org/). It implements application business logic and provides intuitive user interface similar to the interface of such popular desktop GIS applications, as uDIG, QuantumGIS etc. Boundless/OpenGeo architecture was used as a basis for Web-GIS client development. According to general INSPIRE requirements to data visualization Web GIS provides such standard functionality as data overview, image navigation, scrolling, scaling and graphical overlay, displaying map

  17. Report covering examination of parts from downhole steam generators. [Combustor head and sleeve parts

    Energy Technology Data Exchange (ETDEWEB)

    Pettit, F. S.; Meier, G. H.

    1983-08-01

    Combustor head and sleeve parts were examined by using optical and scanning electron metallography after use in oxygen/diesel and air/diesel downhole steam generators. The degradation of the different alloy components is described in terms of reactions with oxygen, sulfur and carbon in the presence of cyclic stresses, all generated by the combustion process. Recommendations are presented for component materials (alloys and coatings) to extend component lives in the downhole steam generators. 9 references, 22 figures, 3 tables.

  18. Human Trafficking in the United States. Part II. Survey of U.S. Government Web Resources for Publications and Data

    Science.gov (United States)

    Panigabutra-Roberts, Anchalee

    2012-01-01

    This second part of a two-part series is a survey of U.S. government web resources on human trafficking in the United States, particularly of the online publications and data included on agencies' websites. Overall, the goal is to provide an introduction, an overview, and a guide on this topic for library staff to use in their research and…

  19. Patient and parent views on a Web 2.0 Diabetes Portal--the management tool, the generator, and the gatekeeper: qualitative study.

    Science.gov (United States)

    Nordfeldt, Sam; Hanberger, Lena; Berterö, Carina

    2010-05-28

    The Internet has undergone rapid development, with significant impact on social life and on modes of communication. Modern management of type 1 diabetes requires that patients have access to continuous support and learning opportunities. Although Web 2.0 resources can provide this support, few pediatric clinics offer it as part of routine diabetes care. We aimed to explore patients' and parents' attitudes toward a local Web 2.0 portal tailored to young patients with type 1 diabetes and their parents, with social networking tools such as message boards and blogs, locally produced self-care and treatment information, and interactive pedagogic devices. Opportunities and obstacles to the implementation of Web 2.0 applications in clinical practice were sought. Participants were 16 mothers, 3 fathers, and 5 young patients (ages 11-18 years; median 14 years) who each wrote an essay on their experience using the portal, irrespective of frequency and/or their success in using it. Two main guiding questions were asked. A qualitative content analysis was conducted of the essays as a whole. Three main categories of portal users' attitudes were found; we named them "the management tool," "the generator," and "the gatekeeper." One category was related to the management tool functionality of the portal, and a wide range of concrete examples was found regarding useful facts and updates. Being enabled to search when necessary and find reliable information provided by local clinicians was regarded as a great advantage, facilitating a feeling of security and being in control. Finding answers to difficult-to-ask questions, questions portal users did not know they had before, and questions focusing on sensitive areas such as anxiety and fear, was also an important feature. A second category was related to the generator function in that visiting the portal could generate more information than expected, which could lead to increased use. Active message boards and chat rooms were found to

  20. Advanced Techniques in Web Intelligence-2 Web User Browsing Behaviour and Preference Analysis

    CERN Document Server

    Palade, Vasile; Jain, Lakhmi

    2013-01-01

    This research volume focuses on analyzing the web user browsing behaviour and preferences in traditional web-based environments, social  networks and web 2.0 applications,  by using advanced  techniques in data acquisition, data processing, pattern extraction and  cognitive science for modeling the human actions.  The book is directed to  graduate students, researchers/scientists and engineers  interested in updating their knowledge with the recent trends in web user analysis, for developing the next generation of web-based systems and applications.

  1. THE IMPORTANCE OF WEB DESIGN: VISUAL DESIGN EVALUATION OF DESTINATION WEB SITES

    OpenAIRE

    Fırlar, Belma; Okat Özdem, Özen

    2013-01-01

    As in the literature, the researchs about web site efficiency are mostly about site context. The analysis about function are mostly superficial. Whereas, controlling every little part of a web site respective is a necessity to show its efficiency. Here in this context in the study of perception and response event web sites that play an important role in visual design criteria are below the lens as featured and the web sites evaulated by heuristic evaluation method.The research focus of this s...

  2. Evaluation of WebEase: An Epilepsy Self-Management Web Site

    Science.gov (United States)

    DiIorio, Colleen; Escoffery, Cam; McCarty, Frances; Yeager, Katherine A.; Henry, Thomas R.; Koganti, Archana; Reisinger, Elizabeth L.; Wexler, Bethany

    2009-01-01

    People with epilepsy have various education needs and must adopt many self-management behaviors in order to control their condition. This study evaluates WebEase, an Internet-based, theory-driven, self-management program for adults with epilepsy. Thirty-five participants took part in a 6-week pilot implementation of WebEase. The main components of…

  3. Implementation of SRPT Scheduling in Web Servers

    National Research Council Canada - National Science Library

    Harchol-Balter, Mor

    2000-01-01

    .... Experiments use the Linux operating system and the Flash web server. All experiments are repeated under a range of server loads and under both trace-based workloads and those generated by a Web workload generator...

  4. Web Accessibility and Guidelines

    Science.gov (United States)

    Harper, Simon; Yesilada, Yeliz

    Access to, and movement around, complex online environments, of which the World Wide Web (Web) is the most popular example, has long been considered an important and major issue in the Web design and usability field. The commonly used slang phrase ‘surfing the Web’ implies rapid and free access, pointing to its importance among designers and users alike. It has also been long established that this potentially complex and difficult access is further complicated, and becomes neither rapid nor free, if the user is disabled. There are millions of people who have disabilities that affect their use of the Web. Web accessibility aims to help these people to perceive, understand, navigate, and interact with, as well as contribute to, the Web, and thereby the society in general. This accessibility is, in part, facilitated by the Web Content Accessibility Guidelines (WCAG) currently moving from version one to two. These guidelines are intended to encourage designers to make sure their sites conform to specifications, and in that conformance enable the assistive technologies of disabled users to better interact with the page content. In this way, it was hoped that accessibility could be supported. While this is in part true, guidelines do not solve all problems and the new WCAG version two guidelines are surrounded by controversy and intrigue. This chapter aims to establish the published literature related to Web accessibility and Web accessibility guidelines, and discuss limitations of the current guidelines and future directions.

  5. Primer on client-side web security

    CERN Document Server

    De Ryck, Philippe; Piessens, Frank; Johns, Martin

    2014-01-01

    This volume illustrates the continuous arms race between attackers and defenders of the Web ecosystem by discussing a wide variety of attacks. In the first part of the book, the foundation of the Web ecosystem is briefly recapped and discussed. Based on this model, the assets of the Web ecosystem are identified, and the set of capabilities an attacker may have are enumerated. In the second part, an overview of the web security vulnerability landscape is constructed. Included are selections of the most representative attack techniques reported in great detail. In addition to descriptions of the

  6. Global Web Accessibility Analysis of National Government Portals and Ministry Web Sites

    DEFF Research Database (Denmark)

    Goodwin, Morten; Susar, Deniz; Nietzio, Annika

    2011-01-01

    Equal access to public information and services for all is an essential part of the United Nations (UN) Declaration of Human Rights. Today, the Web plays an important role in providing information and services to citizens. Unfortunately, many government Web sites are poorly designed and have...... accessibility barriers that prevent people with disabilities from using them. This article combines current Web accessibility benchmarking methodologies with a sound strategy for comparing Web accessibility among countries and continents. Furthermore, the article presents the first global analysis of the Web...... accessibility of 192 United Nation Member States made publically available. The article also identifies common properties of Member States that have accessible and inaccessible Web sites and shows that implementing antidisability discrimination laws is highly beneficial for the accessibility of Web sites, while...

  7. An Open-Source Web-Based Tool for Resource-Agnostic Interactive Translation Prediction

    Directory of Open Access Journals (Sweden)

    Daniel Torregrosa

    2014-09-01

    Full Text Available We present a web-based open-source tool for interactive translation prediction (ITP and describe its underlying architecture. ITP systems assist human translators by making context-based computer-generated suggestions as they type. Most of the ITP systems in literature are strongly coupled with a statistical machine translation system that is conveniently adapted to provide the suggestions. Our system, however, follows a resource-agnostic approach and suggestions are obtained from any unmodified black-box bilingual resource. This paper reviews our ITP method and describes the architecture of Forecat, a web tool, partly based on the recent technology of web components, that eases the use of our ITP approach in any web application requiring this kind of translation assistance. We also evaluate the performance of our method when using an unmodified Moses-based statistical machine translation system as the bilingual resource.

  8. Web Security, Privacy & Commerce

    CERN Document Server

    Garfinkel, Simson

    2011-01-01

    Since the first edition of this classic reference was published, World Wide Web use has exploded and e-commerce has become a daily part of business and personal life. As Web use has grown, so have the threats to our security and privacy--from credit card fraud to routine invasions of privacy by marketers to web site defacements to attacks that shut down popular web sites. Web Security, Privacy & Commerce goes behind the headlines, examines the major security risks facing us today, and explains how we can minimize them. It describes risks for Windows and Unix, Microsoft Internet Exp

  9. Cuypers : a semi-automatic hypermedia generation system

    NARCIS (Netherlands)

    J.R. van Ossenbruggen (Jacco); F.J. Cornelissen; J.P.T.M. Geurts (Joost); L. Rutledge (Lloyd); L. Hardman (Lynda)

    2000-01-01

    textabstractThe report describes the architecture of emph{Cuypers, a system supporting second and third generation Web-based multimedia. First generation Web-content encodes information in handwritten (HTML) Web pages. Second generation Web content generates HTML pages on demand, e.g. by filling in

  10. Semantics-based Automated Web Testing

    Directory of Open Access Journals (Sweden)

    Hai-Feng Guo

    2015-08-01

    Full Text Available We present TAO, a software testing tool performing automated test and oracle generation based on a semantic approach. TAO entangles grammar-based test generation with automated semantics evaluation using a denotational semantics framework. We show how TAO can be incorporated with the Selenium automation tool for automated web testing, and how TAO can be further extended to support automated delta debugging, where a failing web test script can be systematically reduced based on grammar-directed strategies. A real-life parking website is adopted throughout the paper to demonstrate the effectivity of our semantics-based web testing approach.

  11. Cuypers : a semi-automatic hypermedia generation system

    OpenAIRE

    Ossenbruggen, Jacco; Cornelissen, F.J.; Geurts, Joost; Rutledge, Lloyd; Hardman, Lynda

    2000-01-01

    textabstractThe report describes the architecture of emph{Cuypers, a system supporting second and third generation Web-based multimedia. First generation Web-content encodes information in handwritten (HTML) Web pages. Second generation Web content generates HTML pages on demand, e.g. by filling in templates with content retrieved dynamically from a database or transformation of structured documents using style sheets (e.g. XSLT). Third generation Web pages will make use of rich markup (e.g. ...

  12. SaaS ve web designu

    OpenAIRE

    Míka, Filip

    2011-01-01

    This thesis is aimed to evaluate if the current SaaS market is able to meet functional re-quirements of web design in order to appropriately support web design's activities. The theoretical part introduces the web design model which describes web design's functional requirements. The next section presents a research concept that describes model assessment (i.e. solutions delivered as SaaS that support web design) and evaluation process. The results show that the current SaaS market is able to...

  13. WebVR: an interactive web browser for virtual environments

    Science.gov (United States)

    Barsoum, Emad; Kuester, Falko

    2005-03-01

    The pervasive nature of web-based content has lead to the development of applications and user interfaces that port between a broad range of operating systems and databases, while providing intuitive access to static and time-varying information. However, the integration of this vast resource into virtual environments has remained elusive. In this paper we present an implementation of a 3D Web Browser (WebVR) that enables the user to search the internet for arbitrary information and to seamlessly augment this information into virtual environments. WebVR provides access to the standard data input and query mechanisms offered by conventional web browsers, with the difference that it generates active texture-skins of the web contents that can be mapped onto arbitrary surfaces within the environment. Once mapped, the corresponding texture functions as a fully integrated web-browser that will respond to traditional events such as the selection of links or text input. As a result, any surface within the environment can be turned into a web-enabled resource that provides access to user-definable data. In order to leverage from the continuous advancement of browser technology and to support both static as well as streamed content, WebVR uses ActiveX controls to extract the desired texture skin from industry strength browsers, providing a unique mechanism for data fusion and extensibility.

  14. Lossy compression for Animated Web Visualisation

    Science.gov (United States)

    Prudden, R.; Tomlinson, J.; Robinson, N.; Arribas, A.

    2017-12-01

    This talk will discuss an technique for lossy data compression specialised for web animation. We set ourselves the challenge of visualising a full forecast weather field as an animated 3D web page visualisation. This data is richly spatiotemporal, however it is routinely communicated to the public as a 2D map, and scientists are largely limited to visualising data via static 2D maps or 1D scatter plots. We wanted to present Met Office weather forecasts in a way that represents all the generated data. Our approach was to repurpose the technology used to stream high definition videos. This enabled us to achieve high rates of compression, while being compatible with both web browsers and GPU processing. Since lossy compression necessarily involves discarding information, evaluating the results is an important and difficult problem. This is essentially a problem of forecast verification. The difficulty lies in deciding what it means for two weather fields to be "similar", as simple definitions such as mean squared error often lead to undesirable results. In the second part of the talk, I will briefly discuss some ideas for alternative measures of similarity.

  15. Web X-Ray: Developing and Adopting Web Best Practices in Enterprises

    Directory of Open Access Journals (Sweden)

    Reinaldo Ferreira

    2016-12-01

    Full Text Available The adoption of Semantic Web technologies constitutes a promising approach to data structuring and integration, both for public and private usage. While these technologies have been around for some time, their adoption is behind overall expectations, particularly in the case of Enterprises. Having that in mind, we developed a Semantic Web Implementation Model that measures and facilitates the implementation of the technology. The advantages of using the model proposed are two-fold: the model serves as a guide for driving the implementation of the Semantic Web as well as it helps to evaluate the impact of the introduction of the technology. The model was adopted by 19 enterprises in an Action Research intervention of one year with promising results: according to the model's scale, in average, all enterprises evolved from a 6% evaluation to 46% during that period. Furthermore, practical implementation recommendations, a typical consulting tool, were developed and adopted during the project by all enterprises, providing important guidelines for the identification of a development path that may be adopted on a larger scale. Meanwhile, the project also outlined that most enterprises were interested in an even broader scope of the Implementation Model and the ambition of a "All Web Technologies" approach arose. One model that could embrace the observable overlapping of different Web generations, namely the Web of Documents, the Social Web, the Web of Data and, ultimately, the Web of Context. One model that could combine the evaluation and guidance for all enterprises to follow. That's the goal of the undergoing "Project Web X-ray" that aims to involve 200 enterprises in the adoption of best practices that may lead to their business development based on Web technologies. This paper presents a case of how Action Research promoted the simultaneous advancement of academic research and enterprise development and introduces the framework and opportunities

  16. The Creative task Creator: a tool for the generation of customized, Web-based creativity tasks.

    Science.gov (United States)

    Pretz, Jean E; Link, John A

    2008-11-01

    This article presents a Web-based tool for the creation of divergent-thinking and open-ended creativity tasks. A Java program generates HTML forms with PHP scripting that run an Alternate Uses Task and/or open-ended response items. Researchers may specify their own instructions, objects, and time limits, or use default settings. Participants can also be prompted to select their best responses to the Alternate Uses Task (Silvia et al., 2008). Minimal programming knowledge is required. The program runs on any server, and responses are recorded in a standard MySQL database. Responses can be scored using the consensual assessment technique (Amabile, 1996) or Torrance's (1998) traditional scoring method. Adoption of this Web-based tool should facilitate creativity research across cultures and access to eminent creators. The Creative Task Creator may be downloaded from the Psychonomic Society's Archive of Norms, Stimuli, and Data, www.psychonomic.org/archive.

  17. PseKRAAC: a flexible web server for generating pseudo K-tuple reduced amino acids composition.

    Science.gov (United States)

    Zuo, Yongchun; Li, Yuan; Chen, Yingli; Li, Guangpeng; Yan, Zhenhe; Yang, Lei

    2017-01-01

    The reduced amino acids perform powerful ability for both simplifying protein complexity and identifying functional conserved regions. However, dealing with different protein problems may need different kinds of cluster methods. Encouraged by the success of pseudo-amino acid composition algorithm, we developed a freely available web server, called PseKRAAC (the pseudo K-tuple reduced amino acids composition). By implementing reduced amino acid alphabets, the protein complexity can be significantly simplified, which leads to decrease chance of overfitting, lower computational handicap and reduce information redundancy. PseKRAAC delivers more capability for protein research by incorporating three crucial parameters that describes protein composition. Users can easily generate many different modes of PseKRAAC tailored to their needs by selecting various reduced amino acids alphabets and other characteristic parameters. It is anticipated that the PseKRAAC web server will become a very useful tool in computational proteomics and protein sequence analysis. Freely available on the web at http://bigdata.imu.edu.cn/psekraac CONTACTS: yczuo@imu.edu.cn or imu.hema@foxmail.com or yanglei_hmu@163.comSupplementary information: Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  18. Differences of Perceived Image Generated through the Web Site: Empirical Evidence Obtained in Spanish Destinations

    Science.gov (United States)

    Blazquez-Resino, Juan J.; Muro-Rodriguez, Ana I.; Perez-Jimenez, Israel R.

    2016-01-01

    In this paper, a study of the perceived destination image created by promotional Web Pages is expounded in an attempt to identify their differences as generators of destination image in the consumers' mind. Specifically, it seeks to analyse whether the web sites of different Spanish regions improve the image that consumers have of the destination, identifying their main dimensions and analysing its effect on satisfaction and intentions of the future behavior of potential visitors. To achieve these objectives and verify the hypotheses, a laboratory experiment was performed, where it was determined what changes are produced in the tourist's previous image after browsing the tourist webs of three different regions. Moreover, it analyses the differences in the effect of the perceived image on satisfaction and potential visitors' future behavioral intentions. The results obtained enable us to identify differences in the composition of the perceived image according to the destination, while confirming the significant effect of different perceived image dimensions regarding satisfaction. The results allow managers to gain a better understanding of the effectiveness of their sites from a consumer perspective as well as suggestions to follow in order to achieve greater efficiency in their communication actions in order to improve the motivation of visitors to go to the destination. PMID:27933027

  19. Differences of perceived image generated through the Web site: Empirical Evidence Obtained in Spanish Destinations

    Directory of Open Access Journals (Sweden)

    Juan Jose Blazquez-Resino

    2016-11-01

    Full Text Available In this paper, a study of the perceived destination image created by promotional Web Pages is expounded in an attempt to identify their differences as generators of destination image in the consumers’ mind. Specifically, it seeks to analyse whether the web sites of different Spanish regions improve the image that consumers have of the destination, identifying their main dimensions and analysing its effect on satisfaction and intentions of the future behaviour of potential visitors. To achieve these objectives and verify the hypotheses, a laboratory experiment was performed, where it was determined what changes are produced in the tourist´s previous image after browsing the tourist webs of three different regions. Moreover, it analyses the differences in the effect of the perceived image on satisfaction and potential visitors´ future behavioural intentions. The results obtained enable us to identify differences in the composition of the perceived image according to the destination, while confirming the significant effect of different perceived image dimensions regarding satisfaction. The results allow managers to gain a better understanding of the effectiveness of their sites from a consumer perspective as well as suggestions to follow in order to achieve greater efficiency in their communication actions in order to improve the motivation of visitors to go to the destination.

  20. It's Time to Use a Wiki as Part of Your Web Site

    Science.gov (United States)

    Ribaric, Tim

    2007-01-01

    Without a doubt, the term "wiki" has leaked into almost every discussion concerning Web 2.0. The real question becomes: Is there a place for a wiki on every library Web site? The answer should be an emphatic "yes." People often praise the wiki because it offers simple page creation and provides instant gratification for amateur Web developers.…

  1. Chemical cleaning an essential part of steam generator asset management

    International Nuclear Information System (INIS)

    Amman, Franz

    2008-01-01

    Chemical Cleaning an essential part of Steam Generator asset management accumulation of deposits is intrinsic for the operation of Steam Generators in PWRs. Such depositions often lead to reduction of thermal performance, loss of component integrity and, in some cases to power restrictions. Accordingly removal of such deposits is an essential part of the asset management of the Steam Generators in a Nuclear Power Plant. Every plant has its individual condition, history and constraints which need to be considered when planning and performing a chemical cleaning. Typical points are: - Sludge load amount and constitution of the deposits - Sludge distribution in the steam generator - Existing or expected corrosion problems - Amount and tendency of fouling for waste treatment Depending on this points the strategy for chemical cleaning shall be evolved. the range of treatment starts with very soft cleanings with a removal of approx 100 kg per steam generator and goes to a full scale cleaning which can remove up to several thousand kilograms of deposits from a steam generator. Depending on the goal to be achieved and the steam generator present an adequate cleaning method shall be selected. This requires flexible and 'customisable' cleaning methods that can be adapted to the individual needs of a plant. Such customizing of chemical cleaning methods is an essential factor for an optimized asset management of the steam generator in a nuclear power plant

  2. Moving toward a universally accessible web: Web accessibility and education.

    Science.gov (United States)

    Kurt, Serhat

    2017-12-08

    The World Wide Web is an extremely powerful source of information, inspiration, ideas, and opportunities. As such, it has become an integral part of daily life for a great majority of people. Yet, for a significant number of others, the internet offers only limited value due to the existence of barriers which make accessing the Web difficult, if not impossible. This article illustrates some of the reasons that achieving equality of access to the online world of education is so critical, explores the current status of Web accessibility, discusses evaluative tools and methods that can help identify accessibility issues in educational websites, and provides practical recommendations and guidelines for resolving some of the obstacles that currently hinder the achievability of the goal of universal Web access.

  3. Executing Medical Guidelines on the Web: Towards Next Generation Healthcare

    Science.gov (United States)

    Argüello, M.; Des, J.; Fernandez-Prieto, M. J.; Perez, R.; Paniagua, H.

    There is still a lack of full integration between current Electronic Health Records (EHRs) and medical guidelines that encapsulate evidence-based medicine. Thus, general practitioners (GPs) and specialised physicians still have to read document-based medical guidelines and decide among various options for managing common non-life-threatening conditions where the selection of the most appropriate therapeutic option for each individual patient can be a difficult task. This paper presents a simulation framework and computational test-bed, called V.A.F. Framework, for supporting simulations of clinical situations that boosted the integration between Health Level Seven (HL7) and Semantic Web technologies (OWL, SWRL, and OWL-S) to achieve content layer interoperability between online clinical cases and medical guidelines, and therefore, it proves that higher integration between EHRs and evidence-based medicine can be accomplished which could lead to a next generation of healthcare systems that provide more support to physicians and increase patients' safety.

  4. Classroom Assessment in Web-Based Instructional Environment: Instructors' Experience

    Directory of Open Access Journals (Sweden)

    Xin Liang

    2004-03-01

    Full Text Available While a great deal has been written on the advantage and benefits of online teaching, little is known on how..assessment is implemented in online classrooms to monitor and inform performance and progress. The..purpose of this study is to investigate the dynamics of WebCT classroom assessment by analyzing the..perceptions and experience of the instructors. Grounded theory method was employed to generate a - process..theory- . The study included 10 faculties who taught WebCT classes, and 216 students in the College of..Education in an urban university in the Mid west. Interviews and classroom observations were undertaken..on line. The findings indicated that, performance-based assessment, writing skills, interactive assessment..and learner autonomy were major assessment aspects to inform teaching and enhance learning. If one of..the major roles of online instruction is to increase self-directed learning, as part of the pedagogical..mechanism, web-based classroom assessment should be designed and practiced to impact learner autonomy.

  5. Web 2.0 Solutions to Wicked Climate Change Problems

    Directory of Open Access Journals (Sweden)

    Alanah Kazlauskas

    2010-01-01

    Full Text Available One of the most pressing ‘wicked problems’ facing humankind is climate change together with its many interrelated environmental concerns. The complexity of this set of problems can be overwhelming as there is such diversity among both the interpretations of the scientific evidence and the viability of possible solutions. Among the social technologies associated with the second generation of the Internet known as Web 2.0, there are tools that allow people to communicate, coordinate and collaborate in ways that reduce their carbon footprint and a potential to become part of the climate change solution. However the way forward is not obvious or easy as Web 2.0, while readily accepted in the chaotic social world, is often treated with suspicion in the more ordered world of business and government. This paper applies a holistic theoretical sense-making framework to research and practice on potential Web 2.0 solutions to climate change problems. The suite of issues, activities and tools involved are viewed as an ecosystem where all elements are dynamic and inter-related. Through such innovative thinking the Information Systems community can make a valuable contribution to a critical global problem and hence find a new relevance as part of the solution.

  6. The Role of Web Interviews as Part of a National Travel Survey

    DEFF Research Database (Denmark)

    Christensen, Linda

    2013-01-01

    Purpose — The paper is analysing the effect of adding a web survey to a traditional telephone-based national travel survey by asking the respondents to check in on the web and answer the questions there (Computer Assisted Web Interview, CAWI). If they are not participating by web they are as usual...... called by telephone (Computer Assisted Telephone Interview, CATI). Design/methodology/approach — Multivariate regression analyses are used to analyse the difference in response rates by the two media and to analyse if respondents’ answering by the two media have different travel patterns. Findings...... — The analyses show that web interviews are saving money, even though a more intensive post-processing is necessary. The analyses seem to show that the CAWI is resulting in a more careful answering which results in more trips reported. A CAWI is increasing the participation of children in the survey...

  7. Reliable execution based on CPN and skyline optimization for Web service composition.

    Science.gov (United States)

    Chen, Liping; Ha, Weitao; Zhang, Guojun

    2013-01-01

    With development of SOA, the complex problem can be solved by combining available individual services and ordering them to best suit user's requirements. Web services composition is widely used in business environment. With the features of inherent autonomy and heterogeneity for component web services, it is difficult to predict the behavior of the overall composite service. Therefore, transactional properties and nonfunctional quality of service (QoS) properties are crucial for selecting the web services to take part in the composition. Transactional properties ensure reliability of composite Web service, and QoS properties can identify the best candidate web services from a set of functionally equivalent services. In this paper we define a Colored Petri Net (CPN) model which involves transactional properties of web services in the composition process. To ensure reliable and correct execution, unfolding processes of the CPN are followed. The execution of transactional composition Web service (TCWS) is formalized by CPN properties. To identify the best services of QoS properties from candidate service sets formed in the TCSW-CPN, we use skyline computation to retrieve dominant Web service. It can overcome that the reduction of individual scores to an overall similarity leads to significant information loss. We evaluate our approach experimentally using both real and synthetically generated datasets.

  8. Exponentiation for products of Wilson lines within the generating function approach

    International Nuclear Information System (INIS)

    Vladimirov, A.A.

    2015-01-01

    We present the generating function approach to the perturbative exponentiation of correlators of a product of Wilson lines and loops. The exponentiated expression is presented in closed form as an algebraic function of correlators of known operators, which can be seen as a generating function for web diagrams. The expression is naturally split onto two parts: the exponentiation kernel, which accumulates all non-trivial information about web diagrams, and the defect of exponentiation, which reconstructs the matrix exponent and is a function of the exponentiation kernel. The detailed comparison of the presented approach with existing approaches to exponentiation is presented as well. We also give examples of calculations within the generating function exponentiation, namely, we consider different configurations of light-like Wilson lines in the multi-gluon-exchange-webs (MGEW) approximation. Within this approximation the corresponding correlators can be calculated exactly at any order of perturbative expansion by only algebraic manipulations. The MGEW approximation shows violation of the dipole formula for infrared singularities at three-loop order.

  9. APFEL Web a web-based application for the graphical visualization of parton distribution functions

    CERN Document Server

    Carrazza, Stefano; Palazzo, Daniele; Rojo, Juan

    2015-01-01

    We present APFEL Web, a web-based application designed to provide a flexible user-friendly tool for the graphical visualization of parton distribution functions (PDFs). In this note we describe the technical design of the APFEL Web application, motivating the choices and the framework used for the development of this project. We document the basic usage of APFEL Web and show how it can be used to provide useful input for a variety of collider phenomenological studies. Finally we provide some examples showing the output generated by the application.

  10. APFEL Web: a web-based application for the graphical visualization of parton distribution functions

    International Nuclear Information System (INIS)

    Carrazza, Stefano; Ferrara, Alfio; Palazzo, Daniele; Rojo, Juan

    2015-01-01

    We present APFEL Web, a Web-based application designed to provide a flexible user-friendly tool for the graphical visualization of parton distribution functions. In this note we describe the technical design of the APFEL Web application, motivating the choices and the framework used for the development of this project. We document the basic usage of APFEL Web and show how it can be used to provide useful input for a variety of collider phenomenological studies. Finally we provide some examples showing the output generated by the application. (note)

  11. Building Social Web Applications

    CERN Document Server

    Bell, Gavin

    2009-01-01

    Building a web application that attracts and retains regular visitors is tricky enough, but creating a social application that encourages visitors to interact with one another requires careful planning. This book provides practical solutions to the tough questions you'll face when building an effective community site -- one that makes visitors feel like they've found a new home on the Web. If your company is ready to take part in the social web, this book will help you get started. Whether you're creating a new site from scratch or reworking an existing site, Building Social Web Applications

  12. 8th Chinese Conference on The Semantic Web and Web Science

    CERN Document Server

    Du, Jianfeng; Wang, Haofen; Wang, Peng; Ji, Donghong; Pan, Jeff Z; CSWS 2014

    2014-01-01

    This book constitutes the thoroughly refereed papers of the 8th Chinese Conference on The Semantic Web and Web Science, CSWS 2014, held in Wuhan, China, in August 2014. The 22 research papers presented were carefully reviewed and selected from 61 submissions. The papers are organized in topical sections such as ontology reasoning and learning; semantic data generation and management; and semantic technology and applications.

  13. Post-Web 2.0 Pedagogy: From Student-Generated Content to International Co-Production Enabled by Mobile Social Media

    Science.gov (United States)

    Cochrane, Thomas; Antonczak, Laurent; Wagner, Daniel

    2013-01-01

    The advent of web 2.0 has enabled new forms of collaboration centred upon user-generated content, however, mobile social media is enabling a new wave of social collaboration. Mobile devices have disrupted and reinvented traditional media markets and distribution: iTunes, Google Play and Amazon now dominate music industry distribution channels,…

  14. Web-Mediated Knowledge Synthesis for Educators

    Science.gov (United States)

    DeSchryver, Michael

    2015-01-01

    Ubiquitous and instant access to information on the Web is challenging what constitutes 21st century literacies. This article explores the notion of Web-mediated knowledge synthesis, an approach to integrating Web-based learning that may result in generative synthesis of ideas. This article describes the skills and strategies that may support…

  15. Exploring the academic invisible web

    OpenAIRE

    Lewandowski, Dirk; Mayr, Philipp

    2006-01-01

    Purpose: To provide a critical review of Bergman’s 2001 study on the Deep Web. In addition, we bring a new concept into the discussion, the Academic Invisible Web (AIW). We define the Academic Invisible Web as consisting of all databases and collections relevant to academia but not searchable by the general-purpose internet search engines. Indexing this part of the Invisible Web is central to scientific search engines. We provide an overview of approaches followed thus far. Design/methodol...

  16. Personalized Metaheuristic Clustering Onto Web Documents

    Institute of Scientific and Technical Information of China (English)

    Wookey Lee

    2004-01-01

    Optimal clustering for the web documents is known to complicated cornbinatorial Optimization problem and it is hard to develop a generally applicable oplimal algorithm. An accelerated simuIated arlneaIing aIgorithm is developed for automatic web document classification. The web document classification problem is addressed as the problem of best describing a match between a web query and a hypothesized web object. The normalized term frequency and inverse document frequency coefficient is used as a measure of the match. Test beds are generated on - line during the search by transforming model web sites. As a result, web sites can be clustered optimally in terms of keyword vectofs of corresponding web documents.

  17. PERANCANGAN SISTEM PEMESANAN BARANG BERBASIS WEB DI TOKO ZENITH KOMPUTER DI PEKANBARU

    Directory of Open Access Journals (Sweden)

    Fery Wongso Johan Wyanaputra

    2016-03-01

    Full Text Available Abstrak : Sistem Informasi Pemesananberbasis web merupakan bagian dari sistem informasi pemasaran yang dikembangkan untuk mengumpulkan, mengolah data sehingga data tersebut dapat dilihat kembali untuk disalurkan sebagai suatu informasi yang berguna. Wujud dari pengembangan Sistem Informasi Pemesananberbasis web ini adalah pembuatan aplikasi Komputer yang mampu mewakili sistem informasi yang dirancang secara keseluruhan.Aplikasi Sistem Informasi pemesanan yang dihasilkan mampu mengelola data pemesanan secara terorgasisasi, serta menghasilkan laporan yang lengkap, akurat dan selalu aktual untuk setiap tingkatan manajemen. Perancangan sistemnya menggunakan PHP (Personal Home Page dan rancangan databasenya menggunakan Xamp Server. Hasil dari perancangan aplikasi Sistem Informasi Pemesananberbasis web menunjukkan bahwa peranan aplikasi Komputer dalam sistem informasi sangat penting sebagai penunjang dalam meningkatkan kualitas kegiatan Pemesanan dan pelayanan di lingkungan toko Zenith Komputer. Abstract: Information Systems is a web Pemesananberbasis Part Of The Marketing Information System was developed to review collect, process the data so that the data can be Seen Back to review the information supplied as a useful thing. From the form of development of Information Systems web Pemesananberbasis Singer Was Able Computer Application Development Information System Designed represent keseluruhan.Aplikasi Operating System Generated Ordering Information Ability to Manage Booking Data Operating terorgasisasi, as well as generate reports The complete, Accurate And Always Currents for EVERY level management review, The design of the system using PHP (Personal Home Page and database design using XAMP Server. Results From designing web applications Pemesananberbasis Information System showed that Role of Information Systems Computer Application hearts hearts Sangat as supporting activities improving QUALITY Booking And Environmental Services at Zenith

  18. Client-Side Event Processing for Personalized Web Advertisement

    Science.gov (United States)

    Stühmer, Roland; Anicic, Darko; Sen, Sinan; Ma, Jun; Schmidt, Kay-Uwe; Stojanovic, Nenad

    The market for Web advertisement is continuously growing and correspondingly, the number of approaches that can be used for realizing Web advertisement are increasing. However, current approaches fail to generate very personalized ads for a current Web user that is visiting a particular Web content. They mainly try to develop a profile based on the content of that Web page or on a long-term user's profile, by not taking into account current user's preferences. We argue that by discovering a user's interest from his current Web behavior we can support the process of ad generation, especially the relevance of an ad for the user. In this paper we present the conceptual architecture and implementation of such an approach. The approach is based on the extraction of simple events from the user interaction with a Web page and their combination in order to discover the user's interests. We use semantic technologies in order to build such an interpretation out of many simple events. We present results from preliminary evaluation studies. The main contribution of the paper is a very efficient, semantic-based client-side architecture for generating and combining Web events. The architecture ensures the agility of the whole advertisement system, by complexly processing events on the client. In general, this work contributes to the realization of new, event-driven applications for the (Semantic) Web.

  19. Emergent web intelligence advanced information retrieval

    CERN Document Server

    Badr, Youakim; Abraham, Ajith; Hassanien, Aboul-Ella

    2010-01-01

    Web Intelligence explores the impact of artificial intelligence and advanced information technologies representing the next generation of Web-based systems, services, and environments, and designing hybrid web systems that serve wired and wireless users more efficiently. Multimedia and XML-based data are produced regularly and in increasing way in our daily digital activities, and their retrieval must be explored and studied in this emergent web-based era. 'Emergent Web Intelligence: Advanced information retrieval, provides reviews of the related cutting-edge technologies and insights. It is v

  20. Connecting small and medium enterprises to the new consumer: The Web 2.0 as marketing tool

    NARCIS (Netherlands)

    Constantinides, Efthymios; Bharati, P.; Lee, I.

    2010-01-01

    This chapter explains the nature, effects and current standing of the new generation of Internet applications, commonly known as Social Media or Web 2.0, reviews their role as marketing instruments and identifies opportunities for SMEs for engaging them as part of their marketing strategy. The

  1. Caught in the Web

    Energy Technology Data Exchange (ETDEWEB)

    Gillies, James

    1995-06-15

    The World-Wide Web may have taken the Internet by storm, but many people would be surprised to learn that it owes its existence to CERN. Around half the world's particle physicists come to CERN for their experiments, and the Web is the result of their need to share information quickly and easily on a global scale. Six years after Tim Berners-Lee's inspired idea to marry hypertext to the Internet in 1989, CERN is handing over future Web development to the World-Wide Web Consortium, run by the French National Institute for Research in Computer Science and Control, INRIA, and the Laboratory for Computer Science of the Massachusetts Institute of Technology, MIT, leaving itself free to concentrate on physics. The Laboratory marked this transition with a conference designed to give a taste of what the Web can do, whilst firmly stamping it with the label ''Made in CERN''. Over 200 European journalists and educationalists came to CERN on 8 - 9 March for the World-Wide Web Days, resulting in wide media coverage. The conference was opened by UK Science Minister David Hunt who stressed the importance of fundamental research in generating new ideas. ''Who could have guessed 10 years ago'', he said, ''that particle physics research would lead to a communication system which would allow every school to have the biggest library in the world in a single computer?''. In his introduction, the Minister also pointed out that ''CERN and other basic research laboratories help to break new technological ground and sow the seeds of what will become mainstream manufacturing in the future.'' Learning the jargon is often the hardest part of coming to grips with any new invention, so CERN put it at the top of the agenda. Jacques Altaber, who helped introduce the Internet to CERN in the early 1980s, explained that without the Internet, the Web couldn't exist. The Internet began as a US Defense Department research project in the 1970s and has grown into a global network-ofnetworks linking some

  2. The RCSB Protein Data Bank: redesigned web site and web services.

    Science.gov (United States)

    Rose, Peter W; Beran, Bojan; Bi, Chunxiao; Bluhm, Wolfgang F; Dimitropoulos, Dimitris; Goodsell, David S; Prlic, Andreas; Quesada, Martha; Quinn, Gregory B; Westbrook, John D; Young, Jasmine; Yukich, Benjamin; Zardecki, Christine; Berman, Helen M; Bourne, Philip E

    2011-01-01

    The RCSB Protein Data Bank (RCSB PDB) web site (http://www.pdb.org) has been redesigned to increase usability and to cater to a larger and more diverse user base. This article describes key enhancements and new features that fall into the following categories: (i) query and analysis tools for chemical structure searching, query refinement, tabulation and export of query results; (ii) web site customization and new structure alerts; (iii) pair-wise and representative protein structure alignments; (iv) visualization of large assemblies; (v) integration of structural data with the open access literature and binding affinity data; and (vi) web services and web widgets to facilitate integration of PDB data and tools with other resources. These improvements enable a range of new possibilities to analyze and understand structure data. The next generation of the RCSB PDB web site, as described here, provides a rich resource for research and education.

  3. Geo-communication and web-based infrastructure

    DEFF Research Database (Denmark)

    Brodersen, Lars; Nielsen, Anders

    2005-01-01

    The role of geo-information and the distribution of geo-information have changed dramatically since the introduction of web-services on the Internet. In the framework of web-services maps should be seen as an index to further geo-information. Maps are no longer an aim in themselves. In this context...... web-services perform the function as index-portals on the basis of geoinformation. The introduction of web-services as index-portals based on geoinformation has changed the conditions for both content and form of geocommunication. A high number of players and interactions (as well as a very high...... number of all kinds of information and combinations of these) characterize web-services, where maps are only a part of the whole. These new conditions demand new ways of modelling the processes leading to geo-communication. One new aspect is the fact that the service providers have become a part...

  4. Next-Generation Web Frameworks in Python

    CERN Document Server

    Daly, Liza

    2007-01-01

    With its flexibility, readability, and maturecode libraries, Python is a naturalchoice for developing agile and maintainableweb applications. Severalframeworks have emerged in the last fewyears that share ideas with Ruby on Railsand leverage the expressive nature of Python.This Short Cut will tell you whatyou need to know about the hottest fullstackframeworks: Django, Pylons, andTurboGears. Their philosophies, relativestrengths, and development status aredescribed in detail. What you won't find out is, "Which oneshould I use?" The short answer is thatall of them can be used to build web appl

  5. Collaborative web hosting challenges and research directions

    CERN Document Server

    Ahmed, Reaz

    2014-01-01

    This brief presents a peer-to-peer (P2P) web-hosting infrastructure (named pWeb) that can transform networked, home-entertainment devices into lightweight collaborating Web servers for persistently storing and serving multimedia and web content. The issues addressed include ensuring content availability, Plexus routing and indexing, naming schemes, web ID, collaborative web search, network architecture and content indexing. In pWeb, user-generated voluminous multimedia content is proactively uploaded to a nearby network location (preferably within the same LAN or at least, within the same ISP)

  6. Significant Benefits from Libraries in Web 3.0 Environment

    African Journals Online (AJOL)

    pc

    2018-03-05

    Mar 5, 2018 ... Keywords- Web 3.0, library 3.0, Web 3.0 Applications, Semantic. Web ... providing virtual information services, and other services cannot be ... web third generation, definition, beginning, and retrieve system. The study ...

  7. Automating Visualization Service Generation with the WATT Compiler

    Science.gov (United States)

    Bollig, E. F.; Lyness, M. D.; Erlebacher, G.; Yuen, D. A.

    2007-12-01

    As tasks and workflows become increasingly complex, software developers are devoting increasing attention to automation tools. Among many examples, the Automator tool from Apple collects components of a workflow into a single script, with very little effort on the part of the user. Tasks are most often described as a series of instructions. The granularity of the tasks dictates the tools to use. Compilers translate fine-grained instructions to assembler code, while scripting languages (ruby, perl) are used to describe a series of tasks at a higher level. Compilers can also be viewed as transformational tools: a cross-compiler can translate executable code written on one computer to assembler code understood on another, while transformational tools can translate from one high-level language to another. We are interested in creating visualization web services automatically, starting from stand-alone VTK (Visualization Toolkit) code written in Tcl. To this end, using the OCaml programming language, we have developed a compiler that translates Tcl into C++, including all the stubs, classes and methods to interface with gSOAP, a C++ implementation of the Soap 1.1/1.2 protocols. This compiler, referred to as the Web Automation and Translation Toolkit (WATT), is the first step towards automated creation of specialized visualization web services without input from the user. The WATT compiler seeks to automate all aspects of web service generation, including the transport layer, the division of labor and the details related to interface generation. The WATT compiler is part of ongoing efforts within the NSF funded VLab consortium [1] to facilitate and automate time-consuming tasks for the science related to understanding planetary materials. Through examples of services produced by WATT for the VLab portal, we will illustrate features, limitations and the improvements necessary to achieve the ultimate goal of complete and transparent automation in the generation of web

  8. Web-video-mining-supported workflow modeling for laparoscopic surgeries.

    Science.gov (United States)

    Liu, Rui; Zhang, Xiaoli; Zhang, Hao

    2016-11-01

    As quality assurance is of strong concern in advanced surgeries, intelligent surgical systems are expected to have knowledge such as the knowledge of the surgical workflow model (SWM) to support their intuitive cooperation with surgeons. For generating a robust and reliable SWM, a large amount of training data is required. However, training data collected by physically recording surgery operations is often limited and data collection is time-consuming and labor-intensive, severely influencing knowledge scalability of the surgical systems. The objective of this research is to solve the knowledge scalability problem in surgical workflow modeling with a low cost and labor efficient way. A novel web-video-mining-supported surgical workflow modeling (webSWM) method is developed. A novel video quality analysis method based on topic analysis and sentiment analysis techniques is developed to select high-quality videos from abundant and noisy web videos. A statistical learning method is then used to build the workflow model based on the selected videos. To test the effectiveness of the webSWM method, 250 web videos were mined to generate a surgical workflow for the robotic cholecystectomy surgery. The generated workflow was evaluated by 4 web-retrieved videos and 4 operation-room-recorded videos, respectively. The evaluation results (video selection consistency n-index ≥0.60; surgical workflow matching degree ≥0.84) proved the effectiveness of the webSWM method in generating robust and reliable SWM knowledge by mining web videos. With the webSWM method, abundant web videos were selected and a reliable SWM was modeled in a short time with low labor cost. Satisfied performances in mining web videos and learning surgery-related knowledge show that the webSWM method is promising in scaling knowledge for intelligent surgical systems. Copyright © 2016 Elsevier B.V. All rights reserved.

  9. A web based tool for storing and visualising data generated within a smart home.

    Science.gov (United States)

    McDonald, H A; Nugent, C D; Moore, G; Finlay, D D; Hallberg, J

    2011-01-01

    There is a growing need to re-assess the current approaches available to researchers for storing and managing heterogeneous data generated within a smart home environment. In our current work we have developed the homeML Application; a web based tool to support researchers engaged in the area of smart home research as they perform experiments. Within this paper the homeML Application is presented which includes the fundamental components of the homeML Repository and the homeML Toolkit. Results from a usability study conducted by 10 computer science researchers are presented; the initial results of which have been positive.

  10. WebMGA: a customizable web server for fast metagenomic sequence analysis.

    Science.gov (United States)

    Wu, Sitao; Zhu, Zhengwei; Fu, Liming; Niu, Beifang; Li, Weizhong

    2011-09-07

    The new field of metagenomics studies microorganism communities by culture-independent sequencing. With the advances in next-generation sequencing techniques, researchers are facing tremendous challenges in metagenomic data analysis due to huge quantity and high complexity of sequence data. Analyzing large datasets is extremely time-consuming; also metagenomic annotation involves a wide range of computational tools, which are difficult to be installed and maintained by common users. The tools provided by the few available web servers are also limited and have various constraints such as login requirement, long waiting time, inability to configure pipelines etc. We developed WebMGA, a customizable web server for fast metagenomic analysis. WebMGA includes over 20 commonly used tools such as ORF calling, sequence clustering, quality control of raw reads, removal of sequencing artifacts and contaminations, taxonomic analysis, functional annotation etc. WebMGA provides users with rapid metagenomic data analysis using fast and effective tools, which have been implemented to run in parallel on our local computer cluster. Users can access WebMGA through web browsers or programming scripts to perform individual analysis or to configure and run customized pipelines. WebMGA is freely available at http://weizhongli-lab.org/metagenomic-analysis. WebMGA offers to researchers many fast and unique tools and great flexibility for complex metagenomic data analysis.

  11. WebMGA: a customizable web server for fast metagenomic sequence analysis

    Directory of Open Access Journals (Sweden)

    Niu Beifang

    2011-09-01

    Full Text Available Abstract Background The new field of metagenomics studies microorganism communities by culture-independent sequencing. With the advances in next-generation sequencing techniques, researchers are facing tremendous challenges in metagenomic data analysis due to huge quantity and high complexity of sequence data. Analyzing large datasets is extremely time-consuming; also metagenomic annotation involves a wide range of computational tools, which are difficult to be installed and maintained by common users. The tools provided by the few available web servers are also limited and have various constraints such as login requirement, long waiting time, inability to configure pipelines etc. Results We developed WebMGA, a customizable web server for fast metagenomic analysis. WebMGA includes over 20 commonly used tools such as ORF calling, sequence clustering, quality control of raw reads, removal of sequencing artifacts and contaminations, taxonomic analysis, functional annotation etc. WebMGA provides users with rapid metagenomic data analysis using fast and effective tools, which have been implemented to run in parallel on our local computer cluster. Users can access WebMGA through web browsers or programming scripts to perform individual analysis or to configure and run customized pipelines. WebMGA is freely available at http://weizhongli-lab.org/metagenomic-analysis. Conclusions WebMGA offers to researchers many fast and unique tools and great flexibility for complex metagenomic data analysis.

  12. A guide: Sale of power opportunities for distributed generators. Part A: the summary

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2005-07-01

    A lengthy guide has been written for the benefit of prospective distributed generators for the electric power market in Great Britain. The growing range of generation methods and the government's emissions reduction programme were the reasons for the study. The document consists of three parts. Part A is an introductory guide for distributed generators in Britain, Part B is a guide to the sale of power opportunities in Britain, and Part C gives additional information. The guide represents a basic but high-level overview of the sale of electric power but anyone with a serious interest in the subject matter is advised to also seek professional help before embarking on an expensive or long-term project. The guide was prepared by SOHN Associates under contract to the DTI.

  13. The Semantic Web and Educational Technology

    Science.gov (United States)

    Maddux, Cleborne D., Ed.

    2008-01-01

    The "Semantic Web" is an idea proposed by Tim Berners-Lee, the inventor of the "World Wide Web." The topic has been generating a great deal of interest and enthusiasm, and there is a rapidly growing body of literature dealing with it. This article attempts to explain how the Semantic Web would work, and explores short-term and long-term…

  14. The Geospatial Web and Local Geographical Education

    Science.gov (United States)

    Harris, Trevor M.; Rouse, L. Jesse; Bergeron, Susan J.

    2010-01-01

    Recent innovations in the Geospatial Web represent a paradigm shift in Web mapping by enabling educators to explore geography in the classroom by dynamically using a rapidly growing suite of impressive online geospatial tools. Coupled with access to spatial data repositories and User-Generated Content, the Geospatial Web provides a powerful…

  15. Web traffic and firm performance

    DEFF Research Database (Denmark)

    Farooq, Omar; Aguenaou, Samir

    2013-01-01

    Does the traffic generated by websites of firms signal anything to stock market participants? Does higher web-traffic translate into availability of more information and therefore lower agency problems? And if answers to above questions are in affirmative, does higher web-traffic traffic translate...... into better firm performance? This paper aims to answer these questions by documenting a positive relationship between the extent of web-traffic and firm performance in the MENA region during the 2010. We argue that higher web-traffic lowers the agency problems in firms by disseminating more information...... to stock market participants. Consequently, lower agency problems translate into better performance. Furthermore, we also show that agency reducing role of web-traffic is more pronounced in regimes where information environment is already bad. For example, our results show stronger impact of web...

  16. Web 2.0 (and Beyond)

    NARCIS (Netherlands)

    P.A. Arora (Payal)

    2015-01-01

    textabstractWeb 2.0 is a term coined to mark a new era of Internet usage driven by user interactivity and collaboration in generating content, moving away from the static information dissemination model associated with Web 1.0. It became common in early 2000 with the growth of social network sites,

  17. SED-ML web tools: generate, modify and export standard-compliant simulation studies.

    Science.gov (United States)

    Bergmann, Frank T; Nickerson, David; Waltemath, Dagmar; Scharm, Martin

    2017-04-15

    The Simulation Experiment Description Markup Language (SED-ML) is a standardized format for exchanging simulation studies independently of software tools. We present the SED-ML Web Tools, an online application for creating, editing, simulating and validating SED-ML documents. The Web Tools implement all current SED-ML specifications and, thus, support complex modifications and co-simulation of models in SBML and CellML formats. Ultimately, the Web Tools lower the bar on working with SED-ML documents and help users create valid simulation descriptions. http://sysbioapps.dyndns.org/SED-ML_Web_Tools/ . fbergman@caltech.edu . © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  18. WebWise 2.0: The Power of Community. WebWise Conference on Libraries and Museums in the Digital World Proceedings (9th, Miami Beach, Florida, March 5-7, 2008)

    Science.gov (United States)

    Green, David

    2009-01-01

    Since it was coined by Tim O'Reilly in formulating the first Web 2.0 Conference in 2004, the term "Web 2.0" has definitely caught on as a designation of a second generation of Web design and experience that emphasizes a high degree of interaction with, and among, users. Rather than simply consulting and reading Web pages, the Web 2.0 generation is…

  19. Prediction of toxicity and comparison of alternatives using WebTEST (Web-services Toxicity Estimation Software Tool)

    Science.gov (United States)

    A Java-based web service is being developed within the US EPA’s Chemistry Dashboard to provide real time estimates of toxicity values and physical properties. WebTEST can generate toxicity predictions directly from a simple URL which includes the endpoint, QSAR method, and ...

  20. Development of STEP-NC Adaptor for Advanced Web Manufacturing System

    Science.gov (United States)

    Ajay Konapala, Mr.; Koona, Ramji, Dr.

    2017-08-01

    Information systems play a key role in the modern era of Information Technology. Rapid developments in IT & global competition calls for many changes in basic CAD/CAM/CAPP/CNC manufacturing chain of operations. ‘STEP-NC’ an enhancement to STEP for operating CNC machines, creating new opportunities for collaborative, concurrent, adaptive works across the manufacturing chain of operations. Schemas and data models defined by ISO14649 in liaison with ISO10303 standards made STEP-NC file rich with feature based, rather than mere point to point information of G/M Code format. But one needs to have a suitable information system to understand and modify these files. Various STEP-NC information systems are reviewed to understand the suitability of STEP-NC for web manufacturing. Present work also deals with the development of an adaptor which imports STEP-NC file, organizes its information, allowing modifications to entity values and finally generates a new STEP-NC file to export. The system is designed and developed to work on web to avail additional benefits through the web and also to be part of a proposed ‘Web based STEP-NC manufacturing platform’ which is under development and explained as future scope.

  1. An Automatic Web Service Composition Framework Using QoS-Based Web Service Ranking Algorithm.

    Science.gov (United States)

    Mallayya, Deivamani; Ramachandran, Baskaran; Viswanathan, Suganya

    2015-01-01

    Web service has become the technology of choice for service oriented computing to meet the interoperability demands in web applications. In the Internet era, the exponential addition of web services nominates the "quality of service" as essential parameter in discriminating the web services. In this paper, a user preference based web service ranking (UPWSR) algorithm is proposed to rank web services based on user preferences and QoS aspect of the web service. When the user's request cannot be fulfilled by a single atomic service, several existing services should be composed and delivered as a composition. The proposed framework allows the user to specify the local and global constraints for composite web services which improves flexibility. UPWSR algorithm identifies best fit services for each task in the user request and, by choosing the number of candidate services for each task, reduces the time to generate the composition plans. To tackle the problem of web service composition, QoS aware automatic web service composition (QAWSC) algorithm proposed in this paper is based on the QoS aspects of the web services and user preferences. The proposed framework allows user to provide feedback about the composite service which improves the reputation of the services.

  2. SEMANTIC WEB SERVICES – DISCOVERY, SELECTION AND COMPOSITION TECHNIQUES

    OpenAIRE

    Sowmya Kamath S; Ananthanarayana V.S

    2013-01-01

    Web services are already one of the most important resources on the Internet. As an integrated solution for realizing the vision of the Next Generation Web, semantic web services combine semantic web technology with web service technology, envisioning automated life cycle management of web services. This paper discusses the significance and importance of service discovery & selection to business logic, and the requisite current research in the various phases of the semantic web...

  3. HEPWEB - WEB-portal for Monte Carlo simulations in high-energy physics

    International Nuclear Information System (INIS)

    Aleksandrov, E.I.; Kotov, V.M.; Uzhinsky, V.V.; Zrelov, P.V.

    2011-01-01

    A WEB-portal HepWeb allows users to perform the most popular calculations in high-energy physics - calculations of hadron-hadron, hadron-nucleus, and nucleus-nucleus interaction cross sections as well as calculations of secondary-particle characteristics in the interactions using Monte Carlo event generators. The list of the generators includes Dubna version of the intranuclear cascade model (CASCADE), FRITIOF model, ultrarelativistic quantum molecular dynamics model (UrQMD), HIJING model, and AMPT model. Setting up the colliding particles/nucleus properties (collision energy, mass numbers and charges of nuclei, impact parameters of interactions, and number of generated events) is realized by a WEB-interface. A query is processed by a server, and results are presented to the user as a WEB-page. Short descriptions of the installed generators, the WEB-interface implementation and the server operation are given

  4. HEPWEB - WEB-portal for Monte Carlo simulations in high-energy physics

    Energy Technology Data Exchange (ETDEWEB)

    Aleksandrov, E I; Kotov, V M; Uzhinsky, V V; Zrelov, P V

    2011-07-01

    A WEB-portal HepWeb allows users to perform the most popular calculations in high-energy physics - calculations of hadron-hadron, hadron-nucleus, and nucleus-nucleus interaction cross sections as well as calculations of secondary-particle characteristics in the interactions using Monte Carlo event generators. The list of the generators includes Dubna version of the intranuclear cascade model (CASCADE), FRITIOF model, ultrarelativistic quantum molecular dynamics model (UrQMD), HIJING model, and AMPT model. Setting up the colliding particles/nucleus properties (collision energy, mass numbers and charges of nuclei, impact parameters of interactions, and number of generated events) is realized by a WEB-interface. A query is processed by a server, and results are presented to the user as a WEB-page. Short descriptions of the installed generators, the WEB-interface implementation and the server operation are given.

  5. Being part of a multi-generational medical practice team.

    Science.gov (United States)

    Hills, Laura

    2010-01-01

    What happens when you find yourself working in your medical practice every day with co-workers who are the ages of your parents or children? Do you find yourself reverting to age-related roles? Do you become exasperated with or bewildered by the values and behaviors of older or younger colleagues? This article explores the challenges and opportunities the medical practice staff member faces when he or she is part of a multi-generational medical practice team. It describes the tensions that often occur when a medical practice staff runs the gamut from those who remember using a library card catalog and those who can't remember the days before Google. It describes the core values, career goals, key formative events, and attitudes that may have shaped the thinking and behavior of the four generations that may work in the medical practice today: Traditionalists, Baby Boomers, Generation X, and Generation Y. It suggests preferred communication and learning methods for staff members of different generations. Finally, this article offers 10 best practices for working in a multi-generational staff and for creating a supportive multi-generational medical practice culture.

  6. DATA EXTRACTION AND LABEL ASSIGNMENT FOR WEB DATABASES

    OpenAIRE

    T. Rajesh; T. Prathap; S.Naveen Nambi; A.R. Arunachalam

    2015-01-01

    Deep Web contents are accessed by queries submitted to Web databases and the returned data records are en wrapped in dynamically generated Web pages (they will be called deep Web pages in this paper). The structured data that Extracting from deep Web pages is a challenging problem due to the underlying intricate structures of such pages. Until now, a too many number of techniques have been proposed to address this problem, but all of them have limitations because they are Web-page-programming...

  7. A Sensor Web and Web Service-Based Approach for Active Hydrological Disaster Monitoring

    Directory of Open Access Journals (Sweden)

    Xi Zhai

    2016-09-01

    Full Text Available Rapid advancements in Earth-observing sensor systems have led to the generation of large amounts of remote sensing data that can be used for the dynamic monitoring and analysis of hydrological disasters. The management and analysis of these data could take advantage of distributed information infrastructure technologies such as Web service and Sensor Web technologies, which have shown great potential in facilitating the use of observed big data in an interoperable, flexible and on-demand way. However, it remains a challenge to achieve timely response to hydrological disaster events and to automate the geoprocessing of hydrological disaster observations. This article proposes a Sensor Web and Web service-based approach to support active hydrological disaster monitoring. This approach integrates an event-driven mechanism, Web services, and a Sensor Web and coordinates them using workflow technologies to facilitate the Web-based sharing and processing of hydrological hazard information. The design and implementation of hydrological Web services for conducting various hydrological analysis tasks on the Web using dynamically updating sensor observation data are presented. An application example is provided to demonstrate the benefits of the proposed approach over the traditional approach. The results confirm the effectiveness and practicality of the proposed approach in cases of hydrological disaster.

  8. Caught in the Web

    International Nuclear Information System (INIS)

    Gillies, James

    1995-01-01

    The World-Wide Web may have taken the Internet by storm, but many people would be surprised to learn that it owes its existence to CERN. Around half the world's particle physicists come to CERN for their experiments, and the Web is the result of their need to share information quickly and easily on a global scale. Six years after Tim Berners-Lee's inspired idea to marry hypertext to the Internet in 1989, CERN is handing over future Web development to the World-Wide Web Consortium, run by the French National Institute for Research in Computer Science and Control, INRIA, and the Laboratory for Computer Science of the Massachusetts Institute of Technology, MIT, leaving itself free to concentrate on physics. The Laboratory marked this transition with a conference designed to give a taste of what the Web can do, whilst firmly stamping it with the label ''Made in CERN''. Over 200 European journalists and educationalists came to CERN on 8 - 9 March for the World-Wide Web Days, resulting in wide media coverage. The conference was opened by UK Science Minister David Hunt who stressed the importance of fundamental research in generating new ideas. ''Who could have guessed 10 years ago'', he said, ''that particle physics research would lead to a communication system which would allow every school to have the biggest library in the world in a single computer?''. In his introduction, the Minister also pointed out that ''CERN and other basic research laboratories help to break new technological ground and sow the seeds of what will become mainstream manufacturing in the future.'' Learning the jargon is often the hardest part of coming to grips with any new invention, so CERN put it at the top of the agenda. Jacques Altaber, who helped introduce the Internet to CERN in the early 1980s, explained that without the Internet, the Web couldn't exist. The Internet began as a US Defense

  9. Web party effect: a cocktail party effect in the web environment.

    Science.gov (United States)

    Rigutti, Sara; Fantoni, Carlo; Gerbino, Walter

    2015-01-01

    In goal-directed web navigation, labels compete for selection: this process often involves knowledge integration and requires selective attention to manage the dizziness of web layouts. Here we ask whether the competition for selection depends on all web navigation options or only on those options that are more likely to be useful for information seeking, and provide evidence in favor of the latter alternative. Participants in our experiment navigated a representative set of real websites of variable complexity, in order to reach an information goal located two clicks away from the starting home page. The time needed to reach the goal was accounted for by a novel measure of home page complexity based on a part of (not all) web options: the number of links embedded within web navigation elements weighted by the number and type of embedding elements. Our measure fully mediated the effect of several standard complexity metrics (the overall number of links, words, images, graphical regions, the JPEG file size of home page screenshots) on information seeking time and usability ratings. Furthermore, it predicted the cognitive demand of web navigation, as revealed by the duration judgment ratio (i.e., the ratio of subjective to objective duration of information search). Results demonstrate that focusing on relevant links while ignoring other web objects optimizes the deployment of attentional resources necessary to navigation. This is in line with a web party effect (i.e., a cocktail party effect in the web environment): users tune into web elements that are relevant for the achievement of their navigation goals and tune out all others.

  10. Web Log Pre-processing and Analysis for Generation of Learning Profiles in Adaptive E-learning

    Directory of Open Access Journals (Sweden)

    Radhika M. Pai

    2016-03-01

    Full Text Available Adaptive E-learning Systems (AESs enhance the efficiency of online courses in education by providing personalized contents and user interfaces that changes according to learner’s requirements and usage patterns. This paper presents the approach to generate learning profile of each learner which helps to identify the learning styles and provide Adaptive User Interface which includes adaptive learning components and learning material. The proposed method analyzes the captured web usage data to identify the learning profile of the learners. The learning profiles are identified by an algorithmic approach that is based on the frequency of accessing the materials and the time spent on the various learning components on the portal. The captured log data is pre-processed and converted into standard XML format to generate learners sequence data corresponding to the different sessions and time spent. The learning style model adopted in this approach is Felder-Silverman Learning Style Model (FSLSM. This paper also presents the analysis of learner’s activities, preprocessed XML files and generated sequences.

  11. Web Log Pre-processing and Analysis for Generation of Learning Profiles in Adaptive E-learning

    Directory of Open Access Journals (Sweden)

    Radhika M. Pai

    2016-04-01

    Full Text Available Adaptive E-learning Systems (AESs enhance the efficiency of online courses in education by providing personalized contents and user interfaces that changes according to learner’s requirements and usage patterns. This paper presents the approach to generate learning profile of each learner which helps to identify the learning styles and provide Adaptive User Interface which includes adaptive learning components and learning material. The proposed method analyzes the captured web usage data to identify the learning profile of the learners. The learning profiles are identified by an algorithmic approach that is based on the frequency of accessing the materials and the time spent on the various learning components on the portal. The captured log data is pre-processed and converted into standard XML format to generate learners sequence data corresponding to the different sessions and time spent. The learning style model adopted in this approach is Felder-Silverman Learning Style Model (FSLSM. This paper also presents the analysis of learner’s activities, preprocessed XML files and generated sequences.

  12. Modelo de web semántica para universidades

    Directory of Open Access Journals (Sweden)

    Karla Abad

    2015-12-01

    Full Text Available A raíz del estudio de estado actual de micrositios y repositorios en la Universidad Estatal Península de Santa Elena se encontró que su información carecía de semántica óptima y adecuada. Bajo estas circunstancias, se plantea entonces la necesidad de crear un modelo de estructura de web semántica para Universidades, el cual posteriormente fue aplicado a micrositios y repositorio digital de la UPSE, como caso de prueba. Parte de este proyecto incluye la instalación de módulos de software con sus respectivas configuraciones y la utilización de estándares de metadatos como DUBLIN CORE, para la mejora del SEO (optimización en motores de búsqueda; con ello se ha logrado la generación de metadatos estandarizados y la creación de políticas para la subida de información. El uso de metadatos transforma datos simples en estructuras bien organizadas que aportan información y conocimiento para generar resultados en buscadores web. Al culminar la implementación del modelo de web semántica es posible decir que la universidad ha mejorado su presencia y visibilidad en la web a través del indexamiento de información en diferentes motores de búsqueda y posicionamiento en la categorización de universidades y de repositorios de Webometrics (ranking que proporciona clasificación de universidades de todo el mundo.   Abstract After examining the current microsites and repositories situation in University, Peninsula of Santa Elena´s, it was found that information lacked optimal and appropriate semantic. Under these circumstances, there is a need to create a semantic web structure model for Universities, which was subsequently applied to UPSE´s microsites and digital repositories, as a test study case. Part of this project includes the installation of software modules with their respective configurations and the use of metadata standards such as DUBLIN CORE, to improve the SEO (Search Engine Optimization; with these applications, it was

  13. Forensic web watch.

    Science.gov (United States)

    Abbas, Ali; N Rutty, Guy

    2003-06-01

    When one thinks of print identification techniques one automatically considers fingerprints. Although finger prints have been in use now for over 100 years there is in fact an older type of identification technique related to prints left at scenes of crime and the anatomy of human body parts. This is the world of ear prints. This short web review considers web sites related to ear print identification particularly the continuing controversy as to whether or not an ear print is unique.

  14. The Partial Mapping of the Web Graph

    Directory of Open Access Journals (Sweden)

    Kristina Machova

    2009-06-01

    Full Text Available The paper presents an approach to partial mapping of a web sub-graph. This sub-graph contains the nearest surroundings of an actual web page. Our work deals with acquiring relevant Hyperlinks of a base web site, generation of adjacency matrix, the nearest distance matrix and matrix of converted distances of Hyperlinks, detection of compactness of web representation, and visualization of its graphical representation. The paper introduces an LWP algorithm – a technique for Hyperlink filtration.  This work attempts to help users with the orientation within the web graph.

  15. WebBio, a web-based management and analysis system for patient data of biological products in hospital.

    Science.gov (United States)

    Lu, Ying-Hao; Kuo, Chen-Chun; Huang, Yaw-Bin

    2011-08-01

    We selected HTML, PHP and JavaScript as the programming languages to build "WebBio", a web-based system for patient data of biological products and used MySQL as database. WebBio is based on the PHP-MySQL suite and is run by Apache server on Linux machine. WebBio provides the functions of data management, searching function and data analysis for 20 kinds of biological products (plasma expanders, human immunoglobulin and hematological products). There are two particular features in WebBio: (1) pharmacists can rapidly find out whose patients used contaminated products for medication safety, and (2) the statistics charts for a specific product can be automatically generated to reduce pharmacist's work loading. WebBio has successfully turned traditional paper work into web-based data management.

  16. Aspectos de seguridad en Web 2.0 y redes sociales

    OpenAIRE

    Caballero Velasco, María Ángeles

    2011-01-01

    Este proyecto final de carrera de universidad va enfocado a la seguridad en la Web 2.0 y las redes sociales. El documento se divide en dos partes. La primera parte define la visión principal del mundo 2.0, cómo nació y que implicaciones tiene hoy en día en la sociedad. Se definen las características de la Web 2.0 y su historia. Se analizan las diferencias entre la Web 2.0 y la Web 1.0, llegando hasta la Web 3.0. Se clasifican los tipos de Web 2.0 por uso y por aplicación y se analizan todas l...

  17. Aesthetics and function in web design

    DEFF Research Database (Denmark)

    Thorlacius, Lisbeth

    2004-01-01

    Since the origin of the web site in the first part of the 90’s there has been discussions regarding the relative weighting of function and aesthetics. A renewed discussion is needed, however, to clarify what exactly is meant by aesthetics in web design. Moreover the balance between aesthetics...... and function ought to be considered more in respect to the target group and the genre of web site....

  18. Adaptive web data extraction policies

    Directory of Open Access Journals (Sweden)

    Provetti, Alessandro

    2008-12-01

    Full Text Available Web data extraction is concerned, among other things, with routine data accessing and downloading from continuously-updated dynamic Web pages. There is a relevant trade-off between the rate at which the external Web sites are accessed and the computational burden on the accessing client. We address the problem by proposing a predictive model, typical of the Operating Systems literature, of the rate-of-update of each Web source. The presented model has been implemented into a new version of the Dynamo project: a middleware that assists in generating informative RSS feeds out of traditional HTML Web sites. To be effective, i.e., make RSS feeds be timely and informative and to be scalable, Dynamo needs a careful tuning and customization of its polling policies, which are described in detail.

  19. Transonic Airfoil Flow Simulation. Part I: Mesh Generation and Inviscid Method

    Directory of Open Access Journals (Sweden)

    Vladimir CARDOS

    2010-06-01

    Full Text Available A calculation method for the subsonic and transonic viscous flow over airfoil using thedisplacement surface concept is described. Part I presents a mesh generation method forcomputational grid and a finite volume method for the time-dependent Euler equations. The inviscidsolution is used for the inviscid-viscous coupling procedure presented in the Part II.

  20. A Typology for Web 2.0

    DEFF Research Database (Denmark)

    Dalsgaard, Christian; Sorensen, Elsebeth Korsgaard

    2008-01-01

    of a learning environment: 1) organizing communicative processes and 2) organizing resources. Organizing communicative processes is supported by Web 2.0’s ability to provide a range of communicative tools that can be organized flexibly by students. Web 2.0 provides opportunities for communities and groups...... to organize their own communicative processes. Further, Web 2.0 supports organization of resources by empowering students to create, construct, manage and share content themselves. However, the main potential lies within collaborative creation and sharing in networks. Potentially, networking tools......Web 2.0 is a term used to describe recent developments on the World Wide Web. The term is often used to describe the increased use of the web for user-generated content, collaboration, and social networking. However, Web 2.0 is a weakly defined concept, and it is unclear exactly what kind...

  1. Hypermedia and the Semantic Web: A Research Agenda

    NARCIS (Netherlands)

    J.R. van Ossenbruggen (Jacco); L. Hardman (Lynda); L. Rutledge (Lloyd)

    2002-01-01

    textabstractUntil recently, the Semantic Web was little more than a name for the next generation Web infrastructure as envisioned by its inventor, Tim Berners-Lee. Now, with the introduction of XML and RDF, and new developments such as RDF Schema and DAML+OIL, the Semantic Web is rapidly taking

  2. Hypermedia and the semantic web: a research agenda

    NARCIS (Netherlands)

    J.R. van Ossenbruggen (Jacco); L. Hardman (Lynda); L. Rutledge (Lloyd)

    2001-01-01

    textabstractUntil recently, the Semantic Web was little more than a name for the next generation Web infrastructure as envisioned by its inventor, Tim Berners-Lee. Now, with the introduction of XML and RDF, and new developments such as RDF Schema and DAML+OIL, the Semantic Web is rapidly taking

  3. Web party effect: a cocktail party effect in the web environment

    Directory of Open Access Journals (Sweden)

    Sara Rigutti

    2015-03-01

    Full Text Available In goal-directed web navigation, labels compete for selection: this process often involves knowledge integration and requires selective attention to manage the dizziness of web layouts. Here we ask whether the competition for selection depends on all web navigation options or only on those options that are more likely to be useful for information seeking, and provide evidence in favor of the latter alternative. Participants in our experiment navigated a representative set of real websites of variable complexity, in order to reach an information goal located two clicks away from the starting home page. The time needed to reach the goal was accounted for by a novel measure of home page complexity based on a part of (not all web options: the number of links embedded within web navigation elements weighted by the number and type of embedding elements. Our measure fully mediated the effect of several standard complexity metrics (the overall number of links, words, images, graphical regions, the JPEG file size of home page screenshots on information seeking time and usability ratings. Furthermore, it predicted the cognitive demand of web navigation, as revealed by the duration judgment ratio (i.e., the ratio of subjective to objective duration of information search. Results demonstrate that focusing on relevant links while ignoring other web objects optimizes the deployment of attentional resources necessary to navigation. This is in line with a web party effect (i.e., a cocktail party effect in the web environment: users tune into web elements that are relevant for the achievement of their navigation goals and tune out all others.

  4. Web party effect: a cocktail party effect in the web environment

    Science.gov (United States)

    Gerbino, Walter

    2015-01-01

    In goal-directed web navigation, labels compete for selection: this process often involves knowledge integration and requires selective attention to manage the dizziness of web layouts. Here we ask whether the competition for selection depends on all web navigation options or only on those options that are more likely to be useful for information seeking, and provide evidence in favor of the latter alternative. Participants in our experiment navigated a representative set of real websites of variable complexity, in order to reach an information goal located two clicks away from the starting home page. The time needed to reach the goal was accounted for by a novel measure of home page complexity based on a part of (not all) web options: the number of links embedded within web navigation elements weighted by the number and type of embedding elements. Our measure fully mediated the effect of several standard complexity metrics (the overall number of links, words, images, graphical regions, the JPEG file size of home page screenshots) on information seeking time and usability ratings. Furthermore, it predicted the cognitive demand of web navigation, as revealed by the duration judgment ratio (i.e., the ratio of subjective to objective duration of information search). Results demonstrate that focusing on relevant links while ignoring other web objects optimizes the deployment of attentional resources necessary to navigation. This is in line with a web party effect (i.e., a cocktail party effect in the web environment): users tune into web elements that are relevant for the achievement of their navigation goals and tune out all others. PMID:25802803

  5. Data management on the spatial web

    DEFF Research Database (Denmark)

    Jensen, Christian S.

    2012-01-01

    Due in part to the increasing mobile use of the web and the proliferation of geo-positioning, the web is fast acquiring a significant spatial aspect. Content and users are being augmented with locations that are used increasingly by location-based services. Studies suggest that each week, several...... billion web queries are issued that have local intent and target spatial web objects. These are points of interest with a web presence, and they thus have locations as well as textual descriptions. This development has given prominence to spatial web data management, an area ripe with new and exciting...... opportunities and challenges. The research community has embarked on inventing and supporting new query functionality for the spatial web. Different kinds of spatial web queries return objects that are near a location argument and are relevant to a text argument. To support such queries, it is important...

  6. An Algebraic Specification of the Semantic Web

    OpenAIRE

    Ksystra, Katerina; Triantafyllou, Nikolaos; Stefaneas, Petros; Frangos, Panayiotis

    2011-01-01

    We present a formal specification of the Semantic Web, as an extension of the World Wide Web using the well known algebraic specification language CafeOBJ. Our approach allows the description of the key elements of the Semantic Web technologies, in order to give a better understanding of the system, without getting involved with their implementation details that might not yet be standardized. This specification is part of our work in progress concerning the modeling the Social Semantic Web.

  7. A simple method for serving Web hypermaps with dynamic database drill-down

    Directory of Open Access Journals (Sweden)

    Carson Ewart R

    2002-08-01

    Full Text Available Abstract Background HealthCyberMap http://healthcybermap.semanticweb.org aims at mapping parts of health information cyberspace in novel ways to deliver a semantically superior user experience. This is achieved through "intelligent" categorisation and interactive hypermedia visualisation of health resources using metadata, clinical codes and GIS. HealthCyberMap is an ArcView 3.1 project. WebView, the Internet extension to ArcView, publishes HealthCyberMap ArcView Views as Web client-side imagemaps. The basic WebView set-up does not support any GIS database connection, and published Web maps become disconnected from the original project. A dedicated Internet map server would be the best way to serve HealthCyberMap database-driven interactive Web maps, but is an expensive and complex solution to acquire, run and maintain. This paper describes HealthCyberMap simple, low-cost method for "patching" WebView to serve hypermaps with dynamic database drill-down functionality on the Web. Results The proposed solution is currently used for publishing HealthCyberMap GIS-generated navigational information maps on the Web while maintaining their links with the underlying resource metadata base. Conclusion The authors believe their map serving approach as adopted in HealthCyberMap has been very successful, especially in cases when only map attribute data change without a corresponding effect on map appearance. It should be also possible to use the same solution to publish other interactive GIS-driven maps on the Web, e.g., maps of real world health problems.

  8. 60. The World-Wide Inaccessible Web, Part 1: Browsing

    Science.gov (United States)

    Baggaley, Jon; Batpurev, Batchuluun

    2007-01-01

    Two studies are reported, comparing the browser loading times of webpages created using common Web development techniques. The loading speeds were estimated in 12 Asian countries by members of the "PANdora" network, funded by the International Development Research Centre (IDRC) to conduct collaborative research in the development of…

  9. Persistent Web References – Best Practices and New Suggestions

    DEFF Research Database (Denmark)

    Zierau, Eld; Nyvang, Caroline; Kromann, Thomas Hvid

    In this paper, we suggest adjustments to best practices for persistent web referencing; adjustments that aim at preservation and long time accessibility of web referenced resources in general, but with focus on web references in web archives. Web referencing is highly relevant and crucial...... refer to archive URLs which depends on the web archives access implementations. A major part of the suggested adjustments is a new web reference standard for archived web references (called wPID), which is a supplement to the current practices. The purpose of the standard is to support general, global...

  10. A study on the Web intelligence

    Institute of Scientific and Technical Information of China (English)

    Sang-Geun Kim

    2004-01-01

    This paper surveys important aspects of Web Intelligence (WI). WI explores the fundamental roles as well as practical impacts of Artificial Intelligence (AI) and advanced Information Technology (IT) on the next generation of Web - related products, systens, and activities. As a direction for scientific research and devlopment, WI can be extremely beneficial for the field of Artificial Intelligence in Education (AIED). This paper covers these issues only very briefly. It focuses more on other issues in WI, such as intelligent Web services, and semantic web, and proposes how to use them as basis for tackling new and challenging research problems in AIED.

  11. Detection And Classification Of Web Robots With Honeypots

    Science.gov (United States)

    2016-03-01

    Web robots are valuable tools for indexing content on the Web, they can also be malicious through phishing , spamming, or performing targeted attacks...indexing content on the Web, they can also be malicious through phishing , spamming, or performing targeted attacks. In this thesis, we study an approach...programs has been attributed to the explosion in content and user-generated social media on the Internet. The Web search engines like Google require

  12. User interaction in modern web information systems

    NARCIS (Netherlands)

    Barna, P.; Houben, G.J.P.M.; De Bra, P.M.E.

    2003-01-01

    Modern Information Systems based on Web technologies (Web-based Information Systems - WIS) typically generate hypermedia presentations according to the user needs. Hera is our model-driven methodology specifying the design cycle and the architecture framework for WIS. To avoid additional expensive

  13. Prediction of toxicity and comparison of alternatives using WebTEST (Web-services Toxicity Estimation Software Tool)(Bled Slovenia)

    Science.gov (United States)

    A Java-based web service is being developed within the US EPA’s Chemistry Dashboard to provide real time estimates of toxicity values and physical properties. WebTEST can generate toxicity predictions directly from a simple URL which includes the endpoint, QSAR method, and ...

  14. Public health and Web 2.0.

    Science.gov (United States)

    Hardey, Michael

    2008-07-01

    This article examines the nature and role of Web 2.0 resources and their impact on health information made available though the Internet. The transition of the Web from version one to Web 2.0 is described and the main features of the new Web examined. Two characteristic Web 2.0 resources are explored and the implications for the public and practitioners examined. First, what are known as 'user reviews' or 'user testimonials', which allow people to comment on the health services delivered to them, are described. Second, new mapping applications that take advantage of the interactive potential of Web 2.0 and provide tools to visualize complex data are examined. Following a discussion of the potential of Web 2.0, it is concluded that it offers considerable opportunities for disseminating health information and creating new sources of data, as well as generating new questions and dilemmas.

  15. XML databases and the semantic web

    CERN Document Server

    Thuraisingham, Bhavani

    2002-01-01

    Efficient access to data, sharing data, extracting information from data, and making use of the information have become urgent needs for today''s corporations. With so much data on the Web, managing it with conventional tools is becoming almost impossible. New tools and techniques are necessary to provide interoperability as well as warehousing between multiple data sources and systems, and to extract information from the databases. XML Databases and the Semantic Web focuses on critical and new Web technologies needed for organizations to carry out transactions on the Web, to understand how to use the Web effectively, and to exchange complex documents on the Web.This reference for database administrators, database designers, and Web designers working in tandem with database technologists covers three emerging technologies of significant impact for electronic business: Extensible Markup Language (XML), semi-structured databases, and the semantic Web. The first two parts of the book explore these emerging techn...

  16. WebQuest y anotaciones semánticas WebQuest and semantic annotations

    Directory of Open Access Journals (Sweden)

    Santiago Blanco Suárez

    2007-03-01

    Full Text Available En este artículo se presenta un sistema de búsqueda y recuperación de metadatos de actividades educativas que siguen el modelo WebQuest. Se trata de una base de datos relacional, accesible a través del web, que se complementa con un módulo que permite realizar anotaciones semánticas y cuyo objetivo es capturar y enriquecer el conocimiento acerca del uso de dichos ejercicios por parte de la comunidad de docentes que experimentan con ellos, así como documentar los recursos o sitios web de interés didáctico buscando construir un repositorio de enlaces educativos de calidad. This paper presents a system of searching and recovering educational activities that follow the Web-Quest model through the web, complemented with a module to make semantic annotations aimed at getting and enriching the knowledge on the use of these exercises by the teaching community. It also tries to document the resources or websites with didactic interest in order to build a qualified account of educational links.

  17. Hera presentation generator (Poster)

    NARCIS (Netherlands)

    Frasincar, F.; Houben, G.J.P.M.; Barna, P.; Ellis, A.; Hagino, T.

    2005-01-01

    Semantic Web Information Systems (SWIS) are Web Information Systems that use Semantic Web technologies. Hera is a model-driven design methodology for SWIS. In Hera, models are represented in RDFS and model instances in RDF. The Hera Presentation Generator (HPG) is an integrated development

  18. Web 2.0. Nettet holder liv i grusomme borgerkrige

    DEFF Research Database (Denmark)

    Schmidt, Søren

    2017-01-01

    The article explains the reasons, timing and character of on-going civil wars in the Middle East. The web 2.0 explains many parts of this.......The article explains the reasons, timing and character of on-going civil wars in the Middle East. The web 2.0 explains many parts of this....

  19. Security scanning of Web sites at CERN

    CERN Multimedia

    IT Department

    2010-01-01

    As of early 2010, the CERN Computer Security Team will start regular scanning of all Web sites and Web applications at CERN, visible on the Internet, or on the General Purpose Network (office network). The goal of this scanning is to improve the quality of CERN Web sites. All deficits found will be reported by e-mail to the relevant Web site owners, and must be fixed in a timely manner. Web site owners may also request one-off scans of their Web site or Web application, by sending an e-mail to Computer.Security@cern.ch. These Web scans are designed to limit the impact on the scanned Web sites. Nevertheless, in very rare cases scans may cause undesired side-effects, e.g. generate a large number of log entries, or cause particularly badly designed or less robust Web applications to crash. If a Web site is affected by these security scans, it will also be susceptible to any more aggressive scan that can be performed any time by a malicious attacker. Such Web applications should be fixed, and also additionally...

  20. Manufacture of plastic parts by radiation molding

    International Nuclear Information System (INIS)

    Leszyk, G.M.; Morrison, E.D.; Williams, R.F. Jr.

    1977-01-01

    Thin plastic parts which can have precise tolerances and can be of complex shape are prepared by casting a viscous radiation-curable composition onto a support, such as a moving web of polymeric material, in the shape of the desired part and then irradiating, for example with ultraviolet radiation or high energy electrons, to cause curing of the composition to a solid plastic. The radiation-curable composition is formulated with viscosity and flow characteristics it to be cast in the exact shape of the part desired yet retain this shape during curing while supported only by the surface on which it has been cast. Plastic parts made by this method can be formed entirely of the radiation-curable composition by casting onto a web having a release surface from which the part can be stripped subsequent to curing or can be formed partially from a web material and partially from the radiation-curable composition by casting onto a web to which the composition will bond and subsequently cutting the web into discrete portions which include the cured composition

  1. Penilaian Risiko Aplikasi Web Menggunakan Model DREAD

    Directory of Open Access Journals (Sweden)

    Didit Suprihanto

    2016-01-01

    Full Text Available Application that  is developed by web based, beside has surplus in WWW technology, it has susceptibility side that can be threat too. Susceptibility generate risk and can bring out big trouble even effect big disadvantage. The goal of this research is design and build document risk assessment system of threat level and prevention advice. It use DREAD model as method to solve trouble by giving qualified information. This information are used to produce risk level in web application. The result of this research is web application risk assessment system by using DREAD model to know risk threat level and equate perception of web threat risk to application developer, minimize of threat risk and maximize performance of web application.   Keywords : DREAD model, web threat risk, web risk assessment system

  2. A Typology for Web 2.0

    DEFF Research Database (Denmark)

    Dalsgaard, Christian; Sorensen, Elsebeth Korsgaard

    2008-01-01

    Web 2.0 is a term used to describe recent developments on the World Wide Web. The term is often used to describe the increased use of the web for user-generated content, collaboration, and social networking. However, Web 2.0 is a weakly defined concept, and it is unclear exactly what kind...... of technologies it covers. The objective of the paper is to develop a typology that can be used to categorize Web 2.0 technologies. Further, the paper will discuss which of these technologies are unique to Web 2.0. Often, Web 2.0 is described by way of different kinds of software; for instance, blogs, wikis......, podcasts, RSS, and social networking sites. The problem with this type of description is that it fails to distinguish between different types or categories of technologies. As an alternative, the typology developed in the paper distinguishes between technologies on basis of, how - and in which contexts...

  3. A guide: sale of power opportunities for distributed generators. Part B. The guide

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2005-07-01

    This report is Part B of a series of three produced to help prospective distributed generators familiarise themselves with the electricity market in Great Britain (England, Scotland and Wales) and to assist potential generators to identify options for the sale of their electrical output. Part A is an introductory guide and Part C lists sources of further information, useful contact details and a glossary. Part B covers the key issues in some depth and is designed for those with some background knowledge. It examines: the electricity supply industry in Great Britain including the physical and commercial electricity supply chains and the commercial arrangements for trading; the legislative framework including networks, price controls, licensing and exemptions; the value of power produced by small generators including factors affecting the value of power, wholesale market value, Renewables Obligation Certificates (ROCs), Climate Change Levy exemption certificates and the EU Emissions Trading Scheme; options for the sale of power including market trading, supplier and consolidator contracts, exempt and licensed supply, and finance; and the costs and benefits of power sales options including the initial and ongoing costs of selling power.

  4. Theoretical Foundations of the Web: Cognition, Communication, and Co-Operation. Towards an Understanding of Web 1.0, 2.0, 3.0

    Directory of Open Access Journals (Sweden)

    Robert Bichler

    2010-02-01

    Full Text Available Currently, there is much talk of Web 2.0 and Social Software. A common understanding of these notions is not yet in existence. The question of what makes Social Software social has thus far also remained unacknowledged. In this paper we provide a theoretical understanding of these notions by outlining a model of the Web as a techno-social system that enhances human cognition towards communication and co-operation. According to this understanding, we identify three qualities of the Web, namely Web 1.0 as a Web of cognition, Web 2.0 as a Web of human communication, and Web 3.0 as a Web of co-operation. We use the terms Web 1.0, Web 2.0, Web 3.0 not in a technical sense, but for describing and characterizing the social dynamics and information processes that are part of the Internet.

  5. USING WEB MINING IN E-COMMERCE APPLICATIONS

    Directory of Open Access Journals (Sweden)

    Claudia Elena Dinucă

    2011-09-01

    Full Text Available Nowadays, the web is an important part of our daily life. The web is now the best medium of doing business. Large companies rethink their business strategy using the web to improve business. Business carried on the Web offers the opportunity to potential customers or partners where their products and specific business can be found. Business presence through a company web site has several advantages as it breaks the barrier of time and space compared with the existence of a physical office. To differentiate through the Internet economy, winning companies have realized that e-commerce transactions is more than just buying / selling, appropriate strategies are key to improve competitive power. One effective technique used for this purpose is data mining. Data mining is the process of extracting interesting knowledge from data. Web mining is the use of data mining techniques to extract information from web data. This article presents the three components of web mining: web usage mining, web structure mining and web content mining.

  6. Selecting a Web 2.0 Presentation Tool

    Science.gov (United States)

    Hodges, Charles B.; Clark, Kenneth

    2011-01-01

    Web-based presentation tools are sometimes referred to as "next generation presentation tools" (EDUCAUSE, 2010). At the most basic level, these tools are simply online versions of traditional presentation software, such as Microsoft's PowerPoint or Apple's Keynote, but some services offer features like web-based collaboration, online presentation…

  7. Web Intelligence and Artificial Intelligence in Education

    Science.gov (United States)

    Devedzic, Vladan

    2004-01-01

    This paper surveys important aspects of Web Intelligence (WI) in the context of Artificial Intelligence in Education (AIED) research. WI explores the fundamental roles as well as practical impacts of Artificial Intelligence (AI) and advanced Information Technology (IT) on the next generation of Web-related products, systems, services, and…

  8. Wikis, blogs and podcasts: a new generation of Web-based tools for virtual collaborative clinical practice and education

    Directory of Open Access Journals (Sweden)

    Maramba Inocencio

    2006-08-01

    Full Text Available Abstract Background We have witnessed a rapid increase in the use of Web-based 'collaborationware' in recent years. These Web 2.0 applications, particularly wikis, blogs and podcasts, have been increasingly adopted by many online health-related professional and educational services. Because of their ease of use and rapidity of deployment, they offer the opportunity for powerful information sharing and ease of collaboration. Wikis are Web sites that can be edited by anyone who has access to them. The word 'blog' is a contraction of 'Web Log' – an online Web journal that can offer a resource rich multimedia environment. Podcasts are repositories of audio and video materials that can be "pushed" to subscribers, even without user intervention. These audio and video files can be downloaded to portable media players that can be taken anywhere, providing the potential for "anytime, anywhere" learning experiences (mobile learning. Discussion Wikis, blogs and podcasts are all relatively easy to use, which partly accounts for their proliferation. The fact that there are many free and Open Source versions of these tools may also be responsible for their explosive growth. Thus it would be relatively easy to implement any or all within a Health Professions' Educational Environment. Paradoxically, some of their disadvantages also relate to their openness and ease of use. With virtually anybody able to alter, edit or otherwise contribute to the collaborative Web pages, it can be problematic to gauge the reliability and accuracy of such resources. While arguably, the very process of collaboration leads to a Darwinian type 'survival of the fittest' content within a Web page, the veracity of these resources can be assured through careful monitoring, moderation, and operation of the collaborationware in a closed and secure digital environment. Empirical research is still needed to build our pedagogic evidence base about the different aspects of these tools in

  9. DISTANCE LEARNING ONLINE WEB 3 .0

    Directory of Open Access Journals (Sweden)

    S. M. Petryk

    2015-05-01

    Full Text Available This article analyzes the existing methods of identification information in the semantic web, outlines the main problems of its implementation and researches the use of Semantic Web as the part of distance learning. Proposed alternative variant of identification and relationship construction of information and acquired knowledge based on the developed method “spectrum of knowledge”

  10. Introduction to Webometrics Quantitative Web Research for the Social Sciences

    CERN Document Server

    Thelwall, Michael

    2009-01-01

    Webometrics is concerned with measuring aspects of the web: web sites, web pages, parts of web pages, words in web pages, hyperlinks, web search engine results. The importance of the web itself as a communication medium and for hosting an increasingly wide array of documents, from journal articles to holiday brochures, needs no introduction. Given this huge and easily accessible source of information, there are limitless possibilities for measuring or counting on a huge scale (e.g., the number of web sites, the number of web pages, the number of blogs) or on a smaller scale (e.g., the number o

  11. A Web Server for MACCS Magnetometer Data

    Science.gov (United States)

    Engebretson, Mark J.

    1998-01-01

    NASA Grant NAG5-3719 was provided to Augsburg College to support the development of a web server for the Magnetometer Array for Cusp and Cleft Studies (MACCS), a two-dimensional array of fluxgate magnetometers located at cusp latitudes in Arctic Canada. MACCS was developed as part of the National Science Foundation's GEM (Geospace Environment Modeling) Program, which was designed in part to complement NASA's Global Geospace Science programs during the decade of the 1990s. This report describes the successful use of these grant funds to support a working web page that provides both daily plots and file access to any user accessing the worldwide web. The MACCS home page can be accessed at http://space.augsburg.edu/space/MaccsHome.html.

  12. Energy balance and flow in steam generator part with sodium-water reaction

    International Nuclear Information System (INIS)

    Matal, O.

    1980-01-01

    Relations were derived for the calculation of heat liberated during the sodium water reaction in a tube failure in different parts of a steam generator. The results are graphically shown in i-T diagrams. Heat removal is described from the reaction zone to water and steam in undisturbed tubes and to the steam generator metal structure. (author)

  13. A Novel Personalized Web Search Model

    Institute of Scientific and Technical Information of China (English)

    ZHU Zhengyu; XU Jingqiu; TIAN Yunyan; REN Xiang

    2007-01-01

    A novel personalized Web search model is proposed.The new system, as a middleware between a user and a Web search engine, is set up on the client machine. It can learn a user's preference implicitly and then generate the user profile automatically. When the user inputs query keywords, the system can automatically generate a few personalized expansion words by computing the term-term associations according to the current user profile, and then these words together with the query keywords are submitted to a popular search engine such as Yahoo or Google.These expansion words help to express accurately the user's search intention. The new Web search model can make a common search engine personalized, that is, the search engine can return different search results to different users who input the same keywords. The experimental results show the feasibility and applicability of the presented work.

  14. Towards a Pattern Language for Adaptive Web-Based Educational Systems

    NARCIS (Netherlands)

    Avgeriou, P.; Vogiatzis, D.; Tzanavari, A.; Retalis, S.

    2004-01-01

    Adaptive Web-based Educational Systems represent an emerging technology that provides a unique advantage over traditional Web-based Educational Systems; that is the ability to adapt to the user's needs, goals, preferences etc. Adaptive Web-based Educational Systems are increasingly becoming part of

  15. Design Patterns in Adaptive Web-Based Educational Systems : An Overview

    NARCIS (Netherlands)

    Avgeriou, Paris; Vogiatzis, Dimitrios; Tzanavari, Aimilia; Retalis, Symeon

    2004-01-01

    Adaptive Web-based Educational Systems represent an emerging technology that provides a unique advantage over traditional Web-based Educational Systems; that is the ability to adapt to the user's needs, goals, preferences etc. Adaptive Web-based Educational Systems are increasingly becoming part of

  16. Web User Profiling Based on Browsing Behavior Analysis

    OpenAIRE

    Fan , Xiao-Xi; Chow , Kam-Pui; Xu , Fei

    2014-01-01

    Part 1: Internet Crime Investigations; International audience; Determining the source of criminal activity requires a reliable means to estimate a criminal’s identity. One way to do this is to use web browsing history to build a profile of an anonymous user. Since an individual’s web use is unique, matching the web use profile to known samples provides a means to identify an unknown user. This paper describes a model for web user profiling and identification. Two aspects of browsing behavior ...

  17. THE GENERATIVE REPRODUCTIVE CHARACTERISTICS OF RED DRAGON FRUIT (Hylocereus.polyrhizus (Web. Britton & Rose, CACTACEAE

    Directory of Open Access Journals (Sweden)

    Eniek Kriswiyanti

    2013-04-01

    Full Text Available This study aimed to investigate the generative reproductive characteristics of red dragon fruit (Hylocereus polyrhizus (Web. Britton & Rose, that is the characteristics and the development of male and female reproductive system and the embryo. The flowering morphology was observed through the gametogenesis process and the process of embryogenesis was observed using the methods of squash, cytolysis, and embedding of flower before anthesis, during and after anthesis. The results showed that the flower funnelled form with many calyxes, stamens and crowns. The pollen was circular, sulcus, trilate, reticulate, and spheroid. The development of microgametophyte was at the 2nd stages with three nucleuses and the pistil developed after anthesis.  Anthesis was taken place at night time, fertilization and pollen developed before anthesis and has not germinating. A single pistil, the head of pistil with many branches, the style longer than the stamen, open type.  Seeds anathrophus, endosperm and embryo with the globular shaped, developed 5 days after anthesis, the embryo torpedo likes and the differentiation of primer tissues seen in the seeds 7 days after anthesis, and arilus generated from development of megasporangium. Keywords: embedding, self incompatibility, sferoidal, anatrophus, arillus.

  18. Automatic Conversion of a Conceptual Model to a Standard Multi-view Web Services Definition

    Directory of Open Access Journals (Sweden)

    Anass Misbah

    2018-03-01

    Full Text Available Information systems are becoming more and more heterogeneous and here comes the need to have more generic transformation algorithms and more automatic generation Meta rules. In fact, the large number of terminals, devices, operating systems, platforms and environments require a high level of adaptation. Therefore, it is becoming more and more difficult to validate, generate and implement manually models, designs and codes.Web services are one of the technologies that are used massively nowadays; hence, it is considered as one of technologies that require the most automatic rules of validation and automation. Many previous works have dealt with Web services by proposing new concepts such as Multi-view Web services, standard WSDL implementation of Multi-view Web services and even further Generic Meta rules for automatic generation of Multi-view Web services.In this work we will propose a new way of generating Multi-view Web ser-vices, which is based on an engine algorithm that takes as input both an initial Conceptual Model and user’s matrix and then unroll a generic algorithm to gen-erate dynamically a validated set of points of view. This set of points of view will be transformed to a standard WSDL implementation of Multi-view Web services by means of the automatic transformation Meta rules.

  19. Semantic Web Services Challenge, Results from the First Year. Series: Semantic Web And Beyond, Volume 8.

    Science.gov (United States)

    Petrie, C.; Margaria, T.; Lausen, H.; Zaremba, M.

    Explores trade-offs among existing approaches. Reveals strengths and weaknesses of proposed approaches, as well as which aspects of the problem are not yet covered. Introduces software engineering approach to evaluating semantic web services. Service-Oriented Computing is one of the most promising software engineering trends because of the potential to reduce the programming effort for future distributed industrial systems. However, only a small part of this potential rests on the standardization of tools offered by the web services stack. The larger part of this potential rests upon the development of sufficient semantics to automate service orchestration. Currently there are many different approaches to semantic web service descriptions and many frameworks built around them. A common understanding, evaluation scheme, and test bed to compare and classify these frameworks in terms of their capabilities and shortcomings, is necessary to make progress in developing the full potential of Service-Oriented Computing. The Semantic Web Services Challenge is an open source initiative that provides a public evaluation and certification of multiple frameworks on common industrially-relevant problem sets. This edited volume reports on the first results in developing common understanding of the various technologies intended to facilitate the automation of mediation, choreography and discovery for Web Services using semantic annotations. Semantic Web Services Challenge: Results from the First Year is designed for a professional audience composed of practitioners and researchers in industry. Professionals can use this book to evaluate SWS technology for their potential practical use. The book is also suitable for advanced-level students in computer science.

  20. Teaching PBL with Web 2.0

    DEFF Research Database (Denmark)

    Dohn, Nina Bonderup; Buus, Lillian

    2013-01-01

    Web 2.0-mediated learning activities are associated with a number of potentials for learning. Among these potentials are pedagogical advantages connected with bottom-up user-generation of content; opportunities for organizing flexible learning; and possibilities of supporting the development...... of competences necessary for participating in contemporary society. In this article we report on a study of three cases where web 2.0 activities have been introduced into PBL settings. Our focus is the unexpected challenges teachers may face due to inherent tensions between the practice logic of the educational...... system and that of web 2.0-practices. The study documents the development of viable ways of making use of web 2.0-mediated activities in support of a PBL approach and thus provides existence proof that web 2.0 activities may contribute constructively and productively to learning within a PBL perspective...

  1. Effect of Temporal Relationships in Associative Rule Mining for Web Log Data

    Science.gov (United States)

    Mohd Khairudin, Nazli; Mustapha, Aida

    2014-01-01

    The advent of web-based applications and services has created such diverse and voluminous web log data stored in web servers, proxy servers, client machines, or organizational databases. This paper attempts to investigate the effect of temporal attribute in relational rule mining for web log data. We incorporated the characteristics of time in the rule mining process and analysed the effect of various temporal parameters. The rules generated from temporal relational rule mining are then compared against the rules generated from the classical rule mining approach such as the Apriori and FP-Growth algorithms. The results showed that by incorporating the temporal attribute via time, the number of rules generated is subsequently smaller but is comparable in terms of quality. PMID:24587757

  2. Multilingual natural language generation as part of a medical terminology server.

    Science.gov (United States)

    Wagner, J C; Solomon, W D; Michel, P A; Juge, C; Baud, R H; Rector, A L; Scherrer, J R

    1995-01-01

    Re-usable and sharable, and therefore language-independent concept models are of increasing importance in the medical domain. The GALEN project (Generalized Architecture for Languages Encyclopedias and Nomenclatures in Medicine) aims at developing language-independent concept representation systems as the foundations for the next generation of multilingual coding systems. For use within clinical applications, the content of the model has to be mapped to natural language. A so-called Multilingual Information Module (MM) establishes the link between the language-independent concept model and different natural languages. This text generation software must be versatile enough to cope at the same time with different languages and with different parts of a compositional model. It has to meet, on the one hand, the properties of the language as used in the medical domain and, on the other hand, the specific characteristics of the underlying model and its representation formalism. We propose a semantic-oriented approach to natural language generation that is based on linguistic annotations to a concept model. This approach is realized as an integral part of a Terminology Server, built around the concept model and offering different terminological services for clinical applications.

  3. Ten years for the public Web

    CERN Multimedia

    2003-01-01

    Ten years ago, CERN issued a statement declaring that a little known piece of software called the World Wide Web was in the public domain. Nowadays, the Web is an indispensable part of modern communications. The idea for the Web goes back to March 1989 when CERN Computer scientist Tim Berners-Lee wrote a proposal for a 'Distributed Information Management System' for the high-energy physics community. The Web was originaly conceived and developed to meet the demand for information sharing between scientists working all over the world. There were many obstacles in the 1980s to the effective exchange of information. There was, for example a great variety of computer and network systems, with hardly any common features. The main purpose of the web was to allow scientists to access information from any source in a consistent and simple way. By Christmas 1990, Berners-Lee's idea had become the World Wide Web, with its first server and browser running at CERN. Through 1991, the Web spread to other particle physics ...

  4. Development of Web-Based Learning Application for Generation Z

    Science.gov (United States)

    Hariadi, Bambang; Dewiyani Sunarto, M. J.; Sudarmaningtyas, Pantjawati

    2016-01-01

    This study aimed to develop a web-based learning application as a form of learning revolution. The form of learning revolution includes the provision of unlimited teaching materials, real time class organization, and is not limited by time or place. The implementation of this application is in the form of hybrid learning by using Google Apps for…

  5. A rare association between dextrogastria, duodenal web, and ...

    African Journals Online (AJOL)

    ... the radiologic investigation for bilious vomiting and feeding intolerance, revealing congenital duodenal stenosis and dextrogastria. During surgery, the association of the dextrogastria with the duodenal web situated in the second part of the duodenum was established. Keywords: dextrogastria, duodenal web, malrotation ...

  6. Design and Analysis of Web Application Frameworks

    DEFF Research Database (Denmark)

    Schwarz, Mathias Romme

    -state manipulation vulnerabilities. The hypothesis of this dissertation is that we can design frameworks and static analyses that aid the programmer to avoid such errors. First, we present the JWIG web application framework for writing secure and maintainable web applications. We discuss how this framework solves...... some of the common errors through an API that is designed to be safe by default. Second, we present a novel technique for checking HTML validity for output that is generated by web applications. Through string analysis, we approximate the output of web applications as context-free grammars. We model......Numerous web application frameworks have been developed in recent years. These frameworks enable programmers to reuse common components and to avoid typical pitfalls in web application development. Although such frameworks help the programmer to avoid many common errors, we nd...

  7. Lost but Not Forgotten: Finding Pages on the Unarchived Web

    NARCIS (Netherlands)

    Huurdeman, H.C.; Kamps, J.; Samar, T.; de Vries, A.P.; Ben-David, A.; Rogers, R.A.

    2015-01-01

    Web archives attempt to preserve the fast changing web, yet they will always be incomplete. Due to restrictions in crawling depth, crawling frequency, and restrictive selection policies, large parts of the Web are unarchived and, therefore, lost to posterity. In this paper, we propose an approach to

  8. Lost but not forgotten: finding pages on the unarchived web

    NARCIS (Netherlands)

    H.C. Huurdeman; J. Kamps; T. Samar (Thaer); A.P. de Vries (Arjen); A. Ben-David; R.A. Rogers (Richard)

    2015-01-01

    htmlabstractWeb archives attempt to preserve the fast changing web, yet they will always be incomplete. Due to restrictions in crawling depth, crawling frequency, and restrictive selection policies, large parts of the Web are unarchived and, therefore, lost to posterity. In this paper, we propose an

  9. MEAN STACK WEB DEVELOPMENT

    OpenAIRE

    Le Thanh, Nghi

    2017-01-01

    The aim of the thesis is to provide a universal website using JavaScript as the main programming language. It also shows the basic parts anyone need to create a web application. The thesis creates a simple CMS using MEAN stack. MEAN is a collection of JavaScript based technologies used to develop web application. It is an acronym for MongoDB, Express, AngularJS and Node.js. It also allows non-technical users to easily update and manage a website’s content. But the application also lets o...

  10. Sustainable web ecosystem design

    CERN Document Server

    O'Toole, Greg

    2013-01-01

    This book is about the process of creating web-based systems (i.e., websites, content, etc.) that consider each of the parts, the modules, the organisms - binary or otherwise - that make up a balanced, sustainable web ecosystem. In the current media-rich environment, a website is more than a collection of relative html documents of text and images on a static desktop computer monitor. There is now an unlimited combination of screens, devices, platforms, browsers, locations, versions, users, and exabytes of data with which to interact. Written in a highly approachable, practical style, this boo

  11. Classroom Web Pages: A "How-To" Guide for Educators.

    Science.gov (United States)

    Fehling, Eric E.

    This manual provides teachers, with very little or no technology experience, with a step-by-step guide for developing the necessary skills for creating a class Web Page. The first part of the manual is devoted to the thought processes preceding the actual creation of the Web Page. These include looking at other Web Pages, deciding what should be…

  12. MACHINE LEARNING IMPLEMENTATION FOR THE CLASSIFICATION OF ATTACKS ON WEB SYSTEMS. PART 2

    Directory of Open Access Journals (Sweden)

    K. Smirnova

    2017-11-01

    Full Text Available The possibility of applying machine learning for the classification of malicious requests to aWeb application is considered. This approach excludes the use of deterministic analysis systems (for example, expert systems,and is based on the application of a cascade of neural networks or perceptrons on an approximate model to the real humanbrain. The main idea of the work is to enable to describe complex attack vectors consisting of feature sets, abstract terms forcompiling a training sample, controlling the quality of recognition and classifying each of the layers (networks participatingin the work, with the ability to adjust not the entire network, but only a small part of it, in the training of which a mistake orinaccuracy crept in. The design of the developed network can be described as a cascaded, scalable neural network.When using neural networks to detect attacks on web systems, the issue of vectorization and normalization of features isacute. The most commonly used methods for solving these problems are not designed for the case of deliberate distortion ofthe signs of an attack.The proposed approach makes it possible to obtain a neural network that has been studied in more detail by small features,and also to eliminate the normalization issues in order to avoid deliberately bypassing the intrusion detection system. Byisolating one more group of neurons in the network and teaching it to samples containing various variants of circumvention ofthe attack classification, the developed intrusion detection system remains able to classify any types of attacks as well as theiraggregates, putting forward more stringent measures to counteract attacks. This allows you to follow the life cycle of theattack in more detail: from the starting trial attack to deliberate sophisticated attempts to bypass the system and introducemore decisive measures to actively counteract the attack, eliminating the chances of a false alarm system.

  13. Defrosting the digital library: bibliographic tools for the next generation web.

    Science.gov (United States)

    Hull, Duncan; Pettifer, Steve R; Kell, Douglas B

    2008-10-01

    Many scientists now manage the bulk of their bibliographic information electronically, thereby organizing their publications and citation material from digital libraries. However, a library has been described as "thought in cold storage," and unfortunately many digital libraries can be cold, impersonal, isolated, and inaccessible places. In this Review, we discuss the current chilly state of digital libraries for the computational biologist, including PubMed, IEEE Xplore, the ACM digital library, ISI Web of Knowledge, Scopus, Citeseer, arXiv, DBLP, and Google Scholar. We illustrate the current process of using these libraries with a typical workflow, and highlight problems with managing data and metadata using URIs. We then examine a range of new applications such as Zotero, Mendeley, Mekentosj Papers, MyNCBI, CiteULike, Connotea, and HubMed that exploit the Web to make these digital libraries more personal, sociable, integrated, and accessible places. We conclude with how these applications may begin to help achieve a digital defrost, and discuss some of the issues that will help or hinder this in terms of making libraries on the Web warmer places in the future, becoming resources that are considerably more useful to both humans and machines.

  14. Defrosting the digital library: bibliographic tools for the next generation web.

    Directory of Open Access Journals (Sweden)

    Duncan Hull

    2008-10-01

    Full Text Available Many scientists now manage the bulk of their bibliographic information electronically, thereby organizing their publications and citation material from digital libraries. However, a library has been described as "thought in cold storage," and unfortunately many digital libraries can be cold, impersonal, isolated, and inaccessible places. In this Review, we discuss the current chilly state of digital libraries for the computational biologist, including PubMed, IEEE Xplore, the ACM digital library, ISI Web of Knowledge, Scopus, Citeseer, arXiv, DBLP, and Google Scholar. We illustrate the current process of using these libraries with a typical workflow, and highlight problems with managing data and metadata using URIs. We then examine a range of new applications such as Zotero, Mendeley, Mekentosj Papers, MyNCBI, CiteULike, Connotea, and HubMed that exploit the Web to make these digital libraries more personal, sociable, integrated, and accessible places. We conclude with how these applications may begin to help achieve a digital defrost, and discuss some of the issues that will help or hinder this in terms of making libraries on the Web warmer places in the future, becoming resources that are considerably more useful to both humans and machines.

  15. Human Activity in the Web

    OpenAIRE

    Radicchi, Filippo

    2009-01-01

    The recent information technology revolution has enabled the analysis and processing of large-scale datasets describing human activities. The main source of data is represented by the Web, where humans generally use to spend a relevant part of their day. Here we study three large datasets containing the information about Web human activities in different contexts. We study in details inter-event and waiting time statistics. In both cases, the number of subsequent operations which differ by ta...

  16. Web 2.0 and pharmacy education.

    Science.gov (United States)

    Cain, Jeff; Fox, Brent I

    2009-11-12

    New types of social Internet applications (often referred to as Web 2.0) are becoming increasingly popular within higher education environments. Although developed primarily for entertainment and social communication within the general population, applications such as blogs, social video sites, and virtual worlds are being adopted by higher education institutions. These newer applications differ from standard Web sites in that they involve the users in creating and distributing information, hence effectively changing how the Web is used for knowledge generation and dispersion. Although Web 2.0 applications offer exciting new ways to teach, they should not be the core of instructional planning, but rather selected only after learning objectives and instructional strategies have been identified. This paper provides an overview of prominent Web 2.0 applications, explains how they are being used within education environments, and elaborates on some of the potential opportunities and challenges that these applications present.

  17. Web 2.0 and Pharmacy Education

    Science.gov (United States)

    Fox, Brent I.

    2009-01-01

    New types of social Internet applications (often referred to as Web 2.0) are becoming increasingly popular within higher education environments. Although developed primarily for entertainment and social communication within the general population, applications such as blogs, social video sites, and virtual worlds are being adopted by higher education institutions. These newer applications differ from standard Web sites in that they involve the users in creating and distributing information, hence effectively changing how the Web is used for knowledge generation and dispersion. Although Web 2.0 applications offer exciting new ways to teach, they should not be the core of instructional planning, but rather selected only after learning objectives and instructional strategies have been identified. This paper provides an overview of prominent Web 2.0 applications, explains how they are being used within education environments, and elaborates on some of the potential opportunities and challenges that these applications present. PMID:19960079

  18. Digital libraries and World Wide Web sites and page persistence.

    Directory of Open Access Journals (Sweden)

    Wallace Koehler

    1999-01-01

    Full Text Available Web pages and Web sites, some argue, can either be collected as elements of digital or hybrid libraries, or, as others would have it, the WWW is itself a library. We begin with the assumption that Web pages and Web sites can be collected and categorized. The paper explores the proposition that the WWW constitutes a library. We conclude that the Web is not a digital library. However, its component parts can be aggregated and included as parts of digital library collections. These, in turn, can be incorporated into "hybrid libraries." These are libraries with both traditional and digital collections. Material on the Web can be organized and managed. Native documents can be collected in situ, disseminated, distributed, catalogueed, indexed, controlled, in traditional library fashion. The Web therefore is not a library, but material for library collections is selected from the Web. That said, the Web and its component parts are dynamic. Web documents undergo two kinds of change. The first type, the type addressed in this paper, is "persistence" or the existence or disappearance of Web pages and sites, or in a word the lifecycle of Web documents. "Intermittence" is a variant of persistence, and is defined as the disappearance but reappearance of Web documents. At any given time, about five percent of Web pages are intermittent, which is to say they are gone but will return. Over time a Web collection erodes. Based on a 120-week longitudinal study of a sample of Web documents, it appears that the half-life of a Web page is somewhat less than two years and the half-life of a Web site is somewhat more than two years. That is to say, an unweeded Web document collection created two years ago would contain the same number of URLs, but only half of those URLs point to content. The second type of change Web documents experience is change in Web page or Web site content. Again based on the Web document samples, very nearly all Web pages and sites undergo some

  19. Comparing cosmic web classifiers using information theory

    Energy Technology Data Exchange (ETDEWEB)

    Leclercq, Florent [Institute of Cosmology and Gravitation (ICG), University of Portsmouth, Dennis Sciama Building, Burnaby Road, Portsmouth PO1 3FX (United Kingdom); Lavaux, Guilhem; Wandelt, Benjamin [Institut d' Astrophysique de Paris (IAP), UMR 7095, CNRS – UPMC Université Paris 6, Sorbonne Universités, 98bis boulevard Arago, F-75014 Paris (France); Jasche, Jens, E-mail: florent.leclercq@polytechnique.org, E-mail: lavaux@iap.fr, E-mail: j.jasche@tum.de, E-mail: wandelt@iap.fr [Excellence Cluster Universe, Technische Universität München, Boltzmannstrasse 2, D-85748 Garching (Germany)

    2016-08-01

    We introduce a decision scheme for optimally choosing a classifier, which segments the cosmic web into different structure types (voids, sheets, filaments, and clusters). Our framework, based on information theory, accounts for the design aims of different classes of possible applications: (i) parameter inference, (ii) model selection, and (iii) prediction of new observations. As an illustration, we use cosmographic maps of web-types in the Sloan Digital Sky Survey to assess the relative performance of the classifiers T-WEB, DIVA and ORIGAMI for: (i) analyzing the morphology of the cosmic web, (ii) discriminating dark energy models, and (iii) predicting galaxy colors. Our study substantiates a data-supported connection between cosmic web analysis and information theory, and paves the path towards principled design of analysis procedures for the next generation of galaxy surveys. We have made the cosmic web maps, galaxy catalog, and analysis scripts used in this work publicly available.

  20. Comparing cosmic web classifiers using information theory

    International Nuclear Information System (INIS)

    Leclercq, Florent; Lavaux, Guilhem; Wandelt, Benjamin; Jasche, Jens

    2016-01-01

    We introduce a decision scheme for optimally choosing a classifier, which segments the cosmic web into different structure types (voids, sheets, filaments, and clusters). Our framework, based on information theory, accounts for the design aims of different classes of possible applications: (i) parameter inference, (ii) model selection, and (iii) prediction of new observations. As an illustration, we use cosmographic maps of web-types in the Sloan Digital Sky Survey to assess the relative performance of the classifiers T-WEB, DIVA and ORIGAMI for: (i) analyzing the morphology of the cosmic web, (ii) discriminating dark energy models, and (iii) predicting galaxy colors. Our study substantiates a data-supported connection between cosmic web analysis and information theory, and paves the path towards principled design of analysis procedures for the next generation of galaxy surveys. We have made the cosmic web maps, galaxy catalog, and analysis scripts used in this work publicly available.

  1. Commonsense in parts : Mining part-whole relations from theweb and image tags

    NARCIS (Netherlands)

    Tandon, Niket; Hariman, Charles; Urbani, Jacopo; Rohrbach, Anna; Rohrbach, Marcus; Weikum, Gerhard

    2016-01-01

    Commonsense knowledge about part-whole relations (e.g., screen partOf notebook) is important for interpreting user input in web search and question answering, or for object detection in images. Prior work on knowledge base construction has compiled part-whole assertions, but with substantial

  2. Ramakrishnan: Semantics on the Web

    Data.gov (United States)

    National Aeronautics and Space Administration — It is becoming increasingly clear that the next generation of web search and advertising will rely on a deeper understanding of user intent and task modeling, and a...

  3. Food web framework for size-structured populations

    DEFF Research Database (Denmark)

    Hartvig, Martin; Andersen, Ken Haste; Beyer, Jan

    2011-01-01

    . Parameter values are determined from cross-species analysis of fish communities as life-history omnivory is widespread in aquatic systems, but may be reparameterised for other systems. An ensemble of food webs is generated and the resulting communities are analysed at four levels of organisation: community......We synthesise traditional unstructured food webs, allometric body size scaling, trait-based modelling, and physiologically structured modelling to provide a novel and ecologically relevant tool for size-structured food webs. The framework allows food web models to include ontogenetic growth...... level, species level, trait level, and individual level. The model may be solved analytically by assuming that the community spectrum follows a power law. The analytical solution provides a baseline expectation of the results of complex food web simulations, and agrees well with the predictions...

  4. Development of a laboratory niche Web site.

    Science.gov (United States)

    Dimenstein, Izak B; Dimenstein, Simon I

    2013-10-01

    This technical note presents the development of a methodological laboratory niche Web site. The "Grossing Technology in Surgical Pathology" (www.grossing-technology.com) Web site is used as an example. Although common steps in creation of most Web sites are followed, there are particular requirements for structuring the template's menu on methodological laboratory Web sites. The "nested doll principle," in which one object is placed inside another, most adequately describes the methodological approach to laboratory Web site design. Fragmentation in presenting the Web site's material highlights the discrete parts of the laboratory procedure. An optimally minimal triad of components can be recommended for the creation of a laboratory niche Web site: a main set of media, a blog, and an ancillary component (host, contact, and links). The inclusion of a blog makes the Web site a dynamic forum for professional communication. By forming links and portals, cloud computing opens opportunities for connecting a niche Web site with other Web sites and professional organizations. As an additional source of information exchange, methodological laboratory niche Web sites are destined to parallel both traditional and new forms, such as books, journals, seminars, webinars, and internal educational materials. Copyright © 2013 Elsevier Inc. All rights reserved.

  5. Sentiment Analysis of Web Sites Related to Vaginal Mesh Use in Pelvic Reconstructive Surgery.

    Science.gov (United States)

    Hobson, Deslyn T G; Meriwether, Kate V; Francis, Sean L; Kinman, Casey L; Stewart, J Ryan

    2018-05-02

    The purpose of this study was to utilize sentiment analysis to describe online opinions toward vaginal mesh. We hypothesized that sentiment in legal Web sites would be more negative than that in medical and reference Web sites. We generated a list of relevant key words related to vaginal mesh and searched Web sites using the Google search engine. Each unique uniform resource locator (URL) was sorted into 1 of 6 categories: "medical", "legal", "news/media", "patient generated", "reference", or "unrelated". Sentiment of relevant Web sites, the primary outcome, was scored on a scale of -1 to +1, and mean sentiment was compared across all categories using 1-way analysis of variance. Tukey test evaluated differences between category pairs. Google searches of 464 unique key words resulted in 11,405 URLs. Sentiment analysis was performed on 8029 relevant URLs (3472 legal, 1625 "medical", 1774 "reference", 666 "news media", 492 "patient generated"). The mean sentiment for all relevant Web sites was +0.01 ± 0.16; analysis of variance revealed significant differences between categories (P Web sites categorized as "legal" and "news/media" had a slightly negative mean sentiment, whereas those categorized as "medical," "reference," and "patient generated" had slightly positive mean sentiments. Tukey test showed differences between all category pairs except the "medical" versus "reference" in comparison with the largest mean difference (-0.13) seen in the "legal" versus "reference" comparison. Web sites related to vaginal mesh have an overall mean neutral sentiment, and Web sites categorized as "medical," "reference," and "patient generated" have significantly higher sentiment scores than related Web sites in "legal" and "news/media" categories.

  6. A Method for Transforming Existing Web Service Descriptions into an Enhanced Semantic Web Service Framework

    Science.gov (United States)

    Du, Xiaofeng; Song, William; Munro, Malcolm

    Web Services as a new distributed system technology has been widely adopted by industries in the areas, such as enterprise application integration (EAI), business process management (BPM), and virtual organisation (VO). However, lack of semantics in the current Web Service standards has been a major barrier in service discovery and composition. In this chapter, we propose an enhanced context-based semantic service description framework (CbSSDF+) that tackles the problem and improves the flexibility of service discovery and the correctness of generated composite services. We also provide an agile transformation method to demonstrate how the various formats of Web Service descriptions on the Web can be managed and renovated step by step into CbSSDF+ based service description without large amount of engineering work. At the end of the chapter, we evaluate the applicability of the transformation method and the effectiveness of CbSSDF+ through a series of experiments.

  7. Pedagogy for teaching and learning cooperatively on the Web: a Web-based pharmacology course.

    Science.gov (United States)

    Tse, Mimi M Y; Pun, Sandra P Y; Chan, Moon Fai

    2007-02-01

    The Internet is becoming a preferred place to find information. Millions of people go online in the search of health and medical information. Likewise, the demand for Web-based courses grows. This article presents the development, utilization and evaluation of a web-based pharmacology course for nursing students. The course was developed based on 150 commonly used drugs. There were 110 year 1 nursing students took part in the course. After attending six hours face to face lecture of pharmacology over three weeks, students were invited to complete a questionnaire (pre-test) about learning pharmacology. The course materials were then uploaded to a WebCT for student's self-directed learning and attempts to pass two scheduled online quizzes. At the end of the semester, students were given the same questionnaire (post-test). There were a significant increase in the understanding compared with memorizing the subject content, the development of problem solving ability in learning pharmacology and becoming an independent learner (p ,0.05). Online quizzes yielded satisfactory results. In the focused group interview, students appreciated the time flexibility and convenience associated with web-based learning, also, they had made good suggestions in enhancing web-based learning. Web-based approach is promising for teaching and learning pharmacology for nurses and other health-care professionals.

  8. Deep Web : acceso, seguridad y análisis de tráfico

    OpenAIRE

    Cagiga Vila, Ignacio

    2017-01-01

    RESUMEN: Este trabajo pretende hacer un análisis técnico de la Deep Web en el ámbito de las redes y las tecnologías de Internet. La parte principal del proyecto puede verse dividida en dos partes: Acceso a la Deep Web como cliente, e implementación de un relay de la Tor Network. La implementación de un relay de la Tor Network permite comprender como se consigue asegurar la anonimidad y seguridad de los usuarios que intentan acceder a la Deep Web a través de esta red. La parte de laboratorio d...

  9. The wireless Web and patient care.

    Science.gov (United States)

    Bergeron, B P

    2001-01-01

    Wireless computing, when integrated with the Web, is poised to revolutionize the practice and teaching of medicine. As vendors introduce wireless Web technologies in the medical community that have been used successfully in the business and consumer markets, clinicians can expect profound increases in the amount of patient data, as well as the ease with which those data are acquired, analyzed, and disseminated. The enabling technologies involved in this transformation to the wireless Web range from the new generation of wireless PDAs, eBooks, and wireless data acquisition peripherals to new wireless network protocols. The rate-limiting step in the application of this technology in medicine is not technology per se but rather how quickly clinicians and their patients come to accept and appreciate the benefits and limitations of the application of wireless Web technology.

  10. Contrasting Web Robot and Human Behaviors with Network Models

    OpenAIRE

    Brown, Kyle; Doran, Derek

    2018-01-01

    The web graph is a commonly-used network representation of the hyperlink structure of a website. A network of similar structure to the web graph, which we call the session graph has properties that reflect the browsing habits of the agents in the web server logs. In this paper, we apply session graphs to compare the activity of humans against web robots or crawlers. Understanding these properties will enable us to improve models of HTTP traffic, which can be used to predict and generate reali...

  11. Usability Evaluation of Public Web Mapping Sites

    Science.gov (United States)

    Wang, C.

    2014-04-01

    Web mapping sites are interactive maps that are accessed via Webpages. With the rapid development of Internet and Geographic Information System (GIS) field, public web mapping sites are not foreign to people. Nowadays, people use these web mapping sites for various reasons, in that increasing maps and related map services of web mapping sites are freely available for end users. Thus, increased users of web mapping sites led to more usability studies. Usability Engineering (UE), for instance, is an approach for analyzing and improving the usability of websites through examining and evaluating an interface. In this research, UE method was employed to explore usability problems of four public web mapping sites, analyze the problems quantitatively and provide guidelines for future design based on the test results. Firstly, the development progress for usability studies were described, and simultaneously several usability evaluation methods such as Usability Engineering (UE), User-Centered Design (UCD) and Human-Computer Interaction (HCI) were generally introduced. Then the method and procedure of experiments for the usability test were presented in detail. In this usability evaluation experiment, four public web mapping sites (Google Maps, Bing maps, Mapquest, Yahoo Maps) were chosen as the testing websites. And 42 people, who having different GIS skills (test users or experts), gender (male or female), age and nationality, participated in this test to complete the several test tasks in different teams. The test comprised three parts: a pretest background information questionnaire, several test tasks for quantitative statistics and progress analysis, and a posttest questionnaire. The pretest and posttest questionnaires focused on gaining the verbal explanation of their actions qualitatively. And the design for test tasks targeted at gathering quantitative data for the errors and problems of the websites. Then, the results mainly from the test part were analyzed. The

  12. Geo-communication and web-based geospatial infrastructure

    DEFF Research Database (Denmark)

    Brodersen, Lars; Nielsen, Anders

    2005-01-01

    The introduction of web-services as index-portals based on geoinformation has changed the conditions for both content and form of geocommunication. A high number of players and interactions (as well as a very high number of all kinds of information and combinations of these) characterize web-services......, where maps are only a part of the whole. These new conditions demand new ways of modelling the processes leading to geo-communication. One new aspect is the fact that the service providers have become a part of the geo-communication process with influence on the content. Another aspect...

  13. La salud de las web universitarias españolas

    Directory of Open Access Journals (Sweden)

    Thelwall, Mike

    2003-09-01

    Full Text Available The Web has become an important tool for universities, and one that is employed in a variety of ways. Examples are: disseminating and publicising research findings and activities; publishing teaching and administrative information for students; and collaborating with other institutions nationally and internationally. But how effectively are Spanish universities using the Web and what information can be gained about online communication patterns through the study of Web links? This paper reports on an investigation into 64 university Web sites that were indexed using a specialist information science Web crawler and analysed using associated software. There were a wide variety of sizes for university Web sites and that universities attracted links from others broadly in proportion to their site size. The Spanish academic Web was found to lag behind those of the four countries that it was compared to. However, the most commonly targeted top-level Internet domains were from non-Spanish speaking high Web using countries around the world, showing a broad international perspective and high degree of multilingualism for Web authors. The most highly targeted pages were mainly those that attracted automatically generated links, but several government ministries were a surprise inclusion.

    La web se ha convertido en una importante herramienta para las universidades, donde se utiliza en una amplia variedad de formas, tales como publicar y diseminar actividades y resultados de investigación, proporcionar información administrativa y académica de interés para los estudiantes o facilitar la colaboración con otras instituciones nacionales e internacionales. Pero, ¿cómo están realmente utilizando la Web las universidades españolas y qué información se puede obtener sobre sus modos de comunicación en línea a través del estudio de los enlaces web?. Para obtener respuestas se han investigado 64 sedes Web de universidades utilizando un robot

  14. The DIRAC Web Portal 2.0

    Science.gov (United States)

    Mathe, Z.; Casajus Ramo, A.; Lazovsky, N.; Stagni, F.

    2015-12-01

    For many years the DIRAC interware (Distributed Infrastructure with Remote Agent Control) has had a web interface, allowing the users to monitor DIRAC activities and also interact with the system. Since then many new web technologies have emerged, therefore a redesign and a new implementation of the DIRAC Web portal were necessary, taking into account the lessons learnt using the old portal. These new technologies allowed to build a more compact, robust and responsive web interface that enables users to have better control over the whole system while keeping a simple interface. The web framework provides a large set of “applications”, each of which can be used for interacting with various parts of the system. Communities can also create their own set of personalised web applications, and can easily extend already existing ones with a minimal effort. Each user can configure and personalise the view for each application and save it using the DIRAC User Profile service as RESTful state provider, instead of using cookies. The owner of a view can share it with other users or within a user community. Compatibility between different browsers is assured, as well as with mobile versions. In this paper, we present the new DIRAC Web framework as well as the LHCb extension of the DIRAC Web portal.

  15. EasyFRAP-web: a web-based tool for the analysis of fluorescence recovery after photobleaching data.

    Science.gov (United States)

    Koulouras, Grigorios; Panagopoulos, Andreas; Rapsomaniki, Maria A; Giakoumakis, Nickolaos N; Taraviras, Stavros; Lygerou, Zoi

    2018-06-13

    Understanding protein dynamics is crucial in order to elucidate protein function and interactions. Advances in modern microscopy facilitate the exploration of the mobility of fluorescently tagged proteins within living cells. Fluorescence recovery after photobleaching (FRAP) is an increasingly popular functional live-cell imaging technique which enables the study of the dynamic properties of proteins at a single-cell level. As an increasing number of labs generate FRAP datasets, there is a need for fast, interactive and user-friendly applications that analyze the resulting data. Here we present easyFRAP-web, a web application that simplifies the qualitative and quantitative analysis of FRAP datasets. EasyFRAP-web permits quick analysis of FRAP datasets through an intuitive web interface with interconnected analysis steps (experimental data assessment, different types of normalization and estimation of curve-derived quantitative parameters). In addition, easyFRAP-web provides dynamic and interactive data visualization and data and figure export for further analysis after every step. We test easyFRAP-web by analyzing FRAP datasets capturing the mobility of the cell cycle regulator Cdt2 in the presence and absence of DNA damage in cultured cells. We show that easyFRAP-web yields results consistent with previous studies and highlights cell-to-cell heterogeneity in the estimated kinetic parameters. EasyFRAP-web is platform-independent and is freely accessible at: https://easyfrap.vmnet.upatras.gr/.

  16. Molecular structure input on the web

    Directory of Open Access Journals (Sweden)

    Ertl Peter

    2010-02-01

    Full Text Available Abstract A molecule editor, that is program for input and editing of molecules, is an indispensable part of every cheminformatics or molecular processing system. This review focuses on a special type of molecule editors, namely those that are used for molecule structure input on the web. Scientific computing is now moving more and more in the direction of web services and cloud computing, with servers scattered all around the Internet. Thus a web browser has become the universal scientific user interface, and a tool to edit molecules directly within the web browser is essential. The review covers a history of web-based structure input, starting with simple text entry boxes and early molecule editors based on clickable maps, before moving to the current situation dominated by Java applets. One typical example - the popular JME Molecule Editor - will be described in more detail. Modern Ajax server-side molecule editors are also presented. And finally, the possible future direction of web-based molecule editing, based on technologies like JavaScript and Flash, is discussed.

  17. Web Based VRML Modelling

    NARCIS (Netherlands)

    Kiss, S.; Sarfraz, M.

    2004-01-01

    Presents a method to connect VRML (Virtual Reality Modeling Language) and Java components in a Web page using EAI (External Authoring Interface), which makes it possible to interactively generate and edit VRML meshes. The meshes used are based on regular grids, to provide an interaction and modeling

  18. Web application security analysis using the Kali Linux operating system

    OpenAIRE

    BABINCEV IVAN M.; VULETIC DEJAN V.

    2016-01-01

    The Kali Linux operating system is described as well as its purpose and possibilities. There are listed groups of tools that Kali Linux has together with the methods of their functioning, as well as a possibility to install and use tools that are not an integral part of Kali. The final part shows a practical testing of web applications using the tools from the Kali Linux operating system. The paper thus shows a part of the possibilities of this operating system in analaysing web applications ...

  19. Untangling Web 2.0: Charting Web 2.0 Tools, the NCSS Guidelines for Effective Use of Technology, and Bloom's Taxonomy

    Science.gov (United States)

    Diacopoulos, Mark M.

    2015-01-01

    The potential for social studies to embrace instructional technology and Web 2.0 applications has become a growing trend in recent social studies research. As part of an ongoing process of collaborative enquiry between an instructional specialist and social studies teachers in a Professional Learning Community, a table of Web 2.0 applications was…

  20. LigParGen web server: an automatic OPLS-AA parameter generator for organic ligands

    Science.gov (United States)

    Dodda, Leela S.

    2017-01-01

    Abstract The accurate calculation of protein/nucleic acid–ligand interactions or condensed phase properties by force field-based methods require a precise description of the energetics of intermolecular interactions. Despite the progress made in force fields, small molecule parameterization remains an open problem due to the magnitude of the chemical space; the most critical issue is the estimation of a balanced set of atomic charges with the ability to reproduce experimental properties. The LigParGen web server provides an intuitive interface for generating OPLS-AA/1.14*CM1A(-LBCC) force field parameters for organic ligands, in the formats of commonly used molecular dynamics and Monte Carlo simulation packages. This server has high value for researchers interested in studying any phenomena based on intermolecular interactions with ligands via molecular mechanics simulations. It is free and open to all at jorgensenresearch.com/ligpargen, and has no login requirements. PMID:28444340

  1. The generation of large networks from web of science data

    NARCIS (Netherlands)

    Leydesdorff, L.; Khan, G.F.; Bornmann, L.

    2014-01-01

    During the 1990s, one of us developed a series of freeware routines (http://www.leydesdorff.net/indicators) that enable the user to organize downloads from the Web of Science (Thomson Reuters) into a relational database, and then to export matrices for further analysis in various formats (for

  2. Interactive Web-based e-learning for Studying Flexible Manipulator Systems

    Directory of Open Access Journals (Sweden)

    Abul K. M. Azad

    2008-03-01

    Full Text Available Abstract— This paper presents a web-based e-leaning facility for simulation, modeling, and control of flexible manipulator systems. The simulation and modeling part includes finite difference and finite element simulations along with neural network and genetic algorithm based modeling strategies for flexible manipulator systems. The controller part constitutes a number of open-loop and closed-loop designs. Closed loop control designs include the classical, adaptive, and neuro-model based strategies. Matlab software package and its associated toolboxes are used to implement these. The Matlab web server is used as the gateway between the facility and web-access. ASP.NET technology and SQL database are utilized to develop web applications for access control, user account and password maintenance, administrative management, and facility utilization monitoring. The reported facility provides a flexible but effective approach of web-based interactive e-learning facility of an engineering system. This can be extended to incorporate additional engineering systems within the e-learning framework.

  3. Exploiting Multimedia in Creating and Analysing Multimedia Web Archives

    Directory of Open Access Journals (Sweden)

    Jonathon S. Hare

    2014-04-01

    Full Text Available The data contained on the web and the social web are inherently multimedia and consist of a mixture of textual, visual and audio modalities. Community memories embodied on the web and social web contain a rich mixture of data from these modalities. In many ways, the web is the greatest resource ever created by human-kind. However, due to the dynamic and distributed nature of the web, its content changes, appears and disappears on a daily basis. Web archiving provides a way of capturing snapshots of (parts of the web for preservation and future analysis. This paper provides an overview of techniques we have developed within the context of the EU funded ARCOMEM (ARchiving COmmunity MEMories project to allow multimedia web content to be leveraged during the archival process and for post-archival analysis. Through a set of use cases, we explore several practical applications of multimedia analytics within the realm of web archiving, web archive analysis and multimedia data on the web in general.

  4. Mobile Web: the democratisation of an essential tool

    CERN Multimedia

    Laëtitia Pedroso

    2011-01-01

    For many of us, using the Web is a natural and even indispensable part of our daily lives. But only 20% of the world’s population have access to it. Tim Berners-Lee, the Web's inventor, created the Web Foundation in 2007 with the aim of accelerating access to the Web for the rest of the world's population. Showcased at the Sharing Knowledge conference, the Mobile Web is one of the Web Foundation’s projects in which members of CERN are involved.   Virtually no access to the Web but a very extensive GSM network: that's the situation that many developing countries especially in Africa find themselves in. “Owing to its size, its unstable soils and its limited infrastructure, it is technically very difficult to bring optic fibres for Internet connections to all regions of Africa. The idea of the Mobile Web project is therefore to be able to use the GSM network to access the Web,” explains Silvano de Gennaro, a member of the video team within CERN's Communication Gro...

  5. LHCb : The DIRAC Web Portal 2.0

    CERN Multimedia

    Mathe, Zoltan; Lazovsky, N; Stagni, Federico

    2015-01-01

    For many years the DIRAC interware (Distributed Infrastructure with Remote Agent Control) has had a web interface, allowing the users to monitor DIRAC activities and also interact with the system. Since then many new web technologies have emerged, therefore a redesign and a new implementation of the DIRAC Web portal were necessary, taking into account the lessons learnt using the old portal. These new technologies allowed to build a more compact and more responsive web interface that is robust and that enables users to have more control over the whole system while keeping a simple interface. The framework provides a large set of "applications", each of which can be used for interacting with various parts of the system. Communities can also create their own set of personalised web applications, and can easily extend already existing web applications with a minimal effort. Each user can configure and personalise the view for each application and save it using the DIRAC User Profile service as RESTful state prov...

  6. Association and Sequence Mining in Web Usage

    Directory of Open Access Journals (Sweden)

    Claudia Elena DINUCA

    2011-06-01

    Full Text Available Web servers worldwide generate a vast amount of information on web users’ browsing activities. Several researchers have studied these so-called clickstream or web access log data to better understand and characterize web users. Clickstream data can be enriched with information about the content of visited pages and the origin (e.g., geographic, organizational of the requests. The goal of this project is to analyse user behaviour by mining enriched web access log data. With the continued growth and proliferation of e-commerce, Web services, and Web-based information systems, the volumes of click stream and user data collected by Web-based organizations in their daily operations has reached astronomical proportions. This information can be exploited in various ways, such as enhancing the effectiveness of websites or developing directed web marketing campaigns. The discovered patterns are usually represented as collections of pages, objects, or re-sources that are frequently accessed by groups of users with common needs or interests. The focus of this paper is to provide an overview how to use frequent pattern techniques for discovering different types of patterns in a Web log database. In this paper we will focus on finding association as a data mining technique to extract potentially useful knowledge from web usage data. I implemented in Java, using NetBeans IDE, a program for identification of pages’ association from sessions. For exemplification, we used the log files from a commercial web site.

  7. Applying Semantic Web technologies to improve the retrieval, credibility and use of health-related web resources.

    Science.gov (United States)

    Mayer, Miguel A; Karampiperis, Pythagoras; Kukurikos, Antonis; Karkaletsis, Vangelis; Stamatakis, Kostas; Villarroel, Dagmar; Leis, Angela

    2011-06-01

    The number of health-related websites is increasing day-by-day; however, their quality is variable and difficult to assess. Various "trust marks" and filtering portals have been created in order to assist consumers in retrieving quality medical information. Consumers are using search engines as the main tool to get health information; however, the major problem is that the meaning of the web content is not machine-readable in the sense that computers cannot understand words and sentences as humans can. In addition, trust marks are invisible to search engines, thus limiting their usefulness in practice. During the last five years there have been different attempts to use Semantic Web tools to label health-related web resources to help internet users identify trustworthy resources. This paper discusses how Semantic Web technologies can be applied in practice to generate machine-readable labels and display their content, as well as to empower end-users by providing them with the infrastructure for expressing and sharing their opinions on the quality of health-related web resources.

  8. A web service system supporting three-dimensional post-processing of medical images based on WADO protocol.

    Science.gov (United States)

    He, Longjun; Xu, Lang; Ming, Xing; Liu, Qian

    2015-02-01

    Three-dimensional post-processing operations on the volume data generated by a series of CT or MR images had important significance on image reading and diagnosis. As a part of the DIOCM standard, WADO service defined how to access DICOM objects on the Web, but it didn't involve three-dimensional post-processing operations on the series images. This paper analyzed the technical features of three-dimensional post-processing operations on the volume data, and then designed and implemented a web service system for three-dimensional post-processing operations of medical images based on the WADO protocol. In order to improve the scalability of the proposed system, the business tasks and calculation operations were separated into two modules. As results, it was proved that the proposed system could support three-dimensional post-processing service of medical images for multiple clients at the same moment, which met the demand of accessing three-dimensional post-processing operations on the volume data on the web.

  9. MCM generator: a Java-based tool for generating medical metadata.

    Science.gov (United States)

    Munoz, F; Hersh, W

    1998-01-01

    In a previous paper we introduced the need to implement a mechanism to facilitate the discovery of relevant Web medical documents. We maintained that the use of META tags, specifically ones that define the medical subject and resource type of a document, help towards this goal. We have now developed a tool to facilitate the generation of these tags for the authors of medical documents. Written entirely in Java, this tool makes use of the SAPHIRE server, and helps the author identify the Medical Subject Heading terms that most appropriately describe the subject of the document. Furthermore, it allows the author to generate metadata tags for the 15 elements that the Dublin Core considers as core elements in the description of a document. This paper describes the use of this tool in the cataloguing of Web and non-Web medical documents, such as images, movie, and sound files.

  10. Power Plants, Steam and Gas Turbines WebQuest

    Science.gov (United States)

    Ulloa, Carlos; Rey, Guillermo D.; Sánchez, Ángel; Cancela, Ángeles

    2012-01-01

    A WebQuest is an Internet-based and inquiry-oriented learning activity. The aim of this work is to outline the creation of a WebQuest entitled "Power Generation Plants: Steam and Gas Turbines." This is one of the topics covered in the course "Thermodynamics and Heat Transfer," which is offered in the second year of Mechanical…

  11. Characteristics of scientific web publications

    DEFF Research Database (Denmark)

    Thorlund Jepsen, Erik; Seiden, Piet; Ingwersen, Peter Emil Rerup

    2004-01-01

    were generated based on specifically selected domain topics that are searched for in three publicly accessible search engines (Google, AllTheWeb, and AltaVista). A sample of the retrieved hits was analyzed with regard to how various publication attributes correlated with the scientific quality...... of the content and whether this information could be employed to harvest, filter, and rank Web publications. The attributes analyzed were inlinks, outlinks, bibliographic references, file format, language, search engine overlap, structural position (according to site structure), and the occurrence of various...... types of metadata. As could be expected, the ranked output differs between the three search engines. Apparently, this is caused by differences in ranking algorithms rather than the databases themselves. In fact, because scientific Web content in this subject domain receives few inlinks, both Alta...

  12. web cellHTS2: A web-application for the analysis of high-throughput screening data

    Directory of Open Access Journals (Sweden)

    Boutros Michael

    2010-04-01

    Full Text Available Abstract Background The analysis of high-throughput screening data sets is an expanding field in bioinformatics. High-throughput screens by RNAi generate large primary data sets which need to be analyzed and annotated to identify relevant phenotypic hits. Large-scale RNAi screens are frequently used to identify novel factors that influence a broad range of cellular processes, including signaling pathway activity, cell proliferation, and host cell infection. Here, we present a web-based application utility for the end-to-end analysis of large cell-based screening experiments by cellHTS2. Results The software guides the user through the configuration steps that are required for the analysis of single or multi-channel experiments. The web-application provides options for various standardization and normalization methods, annotation of data sets and a comprehensive HTML report of the screening data analysis, including a ranked hit list. Sessions can be saved and restored for later re-analysis. The web frontend for the cellHTS2 R/Bioconductor package interacts with it through an R-server implementation that enables highly parallel analysis of screening data sets. web cellHTS2 further provides a file import and configuration module for common file formats. Conclusions The implemented web-application facilitates the analysis of high-throughput data sets and provides a user-friendly interface. web cellHTS2 is accessible online at http://web-cellHTS2.dkfz.de. A standalone version as a virtual appliance and source code for platforms supporting Java 1.5.0 can be downloaded from the web cellHTS2 page. web cellHTS2 is freely distributed under GPL.

  13. Web components and the semantic web

    OpenAIRE

    Casey, Maire; Pahl, Claus

    2003-01-01

    Component-based software engineering on the Web differs from traditional component and software engineering. We investigate Web component engineering activites that are crucial for the development,com position, and deployment of components on the Web. The current Web Services and Semantic Web initiatives strongly influence our work. Focussing on Web component composition we develop description and reasoning techniques that support a component developer in the composition activities,fo cussing...

  14. Physical Webbing: Collaborative Kinesthetic Three-Dimensional Mind Maps[R

    Science.gov (United States)

    Williams, Marian H.

    2012-01-01

    Mind Mapping has predominantly been used by individuals or collaboratively in groups as a paper-based or computer-generated learning strategy. In an effort to make Mind Mapping kinesthetic, collaborative, and three-dimensional, an innovative pedagogical strategy, termed Physical Webbing, was devised. In the Physical Web activity, groups…

  15. Life as an emergent phenomenon: studies from a large-scale boid simulation and web data

    Science.gov (United States)

    Ikegami, Takashi; Mototake, Yoh-ichi; Kobori, Shintaro; Oka, Mizuki; Hashimoto, Yasuhiro

    2017-11-01

    A large group with a special structure can become the mother of emergence. We discuss this hypothesis in relation to large-scale boid simulations and web data. In the boid swarm simulations, the nucleation, organization and collapse dynamics were found to be more diverse in larger flocks than in smaller flocks. In the second analysis, large web data, consisting of shared photos with descriptive tags, tended to group together users with similar tendencies, allowing the network to develop a core-periphery structure. We show that the generation rate of novel tags and their usage frequencies are high in the higher-order cliques. In this case, novelty is not considered to arise randomly; rather, it is generated as a result of a large and structured network. We contextualize these results in terms of adjacent possible theory and as a new way to understand collective intelligence. We argue that excessive information and material flow can become a source of innovation. This article is part of the themed issue 'Reconceptualizing the origins of life'.

  16. Control Freaks: How User-Generated Content is Managed in Advertising Campaigns. The Romanian Perspective

    Directory of Open Access Journals (Sweden)

    Rodica Maria SĂVULESCU

    2014-10-01

    Full Text Available The advent of Web 2.0 has brought about a paradigm shift in communication: from an informational Web to an editable one. Consequently, the “one-to-many” communication model of the mass media industry has been replaced by one that involves “many-to-many”. Web 2.0 enables those who were formerly conceptualized as audiences/consumers to be part of an open-source movement that means participating into the production of brands and their meanings. It is an era of convergence, of overlapping roles between producers and users/consumers – in short, of “produsage”, “prosumption” and “user-generated content” (UGC. This article investigates how this new paradigm influences marketing communication by addressing the specific case of the Romanian advertising industry. By using in depth-interviews with 20 advertising professionals, it points out to the fact that in communication campaigns with user-generated content there is still a high degree of control from the brand management and agency teams that cannot quite grasp the concept of brand communication democratization.

  17. Semi-automatic scene generation using the Digital Anatomist Foundational Model.

    Science.gov (United States)

    Wong, B A; Rosse, C; Brinkley, J F

    1999-01-01

    A recent survey shows that a major impediment to more widespread use of computers in anatomy education is the inability to directly manipulate 3-D models, and to relate these to corresponding textual information. In the University of Washington Digital Anatomist Project we have developed a prototype Web-based scene generation program that combines the symbolic Foundational Model of Anatomy with 3-D models. A Web user can browse the Foundational Model (FM), then click to request that a 3-D scene be created of an object and its parts or branches. The scene is rendered by a graphics server, and a snapshot is sent to the Web client. The user can then manipulate the scene, adding new structures, deleting structures, rotating the scene, zooming, and saving the scene as a VRML file. Applications such as this, when fully realized with fast rendering and more anatomical content, have the potential to significantly change the way computers are used in anatomy education.

  18. Linked data: un nuovo alfabeto del web semantico

    Directory of Open Access Journals (Sweden)

    Mauro Guerrini

    2013-01-01

    Full Text Available The paper defines the linked data as a set of best practices that are used to publish data on the web using a machine; the technology (or mode of realization of linked data is associated with the concept of the semantic web. It is the area of the semantic web, or web of data, as defined by Tim Berners-Lee "A web of things in the world, described by data on the web". The paper highlights the continuities and differences between semantic web and web traditional, or web documents. The analysis of linked data takes place within the world of libraries, archives and museums, traditionally committed to high standards for structuring and sharing of data. The data, in fact, assume the role of generating quality information for the network. The production of linked data requires compliance with rules and the use of specific technologies and languages, especially in the case of publication of linked data in open mode. The production cycle of linked data may be the track, or a guideline, for institutions that wish to join projects to publish their data. Data quality is assessed through a rating system designed by Tim Berners-Lee.

  19. A teen's guide to creating web pages and blogs

    CERN Document Server

    Selfridge, Peter; Osburn, Jennifer

    2008-01-01

    Whether using a social networking site like MySpace or Facebook or building a Web page from scratch, millions of teens are actively creating a vibrant part of the Internet. This is the definitive teen''s guide to publishing exciting web pages and blogs on the Web. This easy-to-follow guide shows teenagers how to: Create great MySpace and Facebook pages Build their own unique, personalized Web site Share the latest news with exciting blogging ideas Protect themselves online with cyber-safety tips Written by a teenager for other teens, this book leads readers step-by-step through the basics of web and blog design. In this book, teens learn to go beyond clicking through web sites to learning winning strategies for web design and great ideas for writing blogs that attract attention and readership.

  20. WebVis: a hierarchical web homepage visualizer

    Science.gov (United States)

    Renteria, Jose C.; Lodha, Suresh K.

    2000-02-01

    WebVis, the Hierarchical Web Home Page Visualizer, is a tool for managing home web pages. The user can access this tool via the WWW and obtain a hierarchical visualization of one's home web pages. WebVis is a real time interactive tool that supports many different queries on the statistics of internal files such as sizes, age, and type. In addition, statistics on embedded information such as VRML files, Java applets, images and sound files can be extracted and queried. Results of these queries are visualized using color, shape and size of different nodes of the hierarchy. The visualization assists the user in a variety of task, such as quickly finding outdated information or locate large files. WebVIs is one solution to the growing web space maintenance problem. Implementation of WebVis is realized with Perl and Java. Perl pattern matching and file handling routines are used to collect and process web space linkage information and web document information. Java utilizes the collected information to produce visualization of the web space. Java also provides WebVis with real time interactivity, while running off the WWW. Some WebVis examples of home web page visualization are presented.

  1. Enhancing the AliEn Web Service Authentication

    International Nuclear Information System (INIS)

    Zhu Jianlin; Zhou Daicui; Zhang Guoping; Saiz, Pablo; Carminati, Federico; Betev, Latchezar; Lorenzo, Patricia Mendez; Grigoras, Alina Gabriela; Grigoras, Costin; Furano, Fabrizio; Schreiner, Steffen; Datskova, Olga Vladimirovna; Banerjee, Subho Sankar

    2011-01-01

    Web Services are an XML based technology that allow applications to communicate with each other across disparate systems. Web Services are becoming the de facto standard that enable inter operability between heterogeneous processes and systems. AliEn2 is a grid environment based on web services. The AliEn2 services can be divided in three categories: Central services, deployed once per organization; Site services, deployed on each of the participating centers; Job Agents running on the worker nodes automatically. A security model to protect these services is essential for the whole system. Current implementations of web server, such as Apache, are not suitable to be used within the grid environment. Apache with the mod s sl and OpenSSL only supports the X.509 certificates. But in the grid environment, the common credential is the proxy certificate for the purpose of providing restricted proxy and delegation. An Authentication framework was taken for AliEn2 web services to add the ability to accept X.509 certificates and proxy certificates from client-side to Apache Web Server. The authentication framework could also allow the generation of access control policies to limit access to the AliEn2 web services.

  2. PHP The Good Parts

    CERN Document Server

    MacIntyre, Peter

    2010-01-01

    Get past all the hype about PHP and dig into the real power of this language. This book explores the most useful features of PHP and how they can speed up the web development process, and explains why the most commonly used PHP elements are often misused or misapplied. You'll learn which parts add strength to object-oriented programming, and how to use certain features to integrate your application with databases. Written by a longtime member of the PHP community, PHP: The Good Parts is ideal for new PHP programmers, as well as web developers switching from other languages. Become familiar w

  3. Realization of a system to demand and acquire information via world wide web; Realizzazione di un sistema per l'accesso e l'acquisizione di informazioni via web in un ente complesso

    Energy Technology Data Exchange (ETDEWEB)

    Bongiovanni, G [Rome Univ. La Sapienza, Rome (Italy); Di Marco, R A [ENEA, Sede Centrale, Rome (Italy). Funzione Centrale Informatica; Cappitelli, A

    1999-07-01

    The project realized within this thesis exactly regards interactive web and shows how it is possible to obtain particular functionalities opportunely fixed tools. Such project consists in the realization of a system which allows users with opportune account to demand and to acquire information, and to manage these users and a part of the informative system by whom holds administration tasks. Developed software contains an experimental part and a section dedicated to the public key cryptography, which has been employed to carry out sure transactions via web. [Italian] Il rapporto descrive un progetto di realizzazione di un web interattivo e mostra come sia possibile ottenere particolari funzionalita' impiegando opportunamente determinati strumenti come Java, specifica CGI. Il sistema permette agli utenti, tramite opportuno account, di richiedere e acquisire informazioni. Nel software sviluppato e' presente una parte sperimentale e una parte dedicata alla crittografia a chiave pubblica, impiegata per effettuare transazioni sicure via web.

  4. Standard biological parts knowledgebase.

    Directory of Open Access Journals (Sweden)

    Michal Galdzicki

    2011-02-01

    Full Text Available We have created the Knowledgebase of Standard Biological Parts (SBPkb as a publically accessible Semantic Web resource for synthetic biology (sbolstandard.org. The SBPkb allows researchers to query and retrieve standard biological parts for research and use in synthetic biology. Its initial version includes all of the information about parts stored in the Registry of Standard Biological Parts (partsregistry.org. SBPkb transforms this information so that it is computable, using our semantic framework for synthetic biology parts. This framework, known as SBOL-semantic, was built as part of the Synthetic Biology Open Language (SBOL, a project of the Synthetic Biology Data Exchange Group. SBOL-semantic represents commonly used synthetic biology entities, and its purpose is to improve the distribution and exchange of descriptions of biological parts. In this paper, we describe the data, our methods for transformation to SBPkb, and finally, we demonstrate the value of our knowledgebase with a set of sample queries. We use RDF technology and SPARQL queries to retrieve candidate "promoter" parts that are known to be both negatively and positively regulated. This method provides new web based data access to perform searches for parts that are not currently possible.

  5. Standard Biological Parts Knowledgebase

    Science.gov (United States)

    Galdzicki, Michal; Rodriguez, Cesar; Chandran, Deepak; Sauro, Herbert M.; Gennari, John H.

    2011-01-01

    We have created the Knowledgebase of Standard Biological Parts (SBPkb) as a publically accessible Semantic Web resource for synthetic biology (sbolstandard.org). The SBPkb allows researchers to query and retrieve standard biological parts for research and use in synthetic biology. Its initial version includes all of the information about parts stored in the Registry of Standard Biological Parts (partsregistry.org). SBPkb transforms this information so that it is computable, using our semantic framework for synthetic biology parts. This framework, known as SBOL-semantic, was built as part of the Synthetic Biology Open Language (SBOL), a project of the Synthetic Biology Data Exchange Group. SBOL-semantic represents commonly used synthetic biology entities, and its purpose is to improve the distribution and exchange of descriptions of biological parts. In this paper, we describe the data, our methods for transformation to SBPkb, and finally, we demonstrate the value of our knowledgebase with a set of sample queries. We use RDF technology and SPARQL queries to retrieve candidate “promoter” parts that are known to be both negatively and positively regulated. This method provides new web based data access to perform searches for parts that are not currently possible. PMID:21390321

  6. Standard biological parts knowledgebase.

    Science.gov (United States)

    Galdzicki, Michal; Rodriguez, Cesar; Chandran, Deepak; Sauro, Herbert M; Gennari, John H

    2011-02-24

    We have created the Knowledgebase of Standard Biological Parts (SBPkb) as a publically accessible Semantic Web resource for synthetic biology (sbolstandard.org). The SBPkb allows researchers to query and retrieve standard biological parts for research and use in synthetic biology. Its initial version includes all of the information about parts stored in the Registry of Standard Biological Parts (partsregistry.org). SBPkb transforms this information so that it is computable, using our semantic framework for synthetic biology parts. This framework, known as SBOL-semantic, was built as part of the Synthetic Biology Open Language (SBOL), a project of the Synthetic Biology Data Exchange Group. SBOL-semantic represents commonly used synthetic biology entities, and its purpose is to improve the distribution and exchange of descriptions of biological parts. In this paper, we describe the data, our methods for transformation to SBPkb, and finally, we demonstrate the value of our knowledgebase with a set of sample queries. We use RDF technology and SPARQL queries to retrieve candidate "promoter" parts that are known to be both negatively and positively regulated. This method provides new web based data access to perform searches for parts that are not currently possible.

  7. WebDMS: A Web-Based Data Management System for Environmental Data

    Science.gov (United States)

    Ekstrand, A. L.; Haderman, M.; Chan, A.; Dye, T.; White, J. E.; Parajon, G.

    2015-12-01

    DMS is an environmental Data Management System to manage, quality-control (QC), summarize, document chain-of-custody, and disseminate data from networks ranging in size from a few sites to thousands of sites, instruments, and sensors. The server-client desktop version of DMS is used by local and regional air quality agencies (including the Bay Area Air Quality Management District, the South Coast Air Quality Management District, and the California Air Resources Board), the EPA's AirNow Program, and the EPA's AirNow-International (AirNow-I) program, which offers countries the ability to run an AirNow-like system. As AirNow's core data processing engine, DMS ingests, QCs, and stores real-time data from over 30,000 active sensors at over 5,280 air quality and meteorological sites from over 130 air quality agencies across the United States. As part of the AirNow-I program, several instances of DMS are deployed in China, Mexico, and Taiwan. The U.S. Department of State's StateAir Program also uses DMS for five regions in China and plans to expand to other countries in the future. Recent development has begun to migrate DMS from an onsite desktop application to WebDMS, a web-based application designed to take advantage of cloud hosting and computing services to increase scalability and lower costs. WebDMS will continue to provide easy-to-use data analysis tools, such as time-series graphs, scatterplots, and wind- or pollution-rose diagrams, as well as allowing data to be exported to external systems such as the EPA's Air Quality System (AQS). WebDMS will also provide new GIS analysis features and a suite of web services through a RESTful web API. These changes will better meet air agency needs and allow for broader national and international use (for example, by the AirNow-I partners). We will talk about the challenges and advantages of migrating DMS to the web, modernizing the DMS user interface, and making it more cost-effective to enhance and maintain over time.

  8. Web content a writer's guide

    CERN Document Server

    Mizrahi, Janet

    2013-01-01

    The explosion of electronic sources, whether in the form of news, commentary, sales and marketing, or information, has created boundless opportunities for producing content. Whether you're an entrepreneur with a start-up business who needs a website, an executive who uses social media to connect with various stakeholders, or a content provider blogging about topical issues, you'll need to know how to write for the web and address the unique environment of the digital world. This book will help you produce web content that generates results. Writing for the screen differs from writing for a pri

  9. Semantic Web technologies for the big data in life sciences.

    Science.gov (United States)

    Wu, Hongyan; Yamaguchi, Atsuko

    2014-08-01

    The life sciences field is entering an era of big data with the breakthroughs of science and technology. More and more big data-related projects and activities are being performed in the world. Life sciences data generated by new technologies are continuing to grow in not only size but also variety and complexity, with great speed. To ensure that big data has a major influence in the life sciences, comprehensive data analysis across multiple data sources and even across disciplines is indispensable. The increasing volume of data and the heterogeneous, complex varieties of data are two principal issues mainly discussed in life science informatics. The ever-evolving next-generation Web, characterized as the Semantic Web, is an extension of the current Web, aiming to provide information for not only humans but also computers to semantically process large-scale data. The paper presents a survey of big data in life sciences, big data related projects and Semantic Web technologies. The paper introduces the main Semantic Web technologies and their current situation, and provides a detailed analysis of how Semantic Web technologies address the heterogeneous variety of life sciences big data. The paper helps to understand the role of Semantic Web technologies in the big data era and how they provide a promising solution for the big data in life sciences.

  10. Evaluating Web-Based Learning Systems

    Science.gov (United States)

    Pergola, Teresa M.; Walters, L. Melissa

    2011-01-01

    Accounting educators continuously seek ways to effectively integrate instructional technology into accounting coursework as a means to facilitate active learning environments and address the technology-driven learning preferences of the current generation of students. Most accounting textbook publishers now provide interactive, web-based learning…

  11. Overcoming Legal Limitations in Disseminating Slovene Web Corpora

    Directory of Open Access Journals (Sweden)

    Tomaž Erjavec

    2016-09-01

    Full Text Available Web texts are becoming increasingly relevant sources of information, with web corpora useful for corpus linguistic studies and development of language technologies. Even though web texts are directly accessable, which substantially simplifies the collection procedure compilation of web corpora is still complex, time consuming and expensive. It is crucial that similar endeavours are not repeated, which is why it is necessary to make the created corpora easily and widely accessible both to researchers and a wider audience. While this is logistically and technically a straightforward procedure, legal constraints, such as copyright, privacy and terms of use severely hinder the dissemination of web corpora. This paper discusses legal conditions and actual practice in this area, gives an overview of current practices and proposes a range of mitigation measures on the example of the Janes corpus of Slovene user-generated content in order to ensure free and open dissemination of Slovene web corpora.

  12. Web Enabled DROLS Verity TopicSets

    National Research Council Canada - National Science Library

    Tong, Richard

    1999-01-01

    The focus of this effort has been the design and development of automatically generated TopicSets and HTML pages that provide the basis of the required search and browsing capability for DTIC's Web Enabled DROLS System...

  13. Web System for Data Quality Assessment of Tile Calorimeter During the ATLAS Operation

    International Nuclear Information System (INIS)

    Maidantchik, C; Ferreira, F; Grael, F; Sivolella, A; Balabram, L

    2011-01-01

    TileCal, the barrel hadronic calorimeter of the ATLAS experiment, gathers almost about 10,000 electronic channels. The supervision of the detector behavior is very important in order to ensure proper operation. Collaborators perform analysis over reconstructed data of calibration runs for giving detailed considerations about the equipment status. During the commissioning period, our group has developed seven web systems to support the data quality (DQ) assessment task. Each system covers a part of the process by providing information on the latest runs, displaying the DQ status from the monitoring framework, giving details about power supplies operation, presenting the generated plots and storing the validation outcomes, assisting to write logbook entries, creating and submitting the bad channels list to the conditions database and publishing the equipment performance history. The ATLAS operation increases amount of data that are retrieved, processed and stored by the web systems. In order to accomplish the new requirements, an optimized data model was designed to reduce the number of needed queries. The web systems were reassembled in a unique system in order to provide an integrated view of the validating process. The server load was minimized by using asynchronous requests from the browser.

  14. A Collaborative Writing Project Using the Worldwide Web.

    Science.gov (United States)

    Sylvester, Allen; Essex, Christopher

    A student in a distance education course, as part of a midterm project, set out to build a Web site that had written communication as its main focus. The Web site, "The Global Campfire," was modeled on the old Appalachian tradition of the "Story Tree," where a storyteller begins a story and allows group members to add to it.…

  15. Functional webs for freeform architecture

    KAUST Repository

    Deng, Bailin

    2011-08-01

    Rationalization and construction-aware design dominate the issue of realizability of freeform architecture. The former means the decomposition of an intended shape into parts which are sufficiently simple and efficient to manufacture; the latter refers to a design procedure which already incorporates rationalization. Recent contributions to this topic have been concerned mostly with small-scale parts, for instance with planar faces of meshes. The present paper deals with another important aspect, namely long-range parts and supporting structures. It turns out that from the pure geometry viewpoint this means studying families of curves which cover surfaces in certain well-defined ways. Depending on the application one has in mind, different combinatorial arrangements of curves are required. We here restrict ourselves to so-called hexagonal webs which correspond to a triangular or tri-hex decomposition of a surface. The individual curve may have certain special properties, like being planar, being a geodesic, or being part of a circle. Each of these properties is motivated by manufacturability considerations and imposes constraints on the shape of the surface. We investigate the available degrees of freedom, show numerical methods of optimization, and demonstrate the effectivity of our approach and the variability of construction solutions derived from webs by means of actual architectural designs.

  16. Gaining insight into food webs reconstructed by the inverse method

    NARCIS (Netherlands)

    Kones, J.; Soetaert, K.E.R.; Van Oevelen, D.; Owino, J.; Mavuti, K.

    2006-01-01

    The use of the inverse method to analyze flow patterns of organic components in ecological systems has had wide application in ecological modeling. Through this approach, an infinite number of food web flows describing the food web and satisfying biological constraints are generated, from which one

  17. AdaFF: Adaptive Failure-Handling Framework for Composite Web Services

    Science.gov (United States)

    Kim, Yuna; Lee, Wan Yeon; Kim, Kyong Hoon; Kim, Jong

    In this paper, we propose a novel Web service composition framework which dynamically accommodates various failure recovery requirements. In the proposed framework called Adaptive Failure-handling Framework (AdaFF), failure-handling submodules are prepared during the design of a composite service, and some of them are systematically selected and automatically combined with the composite Web service at service instantiation in accordance with the requirement of individual users. In contrast, existing frameworks cannot adapt the failure-handling behaviors to user's requirements. AdaFF rapidly delivers a composite service supporting the requirement-matched failure handling without manual development, and contributes to a flexible composite Web service design in that service architects never care about failure handling or variable requirements of users. For proof of concept, we implement a prototype system of the AdaFF, which automatically generates a composite service instance with Web Services Business Process Execution Language (WS-BPEL) according to the users' requirement specified in XML format and executes the generated instance on the ActiveBPEL engine.

  18. Generational Influences in Academic Emergency Medicine: Teaching and Learning, Mentoring, and Technology (Part I)

    Science.gov (United States)

    Mohr, Nicholas M.; Moreno-Walton, Lisa; Mills, Angela M.; Brunett, Patrick H.; Promes, Susan B.

    2010-01-01

    For the first time in history, four generations are working together – Traditionalists, Baby Boomers, Generation Xers, and Millennials. Members of each generation carry with them a unique perspective of the world and interact differently with those around them. Through a review of the literature and consensus by modified Delphi methodology of the Society for Academic Emergency Medicine (SAEM) Aging and Generational Issues Task Force, the authors have developed this two-part series to address generational issues present in academic emergency medicine (EM). Understanding generational characteristics and mitigating strategies can help address some common issues encountered in academic EM. Through recognition of the unique characteristics of each of the generations with respect to teaching and learning, mentoring, and technology, academicians have the opportunity to strategically optimize interactions with one another. PMID:21314779

  19. New tool to assemble repetitive regions using next-generation sequencing data

    Science.gov (United States)

    Kuśmirek, Wiktor; Nowak, Robert M.; Neumann, Łukasz

    2017-08-01

    The next generation sequencing techniques produce a large amount of sequencing data. Some part of the genome are composed of repetitive DNA sequences, which are very problematic for the existing genome assemblers. We propose a modification of the algorithm for a DNA assembly, which uses the relative frequency of reads to properly reconstruct repetitive sequences. The new approach was implemented and tested, as a demonstration of the capability of our software we present some results for model organisms. The new implementation, using a three-layer software architecture was selected, where the presentation layer, data processing layer, and data storage layer were kept separate. Source code as well as demo application with web interface and the additional data are available at project web-page: http://dnaasm.sourceforge.net.

  20. Utilizing Web 2.0 Technologies for Library Web Tutorials: An Examination of Instruction on Community College Libraries' Websites Serving Large Student Bodies

    Science.gov (United States)

    Blummer, Barbara; Kenton, Jeffrey M.

    2015-01-01

    This is the second part of a series on Web 2.0 tools available from community college libraries' Websites. The first article appeared in an earlier volume of this journal and it illustrated the wide variety of Web 2.0 tools on community college libraries' Websites serving large student bodies (Blummer and Kenton 2014). The research found many of…

  1. ANALISIS DEL COMPORTAMIENTO DEL USUARIO WEB

    OpenAIRE

    ROMAN ASENJO, PABLO ENRIQUE; ROMAN ASENJO, PABLO ENRIQUE

    2011-01-01

    Desde los orígenes de la Web en el CERN, ha existido una pregunta recurrente entre los investigadores y desarrolladores: ¿Cual es la estructura y contenido correcto para que un sitio Web atraiga y/o retenga a sus visitantes? En parte, la respuesta a esta interrogante, se encuentra fuertemente relacionada con una mayor comprensión de las motivaciones que posee un usuario al visitar un sitio. Tradicionalmente, se han utilizado algoritmos de minería de datos (Machine Learning) para extraer...

  2. ANALISIS DEL COMPORTAMIENTO DEL USUARIO WEB

    OpenAIRE

    ROMAN ASENJO, PABLO ENRIQUE

    2011-01-01

    Desde los orígenes de la Web en el CERN, ha existido una pregunta recurrente entre los investigadores y desarrolladores: ¿Cual es la estructura y contenido correcto para que un sitio Web atraiga y/o retenga a sus visitantes? En parte, la respuesta a esta inteJTogante, se encuentra fuertemente relacionada con una mayor comprensión de las motivaciones que posee un usuario al visitar un sitio. Tradicionalmente, se han utilizado algoritmos de minería de datos (Machine Learning) para extraer pa...

  3. Scientific Workflows and the Sensor Web for Virtual Environmental Observatories

    Science.gov (United States)

    Simonis, I.; Vahed, A.

    2008-12-01

    interfaces. All data sets and sensor communication follow well-defined abstract models and corresponding encodings, mostly developed by the OGC Sensor Web Enablement initiative. Scientific progress is currently accelerated by an emerging new concept called scientific workflows, which organize and manage complex distributed computations. A scientific workflow represents and records the highly complex processes that a domain scientist typically would follow in exploration, discovery and ultimately, transformation of raw data to publishable results. The challenge is now to integrate the benefits of scientific workflows with those provided by the Sensor Web in order to leverage all resources for scientific exploration, problem solving, and knowledge generation. Scientific workflows for the Sensor Web represent the next evolutionary step towards efficient, powerful, and flexible earth observation frameworks and platforms. Those platforms support the entire process from capturing data, sharing and integrating, to requesting additional observations. Multiple sites and organizations will participate on single platforms and scientists from different countries and organizations interact and contribute to large-scale research projects. Simultaneously, the data- and information overload becomes manageable, as multiple layers of abstraction will free scientists to deal with underlying data-, processing or storage peculiarities. The vision are automated investigation and discovery mechanisms that allow scientists to pose queries to the system, which in turn would identify potentially related resources, schedules processing tasks and assembles all parts in workflows that may satisfy the query.

  4. WebAssign: Assessing Your Students' Understanding Continuously

    Science.gov (United States)

    Risley, John S.

    1999-11-01

    Motivating students to learn is a constant challenge for faculty. Technology can play a significant role. One such solution is WebAssign — a web-based homework system that offers new teaching and learning opportunities for educators and their students. WebAssign delivers, collects, grades, and records customized homework assignments over the Internet. Students get immediate feedback with credit and instructors can implement "Just-in-Time" teaching. In this talk, I will describe how assignments can be generated with different numerical values for each question, giving each student a unique problem to solve. This feature encourages independent thinking with the benefit of collaborative learning. Example assignments taken from textbook questions and intellectually engaging Java applet simulations will be shown. Studies and first-hand experience on the educational impact of using WebAssign will also be discussed.

  5. Open Source Web Based Geospatial Processing with OMAR

    Directory of Open Access Journals (Sweden)

    Mark Lucas

    2009-01-01

    Full Text Available The availability of geospatial data sets is exploding. New satellites, aerial platforms, video feeds, global positioning system tagged digital photos, and traditional GIS information are dramatically increasing across the globe. These raw materials need to be dynamically processed, combined and correlated to generate value added information products to answer a wide range of questions. This article provides an overview of OMAR web based geospatial processing. OMAR is part of the Open Source Software Image Map project under the Open Source Geospatial Foundation. The primary contributors of OSSIM make their livings by providing professional services to US Government agencies and programs. OMAR provides one example that open source software solutions are increasingly being deployed in US government agencies. We will also summarize the capabilities of OMAR and its plans for near term development.

  6. Plankton food-webs: to what extent can they be simplified?

    Directory of Open Access Journals (Sweden)

    Domenico D'Alelio

    2016-05-01

    Full Text Available Plankton is a hugely diverse community including both unicellular and multicellular organisms, whose individual dimensions span over seven orders of magnitude. Plankton is a fundamental part of biogeochemical cycles and food-webs in aquatic systems. While knowledge has progressively accumulated at the level of single species and single trophic processes, the overwhelming biological diversity of plankton interactions is insufficiently known and a coherent and unifying trophic framework is virtually lacking. We performed an extensive review of the plankton literature to provide a compilation of data suitable for implementing food-web models including plankton trophic processes at high taxonomic resolution. We identified the components of the plankton community at the Long Term Ecological Research Station MareChiara in the Gulf of Naples. These components represented the sixty-three nodes of a plankton food-web. To each node we attributed biomass and vital rates, i.e. production, consumption, assimilation rates and ratio between autotrophy and heterotrophy in mixotrophic protists. Biomasses and rates values were defined for two opposite system’s conditions; relatively eutrophic and oligotrophic states. We finally identified 817 possible trophic links within the web and provided each of them with a relative weight, in order to define a diet-matrix, valid for both trophic states, which included all consumers, fromn anoflagellates to carnivorous plankton. Vital rates for plankton resulted, as expected, very wide; this strongly contrasts with the narrow ranges considered in plankton system models implemented so far. Moreover, the amount and variety of trophic links highlighted by our review is largely excluded by state-of-the-art biogeochemical and food-web models for aquatic systems. Plankton models could potentially benefit from the integration of the trophic diversity outlined in this paper: first, by using more realistic rates; second, by better

  7. Realization of a system to demand and acquire information via world wide web; Realizzazione di un sistema per l'accesso e l'acquisizione di informazioni via web in un ente complesso

    Energy Technology Data Exchange (ETDEWEB)

    Bongiovanni, G. [Rome Univ. La Sapienza, Rome (Italy); Di Marco, R.A. [ENEA, Sede Centrale, Rome (Italy). Funzione Centrale Informatica; Cappitelli, A.

    1999-07-01

    The project realized within this thesis exactly regards interactive web and shows how it is possible to obtain particular functionalities opportunely fixed tools. Such project consists in the realization of a system which allows users with opportune account to demand and to acquire information, and to manage these users and a part of the informative system by whom holds administration tasks. Developed software contains an experimental part and a section dedicated to the public key cryptography, which has been employed to carry out sure transactions via web. [Italian] Il rapporto descrive un progetto di realizzazione di un web interattivo e mostra come sia possibile ottenere particolari funzionalita' impiegando opportunamente determinati strumenti come Java, specifica CGI. Il sistema permette agli utenti, tramite opportuno account, di richiedere e acquisire informazioni. Nel software sviluppato e' presente una parte sperimentale e una parte dedicata alla crittografia a chiave pubblica, impiegata per effettuare transazioni sicure via web.

  8. Beginning ASPNET Web Pages with WebMatrix

    CERN Document Server

    Brind, Mike

    2011-01-01

    Learn to build dynamic web sites with Microsoft WebMatrix Microsoft WebMatrix is designed to make developing dynamic ASP.NET web sites much easier. This complete Wrox guide shows you what it is, how it works, and how to get the best from it right away. It covers all the basic foundations and also introduces HTML, CSS, and Ajax using jQuery, giving beginning programmers a firm foundation for building dynamic web sites.Examines how WebMatrix is expected to become the new recommended entry-level tool for developing web sites using ASP.NETArms beginning programmers, students, and educators with al

  9. The Social Web and Learning

    NARCIS (Netherlands)

    dr. Guus Wijngaards

    2008-01-01

    On the internet we see a continuously growing generation of web applications enabling anyone to create and publish online content in a simple way, to link content and to share it with others: wellknown instances include MySpace, Facebook, YouTube, Flickr, Wikipedia and Google Earth. The internet has

  10. A new generation of tools for search, recovery and quality evaluation of World Wide Web medical resources.

    Science.gov (United States)

    Aguillo, I

    2000-01-01

    Although the Internet is already a valuable information resource in medicine, there are important challenges to be faced before physicians and general users will have extensive access to this information. As a result of a research effort to compile a health-related Internet directory, new tools and strategies have been developed to solve key problems derived from the explosive growth of medical information on the Net and the great concern over the quality of such critical information. The current Internet search engines lack some important capabilities. We suggest using second generation tools (client-side based) able to deal with large quantities of data and to increase the usability of the records recovered. We tested the capabilities of these programs to solve health-related information problems, recognising six groups according to the kind of topics addressed: Z39.50 clients, downloaders, multisearchers, tracing agents, indexers and mappers. The evaluation of the quality of health information available on the Internet could require a large amount of human effort. A possible solution may be to use quantitative indicators based on the hypertext visibility of the Web sites. The cybermetric measures are valid for quality evaluation if they are derived from indirect peer review by experts with Web pages citing the site. The hypertext links acting as citations need to be extracted from a controlled sample of quality super-sites.

  11. Non-visual Web Browsing: Beyond Web Accessibility.

    Science.gov (United States)

    Ramakrishnan, I V; Ashok, Vikas; Billah, Syed Masum

    2017-07-01

    People with vision impairments typically use screen readers to browse the Web. To facilitate non-visual browsing, web sites must be made accessible to screen readers, i.e., all the visible elements in the web site must be readable by the screen reader. But even if web sites are accessible, screen-reader users may not find them easy to use and/or easy to navigate. For example, they may not be able to locate the desired information without having to listen to a lot of irrelevant contents. These issues go beyond web accessibility and directly impact web usability. Several techniques have been reported in the accessibility literature for making the Web usable for screen reading. This paper is a review of these techniques. Interestingly, the review reveals that understanding the semantics of the web content is the overarching theme that drives these techniques for improving web usability.

  12. Generational influences in academic emergency medicine: teaching and learning, mentoring, and technology (part I).

    Science.gov (United States)

    Mohr, Nicholas M; Moreno-Walton, Lisa; Mills, Angela M; Brunett, Patrick H; Promes, Susan B

    2011-02-01

    For the first time in history, four generations are working together-traditionalists, baby boomers, generation Xers (Gen Xers), and millennials. Members of each generation carry with them a unique perspective of the world and interact differently with those around them. Through a review of the literature and consensus by modified Delphi methodology of the Society for Academic Emergency Medicine Aging and Generational Issues Task Force, the authors have developed this two-part series to address generational issues present in academic emergency medicine (EM). Understanding generational characteristics and mitigating strategies can help address some common issues encountered in academic EM. Through recognition of the unique characteristics of each of the generations with respect to teaching and learning, mentoring, and technology, academicians have the opportunity to strategically optimize interactions with one another. © 2011 by the Society for Academic Emergency Medicine.

  13. Understanding the Web from an Economic Perspective: The Evolution of Business Models and the Web

    Directory of Open Access Journals (Sweden)

    Louis Rinfret

    2014-08-01

    Full Text Available The advent of the World Wide Web is arguably amongst the most important changes that have occurred since the 1990s in the business landscape. It has fueled the rise of new industries, supported the convergence and reshaping of existing ones and enabled the development of new business models. During this time the web has evolved tremendously from a relatively static pagedisplay tool to a massive network of user-generated content, collective intelligence, applications and hypermedia. As technical standards continue to evolve, business models catch-up to the new capabilities. New ways of creating value, distributing it and profiting from it emerge more rapidly than ever. In this paper we explore how the World Wide Web and business models evolve and we identify avenues for future research in light of the web‟s ever-evolving nature and its influence on business models.

  14. Importancia y situación actual de la accesibilidad web para el turismo accesible

    Directory of Open Access Journals (Sweden)

    Gabriel Fontanet Nadal

    2011-04-01

    Full Text Available Accesible Tourism is a kind of Tourism that is specially dedicated to disabled people. This Tourism refers to the removal of physical elements that difficult the disabled people mobility at the destination. The Accesible Tourism should take care of both physical and web accessibility. The Web Accessibility of a web is defined as the capability this web to be accessed by people with any kind of disability. Some organizations generate rules to improve web accessibility. An analysis of Web Accessibility in Tourist Web Sites is shown at this document.

  15. Bibliographic information organization in the semantic web

    CERN Document Server

    Willer, Mirna

    2013-01-01

    New technologies will underpin the future generation of library catalogues. To facilitate their role providing information, serving users, and fulfilling their mission as cultural heritage and memory institutions, libraries must take a technological leap; their standards and services must be transformed to those of the Semantic Web. Bibliographic Information Organization in the Semantic Web explores the technologies that may power future library catalogues, and argues the necessity of such a leap. The text introduces international bibliographic standards and models, and fundamental concepts in

  16. Chemical cleaning as an essential part of steam generator asset management

    International Nuclear Information System (INIS)

    Stiepani, C.; Ammann, F.; Jones, D.; Evans, S.; Harper, K.

    2010-01-01

    Accumulation of deposits is intrinsic for the operation of Steam Generators in PWRs. Such depositions often lead to reduction of thermal performance, loss of component integrity and, in some cases to power restrictions. Accordingly removal of such deposits is an essential part of the asset management of the Steam Generators in a Nuclear Power Plant. Every plant has its individual condition, history and constraints which need to be considered when planning and performing a chemical cleaning. Typical points are: Sludge load amount and constitution of the deposits; Sludge distribution in the steam generator; Existing or expected corrosion problems; Amount and treatment possibilities for the waste generated. Depending on these points the strategy for chemical cleaning shall be evolved. The range of treatment starts with very soft cleanings with a removal of approx 100 kg per steam generator and goes to a full scale cleaning which can remove up to several thousand kilograms of deposits from a steam generator. Depending on the goal to be achieved and the steam generator present an adequate cleaning method shall be selected. Flexible and 'customizable' cleaning methods that can be adapted to the individual needs of a plant are therefore a must. Particular for the application of preventive cleanings where repeated or even regular application are intended, special focus has to be put on low corrosion and easy waste handling. Therefore AREVA has developed the 'C3' concept, Customized Chemical Cleaning concept. This concept covers the entire range of steam generator cleaning. Particular for the preventive maintenance cleanings processes with extreme low corrosion rates and easy waste handling are provided which make repeated applications safe and cost efficient. (author)

  17. DIRAC: Secure web user interface

    International Nuclear Information System (INIS)

    Casajus Ramo, A; Sapunov, M

    2010-01-01

    Traditionally the interaction between users and the Grid is done with command line tools. However, these tools are difficult to use by non-expert users providing minimal help and generating outputs not always easy to understand especially in case of errors. Graphical User Interfaces are typically limited to providing access to the monitoring or accounting information and concentrate on some particular aspects failing to cover the full spectrum of grid control tasks. To make the Grid more user friendly more complete graphical interfaces are needed. Within the DIRAC project we have attempted to construct a Web based User Interface that provides means not only for monitoring the system behavior but also allows to steer the main user activities on the grid. Using DIRAC's web interface a user can easily track jobs and data. It provides access to job information and allows performing actions on jobs such as killing or deleting. Data managers can define and monitor file transfer activity as well as check requests set by jobs. Production managers can define and follow large data productions and react if necessary by stopping or starting them. The Web Portal is build following all the grid security standards and using modern Web 2.0 technologies which allow to achieve the user experience similar to the desktop applications. Details of the DIRAC Web Portal architecture and User Interface will be presented and discussed.

  18. The Concept of “Web Science” in the Social Realm: Building Bridges between a new Interdisciplinary Field and the Cultural “Wealth of Networks”

    Directory of Open Access Journals (Sweden)

    Michael Dick

    2010-06-01

    Full Text Available Discourse in the field of “cyber culture” largely does not take into account the major shift in constituent technology that has begun to advance the Web from one based solely on human-understandable hypertext documents to one based on machine-understandable data. Such innovation includes the refinement of new search engine technology to mine data in Web services applications (the “Deep Web” coupled with the desire to annotate data with mark-up languages that facilitate greater interactivity and infer meaning within either user-created knowledge representation models (“folksonomies” as a part of “Web 2.0” or more rigid ontological structures (part of the “Semantic Web” or “Web 3.0”. In this paper, I consider this overall evident and predicted shift from a “Web of documents” to a “Web of data” to be the central element in the creation of the next-generation of the Web, and the recent drive to study it within an integrated framework known as “Web Science”. Central to this shift is the need to reconsider not only the cultural aspects of the medium, but also the interactions between cultural theory and technical texts. I conclude that with the emergence of certain new technology the entire concept of intellectual property, and more specifically where value ultimately lies in terms of the creation of cultural product, is also changing. Within, I thus focus on alternative frameworks (namely the work of Yochai Benkler to conceptualize knowledge production, in order to re-examine issues of Web-enabled participatory culture. In order to highlight new cultural paradigms, opportunities and challenges, I discuss how the concept of “social production” may foster a “cultural democracy” that transcends traditional hegemonic conditions that encumber publics.

  19. Web 2.0 Technologies: Applications for Community Colleges

    Science.gov (United States)

    Bajt, Susanne K.

    2011-01-01

    The current generation of new students, referred to as the Millennial Generation, brings a new set of challenges to the community college. The influx of these technologically sophisticated students, who interact through the social phenomenon of Web 2.0 technology, bring expectations that may reshape institutions of higher learning. This chapter…

  20. The role of ciliates within the microbial food web in the eutrophicated part of Kaštela Bay (Middle Adriatic Sea

    Directory of Open Access Journals (Sweden)

    Natalia Bojanic

    2006-09-01

    Full Text Available Interactions among phytoplankton, bacterioplankton, heterotrophic nanoflagellates (HNF, ciliated protozoa and copepod nauplii were studied in the eutrophicated part of Kas?tela Bay from May 1998 to November 1999. Special emphasis was placed on relationships between size categories of nonloricate ciliates (NLC and other microbial food web components. Biomasses of phytoplankton and bacteria were primarily influenced by abiotic parameters. Temperature indirectly controlled variation in HNF biomass through the changes in biomass of bacteria and the smaller phytoplankton fraction. Besides HNF, bacterial biomass was affected by the NLC

  1. Bringing Web 2.0 to bioinformatics.

    Science.gov (United States)

    Zhang, Zhang; Cheung, Kei-Hoi; Townsend, Jeffrey P

    2009-01-01

    Enabling deft data integration from numerous, voluminous and heterogeneous data sources is a major bioinformatic challenge. Several approaches have been proposed to address this challenge, including data warehousing and federated databasing. Yet despite the rise of these approaches, integration of data from multiple sources remains problematic and toilsome. These two approaches follow a user-to-computer communication model for data exchange, and do not facilitate a broader concept of data sharing or collaboration among users. In this report, we discuss the potential of Web 2.0 technologies to transcend this model and enhance bioinformatics research. We propose a Web 2.0-based Scientific Social Community (SSC) model for the implementation of these technologies. By establishing a social, collective and collaborative platform for data creation, sharing and integration, we promote a web services-based pipeline featuring web services for computer-to-computer data exchange as users add value. This pipeline aims to simplify data integration and creation, to realize automatic analysis, and to facilitate reuse and sharing of data. SSC can foster collaboration and harness collective intelligence to create and discover new knowledge. In addition to its research potential, we also describe its potential role as an e-learning platform in education. We discuss lessons from information technology, predict the next generation of Web (Web 3.0), and describe its potential impact on the future of bioinformatics studies.

  2. That obscure object of desire: multimedia metadata on the Web, part 2

    NARCIS (Netherlands)

    F.-M. Nack (Frank); J.R. van Ossenbruggen (Jacco); L. Hardman (Lynda)

    2003-01-01

    textabstractThis article discusses the state of the art in metadata for audio-visual media in large semantic networks, such as the Semantic Web. Our discussion is predominantly motivated by the two most widely known approaches towards machine-processable and semantic-based content description,

  3. That obscure object of desire: multimedia metadata on the Web, part 1

    NARCIS (Netherlands)

    F.-M. Nack (Frank); J.R. van Ossenbruggen (Jacco); L. Hardman (Lynda)

    2003-01-01

    textabstractThis article discusses the state of the art in metadata for audio-visual media in large semantic networks, such as the Semantic Web. Our discussion is predominantly motivated by the two most widely known approaches towards machine-processable and semantic-based content description,

  4. Web server for priority ordered multimedia services

    Science.gov (United States)

    Celenk, Mehmet; Godavari, Rakesh K.; Vetnes, Vermund

    2001-10-01

    In this work, our aim is to provide finer priority levels in the design of a general-purpose Web multimedia server with provisions of the CM services. The type of services provided include reading/writing a web page, downloading/uploading an audio/video stream, navigating the Web through browsing, and interactive video teleconferencing. The selected priority encoding levels for such operations follow the order of admin read/write, hot page CM and Web multicasting, CM read, Web read, CM write and Web write. Hot pages are the most requested CM streams (e.g., the newest movies, video clips, and HDTV channels) and Web pages (e.g., portal pages of the commercial Internet search engines). Maintaining a list of these hot Web pages and CM streams in a content addressable buffer enables a server to multicast hot streams with lower latency and higher system throughput. Cold Web pages and CM streams are treated as regular Web and CM requests. Interactive CM operations such as pause (P), resume (R), fast-forward (FF), and rewind (RW) have to be executed without allocation of extra resources. The proposed multimedia server model is a part of the distributed network with load balancing schedulers. The SM is connected to an integrated disk scheduler (IDS), which supervises an allocated disk manager. The IDS follows the same priority handling as the SM, and implements a SCAN disk-scheduling method for an improved disk access and a higher throughput. Different disks are used for the Web and CM services in order to meet the QoS requirements of CM services. The IDS ouput is forwarded to an Integrated Transmission Scheduler (ITS). The ITS creates a priority ordered buffering of the retrieved Web pages and CM data streams that are fed into an auto regressive moving average (ARMA) based traffic shaping circuitry before being transmitted through the network.

  5. Maintaining Web Cache Coherency

    Directory of Open Access Journals (Sweden)

    2000-01-01

    Full Text Available Document coherency is a challenging problem for Web caching. Once the documents are cached throughout the Internet, it is often difficult to keep them coherent with the origin document without generating a new traffic that could increase the traffic on the international backbone and overload the popular servers. Several solutions have been proposed to solve this problem, among them two categories have been widely discussed: the strong document coherency and the weak document coherency. The cost and the efficiency of the two categories are still a controversial issue, while in some studies the strong coherency is far too expensive to be used in the Web context, in other studies it could be maintained at a low cost. The accuracy of these analysis is depending very much on how the document updating process is approximated. In this study, we compare some of the coherence methods proposed for Web caching. Among other points, we study the side effects of these methods on the Internet traffic. The ultimate goal is to study the cache behavior under several conditions, which will cover some of the factors that play an important role in the Web cache performance evaluation and quantify their impact on the simulation accuracy. The results presented in this study show indeed some differences in the outcome of the simulation of a Web cache depending on the workload being used, and the probability distribution used to approximate updates on the cached documents. Each experiment shows two case studies that outline the impact of the considered parameter on the performance of the cache.

  6. Web Caching

    Indian Academy of Sciences (India)

    leveraged through Web caching technology. Specifically, Web caching becomes an ... Web routing can improve the overall performance of the Internet. Web caching is similar to memory system caching - a Web cache stores Web resources in ...

  7. Customisable Scientific Web Portal for Fusion Research

    Energy Technology Data Exchange (ETDEWEB)

    Abla, G; Kim, E; Schissel, D; Flannagan, S [General Atomics, San Diego (United States)

    2009-07-01

    The Web browser has become one of the major application interfaces for remotely participating in magnetic fusion. Web portals are used to present very diverse sources of information in a unified way. While a web portal has several benefits over other software interfaces, such as providing single point of access for multiple computational services, and eliminating the need for client software installation, the design and development of a web portal has unique challenges. One of the challenges is that a web portal needs to be fast and interactive despite a high volume of tools and information that it presents. Another challenge is the visual output on the web portal often is overwhelming due to the high volume of data generated by complex scientific instruments and experiments; therefore the applications and information should be customizable depending on the needs of users. An appropriate software architecture and web technologies can meet these problems. A web-portal has been designed to support the experimental activities of DIII-D researchers worldwide. It utilizes a multi-tier software architecture, and web 2.0 technologies, such as AJAX, Django, and Memcached, to develop a highly interactive and customizable user interface. It offers a customizable interface with personalized page layouts and list of services for users to select. Customizable services are: real-time experiment status monitoring, diagnostic data access, interactive data visualization. The web-portal also supports interactive collaborations by providing collaborative logbook, shared visualization and online instant message services. Furthermore, the web portal will provide a mechanism to allow users to create their own applications on the web portal as well as bridging capabilities to external applications such as Twitter and other social networks. In this series of slides, we describe the software architecture of this scientific web portal and our experiences in utilizing web 2.0 technologies. A

  8. Web Program for Development of GUIs for Cluster Computers

    Science.gov (United States)

    Czikmantory, Akos; Cwik, Thomas; Klimeck, Gerhard; Hua, Hook; Oyafuso, Fabiano; Vinyard, Edward

    2003-01-01

    WIGLAF (a Web Interface Generator and Legacy Application Facade) is a computer program that provides a Web-based, distributed, graphical-user-interface (GUI) framework that can be adapted to any of a broad range of application programs, written in any programming language, that are executed remotely on any cluster computer system. WIGLAF enables the rapid development of a GUI for controlling and monitoring a specific application program running on the cluster and for transferring data to and from the application program. The only prerequisite for the execution of WIGLAF is a Web-browser program on a user's personal computer connected with the cluster via the Internet. WIGLAF has a client/server architecture: The server component is executed on the cluster system, where it controls the application program and serves data to the client component. The client component is an applet that runs in the Web browser. WIGLAF utilizes the Extensible Markup Language to hold all data associated with the application software, Java to enable platform-independent execution on the cluster system and the display of a GUI generator through the browser, and the Java Remote Method Invocation software package to provide simple, effective client/server networking.

  9. Web tools for predictive toxicology model building.

    Science.gov (United States)

    Jeliazkova, Nina

    2012-07-01

    The development and use of web tools in chemistry has accumulated more than 15 years of history already. Powered by the advances in the Internet technologies, the current generation of web systems are starting to expand into areas, traditional for desktop applications. The web platforms integrate data storage, cheminformatics and data analysis tools. The ease of use and the collaborative potential of the web is compelling, despite the challenges. The topic of this review is a set of recently published web tools that facilitate predictive toxicology model building. The focus is on software platforms, offering web access to chemical structure-based methods, although some of the frameworks could also provide bioinformatics or hybrid data analysis functionalities. A number of historical and current developments are cited. In order to provide comparable assessment, the following characteristics are considered: support for workflows, descriptor calculations, visualization, modeling algorithms, data management and data sharing capabilities, availability of GUI or programmatic access and implementation details. The success of the Web is largely due to its highly decentralized, yet sufficiently interoperable model for information access. The expected future convergence between cheminformatics and bioinformatics databases provides new challenges toward management and analysis of large data sets. The web tools in predictive toxicology will likely continue to evolve toward the right mix of flexibility, performance, scalability, interoperability, sets of unique features offered, friendly user interfaces, programmatic access for advanced users, platform independence, results reproducibility, curation and crowdsourcing utilities, collaborative sharing and secure access.

  10. Graph-Based Semantic Web Service Composition for Healthcare Data Integration.

    Science.gov (United States)

    Arch-Int, Ngamnij; Arch-Int, Somjit; Sonsilphong, Suphachoke; Wanchai, Paweena

    2017-01-01

    Within the numerous and heterogeneous web services offered through different sources, automatic web services composition is the most convenient method for building complex business processes that permit invocation of multiple existing atomic services. The current solutions in functional web services composition lack autonomous queries of semantic matches within the parameters of web services, which are necessary in the composition of large-scale related services. In this paper, we propose a graph-based Semantic Web Services composition system consisting of two subsystems: management time and run time. The management-time subsystem is responsible for dependency graph preparation in which a dependency graph of related services is generated automatically according to the proposed semantic matchmaking rules. The run-time subsystem is responsible for discovering the potential web services and nonredundant web services composition of a user's query using a graph-based searching algorithm. The proposed approach was applied to healthcare data integration in different health organizations and was evaluated according to two aspects: execution time measurement and correctness measurement.

  11. WebSelF: A Web Scraping Framework

    DEFF Research Database (Denmark)

    Thomsen, Jakob; Ernst, Erik; Brabrand, Claus

    2012-01-01

    We present, WebSelF, a framework for web scraping which models the process of web scraping and decomposes it into four conceptually independent, reusable, and composable constituents. We have validated our framework through a full parameterized implementation that is flexible enough to capture...... previous work on web scraping. We have experimentally evaluated our framework and implementation in an experiment that evaluated several qualitatively different web scraping constituents (including previous work and combinations hereof) on about 11,000 HTML pages on daily versions of 17 web sites over...... a period of more than one year. Our framework solves three concrete problems with current web scraping and our experimental results indicate that com- position of previous and our new techniques achieve a higher degree of accuracy, precision and specificity than existing techniques alone....

  12. Customizable scientific web portal for fusion research

    International Nuclear Information System (INIS)

    Abla, G.; Kim, E.N.; Schissel, D.P.; Flanagan, S.M.

    2010-01-01

    Web browsers have become a major application interface for participating in scientific experiments such as those in magnetic fusion. The recent advances in web technologies motivated the deployment of interactive web applications with rich features. In the scientific world, web applications have been deployed in portal environments. When used in a scientific research environment, such as fusion experiments, web portals can present diverse sources of information in a unified interface. However, the design and development of a scientific web portal has its own challenges. One such challenge is that a web portal needs to be fast and interactive despite the high volume of information and number of tools it presents. Another challenge is that the visual output of the web portal must not be overwhelming to the end users, despite the high volume of data generated by fusion experiments. Therefore, the applications and information should be customizable depending on the needs of end users. In order to meet these challenges, the design and implementation of a web portal needs to support high interactivity and user customization. A web portal has been designed to support the experimental activities of DIII-D researchers worldwide by providing multiple services, such as real-time experiment status monitoring, diagnostic data access and interactive data visualization. The web portal also supports interactive collaborations by providing a collaborative logbook, shared visualization and online instant messaging services. The portal's design utilizes the multi-tier software architecture and has been implemented utilizing web 2.0 technologies, such as AJAX, Django, and Memcached, to develop a highly interactive and customizable user interface. It offers a customizable interface with personalized page layouts and list of services, which allows users to create a unique, personalized working environment to fit their own needs and interests. This paper describes the software

  13. Customizable scientific web portal for fusion research

    Energy Technology Data Exchange (ETDEWEB)

    Abla, G., E-mail: abla@fusion.gat.co [General Atomics, P.O. Box 85608, San Diego, CA (United States); Kim, E.N.; Schissel, D.P.; Flanagan, S.M. [General Atomics, P.O. Box 85608, San Diego, CA (United States)

    2010-07-15

    Web browsers have become a major application interface for participating in scientific experiments such as those in magnetic fusion. The recent advances in web technologies motivated the deployment of interactive web applications with rich features. In the scientific world, web applications have been deployed in portal environments. When used in a scientific research environment, such as fusion experiments, web portals can present diverse sources of information in a unified interface. However, the design and development of a scientific web portal has its own challenges. One such challenge is that a web portal needs to be fast and interactive despite the high volume of information and number of tools it presents. Another challenge is that the visual output of the web portal must not be overwhelming to the end users, despite the high volume of data generated by fusion experiments. Therefore, the applications and information should be customizable depending on the needs of end users. In order to meet these challenges, the design and implementation of a web portal needs to support high interactivity and user customization. A web portal has been designed to support the experimental activities of DIII-D researchers worldwide by providing multiple services, such as real-time experiment status monitoring, diagnostic data access and interactive data visualization. The web portal also supports interactive collaborations by providing a collaborative logbook, shared visualization and online instant messaging services. The portal's design utilizes the multi-tier software architecture and has been implemented utilizing web 2.0 technologies, such as AJAX, Django, and Memcached, to develop a highly interactive and customizable user interface. It offers a customizable interface with personalized page layouts and list of services, which allows users to create a unique, personalized working environment to fit their own needs and interests. This paper describes the software

  14. Web 2.0: Inherent tensions and evident challenges for education

    DEFF Research Database (Denmark)

    Dohn, Nina Bonderup

    2009-01-01

    In upper tertiary educational programmes around the world, the new Web-mediated communication practices termed Web 2.0 are introduced as learning activities with the goal of facilitating learning through collaborative knowledge construction. The aim of this paper is to point to discrepancies...... in the views of learning, knowledge, and the goals of the practice implicit in Web 2.0 and educational practices and to argue that these discrepancies lead to theoretical tensions and practical challenges when Web 2.0 practices are utilized for educational purposes. The article is structured into four main...... parts: First, Web 2.0 is characterized from a practice perspective. Second, some conceptual discrepancies between the "practice logics" of Web 2.0 and educational practices are identified. Third, the question of transcending the discrepancies is raised through a discussion of related pedagogical...

  15. Running codes through the web

    International Nuclear Information System (INIS)

    Clark, R.E.H.

    2001-01-01

    Dr. Clark presented a report and demonstration of running atomic physics codes through the WWW. The atomic physics data is generated from Los Alamos National Laboratory (LANL) codes that calculate electron impact excitation, ionization, photoionization, and autoionization, and inversed processes through detailed balance. Samples of Web interfaces, input and output are given in the report

  16. Using Technology of .Net Web Services in the Area of Automation

    Directory of Open Access Journals (Sweden)

    Martin Hnik

    2009-12-01

    Full Text Available This work deals with a technology for data exchange XML Web Services and its application to specific tasks. One of the applications created allows you to monitor and control the real thermal process through a number of client devices, independent of the operating system, the type or their location. The thermal process can be controlled, for example, by another process, a website or a mobile phone. The system is designed from its base and contains three main parts. The hardware part consists from a measuring card, actuators and temperature sensors. The core application is a server that is running the XML Web Service, Windows Service and SQL Server. Client software for mobile phones and web sites was also created.

  17. Faculty Recommendations for Web Tools: Implications for Course Management Systems

    Science.gov (United States)

    Oliver, Kevin; Moore, John

    2008-01-01

    A gap analysis of web tools in Engineering was undertaken as one part of the Digital Library Network for Engineering and Technology (DLNET) grant funded by NSF (DUE-0085849). DLNET represents a Web portal and an online review process to archive quality knowledge objects in Engineering and Technology disciplines. The gap analysis coincided with the…

  18. 07051 Executive Summary -- Programming Paradigms for the Web: Web Programming and Web Services

    OpenAIRE

    Hull, Richard; Thiemann, Peter; Wadler, Philip

    2007-01-01

    The world-wide web raises a variety of new programming challenges. To name a few: programming at the level of the web browser, data-centric approaches, and attempts to automatically discover and compose web services. This seminar brought together researchers from the web programming and web services communities and strove to engage them in communication with each other. The seminar was held in an unusual style, in a mixture of short presentations and in-depth discussio...

  19. A thermoelectric power generating heat exchanger: Part II – Numerical modeling and optimization

    DEFF Research Database (Denmark)

    Sarhadi, Ali; Bjørk, Rasmus; Lindeburg, N.

    2016-01-01

    In Part I of this study, the performance of an experimental integrated thermoelectric generator (TEG)-heat exchanger was presented. In the current study, Part II, the obtained experimental results are compared with those predicted by a finite element (FE) model. In the simulation of the integrated...... TEG-heat exchanger, the thermal contact resistance between the TEG and the heat exchanger is modeled assuming either an ideal thermal contact or using a combined Cooper–Mikic–Yovanovich (CMY) and parallel plate gap formulation, which takes into account the contact pressure, roughness and hardness...

  20. Research and implementation of a Web-based remote desktop image monitoring system

    International Nuclear Information System (INIS)

    Ren Weijuan; Li Luofeng; Wang Chunhong

    2010-01-01

    It studied and implemented an ISS (Image Snapshot Server) system based on Web, using Java Web technology. The ISS system consisted of client web browser and server. The server part could be divided into three modules as the screen shots software, web server and Oracle database. Screen shots software intercepted the desktop environment of the remote monitored PC and sent these pictures to a Tomcat web server for displaying on the web at real time. At the same time, these pictures were also saved in an Oracle database. Through the web browser, monitor person can view the real-time and historical desktop pictures of the monitored PC during some period. It is very convenient for any user to monitor the desktop image of remote monitoring PC. (authors)

  1. Examination of the 'web mode effect'

    DEFF Research Database (Denmark)

    Clement, Sanne Lund; Shamshiri-Petersen, Ditte

    Declining response rates is one of the most significant challenges for survey based research today. Seen in isolation, traditional interviewer based data collection methods are still the most effective but also the most expensive, especially the greater difficulty in gaining responses taken...... into account. As a solution, mixed-mode designs have been employed as a way to achieve higher response rates, while keeping the overall costs low. In particular, the use of web based surveys has expanded considerably during the last few years, both as a single data collection method and as a component in mixed...... with telephone surveys, not enabling determination of a “web mode effect”. In this case, differences might as well be due to differences between self-administered and interviewer-administered collection methods. Other parts of literature on mixed-mode design including a web option are using stratified sampling...

  2. Semantic Web Requirements through Web Mining Techniques

    OpenAIRE

    Hassanzadeh, Hamed; Keyvanpour, Mohammad Reza

    2012-01-01

    In recent years, Semantic web has become a topic of active research in several fields of computer science and has applied in a wide range of domains such as bioinformatics, life sciences, and knowledge management. The two fast-developing research areas semantic web and web mining can complement each other and their different techniques can be used jointly or separately to solve the issues in both areas. In addition, since shifting from current web to semantic web mainly depends on the enhance...

  3. Prey interception drives web invasion and spider size determines successful web takeover in nocturnal orb-web spiders.

    Science.gov (United States)

    Gan, Wenjin; Liu, Shengjie; Yang, Xiaodong; Li, Daiqin; Lei, Chaoliang

    2015-09-24

    A striking feature of web-building spiders is the use of silk to make webs, mainly for prey capture. However, building a web is energetically expensive and increases the risk of predation. To reduce such costs and still have access to abundant prey, some web-building spiders have evolved web invasion behaviour. In general, no consistent patterns of web invasion have emerged and the factors determining web invasion remain largely unexplored. Here we report web invasion among conspecifics in seven nocturnal species of orb-web spiders, and examined the factors determining the probability of webs that could be invaded and taken over by conspecifics. About 36% of webs were invaded by conspecifics, and 25% of invaded webs were taken over by the invaders. A web that was built higher and intercepted more prey was more likely to be invaded. Once a web was invaded, the smaller the size of the resident spider, the more likely its web would be taken over by the invader. This study suggests that web invasion, as a possible way of reducing costs, may be widespread in nocturnal orb-web spiders. © 2015. Published by The Company of Biologists Ltd.

  4. Prey interception drives web invasion and spider size determines successful web takeover in nocturnal orb-web spiders

    Directory of Open Access Journals (Sweden)

    Wenjin Gan

    2015-10-01

    Full Text Available A striking feature of web-building spiders is the use of silk to make webs, mainly for prey capture. However, building a web is energetically expensive and increases the risk of predation. To reduce such costs and still have access to abundant prey, some web-building spiders have evolved web invasion behaviour. In general, no consistent patterns of web invasion have emerged and the factors determining web invasion remain largely unexplored. Here we report web invasion among conspecifics in seven nocturnal species of orb-web spiders, and examined the factors determining the probability of webs that could be invaded and taken over by conspecifics. About 36% of webs were invaded by conspecifics, and 25% of invaded webs were taken over by the invaders. A web that was built higher and intercepted more prey was more likely to be invaded. Once a web was invaded, the smaller the size of the resident spider, the more likely its web would be taken over by the invader. This study suggests that web invasion, as a possible way of reducing costs, may be widespread in nocturnal orb-web spiders.

  5. Grid Computing Das wahre Web 2.0?

    CERN Document Server

    2008-01-01

    'Grid-Computing ist eine Fortentwicklung des World Wide Web, sozusagen die nchste Generation', sagte (1) Franz-Josef Pfreundt (Fraunhofer-Institut fr Techno- und Wirtschaftsmathematik) schon auf der CeBIT 2003 und verwies auf die NASA als Grid-Avantgarde.

  6. Using Open Web APIs in Teaching Web Mining

    Science.gov (United States)

    Chen, Hsinchun; Li, Xin; Chau, M.; Ho, Yi-Jen; Tseng, Chunju

    2009-01-01

    With the advent of the World Wide Web, many business applications that utilize data mining and text mining techniques to extract useful business information on the Web have evolved from Web searching to Web mining. It is important for students to acquire knowledge and hands-on experience in Web mining during their education in information systems…

  7. Design and Application of an Intelligent Agent for Web Information Discovery

    Institute of Scientific and Technical Information of China (English)

    闵君; 冯珊; 唐超; 许立达

    2003-01-01

    With the propagation of applications on the internet, the internet has become a great information source which supplies users with valuable information. But it is hard for users to quickly acquire the right information on the web. This paper an intelligent agent for internet applications to retrieve and extract web information under user's guidance. The intelligent agent is made up of a retrieval script to identify web sources, an extraction script based on the document object model to express extraction process, a data translator to export the extracted information into knowledge bases with frame structures, and a data reasoning to reply users' questions. A GUI tool named Script Writer helps to generate the extraction script visually, and knowledge rule databases help to extract wanted information and to generate the answer to questions.

  8. Using a WebCT to Develop a Research Skills Module

    OpenAIRE

    Bellew Martin, Kelli; Lee, Jennifer

    2003-01-01

    At the start of every academic year, the University of Calgary Library welcomes 1,000 first-year biology students to basic library research skills sessions. These sessions are traditionally taught in lecture format with a PowerPoint presentation and students following along on computers. As part of a pilot project in the Fall of 2002, 200 first-year biology students received the session via WebCT. WebCT is the web-based course management system utilized by the University of Calgary1; it d...

  9. Addressing and Presenting Quality of Satellite Data via Web-Based Services

    Science.gov (United States)

    Leptoukh, Gregory; Lynnes, C.; Ahmad, S.; Fox, P.; Zednik, S.; West, P.

    2011-01-01

    With the recent attention to climate change and proliferation of remote-sensing data utilization, climate model and various environmental monitoring and protection applications have begun to increasingly rely on satellite measurements. Research application users seek good quality satellite data, with uncertainties and biases provided for each data point. However, different communities address remote sensing quality issues rather inconsistently and differently. We describe our attempt to systematically characterize, capture, and provision quality and uncertainty information as it applies to the NASA MODIS Aerosol Optical Depth data product. In particular, we note the semantic differences in quality/bias/uncertainty at the pixel, granule, product, and record levels. We outline various factors contributing to uncertainty or error budget; errors. Web-based science analysis and processing tools allow users to access, analyze, and generate visualizations of data while alleviating users from having directly managing complex data processing operations. These tools provide value by streamlining the data analysis process, but usually shield users from details of the data processing steps, algorithm assumptions, caveats, etc. Correct interpretation of the final analysis requires user understanding of how data has been generated and processed and what potential biases, anomalies, or errors may have been introduced. By providing services that leverage data lineage provenance and domain-expertise, expert systems can be built to aid the user in understanding data sources, processing, and the suitability for use of products generated by the tools. We describe our experiences developing a semantic, provenance-aware, expert-knowledge advisory system applied to NASA Giovanni web-based Earth science data analysis tool as part of the ESTO AIST-funded Multi-sensor Data Synergy Advisor project.

  10. Preenchimento automático de formulários na web oculta

    OpenAIRE

    Gustavo Zanini Kantorski

    2014-01-01

    Muitas informações disponíveis na Web estão armazenadas em bancos de dados on-line e são acessíveis somente após um usuário enviar uma consulta por meio de uma interface de busca. Essas informações estão localizadas em uma parte da Web conhecida como Web Oculta ou Web Profunda e, geralmente, são inacessíveis por máquinas de busca tradicionais. Uma vez que a forma de acessar os dados na Web Oculta se dá por intermédio de submissões de consultas, muitos trabalhos têm focado em como preencher au...

  11. ULSGEN (Uplink Summary Generator)

    Science.gov (United States)

    Wang, Y.-F.; Schrock, M.; Reeve, T.; Nguyen, K.; Smith, B.

    2014-01-01

    Uplink is an important part of spacecraft operations. Ensuring the accuracy of uplink content is essential to mission success. Before commands are radiated to the spacecraft, the command and sequence must be reviewed and verified by various teams. In most cases, this process requires collecting the command data, reviewing the data during a command conference meeting, and providing physical signatures by designated members of various teams to signify approval of the data. If commands or sequences are disapproved for some reason, the whole process must be restarted. Recording data and decision history is important for traceability reasons. Given that many steps and people are involved in this process, an easily accessible software tool for managing the process is vital to reducing human error which could result in uplinking incorrect data to the spacecraft. An uplink summary generator called ULSGEN was developed to assist this uplink content approval process. ULSGEN generates a web-based summary of uplink file content and provides an online review process. Spacecraft operations personnel view this summary as a final check before actual radiation of the uplink data. .

  12. Customizable Scientific Web Portal for Fusion Research

    Energy Technology Data Exchange (ETDEWEB)

    Abla, G; Kim, E; Schissel, D; Flannagan, S [General Atomics, San Diego (United States)

    2009-07-01

    The Web browser has become one of the major application interfaces for remotely participating in magnetic fusion experiments. Recently in other areas, web portals have begun to be deployed. These portals are used to present very diverse sources of information in a unified way. While a web portal has several benefits over other software interfaces, such as providing single point of access for multiple computational services, and eliminating the need for client software installation, the design and development of a web portal has unique challenges. One of the challenges is that a web portal needs to be fast and interactive despite a high volume of tools and information that it presents. Another challenge is the visual output on the web portal often is overwhelming due to the high volume of data generated by complex scientific instruments and experiments; therefore the applications and information should be customizable depending on the needs of users. An appropriate software architecture and web technologies can meet these problems. A web-portal has been designed to support the experimental activities of DIII-D researchers worldwide. It utilizes a multi-tier software architecture, and web 2.0 technologies, such as AJAX, Django, and Memcached, to develop a highly interactive and customizable user interface. It offers a customizable interface with personalized page layouts and list of services for users to select. The users can create a unique personalized working environment to fit their own needs and interests. Customizable services are: real-time experiment status monitoring, diagnostic data access, interactive data visualization. The web-portal also supports interactive collaborations by providing collaborative logbook, shared visualization and online instant message services. Furthermore, the web portal will provide a mechanism to allow users to create their own applications on the web portal as well as bridging capabilities to external applications such as

  13. Correct software in web applications and web services

    CERN Document Server

    Thalheim, Bernhard; Prinz, Andreas; Buchberger, Bruno

    2015-01-01

    The papers in this volume aim at obtaining a common understanding of the challenging research questions in web applications comprising web information systems, web services, and web interoperability; obtaining a common understanding of verification needs in web applications; achieving a common understanding of the available rigorous approaches to system development, and the cases in which they have succeeded; identifying how rigorous software engineering methods can be exploited to develop suitable web applications; and at developing a European-scale research agenda combining theory, methods a

  14. Webmail: an Automated Web Publishing System

    Science.gov (United States)

    Bell, David

    A system for publishing frequently updated information to the World Wide Web will be described. Many documents now hosted by the NOAO Web server require timely posting and frequent updates, but need only minor changes in markup or are in a standard format requiring only conversion to HTML. These include information from outside the organization, such as electronic bulletins, and a number of internal reports, both human and machine generated. Webmail uses procmail and Perl scripts to process incoming email messages in a variety of ways. This processing may include wrapping or conversion to HTML, posting to the Web or internal newsgroups, updating search indices or links on related pages, and sending email notification of the new pages to interested parties. The Webmail system has been in use at NOAO since early 1997 and has steadily grown to include fourteen recipes that together handle about fifty messages per week.

  15. AUTOMATION OF ACCOUNTING THE NUMERIC PARTS AND UNITS AT MANUFACTURING ENTERPRISES

    Directory of Open Access Journals (Sweden)

    E. S. Vasev

    2017-01-01

    Full Text Available Objectives. The substantiation of the need and disclosure of the substantive and technological features of the design and  development of an information system for the registration of numeric parts and units at a manufacturing enterprise.Methods. Case-technology,object-oriented programming.Results. The development mechanism of an information system for registering numeric parts and units at a manufacturing enterprise is  described. The development process begins with an analysis of the  subject area, wherein the requirements for the information system  are defined. Two business processes related to the activity of the  production preparation engineer are identified and the entities and  types of these attributes are indicated. Using the case-tool  DBdesigner 4, a data scheme is constructed and code is generated for scripts used to create a database in the SQL language. The database is implemented on MySQL. The web-interface of the  information system is designed: website scheme, action sequence  diagram, website logical structure and layout. The resulting data  scheme is implemented within MySQL data base management  system (DBMS and a web-based system interface developed using  object-oriented programming using Notepad ++ text editor. The  web-interface allows a multi-user working mode to be provided with the information system. The separation of basic and summary data among different tabs is performed in order to simplify the work of personnel and minimise the human factor associated with the processing of disparate data. The requirements to be met by the developed information system are defined.All software used in the development process is free.Conclusion. The selection of software tools for the development of a system for accounting of numeric parts and units at a manufacturing enterprise is implemented, allowing the conceptual, logical and physical design of a database – as well as its web  interface – to

  16. Hybrid Exploration Agent Platform and Sensor Web System

    Science.gov (United States)

    Stoffel, A. William; VanSteenberg, Michael E.

    2004-01-01

    A sensor web to collect the scientific data needed to further exploration is a major and efficient asset to any exploration effort. This is true not only for lunar and planetary environments, but also for interplanetary and liquid environments. Such a system would also have myriad direct commercial spin-off applications. The Hybrid Exploration Agent Platform and Sensor Web or HEAP-SW like the ANTS concept is a Sensor Web concept. The HEAP-SW is conceptually and practically a very different system. HEAP-SW is applicable to any environment and a huge range of exploration tasks. It is a very robust, low cost, high return, solution to a complex problem. All of the technology for initial development and implementation is currently available. The HEAP Sensor Web or HEAP-SW consists of three major parts, The Hybrid Exploration Agent Platforms or HEAP, the Sensor Web or SW and the immobile Data collection and Uplink units or DU. The HEAP-SW as a whole will refer to any group of mobile agents or robots where each robot is a mobile data collection unit that spends most of its time acting in concert with all other robots, DUs in the web, and the HEAP-SWs overall Command and Control (CC) system. Each DU and robot is, however, capable of acting independently. The three parts of the HEAP-SW system are discussed in this paper. The Goals of the HEAP-SW system are: 1) To maximize the amount of exploration enhancing science data collected; 2) To minimize data loss due to system malfunctions; 3) To minimize or, possibly, eliminate the risk of total system failure; 4) To minimize the size, weight, and power requirements of each HEAP robot; 5) To minimize HEAP-SW system costs. The rest of this paper discusses how these goals are attained.

  17. The pepATTRACT web server for blind, large-scale peptide-protein docking.

    Science.gov (United States)

    de Vries, Sjoerd J; Rey, Julien; Schindler, Christina E M; Zacharias, Martin; Tuffery, Pierre

    2017-07-03

    Peptide-protein interactions are ubiquitous in the cell and form an important part of the interactome. Computational docking methods can complement experimental characterization of these complexes, but current protocols are not applicable on the proteome scale. pepATTRACT is a novel docking protocol that is fully blind, i.e. it does not require any information about the binding site. In various stages of its development, pepATTRACT has participated in CAPRI, making successful predictions for five out of seven protein-peptide targets. Its performance is similar or better than state-of-the-art local docking protocols that do require binding site information. Here we present a novel web server that carries out the rigid-body stage of pepATTRACT. On the peptiDB benchmark, the web server generates a correct model in the top 50 in 34% of the cases. Compared to the full pepATTRACT protocol, this leads to some loss of performance, but the computation time is reduced from ∼18 h to ∼10 min. Combined with the fact that it is fully blind, this makes the web server well-suited for large-scale in silico protein-peptide docking experiments. The rigid-body pepATTRACT server is freely available at http://bioserv.rpbs.univ-paris-diderot.fr/services/pepATTRACT. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  18. Biodiversity maintenance in food webs with regulatory environmental feedbacks.

    Science.gov (United States)

    Bagdassarian, Carey K; Dunham, Amy E; Brown, Christopher G; Rauscher, Daniel

    2007-04-21

    Although the food web is one of the most fundamental and oldest concepts in ecology, elucidating the strategies and structures by which natural communities of species persist remains a challenge to empirical and theoretical ecologists. We show that simple regulatory feedbacks between autotrophs and their environment when embedded within complex and realistic food-web models enhance biodiversity. The food webs are generated through the niche-model algorithm and coupled with predator-prey dynamics, with and without environmental feedbacks at the autotroph level. With high probability and especially at lower, more realistic connectance levels, regulatory environmental feedbacks result in fewer species extinctions, that is, in increased species persistence. These same feedback couplings, however, also sensitize food webs to environmental stresses leading to abrupt collapses in biodiversity with increased forcing. Feedback interactions between species and their material environments anchor food-web persistence, adding another dimension to biodiversity conservation. We suggest that the regulatory features of two natural systems, deep-sea tubeworms with their microbial consortia and a soil ecosystem manifesting adaptive homeostatic changes, can be embedded within niche-model food-web dynamics.

  19. A Chatbot as a Natural Web Interface to Arabic Web QA

    Directory of Open Access Journals (Sweden)

    Bayan Abu Shawar

    2011-03-01

    Full Text Available In this paper, we describe a way to access Arabic Web Question Answering (QA corpus using a chatbot, without the need for sophisticated natural language processing or logical inference. Any Natural Language (NL interface to Question Answer (QA system is constrained to reply with the given answers, so there is no need for NL generation to recreate well-formed answers, or for deep analysis or logical inference to map user input questions onto this logical ontology; simple (but large set of pattern-template matching rules will suffice. In previous research, this approach works properly with English and other European languages. In this paper, we try to see how the same chatbot will react in terms of Arabic Web QA corpus. Initial results shows that 93% of answers were correct, but because of a lot of characteristics related to Arabic language, changing Arabic questions into other forms may lead to no answers.

  20. Constructing the Web of Events from Raw Data in the Web of Things

    Directory of Open Access Journals (Sweden)

    Yunchuan Sun

    2014-01-01

    Full Text Available An exciting paradise of data is emerging into our daily life along with the development of the Web of Things. Nowadays, volumes of heterogeneous raw data are continuously generated and captured by trillions of smart devices like sensors, smart controls, readers and other monitoring devices, while various events occur in the physical world. It is hard for users including people and smart things to master valuable information hidden in the massive data, which is more useful and understandable than raw data for users to get the crucial points for problems-solving. Thus, how to automatically and actively extract the knowledge of events and their internal links from the big data is one key challenge for the future Web of Things. This paper proposes an effective approach to extract events and their internal links from large scale data leveraging predefined event schemas in the Web of Things, which starts with grasping the critical data for useful events by filtering data with well-defined event types in the schema. A case study in the context of smart campus is presented to show the application of proposed approach for the extraction of events and their internal semantic links.

  1. Key Technologies and Applications of Satellite and Sensor Web-coupled Real-time Dynamic Web Geographic Information System

    Directory of Open Access Journals (Sweden)

    CHEN Nengcheng

    2017-10-01

    Full Text Available The geo-spatial information service has failed to reflect the live status of spot and meet the needs of integrated monitoring and real-time information for a long time. To tackle the problems in observation sharing and integrated management of space-borne, air-borne, and ground-based platforms and efficient service of spatio-temporal information, an observation sharing model was proposed. The key technologies in real-time dynamic geographical information system (GIS including maximum spatio-temporal coverage-based optimal layout of earth-observation sensor Web, task-driven and feedback-based control, real-time access of streaming observations, dynamic simulation, warning and decision support were detailed. An real-time dynamic Web geographical information system (WebGIS named GeoSensor and its applications in sensing and management of spatio-temporal information of Yangtze River basin including navigation, flood prevention, and power generation were also introduced.

  2. Ajax, XSLT and SVG: Displaying ATLAS conditions data with new web technologies

    Energy Technology Data Exchange (ETDEWEB)

    Roe, S A, E-mail: shaun.roe@cern.c [CERN, CH-1211 Geneve 23 (Switzerland)

    2010-04-01

    The combination of three relatively recent technologies is described which allows an easy path from database retrieval to interactive web display. SQL queries on an Oracle database can be performed in a manner which directly return an XML description of the result, and Ajax techniques (Asynchronous JavaScript And XML) are used to dynamically inject the data into a web display accompanied by an XSLT transform template which determines how the data will be formatted. By tuning the transform to generate SVG (Scalable Vector Graphics) a direct graphical representation can be produced in the web page while retaining the database data as the XML source, allowing dynamic links to be generated in the web representation, but programmatic use of the data when used from a user application. With the release of the SVG 1.2 Tiny draft specification, the display can also be tailored for display on mobile devices. The technologies are described and a sample application demonstrated, showing conditions data from the ATLAS Semiconductor Tracker.

  3. Ajax, XSLT and SVG: Displaying ATLAS conditions data with new web technologies

    International Nuclear Information System (INIS)

    Roe, S A

    2010-01-01

    The combination of three relatively recent technologies is described which allows an easy path from database retrieval to interactive web display. SQL queries on an Oracle database can be performed in a manner which directly return an XML description of the result, and Ajax techniques (Asynchronous JavaScript And XML) are used to dynamically inject the data into a web display accompanied by an XSLT transform template which determines how the data will be formatted. By tuning the transform to generate SVG (Scalable Vector Graphics) a direct graphical representation can be produced in the web page while retaining the database data as the XML source, allowing dynamic links to be generated in the web representation, but programmatic use of the data when used from a user application. With the release of the SVG 1.2 Tiny draft specification, the display can also be tailored for display on mobile devices. The technologies are described and a sample application demonstrated, showing conditions data from the ATLAS Semiconductor Tracker.

  4. Ajax, XSLT and SVG: Displaying ATLAS conditions data with new web technologies

    CERN Document Server

    Roe, S A

    2010-01-01

    The combination of three relatively recent technologies is described which allows an easy path from database retrieval to interactive web display. SQL queries on an Oracle database can be performed in a manner which directly return an XML description of the result, and Ajax techniques (Asynchronous JavaScript And XML) are used to dynamically inject the data into a web display accompanied by an XSLT transform template which determines how the data will be formatted. By tuning the transform to generate SVG (Scalable Vector Graphics) a direct graphical representation can be produced in the web page while retaining the database data as the XML source, allowing dynamic links to be generated in the web representation, but programmatic use of the data when used from a user application. With the release of the SVG 1.2 Tiny draft specification, the display can also be tailored for display on mobile devices. The technologies are described and a sample application demonstrated, showing conditions data from the ATLAS Sem...

  5. Teaching ocean wave forecasting using computer-generated visualization and animation—Part 1: sea forecasting

    Science.gov (United States)

    Whitford, Dennis J.

    2002-05-01

    Ocean waves are the most recognized phenomena in oceanography. Unfortunately, undergraduate study of ocean wave dynamics and forecasting involves mathematics and physics and therefore can pose difficulties with some students because of the subject's interrelated dependence on time and space. Verbal descriptions and two-dimensional illustrations are often insufficient for student comprehension. Computer-generated visualization and animation offer a visually intuitive and pedagogically sound medium to present geoscience, yet there are very few oceanographic examples. A two-part article series is offered to explain ocean wave forecasting using computer-generated visualization and animation. This paper, Part 1, addresses forecasting of sea wave conditions and serves as the basis for the more difficult topic of swell wave forecasting addressed in Part 2. Computer-aided visualization and animation, accompanied by oral explanation, are a welcome pedagogical supplement to more traditional methods of instruction. In this article, several MATLAB ® software programs have been written to visualize and animate development and comparison of wave spectra, wave interference, and forecasting of sea conditions. These programs also set the stage for the more advanced and difficult animation topics in Part 2. The programs are user-friendly, interactive, easy to modify, and developed as instructional tools. By using these software programs, teachers can enhance their instruction of these topics with colorful visualizations and animation without requiring an extensive background in computer programming.

  6. Quality issues in the management of web information

    CERN Document Server

    Bordogna, Gloria; Jain, Lakhmi

    2013-01-01

    This research volume presents a sample of recent contributions related to the issue of quality-assessment for Web Based information in the context of information access, retrieval, and filtering systems. The advent of the Web and the uncontrolled process of documents' generation have raised the problem of declining quality assessment to information on the Web, by considering both the nature of documents (texts, images, video, sounds, and so on), the genre of documents ( news, geographic information, ontologies, medical records, products records, and so on), the reputation of information sources and sites, and, last but not least the actions performed on documents (content indexing, retrieval and ranking, collaborative filtering, and so on). The volume constitutes a compendium of both heterogeneous approaches and sample applications focusing specific aspects of the quality assessment for Web-based information for researchers, PhD students and practitioners carrying out their research activity in the field of W...

  7. Research of web application based on B/S structure testing

    International Nuclear Information System (INIS)

    Ou Ge; Zhang Hongmei; Song Liming

    2007-01-01

    Software testing is very important method used to assure the quality of Web application. With the fast development of Web application, the old testing techniques can not satisfied the require any more. Because of this people begin to classify different part of the application, find out the content that can be tested by the test tools and studies the structure of testing to enhance his efficiency. This paper analyses the testing based on the feature of Web application, sums up the testing method and gives some improvements of them. (authors)

  8. Medical and Biohazardous Waste Generator's Guide (Revision 2)

    International Nuclear Information System (INIS)

    Waste Management Group

    2006-01-01

    These guidelines describe procedures to comply with all Federal and State laws and regulations and Lawrence Berkeley National Laboratory (LBNL) policy applicable to State-regulated medical and unregulated, but biohazardous, waste (medical/biohazardous waste). These guidelines apply to all LBNL personnel who: (1) generate and/or store medical/biohazardous waste, (2) supervise personnel who generate medical/biohazardous waste, or (3) manage a medical/biohazardous waste pickup location. Personnel generating biohazardous waste at the Joint Genome Institute/Production Genomics Facility (JGI/PGF) are referred to the guidelines contained in Section 9. Section 9 is the only part of these guidelines that apply to JGI/PGF. Medical/biohazardous waste referred to in this Web site includes biohazardous, sharps, pathological and liquid waste. Procedures for proper storage and disposal are summarized in the Solid Medical/Biohazardous Waste Disposal Procedures Chart. Contact the Waste Management Group at 486-7663 if you have any questions regarding medical/biohazardous waste management

  9. Mutation based treatment recommendations from next generation sequencing data: a comparison of web tools.

    Science.gov (United States)

    Patel, Jaymin M; Knopf, Joshua; Reiner, Eric; Bossuyt, Veerle; Epstein, Lianne; DiGiovanna, Michael; Chung, Gina; Silber, Andrea; Sanft, Tara; Hofstatter, Erin; Mougalian, Sarah; Abu-Khalaf, Maysa; Platt, James; Shi, Weiwei; Gershkovich, Peter; Hatzis, Christos; Pusztai, Lajos

    2016-04-19

    Interpretation of complex cancer genome data, generated by tumor target profiling platforms, is key for the success of personalized cancer therapy. How to draw therapeutic conclusions from tumor profiling results is not standardized and may vary among commercial and academically-affiliated recommendation tools. We performed targeted sequencing of 315 genes from 75 metastatic breast cancer biopsies using the FoundationOne assay. Results were run through 4 different web tools including the Drug-Gene Interaction Database (DGidb), My Cancer Genome (MCG), Personalized Cancer Therapy (PCT), and cBioPortal, for drug and clinical trial recommendations. These recommendations were compared amongst each other and to those provided by FoundationOne. The identification of a gene as targetable varied across the different recommendation sources. Only 33% of cases had 4 or more sources recommend the same drug for at least one of the usually several altered genes found in tumor biopsies. These results indicate further development and standardization of broadly applicable software tools that assist in our therapeutic interpretation of genomic data is needed. Existing algorithms for data acquisition, integration and interpretation will likely need to incorporate artificial intelligence tools to improve both content and real-time status.

  10. A Type System for Dynamic Web Documents

    DEFF Research Database (Denmark)

    Schwartzbach, Michael Ignatieff; Sandholm, Anders

    2000-01-01

    Many interactive Web services use the CGI interface for communication with clients. They will dynamically create HTML documents that are presented to the client who then resumes the interaction by submitting data through incorporated form fields. This protocol is difficult to statically type-chec...... system is based on a flow analysis of which we prove soundness. We present an efficient runtime implementation that respects the semantics of only well-typed programs. This work is fully implemented as part of the system for defining interactive Web services.......Many interactive Web services use the CGI interface for communication with clients. They will dynamically create HTML documents that are presented to the client who then resumes the interaction by submitting data through incorporated form fields. This protocol is difficult to statically type...

  11. Building interactive simulations in a Web page design program.

    Science.gov (United States)

    Kootsey, J Mailen; Siriphongs, Daniel; McAuley, Grant

    2004-01-01

    A new Web software architecture, NumberLinX (NLX), has been integrated into a commercial Web design program to produce a drag-and-drop environment for building interactive simulations. NLX is a library of reusable objects written in Java, including input, output, calculation, and control objects. The NLX objects were added to the palette of available objects in the Web design program to be selected and dropped on a page. Inserting an object in a Web page is accomplished by adding a template block of HTML code to the page file. HTML parameters in the block must be set to user-supplied values, so the HTML code is generated dynamically, based on user entries in a popup form. Implementing the object inspector for each object permits the user to edit object attributes in a form window. Except for model definition, the combination of the NLX architecture and the Web design program permits construction of interactive simulation pages without writing or inspecting code.

  12. Usare WebDewey

    OpenAIRE

    Baldi, Paolo

    2016-01-01

    This presentation shows how to use the WebDewey tool. Features of WebDewey. Italian WebDewey compared with American WebDewey. Querying Italian WebDewey. Italian WebDewey and MARC21. Italian WebDewey and UNIMARC. Numbers, captions, "equivalente verbale": Dewey decimal classification in Italian catalogues. Italian WebDewey and Nuovo soggettario. Italian WebDewey and LCSH. Italian WebDewey compared with printed version of Italian Dewey Classification (22. edition): advantages and disadvantages o...

  13. Deploying web-based visual exploration tools on the grid

    Energy Technology Data Exchange (ETDEWEB)

    Jankun-Kelly, T.J.; Kreylos, Oliver; Shalf, John; Ma, Kwan-Liu; Hamann, Bernd; Joy, Kenneth; Bethel, E. Wes

    2002-02-01

    We discuss a web-based portal for the exploration, encapsulation, and dissemination of visualization results over the Grid. This portal integrates three components: an interface client for structured visualization exploration, a visualization web application to manage the generation and capture of the visualization results, and a centralized portal application server to access and manage grid resources. We demonstrate the usefulness of the developed system using an example for Adaptive Mesh Refinement (AMR) data visualization.

  14. Web TA Production (WebTA)

    Data.gov (United States)

    US Agency for International Development — WebTA is a web-based time and attendance system that supports USAID payroll administration functions, and is designed to capture hours worked, leave used and...

  15. Web Services as new phenomenon in the PHP environment

    Directory of Open Access Journals (Sweden)

    Pavel Horovčák

    2006-06-01

    Full Text Available The support of development and exploitation of Web Services (WS is gradually becoming an integral part of current development environments. Beside standard environments connected with the emergence of WS (Java or .NET, the support is presently time realized also in a widely-used environment for the web application development – PHP, in its updated version 5. This contribution is oriented towards the development and utilization of WS within the framework of PHP 5. It deals with the development of standard WS (calculation mode as well as WS in the database mode (using MySQL, SQLite. It compares the structured and object-oriented approach (which is preferred to the server part of the service development.

  16. Web-based tools from AHRQ's National Resource Center.

    Science.gov (United States)

    Cusack, Caitlin M; Shah, Sapna

    2008-11-06

    The Agency for Healthcare Research and Quality (AHRQ) has made an investment of over $216 million in research around health information technology (health IT). As part of their investment, AHRQ has developed the National Resource Center for Health IT (NRC) which includes a public domain Web site. New content for the web site, such as white papers, toolkits, lessons from the health IT portfolio and web-based tools, is developed as needs are identified. Among the tools developed by the NRC are the Compendium of Surveys and the Clinical Decision Support (CDS) Resources. The Compendium of Surveys is a searchable repository of health IT evaluation surveys made available for public use. The CDS Resources contains content which may be used to develop clinical decision support tools, such as rules, reminders and templates. This live demonstration will show the access, use, and content of both these freely available web-based tools.

  17. ER2OWL: Generating OWL Ontology from ER Diagram

    Science.gov (United States)

    Fahad, Muhammad

    Ontology is the fundamental part of Semantic Web. The goal of W3C is to bring the web into (its full potential) a semantic web with reusing previous systems and artifacts. Most legacy systems have been documented in structural analysis and structured design (SASD), especially in simple or Extended ER Diagram (ERD). Such systems need up-gradation to become the part of semantic web. In this paper, we present ERD to OWL-DL ontology transformation rules at concrete level. These rules facilitate an easy and understandable transformation from ERD to OWL. The set of rules for transformation is tested on a structured analysis and design example. The framework provides OWL ontology for semantic web fundamental. This framework helps software engineers in upgrading the structured analysis and design artifact ERD, to components of semantic web. Moreover our transformation tool, ER2OWL, reduces the cost and time for building OWL ontologies with the reuse of existing entity relationship models.

  18. Is there a “net generation” in veterinary medicine? A comparative study on the use of the Internet and Web 2.0 by students and the veterinary profession

    Science.gov (United States)

    Tenhaven, Christoph; Tipold, Andrea; Fischer, Martin R.; Ehlers, Jan P.

    2013-01-01

    Introduction: Informal and formal lifelong learning is essential at university and in the workplace. Apart from classical learning techniques, Web 2.0 tools can be used. It is controversial whether there is a so-called net generation amongst people under 30. Aims: To test the hypothesis that a net generation among students and young veterinarians exists. Methods: An online survey of students and veterinarians was conducted in the German-speaking countries which was advertised via online media and traditional print media. Results: 1780 people took part in the survey. Students and veterinarians have different usage patterns regarding social networks (91.9% vs. 69%) and IM (55.9% vs. 24.5%). All tools were predominantly used passively and in private, to a lesser extent also professionally and for studying. Outlook: The use of Web 2.0 tools is useful, however, teaching information and media skills, preparing codes of conduct for the internet and verification of user generated content is essential. PMID:23467682

  19. Evaluation of a metal shear web selectively reinforced with filamentary composites for space shuttle application. Phase 1 summary report: Shear web design development

    Science.gov (United States)

    Laakso, J. H.; Zimmerman, D. K.

    1972-01-01

    An advanced composite shear web design concept was developed for the Space Shuttle orbiter main engine thrust beam structure. Various web concepts were synthesized by a computer-aided adaptive random search procedure. A practical concept is identified having a titanium-clad + or - 45 deg boron/epoxy web plate with vertical boron/epoxy reinforced aluminum stiffeners. The boron-epoxy laminate contributes to the strength and stiffness efficiency of the basic web section. The titanium-cladding functions to protect the polymeric laminate parts from damaging environments and is chem-milled to provide reinforcement in selected areas. Detailed design drawings are presented for both boron/epoxy reinforced and all-metal shear webs. The weight saving offered is 24% relative to all-metal construction at an attractive cost per pound of weight saved, based on the detailed designs. Small scale element tests substantiate the boron/epoxy reinforced design details in critical areas. The results show that the titanium-cladding reliably reinforces the web laminate in critical edge load transfer and stiffener fastener hole areas.

  20. Designing Crop Simulation Web Service with Service Oriented Architecture Principle

    Science.gov (United States)

    Chinnachodteeranun, R.; Hung, N. D.; Honda, K.

    2015-12-01

    Crop simulation models are efficient tools for simulating crop growth processes and yield. Running crop models requires data from various sources as well as time-consuming data processing, such as data quality checking and data formatting, before those data can be inputted to the model. It makes the use of crop modeling limited only to crop modelers. We aim to make running crop models convenient for various users so that the utilization of crop models will be expanded, which will directly improve agricultural applications. As the first step, we had developed a prototype that runs DSSAT on Web called as Tomorrow's Rice (v. 1). It predicts rice yields based on a planting date, rice's variety and soil characteristics using DSSAT crop model. A user only needs to select a planting location on the Web GUI then the system queried historical weather data from available sources and expected yield is returned. Currently, we are working on weather data connection via Sensor Observation Service (SOS) interface defined by Open Geospatial Consortium (OGC). Weather data can be automatically connected to a weather generator for generating weather scenarios for running the crop model. In order to expand these services further, we are designing a web service framework consisting of layers of web services to support compositions and executions for running crop simulations. This framework allows a third party application to call and cascade each service as it needs for data preparation and running DSSAT model using a dynamic web service mechanism. The framework has a module to manage data format conversion, which means users do not need to spend their time curating the data inputs. Dynamic linking of data sources and services are implemented using the Service Component Architecture (SCA). This agriculture web service platform demonstrates interoperability of weather data using SOS interface, convenient connections between weather data sources and weather generator, and connecting

  1. Development and Validation of WebQuests in Teaching Epics

    Directory of Open Access Journals (Sweden)

    Ronald Candy Santos Lasaten

    2017-05-01

    Full Text Available Using the Research Development (R&D methodology, the study aimed to develop and validate WebQuests which can be used in literature subjects, particularly in the tertiary level to address the need of literature teachers for pedagogy in the teaching of epic s. The development of the Web Quests was anchored on the Theory of Constructivism. Two groups of experts validated the Web Quests – the literature experts and the ICT experts. The Content Validation Checklist, used by the literature experts, was utilized t o evaluate the content of the Web Quests. Meanwhile, the Rubric for Evaluating Web Quests, used by the ICT experts, was utilized to evaluate the design characteristics of the Web Quests. Computed weighted means using range interval of point scores were emp loyed to treat the data gathered from the evaluation conducted by both group of experts. The Web Quests developed contain five major parts which include: 1 introduction; 2 task; 3 process; 4 evaluation; and 5 conclusion. Based on the findings, the con tent of the Web Quests developed are valid in terms of objectives, activities and instructional characteristics. Likewise, the design characteristics of the Web Quests are excellent in terms of introductions, tasks, processes, resources, evaluations, concl usions and overall designs. Thus, the Web Quests developed are acceptable and can be utilized as instructional materials by literature teachers in the teaching of epics.

  2. Politiken, Alt om Ikast Brande (web), Lemvig Folkeblad (Web), Politiken (web), Dabladet Ringkjøbing Skjern (web)

    DEFF Research Database (Denmark)

    Lauritsen, Jens

    2014-01-01

    Politiken 01.01.2014 14:16 Danskerne skød nytåret ind med et brag, men for enkeltes vedkommende gik det galt, da nytårskrudtet blev tændt. Skadestuerne har behandlet 73 personer for fyrværkeriskader mellem klokken 18 i aftes og klokken 06 i morges. Det viser en optælling, som Politiken har...... foretaget på baggrund af tal fra Ulykkes Analyse Gruppen på Odense Universitetshospital. Artiklen er også bragt i: Alt om Ikast Brande (web), Lemvig Folkeblad (web), Politiken (web), Dagbladet Ringkjøbing Skjern (web)....

  3. Recommender systems and the social web leveraging tagging data for recommender systems

    CERN Document Server

    Gedikli, Fatih

    2013-01-01

    There is an increasing demand for recommender systems due to the information overload users are facing on the Web. The goal of a recommender system is to provide personalized recommendations of products or services to users. With the advent of the Social Web, user-generated content has enriched the social dimension of the Web. As user-provided content data also tells us something about the user, one can learn the user's individual preferences from the Social Web. This opens up completely new opportunities and challenges for recommender systems research. Fatih Gedikli deals with the question of

  4. E-commerce Systems and E-shop Web Sites Security

    OpenAIRE

    Suchánek, Petr

    2009-01-01

    Fruitfulnes of contemporary companies rests on new business model development, elimination of communication obstacles, simplification of industrial processes, possibilities of responding in real-time and above all meeting the floating custom needs. Quite a number of company activities and transactions are realized within the framework of e-business. Business transactions are supported by e-commerce systems. One of the e-commerce system part is web interface (web sites). Present trend is putti...

  5. Migrant life stories and the Web

    DEFF Research Database (Denmark)

    Marselis, Randi

    2013-01-01

    The life stories of migrants are increasingly being told, as part of the work of cultural organizations, and websites are well suited to making such life story projects accessible to the public. However, by using the lives of real people as raw material in a public forum, Web projects raise...

  6. Using Web Server Logs in Evaluating Instructional Web Sites.

    Science.gov (United States)

    Ingram, Albert L.

    2000-01-01

    Web server logs contain a great deal of information about who uses a Web site and how they use it. This article discusses the analysis of Web logs for instructional Web sites; reviews the data stored in most Web server logs; demonstrates what further information can be gleaned from the logs; and discusses analyzing that information for the…

  7. WEB STRUCTURE MINING

    Directory of Open Access Journals (Sweden)

    CLAUDIA ELENA DINUCĂ

    2011-01-01

    Full Text Available The World Wide Web became one of the most valuable resources for information retrievals and knowledge discoveries due to the permanent increasing of the amount of data available online. Taking into consideration the web dimension, the users get easily lost in the web’s rich hyper structure. Application of data mining methods is the right solution for knowledge discovery on the Web. The knowledge extracted from the Web can be used to raise the performances for Web information retrievals, question answering and Web based data warehousing. In this paper, I provide an introduction of Web mining categories and I focus on one of these categories: the Web structure mining. Web structure mining, one of three categories of web mining for data, is a tool used to identify the relationship between Web pages linked by information or direct link connection. It offers information about how different pages are linked together to form this huge web. Web Structure Mining finds hidden basic structures and uses hyperlinks for more web applications such as web search.

  8. DESENVOLVIMENTO DE UMA FERRAMENTA ASSISTENTE PARA CRIAÇÃO DE APLICAÇÕES CRUD EM JAVA NA WEB

    Directory of Open Access Journals (Sweden)

    Carlos Renato de Souza Perri

    2010-12-01

    Full Text Available Due to the need for computerization of business processes, storage of relevant information in databases and making these information available on Internet, this project proposes to develop a tool for generating Web applications written in Java, that build functionalities to perform CRUD (Create, Retrieve, Update, Delete, ie, storage, read, update and deletion. The software is a tool, from which, the programmer inserts the script to create a database and, after setting the parameters into the tool, source code of Java Web applications are generated. There is a need to create Web applications in Java with low production time, because building these applications using common methods for development takes much time. The implementation of this tool is also to change that concept and the way of developing Java Web applications, because this tool will be used as an assistant, becoming easier to create Java Web applications. The generated applications use the technologies Servlets and JSPs, the Hibernate framework and the jQuery JavaScript library.

  9. The ACR-program for automatic finite element model generation for part through cracks

    International Nuclear Information System (INIS)

    Leinonen, M.S.; Mikkola, T.P.J.

    1989-01-01

    The ACR-program (Automatic Finite Element Model Generation for Part Through Cracks) has been developed at the Technical Research Centre of Finland (VTT) for automatic finite element model generation for surface flaws using three dimensional solid elements. Circumferential or axial cracks can be generated on the inner or outer surface of a cylindrical or toroidal geometry. Several crack forms are available including the standard semi-elliptical surface crack. The program can be used in the development of automated systems for fracture mechanical analyses of structures. The tests for the accuracy of the FE-mesh have been started with two-dimensional models. The results indicate that the accuracy of the standard mesh is sufficient for practical analyses. Refinement of the standard mesh is needed in analyses with high load levels well over the limit load of the structure

  10. Services for Graduate Students: A Review of Academic Library Web Sites

    Science.gov (United States)

    Rempel, Hannah Gascho

    2010-01-01

    A library's Web site is well recognized as the gateway to the library for the vast majority of users. Choosing the most user-friendly Web architecture to reflect the many services libraries offer is a complex process, and librarians are still experimenting to find what works best for their users. As part of a redesign of the Oregon State…

  11. Web Engineering

    Energy Technology Data Exchange (ETDEWEB)

    White, Bebo

    2003-06-23

    Web Engineering is the application of systematic, disciplined and quantifiable approaches to development, operation, and maintenance of Web-based applications. It is both a pro-active approach and a growing collection of theoretical and empirical research in Web application development. This paper gives an overview of Web Engineering by addressing the questions: (a) why is it needed? (b) what is its domain of operation? (c) how does it help and what should it do to improve Web application development? and (d) how should it be incorporated in education and training? The paper discusses the significant differences that exist between Web applications and conventional software, the taxonomy of Web applications, the progress made so far and the research issues and experience of creating a specialization at the master's level. The paper reaches a conclusion that Web Engineering at this stage is a moving target since Web technologies are constantly evolving, making new types of applications possible, which in turn may require innovations in how they are built, deployed and maintained.

  12. Web Accessibility in Romania: The Conformance of Municipal Web Sites to Web Content Accessibility Guidelines

    OpenAIRE

    Costin PRIBEANU; Ruxandra-Dora MARINESCU; Paul FOGARASSY-NESZLY; Maria GHEORGHE-MOISII

    2012-01-01

    The accessibility of public administration web sites is a key quality attribute for the successful implementation of the Information Society. The purpose of this paper is to present a second review of municipal web sites in Romania that is based on automated accessibility checking. A number of 60 web sites were evaluated against WCAG 2.0 recommendations. The analysis of results reveals a relatively low web accessibility of municipal web sites and highlights several aspects. Firstly, a slight ...

  13. Making attributes from the Linked Open Data (LOD) cloud a part of ...

    African Journals Online (AJOL)

    This research contributed to bridging the gap between linked data, SDI and web thematic maps and further showed how existing web mapping and OGC technologies can benefit from the Semantic Web. First, the design of a geospatial web service (representing the visible part of an SDI) that accesses attribute data from the ...

  14. An Introduction to XML and Web Technologies

    DEFF Research Database (Denmark)

    Møller, Anders; Schwartzbach, Michael Ignatieff

    , building on top of the early foundations. This book offers a comprehensive introduction to the area. There are two main threads of development, corresponding to the two parts of this book. XML technologies generalize the notion of data on the Web from hypertext documents to arbitrary data, including those...... that have traditionally been the realm of databases. In this book we cover the basic XML technology and the supporting technologies of XPath, DTD, XML Schema, DSD2, RELAX NG, XSLT, XQuery, DOM, JDOM, JAXB, SAX, STX, SDuce, and XACT. Web technologies build on top of the HTTP protocol to provide richer...

  15. Letting go of the words writing web content that works

    CERN Document Server

    Redish, Janice (Ginny)

    2012-01-01

    Web site design and development continues to become more sophisticated an important part of this maturity originates with well laid out and well written content. Ginny Redish is a world-renowned expert on information design and how to produce clear writing in plain language for the web. All of the invaluable information that she  shared in the first edition is included with numerous new examples. New information on content strategy for web sites, search engine optimization (SEO), and social media will enhance the book's content making it once again the only book you need to own to o

  16. 07051 Working Group Outcomes -- Programming Paradigms for the Web: Web Programming and Web Services

    OpenAIRE

    Hull, Richard; Thiemann, Peter; Wadler, Philip

    2007-01-01

    Participants in the seminar broke into groups on ``Patterns and Paradigms'' for web programming, ``Web Services,'' ``Data on the Web,'' ``Software Engineering'' and ``Security.'' Here we give the raw notes recorded during these sessions.

  17. Towards Web Service-Based Educational Systems

    Science.gov (United States)

    Sampson, Demetrios G.

    2005-01-01

    The need for designing the next generation of web service-based educational systems with the ability of integrating components from different tools and platforms is now recognised as the major challenge in advanced learning technologies. In this paper, we discuss this issue and we present the conceptual design of such environment, referred to as…

  18. Embedding the Form Generator in a Content Management System

    Science.gov (United States)

    Delgado, A.; Wicenec, A.; Delmotte, N.; Tejero, A.

    2008-08-01

    Given the tremendous amount of data generated by ESO's telescopes and the rapid evolution of the World Wide Web, the ESO archive web interface needs to offer more flexible services and advanced functionalities to a growing community of users all over the world. To achieve this endeavour, a query form generator is being developed inside a Content Management System. We present here a progress report.

  19. What will Web 3.0 bring to education?

    Directory of Open Access Journals (Sweden)

    Dan Jiang

    2014-08-01

    Full Text Available Every era of technology has, to some extent, formed education in its own image. It is believed that there is a mutually productive convergence between main technological influences on a culture and the contemporary educational theories and practices. As we are stepping into the era of Web 3.0, it is no surprise that t he network has been one part of our daily life and it becomes one of the most important places for us to learn, work, entertain and socialize, especially for the digital natives. For many people, Web 3.0 still may be a new word in education or a future trend for them, but actually the Web 3.0 technology has been applied and it keeps changing the culture, theory and practice in education subtly. The article aims to discuss about the impacts on Web 3.0 on education and try to view the impacts in terms of culture philosophy and sociology.

  20. Web Mining

    Science.gov (United States)

    Fürnkranz, Johannes

    The World-Wide Web provides every internet citizen with access to an abundance of information, but it becomes increasingly difficult to identify the relevant pieces of information. Research in web mining tries to address this problem by applying techniques from data mining and machine learning to Web data and documents. This chapter provides a brief overview of web mining techniques and research areas, most notably hypertext classification, wrapper induction, recommender systems and web usage mining.

  1. The Web and Information Literacy: Scaffolding the use of Web Sources in a Project-Based Curriculum

    Science.gov (United States)

    Walton, Marion; Archer, Arlene

    2004-01-01

    In this article we describe and discuss a three-year case study of a course in web literacy, part of the academic literacy curriculum for first-year engineering students at the University of Cape Town (UCT). Because they are seen as practical knowledge, not theoretical, information skills tend to be devalued at university and rendered invisible to…

  2. Using Web Services and XML Harvesting to Achieve a Dynamic Web Site. Computers in Small Libraries

    Science.gov (United States)

    Roberts, Gary

    2005-01-01

    Exploiting and contextualizing free information is a natural part of library culture. In this column, Gary Roberts, the information systems and reference librarian at Herrick Library, Alfred University in Alfred, NY, describes how to use XML content on a Web site to link to hundreds of free and useful resources. He gives a general overview of the…

  3. Design and Development of a Web Based User Interface

    OpenAIRE

    László, Magda

    2014-01-01

    The first objective of the thesis is to study the technological background of application design and more specifically the Unified Modeling Language (hereinafter UML). Due to this, the research provides deeper understanding of technical aspects of the practical part of the thesis work. The second and third objectives of this thesis are to design and develop a web application and more specifically a Web Based User Interface for Multimodal Observation and Analysis System for Social Interactions...

  4. SPATIO-TEMPORAL CLUSTERING OF MOVEMENT DATA: AN APPLICATION TO TRAJECTORIES GENERATED BY HUMAN-COMPUTER INTERACTION

    Directory of Open Access Journals (Sweden)

    G. McArdle

    2012-07-01

    Full Text Available Advances in ubiquitous positioning technologies and their increasing availability in mobile devices has generated large volumes of movement data. Analysing these datasets is challenging. While data mining techniques can be applied to this data, knowledge of the underlying spatial region can assist interpreting the data. We have developed a geovisual analysis tool for studying movement data. In addition to interactive visualisations, the tool has features for analysing movement trajectories, in terms of their spatial and temporal similarity. The focus in this paper is on mouse trajectories of users interacting with web maps. The results obtained from a user trial can be used as a starting point to determine which parts of a mouse trajectory can assist personalisation of spatial web maps.

  5. Understanding User-Web Interactions via Web Analytics

    CERN Document Server

    Jansen, Bernard J

    2009-01-01

    This lecture presents an overview of the Web analytics process, with a focus on providing insight and actionable outcomes from collecting and analyzing Internet data. The lecture first provides an overview of Web analytics, providing in essence, a condensed version of the entire lecture. The lecture then outlines the theoretical and methodological foundations of Web analytics in order to make obvious the strengths and shortcomings of Web analytics as an approach. These foundational elements include the psychological basis in behaviorism and methodological underpinning of trace data as an empir

  6. Web based machine status display for INDUS-1 And INDUS-2

    International Nuclear Information System (INIS)

    Srivastava, B.S.K.; Fatnani, P.

    2003-01-01

    Web based machine status display for Indus-1 and Indus-2 is designated to provide on-line status of Indus-1 and Indus-2 to the clients located at various places of CAT premises. Presently, this system provides Indus-1 machine status (e.g. beam current, integrated current, beam life-time etc) to the users working in Indus-1 building, but using the web browsers the same information can be accessed throughout the CAT network. This system is basically a part of Indus-1 Control System Web Site which is under construction (partially constructed). (author)

  7. The design and implementation of web mining in web sites security

    Science.gov (United States)

    Li, Jian; Zhang, Guo-Yin; Gu, Guo-Chang; Li, Jian-Li

    2003-06-01

    The backdoor or information leak of Web servers can be detected by using Web Mining techniques on some abnormal Web log and Web application log data. The security of Web servers can be enhanced and the damage of illegal access can be avoided. Firstly, the system for discovering the patterns of information leakages in CGI scripts from Web log data was proposed. Secondly, those patterns for system administrators to modify their codes and enhance their Web site security were provided. The following aspects were described: one is to combine web application log with web log to extract more information, so web data mining could be used to mine web log for discovering the information that firewall and Information Detection System cannot find. Another approach is to propose an operation module of web site to enhance Web site security. In cluster server session, Density-Based Clustering technique is used to reduce resource cost and obtain better efficiency.

  8. Food-Web Complexity in Guaymas Basin Hydrothermal Vents and Cold Seeps.

    Directory of Open Access Journals (Sweden)

    Marie Portail

    Full Text Available In the Guaymas Basin, the presence of cold seeps and hydrothermal vents in close proximity, similar sedimentary settings and comparable depths offers a unique opportunity to assess and compare the functioning of these deep-sea chemosynthetic ecosystems. The food webs of five seep and four vent assemblages were studied using stable carbon and nitrogen isotope analyses. Although the two ecosystems shared similar potential basal sources, their food webs differed: seeps relied predominantly on methanotrophy and thiotrophy via the Calvin-Benson-Bassham (CBB cycle and vents on petroleum-derived organic matter and thiotrophy via the CBB and reductive tricarboxylic acid (rTCA cycles. In contrast to symbiotic species, the heterotrophic fauna exhibited high trophic flexibility among assemblages, suggesting weak trophic links to the metabolic diversity of chemosynthetic primary producers. At both ecosystems, food webs did not appear to be organised through predator-prey links but rather through weak trophic relationships among co-occurring species. Examples of trophic or spatial niche differentiation highlighted the importance of species-sorting processes within chemosynthetic ecosystems. Variability in food web structure, addressed through Bayesian metrics, revealed consistent trends across ecosystems. Food-web complexity significantly decreased with increasing methane concentrations, a common proxy for the intensity of seep and vent fluid fluxes. Although high fluid-fluxes have the potential to enhance primary productivity, they generate environmental constraints that may limit microbial diversity, colonisation of consumers and the structuring role of competitive interactions, leading to an overall reduction of food-web complexity and an increase in trophic redundancy. Heterogeneity provided by foundation species was identified as an additional structuring factor. According to their biological activities, foundation species may have the potential to

  9. A Ubiquitous Sensor Network Platform for Integrating Smart Devices into the Semantic Sensor Web

    Science.gov (United States)

    de Vera, David Díaz Pardo; Izquierdo, Álvaro Sigüenza; Vercher, Jesús Bernat; Gómez, Luis Alfonso Hernández

    2014-01-01

    Ongoing Sensor Web developments make a growing amount of heterogeneous sensor data available to smart devices. This is generating an increasing demand for homogeneous mechanisms to access, publish and share real-world information. This paper discusses, first, an architectural solution based on Next Generation Networks: a pilot Telco Ubiquitous Sensor Network (USN) Platform that embeds several OGC® Sensor Web services. This platform has already been deployed in large scale projects. Second, the USN-Platform is extended to explore a first approach to Semantic Sensor Web principles and technologies, so that smart devices can access Sensor Web data, allowing them also to share richer (semantically interpreted) information. An experimental scenario is presented: a smart car that consumes and produces real-world information which is integrated into the Semantic Sensor Web through a Telco USN-Platform. Performance tests revealed that observation publishing times with our experimental system were well within limits compatible with the adequate operation of smart safety assistance systems in vehicles. On the other hand, response times for complex queries on large repositories may be inappropriate for rapid reaction needs. PMID:24945678

  10. Food Web Response to Habitat Restoration in Various Coastal Wetland Ecosystems

    Science.gov (United States)

    James, W. R.; Nelson, J. A.

    2017-12-01

    Coastal wetland habitats provide important ecosystem services, including supporting coastal food webs. These habitats are being lost rapidly. To combat the effects of these losses, millions of dollars have been invested to restore these habitats. However, the relationship between restoring habitat and restoring ecosystem functioning is poorly understood. Analyzing energy flow through food web comparisons between restored and natural habitats can give insights into ecosystem functioning. Using published stable isotope values from organisms in restored and natural habitats, we assessed the food web response of habitat restoration in salt marsh, mangrove, sea grass, and algal bed ecosystems. We ran Bayesian mixing models to quantify resource use by consumers and generated habitat specific niche hypervolumes for each ecosystem to assess food web differences between restored and natural habitats. Salt marsh, mangrove, and sea grass ecosystems displayed functional differences between restored and natural habitats. Salt marsh and mangrove food webs varied in the amount of each resource used, while the sea grass food web displayed more variation between individual organisms. The algal bed food web showed little variation between restored and natural habitats.

  11. Using Kalman Filter to Guarantee QoS Robustness of Web Server

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    The exponential growth of the Internet coupled with the increasing popularity of dynamically generated content on the World Wide Web, has created the need for more and faster Web servers capable of serving the over 100 million Internet users. To converge the control method has emerged as a promising technique to solve the Web QoS problem. In this paper, a model of adaptive session is presented and a session flow self-regulating algorism based on Kalman Filter are proposed towards Web Server. And a Web QoS self-regulating scheme is advanced. To attain the goal of on-line system identification, the optimized estimation of QoS parameters is fulfilled by utilizing Kalman Filter in full domain. The simulation results shows that the proposed scheme can guarantee the QoS with both robustness and stability .

  12. Personality in cyberspace: personal Web sites as media for personality expressions and impressions.

    Science.gov (United States)

    Marcus, Bernd; Machilek, Franz; Schütz, Astrid

    2006-06-01

    This research examined the personality of owners of personal Web sites based on self-reports, visitors' ratings, and the content of the Web sites. The authors compared a large sample of Web site owners with population-wide samples on the Big Five dimensions of personality. Controlling for demographic differences, the average Web site owner reported being slightly less extraverted and more open to experience. Compared with various other samples, Web site owners did not generally differ on narcissism, self-monitoring, or self-esteem, but gender differences on these traits were often smaller in Web site owners. Self-other agreement was highest with Openness to Experience, but valid judgments of all Big Five dimensions were derived from Web sites providing rich information. Visitors made use of quantifiable features of the Web site to infer personality, and the cues they utilized partly corresponded to self-reported traits. Copyright 2006 APA, all rights reserved.

  13. Web-based research publications on Sub-Saharan Africa's prized ...

    African Journals Online (AJOL)

    The study confirms Africa's deep interest in the grasscutter which is not shared by other parts of the world. We recommend increased publication of research on cane rats in web-based journals to quickly spread the food value of this prized meat rodent to other parts of the world and so attract research interest and funding.

  14. Web 2.0 and competitiveness improvement

    Directory of Open Access Journals (Sweden)

    Bauerová, Danuše

    2011-12-01

    Full Text Available Article shows implementation of A Model of Learning Powered by Technology at university environment. A Positive Digital Identity development on Web2.0 cloud is presented. People are creating their own portfolios - Personal Learning Portfolio and Personal Credit Portfolio. The part of it such proposals are generalizing methodologies to be useful to improve competitiveness of students, pedagogues, scientists but also institutions as well as anybody of lifelong learning activities.

    Este artículo muestra la puesta en marcha de un modelo de aprendizaje potenciado por la tecnología en el entorno de la Universidad. Se presenta un desarrollo de la Identidad Digital Positiva sobre la nube Web2.0. Las personas crean sus propios portafolios –Personal Learning Portfolio y Personal Credit Portfolio. Por otra parte, estas propuestas representan la generalización de metodologías que son útiles para mejorar la competitividad de estudiantes, pedagogos, científicos, pero también de las instituciones, así como de las actividades de aprendizaje a lo largo de la vida.

  15. Even Faster Web Sites Performance Best Practices for Web Developers

    CERN Document Server

    Souders, Steve

    2009-01-01

    Performance is critical to the success of any web site, and yet today's web applications push browsers to their limits with increasing amounts of rich content and heavy use of Ajax. In this book, Steve Souders, web performance evangelist at Google and former Chief Performance Yahoo!, provides valuable techniques to help you optimize your site's performance. Souders' previous book, the bestselling High Performance Web Sites, shocked the web development world by revealing that 80% of the time it takes for a web page to load is on the client side. In Even Faster Web Sites, Souders and eight exp

  16. Dynamic 2D-barcodes for multi-device Web session migration including mobile phones

    DEFF Research Database (Denmark)

    Alapetite, Alexandre

    2010-01-01

    This article introduces a novel Web architecture that supports session migration in multi-device Web applications, particularly the case when a user starts a Web session on a computer and wishes to continue on a mobile phone. The proposed solution for transferring the needed session identifiers...... across devices is to dynamically generate pictures of 2D-barcodes containing a Web address and a session ID in an encoded form. 2D-barcodes are a cheap, fast and robust approach to the problem. They are widely known and used in Japan, and are spreading in other countries. Variations on the topic...

  17. An Application for Data Preprocessing and Models Extractions in Web Usage Mining

    Directory of Open Access Journals (Sweden)

    Claudia Elena DINUCA

    2011-11-01

    Full Text Available Web servers worldwide generate a vast amount of information on web users’ browsing activities. Several researchers have studied these so-called clickstream or web access log data to better understand and characterize web users. The goal of this application is to analyze user behaviour by mining enriched web access log data. With the continued growth and proliferation of e-commerce, Web services, and Web-based information systems, the volumes of click stream and user data collected by Web-based organizations in their daily operations has reached astronomical proportions. This information can be exploited in various ways, such as enhancing the effectiveness of websites or developing directed web marketing campaigns. The discovered patterns are usually represented as collections of pages, objects, or re-sources that are frequently accessed by groups of users with common needs or interests. In this paper we will focus on displaying the way how it was implemented the application for data preprocessing and extracting different data models from web logs data, finding association as a data mining technique to extract potentially useful knowledge from web usage data. We find different data models navigation patterns by analysing the log files of the web-site. I implemented the application in Java using NetBeans IDE. For exemplification, I used the log files data from a commercial web site www.nice-layouts.com.

  18. The rendering context for stereoscopic 3D web

    Science.gov (United States)

    Chen, Qinshui; Wang, Wenmin; Wang, Ronggang

    2014-03-01

    3D technologies on the Web has been studied for many years, but they are basically monoscopic 3D. With the stereoscopic technology gradually maturing, we are researching to integrate the binocular 3D technology into the Web, creating a stereoscopic 3D browser that will provide users with a brand new experience of human-computer interaction. In this paper, we propose a novel approach to apply stereoscopy technologies to the CSS3 3D Transforms. Under our model, each element can create or participate in a stereoscopic 3D rendering context, in which 3D Transforms such as scaling, translation and rotation, can be applied and be perceived in a truly 3D space. We first discuss the underlying principles of stereoscopy. After that we discuss how these principles can be applied to the Web. A stereoscopic 3D browser with backward compatibility is also created for demonstration purposes. We take advantage of the open-source WebKit project, integrating the 3D display ability into the rendering engine of the web browser. For each 3D web page, our 3D browser will create two slightly different images, each representing the left-eye view and right-eye view, both to be combined on the 3D display to generate the illusion of depth. And as the result turns out, elements can be manipulated in a truly 3D space.

  19. Introduction: Life in the Web

    Directory of Open Access Journals (Sweden)

    Oana Mateescu

    2010-10-01

    Full Text Available The article serves as an introduction to the current journal issue on Online Lives. It discusses and connects the research papers here under three different rubrics: the hopes and fears raised by a world where information is inescapable, the potential and risk of online identities and, finally, the forms of knowledge and participation that define the current architecture of a Web dominated by user-generated content.

  20. Efficient Web Harvesting Strategies for Monitoring Deep Web Content

    NARCIS (Netherlands)

    Khelghati, Mohammadreza; Hiemstra, Djoerd; van Keulen, Maurice

    2016-01-01

    The change of the web content is rapid. In Focused Web Harvesting [?], which aims at achieving a complete harvest for a given topic, this dynamic nature of the web creates problems for users who need to access a complete set of related web data to their interesting topics. Whether you are a fan

  1. Enabling Problem Based Learning through Web 2.0 Technologies

    DEFF Research Database (Denmark)

    Tambouris, Efthimios; Panopoulou, Eleni; Tarabanis, Konstantinos

    2012-01-01

    of modern educational systems. Established pedagogical strategies, such as Problem Based Learning (PBL), are being adapted for online use in conjunction with modern Web 2.0 technologies and tools. However, even though Web 2.0 and progressive social-networking technologies are automatically associated......Advances in Information and Communications Technology (ICT), particularly the so-called Web 2.0, are affecting all aspects of our life: how we communicate, how we shop, how we socialise, and how we learn. Facilitating learning through the use of ICT, also known as eLearning, is a vital part...... with ideals such as collaboration, sharing, and active learning, it is also possible to use them in a very conservative, teacher-centred way limiting thus their impact. In this paper, we present a PBL 2.0 framework, i.e., a framework combining PBL practices with Web 2.0 technologies. More specifically, we (a...

  2. Web2Quests: Updating a Popular Web-Based Inquiry-Oriented Activity

    Science.gov (United States)

    Kurt, Serhat

    2009-01-01

    WebQuest is a popular inquiry-oriented activity in which learners use Web resources. Since the creation of the innovation, almost 15 years ago, the Web has changed significantly, while the WebQuest technique has changed little. This article examines possible applications of new Web trends on WebQuest instructional strategy. Some possible…

  3. Exploring default mode and information flow on the web.

    Science.gov (United States)

    Oka, Mizuki; Ikegami, Takashi

    2013-01-01

    Social networking services (e.g., Twitter, Facebook) are now major sources of World Wide Web (called "Web") dynamics, together with Web search services (e.g., Google). These two types of Web services mutually influence each other but generate different dynamics. In this paper, we distinguish two modes of Web dynamics: the reactive mode and the default mode. It is assumed that Twitter messages (called "tweets") and Google search queries react to significant social movements and events, but they also demonstrate signs of becoming self-activated, thereby forming a baseline Web activity. We define the former as the reactive mode and the latter as the default mode of the Web. In this paper, we investigate these reactive and default modes of the Web's dynamics using transfer entropy (TE). The amount of information transferred between a time series of 1,000 frequent keywords in Twitter and the same keywords in Google queries is investigated across an 11-month time period. Study of the information flow on Google and Twitter revealed that information is generally transferred from Twitter to Google, indicating that Twitter time series have some preceding information about Google time series. We also studied the information flow among different Twitter keywords time series by taking keywords as nodes and flow directions as edges of a network. An analysis of this network revealed that frequent keywords tend to become an information source and infrequent keywords tend to become sink for other keywords. Based on these findings, we hypothesize that frequent keywords form the Web's default mode, which becomes an information source for infrequent keywords that generally form the Web's reactive mode. We also found that the Web consists of different time resolutions with respect to TE among Twitter keywords, which will be another focal point of this paper.

  4. Integration of Web mining and web crawler: Relevance and State of Art

    OpenAIRE

    Subhendu kumar pani; Deepak Mohapatra,; Bikram Keshari Ratha

    2010-01-01

    This study presents the role of web crawler in web mining environment. As the growth of the World Wide Web exceeded all expectations,the research on Web mining is growing more and more.web mining research topic which combines two of the activated research areas: Data Mining and World Wide Web .So, the World Wide Web is a very advanced area for data mining research. Search engines that are based on web crawling framework also used in web mining to find theinteracted web pages. This paper discu...

  5. Efficient Web Harvesting Strategies for Monitoring Deep Web Content

    NARCIS (Netherlands)

    Khelghati, Mohammadreza; Hiemstra, Djoerd; van Keulen, Maurice

    2016-01-01

    Web content changes rapidly [18]. In Focused Web Harvesting [17] which aim it is to achieve a complete harvest for a given topic, this dynamic nature of the web creates problems for users who need to access a set of all the relevant web data to their topics of interest. Whether you are a fan

  6. Law on the web a guide for students and practitioners

    CERN Document Server

    Stein, Stuart

    2014-01-01

    Law on the Web is ideal for anyone who wants to access Law Internet resources quickly and efficiently without becoming an IT expert. The emphasis throughout is on the location of high quality law Internet resources for learning, teaching and research, from among the billions of publicly accessible Web pages. The book is structured so that it will be found useful by both beginners and intermediate level users, and be of continuing use over the course of higher education studies. In addition to extensive coverage on locating files and Web sites, Part III provides a substantial and annotated list of high quality resources for law students.

  7. Teaching Hypertext and Hypermedia through the Web.

    Science.gov (United States)

    de Bra, Paul M. E.

    This paper describes a World Wide Web-based introductory course titled "Hypermedia Structures and Systems," offered as an optional part of the curriculum in computing science at the Eindhoven University of Technology (Netherlands). The technical environment for the current (1996) edition of the course is presented, which features…

  8. Teaching hypertext and hypermedia through the web

    NARCIS (Netherlands)

    De Bra, P.M.E.

    1996-01-01

    Since early 1994 the introductory course 2L670, "Hypermedia Structures and Systems", has been available on World Wide Web, and is an optional part of the curriculum in computing science at the Eindhoven University of Technology. The course has since been completed by more than 200 students from

  9. Enjeux communicationnels du Web 2.0 pour les Relations Publiques/Professional and organisational issues of Web 2.0 Public Relations

    Directory of Open Access Journals (Sweden)

    Francine Charest

    2012-12-01

    Full Text Available The Web 2.0, which includes Facebook, Twitter, Youtube and other social medias, is considered as to be one of the strongest communication tools of the early 21st century. TheWeb evolutionhas changed deeply the way Public relations agents operate. In 2009, Charest and Bédard have shown that the Web 2.0 was in fact a reclaim by the internet users of the Web as it was first imagined by Tim Berners-Lee in November 1993 : a tool to exchange and share information. The Web first generation has instead been used by the administrators for dissemination and promotion. Today, in order to appropriate themselves these new medias, PR agents have to findnew business models, even new ways to communicate.RésuméLe Web 2.0, regroupant les Facebook, Twitter, YouTube et autres médias sociaux, est considéré comme l’un des plus puissants outils de communication, en ce début de XX1 siècle. C’est sous l’angle de mutations qu’il induit sur les pratiques des professionnels en relations publiques, qu’il nous intéresse d’étudier les enjeux de l’évolution des usages du Web 2.0. Charest et Bédard ont montré en 2009 que le Web 2.0 était la revanche des internautes qui tentent de se réapproprier le Web tel qu’il avait été conçu par Tim Berners-Lee en novembre 1993, soit comme un outil d’échange et de partage d’information. Il a été clairement montré que la première générationde Web a plutôt été utilisée par les gestionnaires à des fins de diffusion et de promotion. L’appropriation de ces nouveaux médias par les relationnistes passe nécessairement par de nouveaux modèles d’affaires, voire de nouvelles façons de communiquer.

  10. Relational Constraint Driven Test Case Synthesis for Web Applications

    Directory of Open Access Journals (Sweden)

    Xiang Fu

    2010-09-01

    Full Text Available This paper proposes a relational constraint driven technique that synthesizes test cases automatically for web applications. Using a static analysis, servlets can be modeled as relational transducers, which manipulate backend databases. We present a synthesis algorithm that generates a sequence of HTTP requests for simulating a user session. The algorithm relies on backward symbolic image computation for reaching a certain database state, given a code coverage objective. With a slight adaptation, the technique can be used for discovering workflow attacks on web applications.

  11. Large-area sheet task advanced dendritic web growth development

    Science.gov (United States)

    Duncan, C. S.; Seidensticker, R. G.; Mchugh, J. P.

    1984-01-01

    The thermal models used for analyzing dendritic web growth and calculating the thermal stress were reexamined to establish the validity limits imposed by the assumptions of the models. Also, the effects of thermal conduction through the gas phase were evaluated and found to be small. New growth designs, both static and dynamic, were generated using the modeling results. Residual stress effects in dendritic web were examined. In the laboratory, new techniques for the control of temperature distributions in three dimensions were developed. A new maximum undeformed web width of 5.8 cm was achieved. A 58% increase in growth velocity of 150 micrometers thickness was achieved with dynamic hardware. The area throughput goals for transient growth of 30 and 35 sq cm/min were exceeded.

  12. Usage Of Asp.Net Ajax for Binus School Serpong Web Applications

    Directory of Open Access Journals (Sweden)

    Karto Iskandar

    2016-03-01

    Full Text Available Today web applications have become a necessity and many companies use them as a communication tool to keep in touch with their customers. The usage of Web Application in current time increases as the numberof internet users has been rised. For reason of Rich Internet Application, the desktop application developer wasmoved to web application developer with AJAX technology. BINUS School Serpong is a Cambridge Curriculum base International School that uses web application for access every information about the school. By usingAJAX, performance of web application should be improved and the bandwidth usage is decreased. Problems thatoccur at BINUS School Serpong is not all part of the web application that uses AJAX. This paper introducesusage of AJAX in ASP.NET with C# programming language in web application BINUS School Serpong. It is expected by using ASP.NET AJAX, BINUS School Serpong website performance will be faster because of reducing web page reload. The methodology used in this paper is literature study. Results from this study are to prove that the ASP.NET AJAX can be used easily and improve BINUS School Serpong website performance. Conclusion of this paper is the implementation of ASP.NET AJAX improves performance of web application in BINUS School Serpong.

  13. Accesibilidad web en el espacio universitario público argentino

    Directory of Open Access Journals (Sweden)

    Laitano, María Inés

    2015-03-01

    Full Text Available The study presents a first web accessibility diagnosis carried out in 2012 on a sample of pages from the Argentine public university space. The evaluation establishes the compliance with Web Content Accessibility Guidelines (WCAG 2.0, taking into account the methodological recommendations from the World Wide Web Consortium (W3C. The results suggest that the web accessibility barriers encountered are serious for the most part (level A. The most frequent are related to markup language syntax, content presentation, non-text content and visual readability of text. Likewise it is shown that by addressing these barriers, certain groups of people could benefit specifically.El estudio presenta un primer diagnóstico de accesibilidad web realizado en 2012 sobre una muestra de páginas del espacio universitario público argentino. La evaluación comprueba el cumplimiento de las Pautas de Accesibilidad para el Contenido Web (WCAG en su versión 2.0, contemplando las recomendaciones metodológicas del World Wide Web Consortium (W3C. Los resultados sugieren que las barreras de accesibilidad web encontradas son mayoritariamente graves (nivel A. Las más frecuentes están relacionadas con la sintaxis del lenguaje de marcado, con la presentación del contenido, con el contenido no textual y con la legibilidad visual del texto. De igual modo se muestra que ciertos grupos de personas podrían verse particularmente favorecidos por la solución de estas barreras.

  14. Correspondence of Concept Hierarchies in Semantic Web Based upon Global Instances and its Application to Facility Management Database

    Science.gov (United States)

    Takahashi, Hiroki; Nishi, Yuusuke; Gion, Tomohiro; Minami, Shinichi; Fukunaga, Tatsuya; Ogata, Jiro; Yoshie, Osamu

    Semantic Web is the technology which determines the relevance of data over the Web using meta-data and which enables advanced search of global information. It is now desired to develop and apply this technology to many situations of facility management. In facility management, vocabulary should be unified to share the database of facilities for generating optimal maintenance schedule and so on. Under such situations, ontology databases are usually used to describe composition or hierarchy of facility parts. However, these vocabularies used in databases are not unified even between factories of same company, and this situation causes communication hazard between them. Moreover, concept involved in the hierarchy cannot be corresponded each other. There are some methods to correspond concepts of different hierarchy. But these methods have some defects, because they only attend target hierarchy itself and the number of instances. We propose improved method for corresponding concepts between different concepts' hierarchies, which uses other hierarchies all over the world of Web and the distance of instances to identify their relations. Our method can work even if the sets of instances belonging to the concepts are not identical.

  15. Customizable scientific web-portal for DIII-D nuclear fusion experiment

    Energy Technology Data Exchange (ETDEWEB)

    Abla, G; Kim, E N; Schissel, D P, E-mail: abla@fusion.gat.co [General Atomics, PO Box 85608, San Diego, California 92186-5608 (United States)

    2010-04-01

    Increasing utilization of the Internet and convenient web technologies has made the web-portal a major application interface for remote participation and control of scientific instruments. While web-portals have provided a centralized gateway for multiple computational services, the amount of visual output often is overwhelming due to the high volume of data generated by complex scientific instruments and experiments. Since each scientist may have different priorities and areas of interest in the experiment, filtering and organizing information based on the individual user's need can increase the usability and efficiency of a web-portal. DIII-D is the largest magnetic nuclear fusion device in the US. A web-portal has been designed to support the experimental activities of DIII-D researchers worldwide. It offers a customizable interface with personalized page layouts and list of services for users to select. Each individual user can create a unique working environment to fit his own needs and interests. Customizable services are: real-time experiment status monitoring, diagnostic data access, interactive data analysis and visualization. The web-portal also supports interactive collaborations by providing collaborative logbook, and online instant announcement services. The DIII-D web-portal development utilizes multi-tier software architecture, and Web 2.0 technologies and tools, such as AJAX and Django, to develop a highly-interactive and customizable user interface.

  16. Customizable scientific web-portal for DIII-D nuclear fusion experiment

    International Nuclear Information System (INIS)

    Abla, G; Kim, E N; Schissel, D P

    2010-01-01

    Increasing utilization of the Internet and convenient web technologies has made the web-portal a major application interface for remote participation and control of scientific instruments. While web-portals have provided a centralized gateway for multiple computational services, the amount of visual output often is overwhelming due to the high volume of data generated by complex scientific instruments and experiments. Since each scientist may have different priorities and areas of interest in the experiment, filtering and organizing information based on the individual user's need can increase the usability and efficiency of a web-portal. DIII-D is the largest magnetic nuclear fusion device in the US. A web-portal has been designed to support the experimental activities of DIII-D researchers worldwide. It offers a customizable interface with personalized page layouts and list of services for users to select. Each individual user can create a unique working environment to fit his own needs and interests. Customizable services are: real-time experiment status monitoring, diagnostic data access, interactive data analysis and visualization. The web-portal also supports interactive collaborations by providing collaborative logbook, and online instant announcement services. The DIII-D web-portal development utilizes multi-tier software architecture, and Web 2.0 technologies and tools, such as AJAX and Django, to develop a highly-interactive and customizable user interface.

  17. Customizable scientific web-portal for DIII-D nuclear fusion experiment

    Science.gov (United States)

    Abla, G.; Kim, E. N.; Schissel, D. P.

    2010-04-01

    Increasing utilization of the Internet and convenient web technologies has made the web-portal a major application interface for remote participation and control of scientific instruments. While web-portals have provided a centralized gateway for multiple computational services, the amount of visual output often is overwhelming due to the high volume of data generated by complex scientific instruments and experiments. Since each scientist may have different priorities and areas of interest in the experiment, filtering and organizing information based on the individual user's need can increase the usability and efficiency of a web-portal. DIII-D is the largest magnetic nuclear fusion device in the US. A web-portal has been designed to support the experimental activities of DIII-D researchers worldwide. It offers a customizable interface with personalized page layouts and list of services for users to select. Each individual user can create a unique working environment to fit his own needs and interests. Customizable services are: real-time experiment status monitoring, diagnostic data access, interactive data analysis and visualization. The web-portal also supports interactive collaborations by providing collaborative logbook, and online instant announcement services. The DIII-D web-portal development utilizes multi-tier software architecture, and Web 2.0 technologies and tools, such as AJAX and Django, to develop a highly-interactive and customizable user interface.

  18. Web-Enhanced Instruction and Learning: Findings of a Short- and Long-Term Impact Study and Teacher Use of NASA Web Resources

    Science.gov (United States)

    McCarthy, Marianne C.; Grabowski, Barbara L.; Koszalka, Tiffany

    2003-01-01

    Over a three-year period, researchers and educators from the Pennsylvania State University (PSU), University Park, Pennsylvania, and the NASA Dryden Flight Research Center (DFRC), Edwards, California, worked together to analyze, develop, implement and evaluate materials and tools that enable teachers to use NASA Web resources effectively for teaching science, mathematics, technology and geography. Two conference publications and one technical paper have already been published as part of this educational research series on Web-based instruction and learning. This technical paper, Web-Enhanced Instruction and Learning: Findings of a Short- and Long-Term Impact Study, is the culminating report in this educational research series and is based on the final report submitted to NASA. This report describes the broad spectrum of data gathered from teachers about their experiences using NASA Web resources in the classroom. It also describes participating teachers responses and feedback about the use of the NASA Web-Enhanced Learning Environment Strategies reflection tool on their teaching practices. The reflection tool was designed to help teachers merge the vast array of NASA resources with the best teaching methods, taking into consideration grade levels, subject areas and teaching preferences. The teachers described their attitudes toward technology and innovation in the classroom and their experiences and perceptions as they attempted to integrate Web resources into science, mathematics, technology and geography instruction.

  19. Web-based recruitment: effects of information, organizational brand, and attitudes toward a Web site on applicant attraction.

    Science.gov (United States)

    Allen, David G; Mahto, Raj V; Otondo, Robert F

    2007-11-01

    Recruitment theory and research show that objective characteristics, subjective considerations, and critical contact send signals to prospective applicants about the organization and available opportunities. In the generating applicants phase of recruitment, critical contact may consist largely of interactions with recruitment sources (e.g., newspaper ads, job fairs, organization Web sites); however, research has yet to fully address how all 3 types of signaling mechanisms influence early job pursuit decisions in the context of organizational recruitment Web sites. Results based on data from 814 student participants searching actual organization Web sites support and extend signaling and brand equity theories by showing that job information (directly) and organization information (indirectly) are related to intentions to pursue employment when a priori perceptions of image are controlled. A priori organization image is related to pursuit intentions when subsequent information search is controlled, but organization familiarity is not, and attitudes about a recruitment source also influence attraction and partially mediate the effects of organization information. Theoretical and practical implications for recruitment are discussed. (c) 2007 APA

  20. 90Sr-90Y radionuclide generator based on ionex chromatography. Part 1 - project

    International Nuclear Information System (INIS)

    Miler, V.; Budsky, F.; Malek, Z.

    2003-09-01

    This part contains a proposal for the generator column design, materials to be used (chemicals, ionexes) and technological procedures. The proposal was inspired by the 90 Sr- 90 Y generator operated by Zfk Rossendorf. The aim was to develop and launch a generator for the preparation of carrier-free 90 Y in the form of [ 90 Y] chloride solution in dilute hydrochloric acid. The separation of Y from Sr is based on ionex chromatography by sorbing the two radionuclides on a catex. While Sr remains sorbed, 90 Y is eluted with lithium citrate. During this process, 90 Y is bonded in a citrate complex which, having a negative charge, is subsequently trapped by an anex. A guard column is inserted before the anex column to trap any traces of 90 Sr. 90 Y is eluted from the anex in the yttrium chloride form by using dilute hydrochloric acid. The product from the generator can be used for the preparation of [ 90 Y] - Fe colloid injection or [ 90 Y] - yttrium citrate injection for intra-articular application or for the development of monoclonal antibodies and peptides

  1. A Web API ecosystem through feature-based reuse

    NARCIS (Netherlands)

    Verborgh, Ruben; Dumontier, Michel

    2016-01-01

    The current Web API landscape does not scale well: every API requires its own hardcoded clients in an unusually short-lived, tightly coupled relationship of highly subjective quality. This directly leads to inflated development costs, and prevents the design of a more intelligent generation of

  2. Opportunities and Challenges of using Web 2.0 Technologies in Government: A Conceptual Review

    OpenAIRE

    Sivarajah, U; Weerakkody, V; Irani, Z

    2016-01-01

    Public administration has endured signification transformation over the last decade enabled largely through Information and Communication Technology. In recent times, second generation web technologies (Web 2.0) such as social media and net-working sites are increasingly being used by governments for its digital activities rang-ing from public relations to knowledge management. However, as Web 2.0 technolo-gies are more interactive than the traditional models of information provision or crea-...

  3. PERANCANGAN WEB BASED LEARNING SEBAGAI MEDIA PEMBELAJARAN BERBASIS ICT

    OpenAIRE

    Ricky Firmansyah; Iis Saidah

    2016-01-01

    ABSTRACT The media is very important component of communication process. The effectiveness of media is very influential on extent to which a communication role will be accepted by the audience with fast and precise, or vice versa. E-Learning is present as ICT based learning media that allows students and teachers interact in different places. Web Based Learning (WBL) is used as one part of the E-Learning. This study focuses on designing web-based ICT as a learning medium that is used for ...

  4. Medical and Biohazardous Waste Generator's Guide (Revision2)

    Energy Technology Data Exchange (ETDEWEB)

    Waste Management Group

    2006-11-29

    These guidelines describe procedures to comply with all Federal and State laws and regulations and Lawrence Berkeley National Laboratory (LBNL) policy applicable to State-regulated medical and unregulated, but biohazardous, waste (medical/biohazardous waste). These guidelines apply to all LBNL personnel who: (1) generate and/or store medical/biohazardous waste, (2) supervise personnel who generate medical/biohazardous waste, or (3) manage a medical/biohazardous waste pickup location. Personnel generating biohazardous waste at the Joint Genome Institute/Production Genomics Facility (JGI/PGF) are referred to the guidelines contained in Section 9. Section 9 is the only part of these guidelines that apply to JGI/PGF. Medical/biohazardous waste referred to in this Web site includes biohazardous, sharps, pathological and liquid waste. Procedures for proper storage and disposal are summarized in the Solid Medical/Biohazardous Waste Disposal Procedures Chart. Contact the Waste Management Group at 486-7663 if you have any questions regarding medical/biohazardous waste management.

  5. Web 2.0 Technologies for Effective Knowledge Management in Organizations: A Qualitative Analysis

    Science.gov (United States)

    Nath, Anupam Kumar

    2012-01-01

    A new generation of Internet-based collaborative tools, commonly known as Web 2.0, has increased in popularity, availability, and power in the last few years (Kane and Fichman, 2009). Web 2.0 is a set of Internet-based applications that harness network effects by facilitating collaborative and participative computing (O'Reilly, 2006).…

  6. Applying semantic web services to enterprise web

    OpenAIRE

    Hu, Y; Yang, Q P; Sun, X; Wei, P

    2008-01-01

    Enterprise Web provides a convenient, extendable, integrated platform for information sharing and knowledge management. However, it still has many drawbacks due to complexity and increasing information glut, as well as the heterogeneity of the information processed. Research in the field of Semantic Web Services has shown the possibility of adding higher level of semantic functionality onto the top of current Enterprise Web, enhancing usability and usefulness of resource, enabling decision su...

  7. VisBOL: Web-Based Tools for Synthetic Biology Design Visualization.

    Science.gov (United States)

    McLaughlin, James Alastair; Pocock, Matthew; Mısırlı, Göksel; Madsen, Curtis; Wipat, Anil

    2016-08-19

    VisBOL is a Web-based application that allows the rendering of genetic circuit designs, enabling synthetic biologists to visually convey designs in SBOL visual format. VisBOL designs can be exported to formats including PNG and SVG images to be embedded in Web pages, presentations and publications. The VisBOL tool enables the automated generation of visualizations from designs specified using the Synthetic Biology Open Language (SBOL) version 2.0, as well as a range of well-known bioinformatics formats including GenBank and Pigeoncad notation. VisBOL is provided both as a user accessible Web site and as an open-source (BSD) JavaScript library that can be used to embed diagrams within other content and software.

  8. THE PILOT STUDY OF CHARACTERISTICS OF HOUSEHOLD WASTE GENERATED IN SUBURBAN PARTS OF RURAL AREAS

    Directory of Open Access Journals (Sweden)

    Aleksandra Steinhoff-Wrześniewska

    2015-02-01

    Full Text Available The subject of the studies were waste generated in suburban households, in 3-bag system. The sum of wastes generated during the four analyzed seasons (spring, summer, autumn, winter – 1 year, in the households under study, per 1 person, amounted to 170,3 kg (in wet mass basis. For 1 person, most domestic waste was generated in autumn – 45,5 kg per capita and the least in winter – 39,0 kg per capita. The analysis performed of sieved composition (size fraction showed that fractions: >100 mm, 40–100 mm, 20–40 mm constituted totally 80% of the mass of wastes (average in a year. The lowest fraction (<10 mm, whose significant part constitutes ashes, varied depending on the season of year: from 3.5% to 12.8%. In the morphological composition of the households analyzed (on average in 4 seasons, biowastes totally formed over 53% of the whole mass of wastes. A significant part of waste generated were also glass waste (10,7% average per year and disposable nappies (8,3% average per year. The analysis of basic chemical components of biowastes showed that in case of utilizing them for production of compost, it would be necessary to modify (correct the ratios C/N and C/P. Analysis of the chemical composition showed that the biowastes were characterized by very high moisture content and neutral pH.

  9. Food Web Assembly Rules for Generalized Lotka-Volterra Equations.

    Directory of Open Access Journals (Sweden)

    Jan O Haerter

    2016-02-01

    Full Text Available In food webs, many interacting species coexist despite the restrictions imposed by the competitive exclusion principle and apparent competition. For the generalized Lotka-Volterra equations, sustainable coexistence necessitates nonzero determinant of the interaction matrix. Here we show that this requirement is equivalent to demanding that each species be part of a non-overlapping pairing, which substantially constrains the food web structure. We demonstrate that a stable food web can always be obtained if a non-overlapping pairing exists. If it does not, the matrix rank can be used to quantify the lack of niches, corresponding to unpaired species. For the species richness at each trophic level, we derive the food web assembly rules, which specify sustainable combinations. In neighboring levels, these rules allow the higher level to avert competitive exclusion at the lower, thereby incorporating apparent competition. In agreement with data, the assembly rules predict high species numbers at intermediate levels and thinning at the top and bottom. Using comprehensive food web data, we demonstrate how omnivores or parasites with hosts at multiple trophic levels can loosen the constraints and help obtain coexistence in food webs. Hence, omnivory may be the glue that keeps communities intact even under extinction or ecological release of species.

  10. Food Web Assembly Rules for Generalized Lotka-Volterra Equations.

    Science.gov (United States)

    Haerter, Jan O; Mitarai, Namiko; Sneppen, Kim

    2016-02-01

    In food webs, many interacting species coexist despite the restrictions imposed by the competitive exclusion principle and apparent competition. For the generalized Lotka-Volterra equations, sustainable coexistence necessitates nonzero determinant of the interaction matrix. Here we show that this requirement is equivalent to demanding that each species be part of a non-overlapping pairing, which substantially constrains the food web structure. We demonstrate that a stable food web can always be obtained if a non-overlapping pairing exists. If it does not, the matrix rank can be used to quantify the lack of niches, corresponding to unpaired species. For the species richness at each trophic level, we derive the food web assembly rules, which specify sustainable combinations. In neighboring levels, these rules allow the higher level to avert competitive exclusion at the lower, thereby incorporating apparent competition. In agreement with data, the assembly rules predict high species numbers at intermediate levels and thinning at the top and bottom. Using comprehensive food web data, we demonstrate how omnivores or parasites with hosts at multiple trophic levels can loosen the constraints and help obtain coexistence in food webs. Hence, omnivory may be the glue that keeps communities intact even under extinction or ecological release of species.

  11. Non-Relative Value Unit-Generating Activities Represent One-Fifth of Academic Neuroradiologist Productivity.

    Science.gov (United States)

    Wintermark, M; Zeineh, M; Zaharchuk, G; Srivastava, A; Fischbein, N

    2016-07-01

    A neuroradiologist's activity includes many tasks beyond interpreting relative value unit-generating imaging studies. Our aim was to test a simple method to record and quantify the non-relative value unit-generating clinical activity represented by consults and clinical conferences, including tumor boards. Four full-time neuroradiologists, working an average of 50% clinical and 50% academic activity, systematically recorded all the non-relative value unit-generating consults and conferences in which they were involved during 3 months by using a simple, Web-based, computer-based application accessible from smartphones, tablets, or computers. The number and type of imaging studies they interpreted during the same period and the associated relative value units were extracted from our billing system. During 3 months, the 4 neuroradiologists working an average of 50% clinical activity interpreted 4241 relative value unit-generating imaging studies, representing 8152 work relative value units. During the same period, they recorded 792 non-relative value unit-generating study reviews as part of consults and conferences (not including reading room consults), representing 19% of the interpreted relative value unit-generating imaging studies. We propose a simple Web-based smartphone app to record and quantify non-relative value unit-generating activities including consults, clinical conferences, and tumor boards. The quantification of non-relative value unit-generating activities is paramount in this time of a paradigm shift from volume to value. It also represents an important tool for determining staffing levels, which cannot be performed on the basis of relative value unit only, considering the importance of time spent by radiologists on non-relative value unit-generating activities. It may also influence payment models from medical centers to radiology departments or practices. © 2016 by American Journal of Neuroradiology.

  12. Constructing a web recommender system using web usage mining and user’s profiles

    Directory of Open Access Journals (Sweden)

    T. Mombeini

    2014-12-01

    Full Text Available The World Wide Web is a great source of information, which is nowadays being widely used due to the availability of useful information changing, dynamically. However, the large number of webpages often confuses many users and it is hard for them to find information on their interests. Therefore, it is necessary to provide a system capable of guiding users towards their desired choices and services. Recommender systems search among a large collection of user interests and recommend those, which are likely to be favored the most by the user. Web usage mining was designed to function on web server records, which are included in user search results. Therefore, recommender servers use the web usage mining technique to predict users’ browsing patterns and recommend those patterns in the form of a suggestion list. In this article, a recommender system based on web usage mining phases (online and offline was proposed. In the offline phase, the first step is to analyze user access records to identify user sessions. Next, user profiles are built using data from server records based on the frequency of access to pages, the time spent by the user on each page and the date of page view. Date is of importance since it is more possible for users to request new pages more than old ones and old pages are less probable to be viewed, as users mostly look for new information. Following the creation of user profiles, users are categorized in clusters using the Fuzzy C-means clustering algorithm and S(c criterion based on their similarities. In the online phase, a neural network is offered to identify the suggested model while online suggestions are generated using the suggestion module for the active user. Search engines analyze suggestion lists based on rate of user interest in pages and page rank and finally suggest appropriate pages to the active user. Experiments show that the proposed method of predicting user recent requested pages has more accuracy and

  13. Product line based ontology development for semantic web service

    DEFF Research Database (Denmark)

    Zhang, Weishan; Kunz, Thomas

    2006-01-01

    Ontology is recognized as a key technology for the success of the Semantic Web. Building reusable and evolve-able ontologies in order to cope with ontology evolution and requirement changes is increasingly important. But the existing methodologies and tools fail to support effective ontology reuse...... will lead to the initial implementation of the meta-onotologies using design by reuse and with the objective of design for reuse. After that step new ontologies could be generated by reusing these meta-ontologies. We demonstrate our approach with a Semantic Web Service application to show how to build...

  14. Deep Web: aproximaciones a la ciber irresponsabilidad

    Directory of Open Access Journals (Sweden)

    Dulce María Bautista Luzardo

    2015-01-01

    Full Text Available La Deep web o Hard web es una parte gigantesca de las plataformas virtuales indetectables donde ocurren ciberacciones que tienen como precedente el ocultamiento de la identidad del usuario y han dado pie a la tergiversación del concepto de persona y a la utilización de la web de una manera irresponsable —en algunos casos— para causar desazón, para perseguir o a veces hackear bancos, entidades y cuentas privadas. Este es un artículo de reflexión para analizar los alcances de la práctica de esconder acciones en Internet y de modificar el rostro en la cibersociedad contemporánea. Con esta reflexión se pretende llamar la atención acerca de la responsabilidad que tenemos a la hora de entrar en el mundo del Internet y se analiza los peligros que estas prácticas conllevan.

  15. Strategies to address participant misrepresentation for eligibility in Web-based research.

    Science.gov (United States)

    Kramer, Jessica; Rubin, Amy; Coster, Wendy; Helmuth, Eric; Hermos, John; Rosenbloom, David; Moed, Rich; Dooley, Meghan; Kao, Ying-Chia; Liljenquist, Kendra; Brief, Deborah; Enggasser, Justin; Keane, Terence; Roy, Monica; Lachowicz, Mark

    2014-03-01

    Emerging methodological research suggests that the World Wide Web ("Web") is an appropriate venue for survey data collection, and a promising area for delivering behavioral intervention. However, the use of the Web for research raises concerns regarding sample validity, particularly when the Web is used for recruitment and enrollment. The purpose of this paper is to describe the challenges experienced in two different Web-based studies in which participant misrepresentation threatened sample validity: a survey study and an online intervention study. The lessons learned from these experiences generated three types of strategies researchers can use to reduce the likelihood of participant misrepresentation for eligibility in Web-based research. Examples of procedural/design strategies, technical/software strategies and data analytic strategies are provided along with the methodological strengths and limitations of specific strategies. The discussion includes a series of considerations to guide researchers in the selection of strategies that may be most appropriate given the aims, resources and target population of their studies. Copyright © 2014 John Wiley & Sons, Ltd.

  16. Designing Effective Web Forms for Older Web Users

    Science.gov (United States)

    Li, Hui; Rau, Pei-Luen Patrick; Fujimura, Kaori; Gao, Qin; Wang, Lin

    2012-01-01

    This research aims to provide insight for web form design for older users. The effects of task complexity and information structure of web forms on older users' performance were examined. Forty-eight older participants with abundant computer and web experience were recruited. The results showed significant differences in task time and error rate…

  17. Informatics in radiology: automated Web-based graphical dashboard for radiology operational business intelligence.

    Science.gov (United States)

    Nagy, Paul G; Warnock, Max J; Daly, Mark; Toland, Christopher; Meenan, Christopher D; Mezrich, Reuben S

    2009-11-01

    Radiology departments today are faced with many challenges to improve operational efficiency, performance, and quality. Many organizations rely on antiquated, paper-based methods to review their historical performance and understand their operations. With increased workloads, geographically dispersed image acquisition and reading sites, and rapidly changing technologies, this approach is increasingly untenable. A Web-based dashboard was constructed to automate the extraction, processing, and display of indicators and thereby provide useful and current data for twice-monthly departmental operational meetings. The feasibility of extracting specific metrics from clinical information systems was evaluated as part of a longer-term effort to build a radiology business intelligence architecture. Operational data were extracted from clinical information systems and stored in a centralized data warehouse. Higher-level analytics were performed on the centralized data, a process that generated indicators in a dynamic Web-based graphical environment that proved valuable in discussion and root cause analysis. Results aggregated over a 24-month period since implementation suggest that this operational business intelligence reporting system has provided significant data for driving more effective management decisions to improve productivity, performance, and quality of service in the department.

  18. Automatic Hidden-Web Table Interpretation by Sibling Page Comparison

    Science.gov (United States)

    Tao, Cui; Embley, David W.

    The longstanding problem of automatic table interpretation still illudes us. Its solution would not only be an aid to table processing applications such as large volume table conversion, but would also be an aid in solving related problems such as information extraction and semi-structured data management. In this paper, we offer a conceptual modeling solution for the common special case in which so-called sibling pages are available. The sibling pages we consider are pages on the hidden web, commonly generated from underlying databases. We compare them to identify and connect nonvarying components (category labels) and varying components (data values). We tested our solution using more than 2,000 tables in source pages from three different domains—car advertisements, molecular biology, and geopolitical information. Experimental results show that the system can successfully identify sibling tables, generate structure patterns, interpret tables using the generated patterns, and automatically adjust the structure patterns, if necessary, as it processes a sequence of hidden-web pages. For these activities, the system was able to achieve an overall F-measure of 94.5%.

  19. 75 FR 27986 - Electronic Filing System-Web (EFS-Web) Contingency Option

    Science.gov (United States)

    2010-05-19

    ...] Electronic Filing System--Web (EFS-Web) Contingency Option AGENCY: United States Patent and Trademark Office... contingency option when the primary portal to EFS-Web has an unscheduled outage. Previously, the entire EFS-Web system is not available to the users during such an outage. The contingency option in EFS-Web will...

  20. Fermilab joins in global live Web cast

    CERN Multimedia

    Polansek, Tom

    2005-01-01

    From 2 to 3:30 p.m., Lederman, who won the Nobel Prize for physics in 1988, will host his own wacky, science-centered talk show at Fermi National Accelerator Laboratory as part of a lvie, 12-hour, international Web cast celebrating Albert Einstein and the world Year of Physics (2/3 page)

  1. Web-Based Problem-Solving Assignment and Grading System

    Science.gov (United States)

    Brereton, Giles; Rosenberg, Ronald

    2014-11-01

    In engineering courses with very specific learning objectives, such as fluid mechanics and thermodynamics, it is conventional to reinforce concepts and principles with problem-solving assignments and to measure success in problem solving as an indicator of student achievement. While the modern-day ease of copying and searching for online solutions can undermine the value of traditional assignments, web-based technologies also provide opportunities to generate individualized well-posed problems with an infinite number of different combinations of initial/final/boundary conditions, so that the probability of any two students being assigned identical problems in a course is vanishingly small. Such problems can be designed and programmed to be: single or multiple-step, self-grading, allow students single or multiple attempts; provide feedback when incorrect; selectable according to difficulty; incorporated within gaming packages; etc. In this talk, we discuss the use of a homework/exam generating program of this kind in a single-semester course, within a web-based client-server system that ensures secure operation.

  2. Geospatial semantic web

    CERN Document Server

    Zhang, Chuanrong; Li, Weidong

    2015-01-01

    This book covers key issues related to Geospatial Semantic Web, including geospatial web services for spatial data interoperability; geospatial ontology for semantic interoperability; ontology creation, sharing, and integration; querying knowledge and information from heterogeneous data source; interfaces for Geospatial Semantic Web, VGI (Volunteered Geographic Information) and Geospatial Semantic Web; challenges of Geospatial Semantic Web; and development of Geospatial Semantic Web applications. This book also describes state-of-the-art technologies that attempt to solve these problems such as WFS, WMS, RDF, OWL, and GeoSPARQL, and demonstrates how to use the Geospatial Semantic Web technologies to solve practical real-world problems such as spatial data interoperability.

  3. 3D Web-based HMI with WebGL Rendering Performance

    Directory of Open Access Journals (Sweden)

    Muennoi Atitayaporn

    2016-01-01

    Full Text Available An HMI, or Human-Machine Interface, is a software allowing users to communicate with a machine or automation system. It usually serves as a display section in SCADA (Supervisory Control and Data Acquisition system for device monitoring and control. In this papper, a 3D Web-based HMI with WebGL (Web-based Graphics Library rendering performance is presented. The main purpose of this work is to attempt to reduce the limitations of traditional 3D web HMI using the advantage of WebGL. To evaluate the performance, frame rate and frame time metrics were used. The results showed 3D Web-based HMI can maintain the frame rate 60FPS for #cube=0.5K/0.8K, 30FPS for #cube=1.1K/1.6K when it was run on Internet Explorer and Chrome respectively. Moreover, the study found that 3D Web-based HMI using WebGL contains similar frame time in each frame even though the numbers of cubes are up to 5K. This indicated stuttering incurred less in the proposed 3D Web-based HMI compared to the chosen commercial HMI product.

  4. 07051 Abstracts Collection -- Programming Paradigms for the Web: Web Programming and Web Services

    OpenAIRE

    Hull, Richard; Thiemann, Peter; Wadler, Philip

    2007-01-01

    From 28.01. to 02.02.2007, the Dagstuhl Seminar 07051 ``Programming Paradigms for the Web: Web Programming and Web Services'' was held in the International Conference and Research Center (IBFI), Schloss Dagstuhl. During the seminar, several participants presented their current research, and ongoing work and open problems were discussed. Abstracts of the presentations given during the seminar as well as abstracts of seminar results and ideas are put together in this paper. The firs...

  5. Bridging the gap between legacy services and Web Services

    DEFF Research Database (Denmark)

    Bissyandé, Tegawendé; Réveillère, Laurent; Bromberg, Yérom-David

    2010-01-01

    itself. In this paper, we introduce a generative language based approach for constructing wrappers to facilitate the migration of legacy service functionalities to Web Services. To this end, we have designed the Janus domain-specific language, which provides developers with a high-level way to describe......Web Services is an increasingly used instantiation of Service-Oriented Architectures (SOA) that relies on standard Internet protocols to produce services that are highly interoperable. Other types of services, relying on legacy application layer protocols, however, cannot be composed directly....... A promising solution is to implement wrappers to translate between the application layer protocols and the WS protocol. Doing so manually, however, requires a high level of expertise, in the relevant application layer protocols, in low-level network and system programming, and in the Web Service paradigm...

  6. THE DIFFERENCE BETWEEN DEVELOPING SINGLE PAGE APPLICATION AND TRADITIONAL WEB APPLICATION BASED ON MECHATRONICS ROBOT LABORATORY ONAFT APPLICATION

    Directory of Open Access Journals (Sweden)

    V. Solovei

    2018-04-01

    Full Text Available Today most of desktop and mobile applications have analogues in the form of web-based applications.  With evolution of development technologies and web technologies web application increased in functionality to desktop applications. The Web application consists of two parts of the client part and the server part. The client part is responsible for providing the user with visual information through the browser. The server part is responsible for processing and storing data.MPA appeared simultaneously with the Internet. Multiple-page applications work in a "traditional" way. Every change eg. display the data or submit data back to the server. With the advent of AJAX, MPA learned to load not the whole page, but only a part of it, which eventually led to the appearance of the SPA. SPA is the principle of development when only one page is transferred to the client part, and the content is downloaded only to a certain part of the page, without rebooting it, which allows to speed up the application and simplify the user experience of using the application to the level of desktop applications.Based on the SPA, the Mechatronics Robot Laboratory ONAFT application was designed to automate the management process. The application implements the client-server architecture. The server part consists of a RESTful API, which allows you to get unified access to the application functionality, and a database for storing information. Since the client part is a spa, this allows you to reduce the load on the connection to the server and improve the user experience

  7. Developing web map application based on user centered design

    Directory of Open Access Journals (Sweden)

    Petr Voldan

    2012-03-01

    Full Text Available User centred design is an approach in process of development any kind of human product where the main idea is to create a product for the end user. This article presents User centred design method in developing web mapping services. This method can be split into four main phases – user research, creation of concepts, developing with usability research and lunch of product. The article describes each part of this phase with an aim to provide guidelines for developers and primarily with an aim to improve the usability of web mapping services.

  8. BaBar - A Community Web Site in an Organizational Setting

    Energy Technology Data Exchange (ETDEWEB)

    White, Bebo

    2003-07-10

    The BABAR Web site was established in 1993 at the Stanford Linear Accelerator Center (SLAC) to support the BABAR experiment, to report its results, and to facilitate communication among its scientific and engineering collaborators, currently numbering about 600 individuals from 75 collaborating institutions in 10 countries. The BABAR Web site is, therefore, a community Web site. At the same time it is hosted at SLAC and funded by agencies that demand adherence to policies decided under different priorities. Additionally, the BABAR Web administrators deal with the problems that arise during the course of managing users, content, policies, standards, and changing technologies. Desired solutions to some of these problems may be incompatible with the overall administration of the SLAC Web sites and/or the SLAC policies and concerns. There are thus different perspectives of the same Web site and differing expectations in segments of the SLAC population which act as constraints and challenges in any review or re-engineering activities. Web Engineering, which post-dates the BABAR Web, has aimed to provide a comprehensive understanding of all aspects of Web development. This paper reports on the first part of a recent review of application of Web Engineering methods to the BABAR Web site, which has led to explicit user and information models of the BABAR community and how SLAC and the BABAR community relate and react to each other. The paper identifies the issues of a community Web site in a hierarchical, semi-governmental sector and formulates a strategy for periodic reviews of BABAR and similar sites. A separate paper reports on the findings of a user survey and selected interviews with users, along with their implications and recommendations for future.

  9. BaBar - A Community Web Site in an Organizational Setting

    International Nuclear Information System (INIS)

    White, Bebo

    2003-01-01

    The BABAR Web site was established in 1993 at the Stanford Linear Accelerator Center (SLAC) to support the BABAR experiment, to report its results, and to facilitate communication among its scientific and engineering collaborators, currently numbering about 600 individuals from 75 collaborating institutions in 10 countries. The BABAR Web site is, therefore, a community Web site. At the same time it is hosted at SLAC and funded by agencies that demand adherence to policies decided under different priorities. Additionally, the BABAR Web administrators deal with the problems that arise during the course of managing users, content, policies, standards, and changing technologies. Desired solutions to some of these problems may be incompatible with the overall administration of the SLAC Web sites and/or the SLAC policies and concerns. There are thus different perspectives of the same Web site and differing expectations in segments of the SLAC population which act as constraints and challenges in any review or re-engineering activities. Web Engineering, which post-dates the BABAR Web, has aimed to provide a comprehensive understanding of all aspects of Web development. This paper reports on the first part of a recent review of application of Web Engineering methods to the BABAR Web site, which has led to explicit user and information models of the BABAR community and how SLAC and the BABAR community relate and react to each other. The paper identifies the issues of a community Web site in a hierarchical, semi-governmental sector and formulates a strategy for periodic reviews of BABAR and similar sites. A separate paper reports on the findings of a user survey and selected interviews with users, along with their implications and recommendations for future

  10. Flow Webs: Mechanism and Architecture for the Implementation of Sensor Webs

    Science.gov (United States)

    Gorlick, M. M.; Peng, G. S.; Gasster, S. D.; McAtee, M. D.

    2006-12-01

    The sensor web is a distributed, federated infrastructure much like its predecessors, the internet and the world wide web. It will be a federation of many sensor webs, large and small, under many distinct spans of control, that loosely cooperates and share information for many purposes. Realistically, it will grow piecemeal as distinct, individual systems are developed and deployed, some expressly built for a sensor web while many others were created for other purposes. Therefore, the architecture of the sensor web is of fundamental import and architectural strictures that inhibit innovation, experimentation, sharing or scaling may prove fatal. Drawing upon the architectural lessons of the world wide web, we offer a novel system architecture, the flow web, that elevates flows, sequences of messages over a domain of interest and constrained in both time and space, to a position of primacy as a dynamic, real-time, medium of information exchange for computational services. The flow web captures; in a single, uniform architectural style; the conflicting demands of the sensor web including dynamic adaptations to changing conditions, ease of experimentation, rapid recovery from the failures of sensors and models, automated command and control, incremental development and deployment, and integration at multiple levels—in many cases, at different times. Our conception of sensor webs—dynamic amalgamations of sensor webs each constructed within a flow web infrastructure—holds substantial promise for earth science missions in general, and of weather, air quality, and disaster management in particular. Flow webs, are by philosophy, design and implementation a dynamic infrastructure that permits massive adaptation in real-time. Flows may be attached to and detached from services at will, even while information is in transit through the flow. This concept, flow mobility, permits dynamic integration of earth science products and modeling resources in response to real

  11. Information security threats in web-portals on the open journal systems platform

    Directory of Open Access Journals (Sweden)

    Anton A. Abramov

    2018-05-01

    Full Text Available This article addresses the problem of security threats while working with web portals built on the Open Journal Systems platform. The Open Journal Systems (OJS platform was originally developed as part of the Public Knowledge Project and it is one of the most popular open-source platforms for web journals today. Based on the data available in the Public Knowledge Project, there were more than 10,000 active journals using the open journal systems platform by the end of 2016. A migration of a journal to such advanced and complex platform helps to handle the entire workflow over a single web portal. Therefore it is an important move and only peer-reviewed journals that are part of Russian and Worldwide citation systems go for it. At the same time the problem of keeping privacy for a manuscript before it is published is very important for these journals and for authors who submit it to the journal. The paper describes the most common threats for the web portals on the OJS platform as well as a particular model of the security threats, and suggests the measures that could help to neutralize these threats.

  12. Is there a “net generation” in veterinary medicine? A comparative study on the use of the Internet and Web 2.0 by students and the veterinary profession [Gibt es eine „net-Generation“ in der Tiermedizin? Eine Vergleichende Untersuchung zur Nutzung von Internet und Web 2.0 bei Studierenden und der Tierärzteschaft

    Directory of Open Access Journals (Sweden)

    Tenhaven, Christoph

    2013-02-01

    Full Text Available [english] Introduction: Informal and formal lifelong learning is essential at university and in the workplace. Apart from classical learning techniques, Web 2.0 tools can be used. It is controversial whether there is a so-called net generation amongst people under 30.Aims: To test the hypothesis that a net generation among students and young veterinarians exists.Methods: An online survey of students and veterinarians was conducted in the German-speaking countries which was advertised via online media and traditional print media.Results: 1780 people took part in the survey. Students and veterinarians have different usage patterns regarding social networks (91.9% vs. 69% and IM (55.9% vs. 24.5%. All tools were predominantly used passively and in private, to a lesser extent also professionally and for studying.Outlook: The use of Web 2.0 tools is useful, however, teaching information and media skills, preparing codes of conduct for the internet and verification of user generated content is essential.[german] Einleitung: Informelles und formelles lebenslanges Lernen ist im Studium und Beruf essenziell. Dazu können neben der klassischen Fortbildung auch Web 2.0 Tools benutzt werden. Umstritten in der Literatur ist allerdings, ob es unter den „Unter 30 Jährigen“ eine sogenannte Net-Generation gibt. Ziel: Überprüfung der Hypothese, dass eine Net-Generation unter Studierenden und jungen Tierärzten existiert. Methode: Eine Onlineumfrage unter Studierenden und der Tierärzteschaft wurde im deutschsprachigen Raum durchgeführt, die per Onlinemedien und klassischen Printmedien beworben wurde.Ergebnisse: An der Befragung haben 1780 Personen teilgenommen. Es gibt unterschiedliches Nutzungsverhalten von Studierenden und der Tierärzteschaft bei sozialen Netzwerken (91,9% vs. 69% und IMs (55,9% vs. 24,5%. Alle Tools wurden hauptsächlich passiv und privat genutzt, im geringeren Maße auch für den Beruf und das Studium.Ausblick: Der Einsatz von Web 2

  13. Code Generation from Pragmatics Annotated Coloured Petri Nets

    DEFF Research Database (Denmark)

    Simonsen, Kent Inge

    limited work has been done on transforming CPN model to protocol implementations. The goal of the thesis is to be able to automatically generate high-quality implementations of communication protocols based on CPN models. In this thesis, we develop a methodology for generating implementations of protocols...... third party libraries and the code should be easily usable by third party code. Finally, the code should be readable by developers with expertise on the considered platforms. In this thesis, we show that our code generation approach is able to generate code for a wide range of platforms without altering...... such as games and rich web applications. Finally, we conclude the evaluation of the criteria of our approach by using the WebSocket PA-CPN model to show that we are able to verify fairly large protocols....

  14. Search, Read and Write: An Inquiry into Web Accessibility for People with Dyslexia.

    Science.gov (United States)

    Berget, Gerd; Herstad, Jo; Sandnes, Frode Eika

    2016-01-01

    Universal design in context of digitalisation has become an integrated part of international conventions and national legislations. A goal is to make the Web accessible for people of different genders, ages, backgrounds, cultures and physical, sensory and cognitive abilities. Political demands for universally designed solutions have raised questions about how it is achieved in practice. Developers, designers and legislators have looked towards the Web Content Accessibility Guidelines (WCAG) for answers. WCAG 2.0 has become the de facto standard for universal design on the Web. Some of the guidelines are directed at the general population, while others are targeted at more specific user groups, such as the visually impaired or hearing impaired. Issues related to cognitive impairments such as dyslexia receive less attention, although dyslexia is prevalent in at least 5-10% of the population. Navigation and search are two common ways of using the Web. However, while navigation has received a fair amount of attention, search systems are not explicitly included, although search has become an important part of people's daily routines. This paper discusses WCAG in the context of dyslexia for the Web in general and search user interfaces specifically. Although certain guidelines address topics that affect dyslexia, WCAG does not seem to fully accommodate users with dyslexia.

  15. Design, implementation and practice of JBEI-ICE: an open source biological part registry platform and tools.

    Science.gov (United States)

    Ham, Timothy S; Dmytriv, Zinovii; Plahar, Hector; Chen, Joanna; Hillson, Nathan J; Keasling, Jay D

    2012-10-01

    The Joint BioEnergy Institute Inventory of Composable Elements (JBEI-ICEs) is an open source registry platform for managing information about biological parts. It is capable of recording information about 'legacy' parts, such as plasmids, microbial host strains and Arabidopsis seeds, as well as DNA parts in various assembly standards. ICE is built on the idea of a web of registries and thus provides strong support for distributed interconnected use. The information deposited in an ICE installation instance is accessible both via a web browser and through the web application programming interfaces, which allows automated access to parts via third-party programs. JBEI-ICE includes several useful web browser-based graphical applications for sequence annotation, manipulation and analysis that are also open source. As with open source software, users are encouraged to install, use and customize JBEI-ICE and its components for their particular purposes. As a web application programming interface, ICE provides well-developed parts storage functionality for other synthetic biology software projects. A public instance is available at public-registry.jbei.org, where users can try out features, upload parts or simply use it for their projects. The ICE software suite is available via Google Code, a hosting site for community-driven open source projects.

  16. NEW WEB-BASED ACCESS TO NUCLEAR STRUCTURE DATASETS.

    Energy Technology Data Exchange (ETDEWEB)

    WINCHELL,D.F.

    2004-09-26

    As part of an effort to migrate the National Nuclear Data Center (NNDC) databases to a relational platform, a new web interface has been developed for the dissemination of the nuclear structure datasets stored in the Evaluated Nuclear Structure Data File and Experimental Unevaluated Nuclear Data List.

  17. Working with WebQuests: Making the Web Accessible to Students with Disabilities.

    Science.gov (United States)

    Kelly, Rebecca

    2000-01-01

    This article describes how students with disabilities in regular classes are using the WebQuest lesson format to access the Internet. It explains essential WebQuest principles, creating a draft Web page, and WebQuest components. It offers an example of a WebQuest about salvaging the sunken ships, Titanic and Lusitania. A WebQuest planning form is…

  18. WebGIS based on semantic grid model and web services

    Science.gov (United States)

    Zhang, WangFei; Yue, CaiRong; Gao, JianGuo

    2009-10-01

    As the combination point of the network technology and GIS technology, WebGIS has got the fast development in recent years. With the restriction of Web and the characteristics of GIS, traditional WebGIS has some prominent problems existing in development. For example, it can't accomplish the interoperability of heterogeneous spatial databases; it can't accomplish the data access of cross-platform. With the appearance of Web Service and Grid technology, there appeared great change in field of WebGIS. Web Service provided an interface which can give information of different site the ability of data sharing and inter communication. The goal of Grid technology was to make the internet to a large and super computer, with this computer we can efficiently implement the overall sharing of computing resources, storage resource, data resource, information resource, knowledge resources and experts resources. But to WebGIS, we only implement the physically connection of data and information and these is far from the enough. Because of the different understanding of the world, following different professional regulations, different policies and different habits, the experts in different field will get different end when they observed the same geographic phenomenon and the semantic heterogeneity produced. Since these there are large differences to the same concept in different field. If we use the WebGIS without considering of the semantic heterogeneity, we will answer the questions users proposed wrongly or we can't answer the questions users proposed. To solve this problem, this paper put forward and experienced an effective method of combing semantic grid and Web Services technology to develop WebGIS. In this paper, we studied the method to construct ontology and the method to combine Grid technology and Web Services and with the detailed analysis of computing characteristics and application model in the distribution of data, we designed the WebGIS query system driven by

  19. How to open & operate a financially successful web site design business

    CERN Document Server

    Evans, Charlotte

    2009-01-01

    The Pricing & Ethical Guidelines Handbook published by the Graphic Arts Guild reports that the average cost of designing a Web site for a small corporation can range from 7,750 to 15,000. It is incredibly easy to see the enormous profit potential. Web design businesses can be run part- or full-time and can easily be started in your own home. As such, they are one of the fastest growing segments of the Internet economy. Here is the manual you need to cash in on this highly profitable segment of the industry. This book is a comprehensive and detailed study of the business side of Web site des

  20. Web Project Management

    OpenAIRE

    Suralkar, Sunita; Joshi, Nilambari; Meshram, B B

    2013-01-01

    This paper describes about the need for Web project management, fundamentals of project management for web projects: what it is, why projects go wrong, and what's different about web projects. We also discuss Cost Estimation Techniques based on Size Metrics. Though Web project development is similar to traditional software development applications, the special characteristics of Web Application development requires adaption of many software engineering approaches or even development of comple...

  1. deepTools2: a next generation web server for deep-sequencing data analysis.

    Science.gov (United States)

    Ramírez, Fidel; Ryan, Devon P; Grüning, Björn; Bhardwaj, Vivek; Kilpert, Fabian; Richter, Andreas S; Heyne, Steffen; Dündar, Friederike; Manke, Thomas

    2016-07-08

    We present an update to our Galaxy-based web server for processing and visualizing deeply sequenced data. Its core tool set, deepTools, allows users to perform complete bioinformatic workflows ranging from quality controls and normalizations of aligned reads to integrative analyses, including clustering and visualization approaches. Since we first described our deepTools Galaxy server in 2014, we have implemented new solutions for many requests from the community and our users. Here, we introduce significant enhancements and new tools to further improve data visualization and interpretation. deepTools continue to be open to all users and freely available as a web service at deeptools.ie-freiburg.mpg.de The new deepTools2 suite can be easily deployed within any Galaxy framework via the toolshed repository, and we also provide source code for command line usage under Linux and Mac OS X. A public and documented API for access to deepTools functionality is also available. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  2. UniPrime2: a web service providing easier Universal Primer design.

    Science.gov (United States)

    Boutros, Robin; Stokes, Nicola; Bekaert, Michaël; Teeling, Emma C

    2009-07-01

    The UniPrime2 web server is a publicly available online resource which automatically designs large sets of universal primers when given a gene reference ID or Fasta sequence input by a user. UniPrime2 works by automatically retrieving and aligning homologous sequences from GenBank, identifying regions of conservation within the alignment, and generating suitable primers that can be used to amplify variable genomic regions. In essence, UniPrime2 is a suite of publicly available software packages (Blastn, T-Coffee, GramAlign, Primer3), which reduces the laborious process of primer design, by integrating these programs into a single software pipeline. Hence, UniPrime2 differs from previous primer design web services in that all steps are automated, linked, saved and phylogenetically delimited, only requiring a single user-defined gene reference ID or input sequence. We provide an overview of the web service and wet-laboratory validation of the primers generated. The system is freely accessible at: http://uniprime.batlab.eu. UniPrime2 is licenced under a Creative Commons Attribution Noncommercial-Share Alike 3.0 Licence.

  3. An interactive, web-based tool for genealogical entity resolution

    NARCIS (Netherlands)

    Efremova, I.; Ranjbar-Sahraei, B.; Oliehoek, F.A.; Calders, T.G.K.; Tuyls, K.P.

    2013-01-01

    We demonstrate an interactive, web-based tool which helps historians to do Genealogical Entitiy Resolution. This work has two main goals. First, it uses Machine Learning (ML) algorithms to assist humanites researchers to perform Genealogical Entity Resolution. Second, it facilitates the generation

  4. Cytoscape tools for the web age: D3.js and Cytoscape.js exporters.

    Science.gov (United States)

    Ono, Keiichiro; Demchak, Barry; Ideker, Trey

    2014-01-01

    In this paper we present new data export modules for Cytoscape 3 that can generate network files for Cytoscape.js and D3.js. Cytoscape.js exporter is implemented as a core feature of Cytoscape 3, and D3.js exporter is available as a Cytoscape 3 app. These modules enable users to seamlessly export network and table data sets generated in Cytoscape to popular JavaScript library readable formats. In addition, we implemented template web applications for browser-based interactive network visualization that can be used as basis for complex data visualization applications for bioinformatics research. Example web applications created with these tools demonstrate how Cytoscape works in modern data visualization workflows built with traditional desktop tools and emerging web-based technologies. This interactivity enables researchers more flexibility than with static images, thereby greatly improving the quality of insights researchers can gain from them.

  5. Loser Generated Content

    DEFF Research Database (Denmark)

    Petersen, Søren Mørk

    2008-01-01

    In this article [ 1 ] some of the critical aspects of Web 2.0 are mapped in relation to labor and the production of user generated content. For many years the Internet was considered an apt technology for subversion of capitalism by the Italian post–Marxists. What we have witnessed, however...

  6. Fault diagnosis of generation IV nuclear HTGR components – Part II: The area error enthalpy–entropy graph approach

    International Nuclear Information System (INIS)

    Rand, C.P. du; Schoor, G. van

    2012-01-01

    Highlights: ► Different uncorrelated fault signatures are derived for HTGR component faults. ► A multiple classifier ensemble increases confidence in classification accuracy. ► Detailed simulation model of system is not required for fault diagnosis. - Abstract: The second paper in a two part series presents the area error method for generation of representative enthalpy–entropy (h–s) fault signatures to classify malfunctions in generation IV nuclear high temperature gas-cooled reactor (HTGR) components. The second classifier is devised to ultimately address the fault diagnosis (FD) problem via the proposed methods in a multiple classifier (MC) ensemble. FD is realized by way of different input feature sets to the classification algorithm based on the area and trajectory of the residual shift between the fault-free and the actual operating h–s graph models. The application of the proposed technique is specifically demonstrated for 24 single fault transients considered in the main power system (MPS) of the Pebble Bed Modular Reactor (PBMR). The results show that the area error technique produces different fault signatures with low correlation for all the examined component faults. A brief evaluation of the two fault signature generation techniques is presented and the performance of the area error method is documented using the fault classification index (FCI) presented in Part I of the series. The final part of this work reports the application of the proposed approach for classification of an emulated fault transient in data from the prototype Pebble Bed Micro Model (PBMM) plant. Reference data values are calculated for the plant via a thermo-hydraulic simulation model of the MPS. The results show that the correspondence between the fault signatures, generated via experimental plant data and simulated reference values, are generally good. The work presented in the two part series, related to the classification of component faults in the MPS of different

  7. Promoting Teachers' Positive Attitude towards Web Use: A Study in Web Site Development

    Science.gov (United States)

    Akpinar, Yavuz; Bayramoglu, Yusuf

    2008-01-01

    The purpose of the study was to examine effects of a compact training for developing web sites on teachers' web attitude, as composed of: web self efficacy, perceived web enjoyment, perceived web usefulness and behavioral intention to use the web. To measure the related constructs, the Web Attitude Scale was adapted into Turkish and tested with a…

  8. When the New Application Smell Is Gone: Traditional Intranet Best Practices and Existing Web 2.0 Intranet Infrastructures

    Science.gov (United States)

    Yoose, Becky

    2010-01-01

    With the growth of Web 2.0 library intranets in recent years, many libraries are leaving behind legacy, first-generation intranets. As Web 2.0 intranets multiply and mature, how will traditional intranet best practices--especially in the areas of planning, implementation, and evaluation--translate into an existing Web 2.0 intranet infrastructure?…

  9. A Web-based Visualization System for Three Dimensional Geological Model using Open GIS

    Science.gov (United States)

    Nemoto, T.; Masumoto, S.; Nonogaki, S.

    2017-12-01

    A three dimensional geological model is an important information in various fields such as environmental assessment, urban planning, resource development, waste management and disaster mitigation. In this study, we have developed a web-based visualization system for 3D geological model using free and open source software. The system has been successfully implemented by integrating web mapping engine MapServer and geographic information system GRASS. MapServer plays a role of mapping horizontal cross sections of 3D geological model and a topographic map. GRASS provides the core components for management, analysis and image processing of the geological model. Online access to GRASS functions has been enabled using PyWPS that is an implementation of WPS (Web Processing Service) Open Geospatial Consortium (OGC) standard. The system has two main functions. Two dimensional visualization function allows users to generate horizontal and vertical cross sections of 3D geological model. These images are delivered via WMS (Web Map Service) and WPS OGC standards. Horizontal cross sections are overlaid on the topographic map. A vertical cross section is generated by clicking a start point and an end point on the map. Three dimensional visualization function allows users to visualize geological boundary surfaces and a panel diagram. The user can visualize them from various angles by mouse operation. WebGL is utilized for 3D visualization. WebGL is a web technology that brings hardware-accelerated 3D graphics to the browser without installing additional software. The geological boundary surfaces can be downloaded to incorporate the geologic structure in a design on CAD and model for various simulations. This study was supported by JSPS KAKENHI Grant Number JP16K00158.

  10. A Network of Automatic Control Web-Based Laboratories

    Science.gov (United States)

    Vargas, Hector; Sanchez Moreno, J.; Jara, Carlos A.; Candelas, F. A.; Torres, Fernando; Dormido, Sebastian

    2011-01-01

    This article presents an innovative project in the context of remote experimentation applied to control engineering education. Specifically, the authors describe their experience regarding the analysis, design, development, and exploitation of web-based technologies within the scope of automatic control. This work is part of an inter-university…

  11. What and how children search on the web

    NARCIS (Netherlands)

    Duarte Torres, Sergio; Weber, Ingmar

    2011-01-01

    The Internet has become an important part of the daily life of children as a source of information and leisure activities. Nonetheless, given that most of the content available on the web is aimed at the general public, children are constantly exposed to inappropriate content, either because the

  12. Implications of Web Mercator and its Use in Online Mapping

    Science.gov (United States)

    Battersby, Sarah E.; Finn, Michael P.; Usery, E. Lynn; Yamamoto, Kristina H.

    2014-01-01

    Online interactive maps have become a popular means of communicating with spatial data. In most online mapping systems, Web Mercator has become the dominant projection. While the Mercator projection has a long history of discussion about its inappropriateness for general-purpose mapping, particularly at the global scale, and seems to have been virtually phased out for general-purpose global-scale print maps, it has seen a resurgence in popularity in Web Mercator form. This article theorizes on how Web Mercator came to be widely used for online maps and what this might mean in terms of data display, technical aspects of map generation and distribution, design, and cognition of spatial patterns. The authors emphasize details of where the projection excels and where it does not, as well as some of its advantages and disadvantages for cartographic communication, and conclude with some research directions that may help to develop better solutions to the problem of projections for general-purpose, multi-scale Web mapping.

  13. A Web Service for File-Level Access to Disk Images

    Directory of Open Access Journals (Sweden)

    Sunitha Misra

    2014-07-01

    Full Text Available Digital forensics tools have many potential applications in the curation of digital materials in libraries, archives and museums (LAMs. Open source digital forensics tools can help LAM professionals to extract digital contents from born-digital media and make more informed preservation decisions. Many of these tools have ways to display the metadata of the digital media, but few provide file-level access without having to mount the device or use complex command-line utilities. This paper describes a project to develop software that supports access to the contents of digital media without having to mount or download the entire image. The work examines two approaches in creating this tool: First, a graphical user interface running on a local machine. Second, a web-based application running in web browser. The project incorporates existing open source forensics tools and libraries including The Sleuth Kit and libewf along with the Flask web application framework and custom Python scripts to generate web pages supporting disk image browsing.

  14. Web wisdom how to evaluate and create information quality on the Web

    CERN Document Server

    Alexander, Janet E

    1999-01-01

    Web Wisdom is an essential reference for anyone needing to evaluate or establish information quality on the World Wide Web. The book includes easy to use checklists for step-by-step quality evaluations of virtually any Web page. The checklists can also be used by Web authors to help them ensure quality information on their pages. In addition, Web Wisdom addresses other important issues, such as understanding the ways that advertising and sponsorship may affect the quality of Web information. It features: * a detailed discussion of the items involved in evaluating Web information; * checklists

  15. An Analysis of Video Navigation Behavior for Web Leisure

    Directory of Open Access Journals (Sweden)

    Ying-Han Chang

    2012-12-01

    Full Text Available People nowadays put much emphasis on leisure activities, and web video has gradually become one of the main sources for popular leisure. This article introduces the related concepts of leisure and navigation behavior as well as some recent research topics. Moreover, using YouTube as an experimental setting, the authors invited some experienced web video users and conducted an empirical study on their navigating the web videos for leisure purpose. The study used questionnaires, navigation logs, diaries, and interviews to collect data. Major results show: the subjects watched a variety of video content on the web either from traditional media or user-generated video; these videos can meet their leisure needs of both the broad and personal interests; during the navigation process, each subject quite focuses on video leisure, and is willingly to explore unknown videos; however, within a limited amount of time for leisure, a balance between leisure and rest becomes an issue of achieving real relaxation, which is worth of further attention. [Article content in Chinese

  16. Large-area sheet task: Advanced dendritic-web-growth development

    Science.gov (United States)

    Duncan, C. S.; Seidensticker, R. G.; Mchugh, J. P.; Schruben, J.

    1983-01-01

    Thermally generated stresses in the growing web crystal were reduced. These stresses, which if too high cause the ribbon to degenerate, were reduced by a factor of three, resulting in the demonstrated growth of high-quality web crystals to widths of 5.4 cm. This progress was brought about chiefly by the application of thermal models to the development of low-stress growth configurations. A new temperature model was developed which can analyze the thermal effects of much more complex lid and top shield configurations than was possible with the old lumped shield model. Growth experiments which supplied input data such as actual shield temperature and melt levels were used to verify the modeling results. Desirable modifications in the melt level-sensing circuitry were made in the new experimental web growth furnace, and this furnace has been used to carry out growth experiments under steady-state conditions. New growth configurations were tested in long growth runs at Westinghouse AESD which produced wider, lower stress and higher quality web crystals than designs previously used.

  17. Designing next-generation platforms for evaluating scientific output: What scientists can learn from the social web

    Directory of Open Access Journals (Sweden)

    Tal eYarkoni

    2012-10-01

    Full Text Available Traditional pre-publication peer review of scientific output is a slow, inefficient, and unreliable process. Efforts to replace or supplement traditional evaluation models with open evaluation platforms that leverage advances in information technology are slowly gaining traction, but remain in the early stages of design and implementation. Here I discuss a number of considerations relevant to the development of such platforms. I focus particular attention on three core elements that next-generation evaluation platforms should strive to emphasize, including (a open and transparent access to accumulated evaluation data, (b personalized and highly customizable performance metrics, and (c appropriate short-term incentivization of the userbase. Because all of these elements have already been successfully implemented on a large scale in hundreds of existing social web applications, I argue that development of new scientific evaluation platforms should proceed largely by adapting existing techniques rather than engineering entirely new evaluation mechanisms. Successful implementation of open evaluation platforms has the potential to substantially advance both the pace and the quality of scientific publication and evaluation, and the scientific community has a vested interest in shifting towards such models as soon as possible.

  18. WEB 2.0 TOOLS FOR SUPPORTING TEACHING

    Directory of Open Access Journals (Sweden)

    Angelos KONSTANTINIDIS

    2013-10-01

    Full Text Available Web 2.0 tools provide enormous opportunities for teaching and learning, yet their application in education is still underdeveloped. What is more, it is no longer possible for teachers to ignore such a technological advance, while they are expected to provide students with opportunities to take control of their learning. However, teachers are still reluctant with technology integration. This paper introduces four Web 2.0 tools; Blogger, StripGenerator, Go!Animate, and Google Forms, that are free and easy to use, in an effort to motivate teachers with low technological skills in integrating them into their instruction practices. The aforementioned tools comprised the curriculum in a blended-learning professional development course for in-service teachers and attracted many favourable comments from the participants.

  19. Professional WebGL Programming Developing 3D Graphics for the Web

    CERN Document Server

    Anyuru, Andreas

    2012-01-01

    Everything you need to know about developing hardware-accelerated 3D graphics with WebGL! As the newest technology for creating 3D graphics on the web, in both games, applications, and on regular websites, WebGL gives web developers the capability to produce eye-popping graphics. This book teaches you how to use WebGL to create stunning cross-platform apps. The book features several detailed examples that show you how to develop 3D graphics with WebGL, including explanations of code snippets that help you understand the why behind the how. You will also develop a stronger understanding of W

  20. WebCom: A Model for Understanding Web Site Communication

    DEFF Research Database (Denmark)

    Godsk, Mikkel; Petersen, Anja Bechmann

    2008-01-01

    of the approaches' strengths. Furthermore, it is discussed and shortly demonstrated how WebCom can be used for analytical and design purposes with YouTube as an example. The chapter concludes that WebCom is able to serve as a theoretically-based model for understanding complex Web site communication situations...

  1. SELECTION OF ONTOLOGY FOR WEB SERVICE DESCRIPTION LANGUAGE TO ONTOLOGY WEB LANGUAGE CONVERSION

    OpenAIRE

    J. Mannar Mannan; M. Sundarambal; S. Raghul

    2014-01-01

    Semantic web is to extend the current human readable web to encoding some of the semantic of resources in a machine processing form. As a Semantic web component, Semantic Web Services (SWS) uses a mark-up that makes the data into detailed and sophisticated machine readable way. One such language is Ontology Web Language (OWL). Existing conventional web service annotation can be changed to semantic web service by mapping Web Service Description Language (WSDL) with the semantic annotation of O...

  2. Web Viz 2.0: A versatile suite of tools for collaboration and visualization

    Science.gov (United States)

    Spencer, C.; Yuen, D. A.

    2012-12-01

    Most scientific applications on the web fail to realize the full collaborative potential of the internet by not utilizing web 2.0 technology. To relieve users from the struggle with software tools and allow them to focus on their research, new software developed for scientists and researchers must harness the full suite of web technology. For several years WebViz 1.0 enabled researchers with any web accessible device to interact with the peta-scale data generated by the Hierarchical Volume Renderer (HVR) system. We have developed a new iteration of WebViz that can be easily interfaced with many problem domains in addition to HVR by employing the best practices of software engineering and object-oriented programming. This is done by separating the core WebViz system from domain specific code at an interface, leveraging inheritance and polymorphism to allow newly developed modules access to the core services. We employed several design patterns (model-view-controller, singleton, observer, and application controller) to engineer this highly modular system implemented in Java.

  3. World Wide Web Metaphors for Search Mission Data

    Science.gov (United States)

    Norris, Jeffrey S.; Wallick, Michael N.; Joswig, Joseph C.; Powell, Mark W.; Torres, Recaredo J.; Mittman, David S.; Abramyan, Lucy; Crockett, Thomas M.; Shams, Khawaja S.; Fox, Jason M.; hide

    2010-01-01

    A software program that searches and browses mission data emulates a Web browser, containing standard meta - phors for Web browsing. By taking advantage of back-end URLs, users may save and share search states. Also, since a Web interface is familiar to users, training time is reduced. Familiar back and forward buttons move through a local search history. A refresh/reload button regenerates a query, and loads in any new data. URLs can be constructed to save search results. Adding context to the current search is also handled through a familiar Web metaphor. The query is constructed by clicking on hyperlinks that represent new components to the search query. The selection of a link appears to the user as a page change; the choice of links changes to represent the updated search and the results are filtered by the new criteria. Selecting a navigation link changes the current query and also the URL that is associated with it. The back button can be used to return to the previous search state. This software is part of the MSLICE release, which was written in Java. It will run on any current Windows, Macintosh, or Linux system.

  4. PhyloBot: A Web Portal for Automated Phylogenetics, Ancestral Sequence Reconstruction, and Exploration of Mutational Trajectories.

    Directory of Open Access Journals (Sweden)

    Victor Hanson-Smith

    2016-07-01

    Full Text Available The method of phylogenetic ancestral sequence reconstruction is a powerful approach for studying evolutionary relationships among protein sequence, structure, and function. In particular, this approach allows investigators to (1 reconstruct and "resurrect" (that is, synthesize in vivo or in vitro extinct proteins to study how they differ from modern proteins, (2 identify key amino acid changes that, over evolutionary timescales, have altered the function of the protein, and (3 order historical events in the evolution of protein function. Widespread use of this approach has been slow among molecular biologists, in part because the methods require significant computational expertise. Here we present PhyloBot, a web-based software tool that makes ancestral sequence reconstruction easy. Designed for non-experts, it integrates all the necessary software into a single user interface. Additionally, PhyloBot provides interactive tools to explore evolutionary trajectories between ancestors, enabling the rapid generation of hypotheses that can be tested using genetic or biochemical approaches. Early versions of this software were used in previous studies to discover genetic mechanisms underlying the functions of diverse protein families, including V-ATPase ion pumps, DNA-binding transcription regulators, and serine/threonine protein kinases. PhyloBot runs in a web browser, and is available at the following URL: http://www.phylobot.com. The software is implemented in Python using the Django web framework, and runs on elastic cloud computing resources from Amazon Web Services. Users can create and submit jobs on our free server (at the URL listed above, or use our open-source code to launch their own PhyloBot server.

  5. ZOMG - I. How the cosmic web inhibits halo growth and generates assembly bias

    Science.gov (United States)

    Borzyszkowski, Mikolaj; Porciani, Cristiano; Romano-Díaz, Emilio; Garaldi, Enrico

    2017-07-01

    The clustering of dark matter haloes with fixed mass depends on their formation history, an effect known as assembly bias. We use zoom N-body simulations to investigate the origin of this phenomenon. For each halo at redshift z = 0, we determine the time in which the physical volume containing its final mass becomes stable. We consider five examples for which this happens at z ˜ 1.5 and two that do not stabilize by z = 0. The zoom simulations show that early-collapsing haloes do not grow in mass at z = 0 while late-forming ones show a net inflow. The reason is that 'accreting' haloes are located at the nodes of a network of thin filaments feeding them. Conversely, each 'stalled' halo lies within a prominent filament that is thicker than the halo size. Infalling material from the surroundings becomes part of the filament while matter within it recedes from the halo. We conclude that assembly bias originates from quenching halo growth due to tidal forces following the formation of non-linear structures in the cosmic web, as previously conjectured in the literature. Also the internal dynamics of the haloes change: the velocity anisotropy profile is biased towards radial (tangential) orbits in accreting (stalled) haloes. Our findings reveal the cause of the yet unexplained dependence of halo clustering on the anisotropy. Finally, we extend the excursion-set theory to account for these effects. A simple criterion based on the ellipticity of the linear tidal field combined with the spherical-collapse model provides excellent predictions for both classes of haloes.

  6. The Power and Peril of Web 3.0: It's More than Just Semantics

    Science.gov (United States)

    Ohler, Jason

    2010-01-01

    The Information Age has been built, in part, on the belief that more information is always better. True to that sentiment, people have found ways to make a lot of information available to the masses--perhaps more than anyone ever imagined. The goal of the Semantic Web, often called Web 3.0, is for users to spend less time looking for information…

  7. Sistema web para la gestión de proyectos y el control interno de una empresa

    OpenAIRE

    Espinosa Martín, Francisco Javier

    2011-01-01

    [ES] El proyecto tiene su campo de aplicación en la generación de una aplicación web para la gestión de la empresa. El lenguaje de desarrollo PHP con HTML y el servidor web de uso local Appserv. [EN] The project has its field of application in the generation of a web application for managing the company. The PHP development language with HTML and web server for local use Appserv. Proyecto fin de carrera-Universidad de Salamanca, Escuela Politécnica Superior de Zamora, 2011

  8. Virtual Web Services

    OpenAIRE

    Rykowski, Jarogniew

    2007-01-01

    In this paper we propose an application of software agents to provide Virtual Web Services. A Virtual Web Service is a linked collection of several real and/or virtual Web Services, and public and private agents, accessed by the user in the same way as a single real Web Service. A Virtual Web Service allows unrestricted comparison, information merging, pipelining, etc., of data coming from different sources and in different forms. Detailed architecture and functionality of a single Virtual We...

  9. WebScore: An Effective Page Scoring Approach for Uncertain Web Social Networks

    Directory of Open Access Journals (Sweden)

    Shaojie Qiao

    2011-10-01

    Full Text Available To effectively score pages with uncertainty in web social networks, we first proposed a new concept called transition probability matrix and formally defined the uncertainty in web social networks. Second, we proposed a hybrid page scoring algorithm, called WebScore, based on the PageRank algorithm and three centrality measures including degree, betweenness, and closeness. Particularly,WebScore takes into a full consideration of the uncertainty of web social networks by computing the transition probability from one page to another. The basic idea ofWebScore is to: (1 integrate uncertainty into PageRank in order to accurately rank pages, and (2 apply the centrality measures to calculate the importance of pages in web social networks. In order to verify the performance of WebScore, we developed a web social network analysis system which can partition web pages into distinct groups and score them in an effective fashion. Finally, we conducted extensive experiments on real data and the results show that WebScore is effective at scoring uncertain pages with less time deficiency than PageRank and centrality measures based page scoring algorithms.

  10. Beyond the browser web 2.0 and librarianship

    CERN Document Server

    Bridges, Karl

    2012-01-01

    ""Offering concrete information on how librarians will need to respond to these rapid changes, the title first shares an overview of the history of libraries and the Internet. What follows is an easy-to-read investigation of the impact of second-generation web technologies on library organization and services."" - Library Media Connection

  11. A System for Web-based Access to the HSOS Database

    Science.gov (United States)

    Lin, G.

    Huairou Solar Observing Station's (HSOS) magnetogram and dopplergram are world-class instruments. Access to their data has opened to the world. Web-based access to the data will provide a powerful, convenient tool for data searching and solar physics. It is necessary that our data be provided to users via the Web when it is opened to the world. In this presentation, the author describes general design and programming construction of the system. The system will be generated by PHP and MySQL. The author also introduces basic feature of PHP and MySQL.

  12. The structure of the pelagic food web in relation to water column structure in the Skagerrak

    DEFF Research Database (Denmark)

    Kiørboe, Thomas; Kaas, H.; Kruse, B.

    1990-01-01

    by a doming of the pycnocline, with a deep mixed layer along the periphery and a very shallow pycnocline in central parts. Average phytoplankton size increased with the depth of the upper mixed layer, and the central stratified area was characterized by small flagellates while large and chain-forming diatoms...... on particle surface area rather than particle volume or chl a, and showed a distributional pattern that was nearly the inverse of the distribution of copepod activity. That is, peak bacterial growth rates occurred in central, stratified parts and lower rates were found along the margin with a deep mixed layer....... Thus a 'microbial loop' type of food web seemed to be evolving in the central, strongly stratified parts of the Skagerrak, while a shorter 'classical' type of food web appeared to dominate along the margin. The relation between food web structure and vertical mixing processes observed on oceanwide...

  13. Responsive web design workflow

    OpenAIRE

    LAAK, TIMO

    2013-01-01

    Responsive Web Design Workflow is a literature review about Responsive Web Design, a web standards based modern web design paradigm. The goals of this research were to define what responsive web design is, determine its importance in building modern websites and describe a workflow for responsive web design projects. Responsive web design is a paradigm to create adaptive websites, which respond to the properties of the media that is used to render them. The three key elements of responsi...

  14. Learning about the Human Genome. Part 2: Resources for Science Educators. ERIC Digest.

    Science.gov (United States)

    Haury, David L.

    This ERIC Digest identifies how the human genome project fits into the "National Science Education Standards" and lists Human Genome Project Web sites found on the World Wide Web. It is a resource companion to "Learning about the Human Genome. Part 1: Challenge to Science Educators" (Haury 2001). The Web resources and…

  15. Automated Functional Testing based on the Navigation of Web Applications

    Directory of Open Access Journals (Sweden)

    Boni García

    2011-08-01

    Full Text Available Web applications are becoming more and more complex. Testing such applications is an intricate hard and time-consuming activity. Therefore, testing is often poorly performed or skipped by practitioners. Test automation can help to avoid this situation. Hence, this paper presents a novel approach to perform automated software testing for web applications based on its navigation. On the one hand, web navigation is the process of traversing a web application using a browser. On the other hand, functional requirements are actions that an application must do. Therefore, the evaluation of the correct navigation of web applications results in the assessment of the specified functional requirements. The proposed method to perform the automation is done in four levels: test case generation, test data derivation, test case execution, and test case reporting. This method is driven by three kinds of inputs: i UML models; ii Selenium scripts; iii XML files. We have implemented our approach in an open-source testing framework named Automatic Testing Platform. The validation of this work has been carried out by means of a case study, in which the target is a real invoice management system developed using a model-driven approach.

  16. Tracing agents and other automatic sampling procedures for the World Wide Web

    OpenAIRE

    Aguillo, Isidro F.

    1999-01-01

    Many of the search engines and recovery tools are not suitable to make samples of web resources for quantitative analysis. The increasing size of the web and its hypertextual nature offer opportunities for a novel approach. A new generation of recovering tools involving tracing hypertext links from selected sites are very promising. Offering capabilities to automate tasks Extracting large samples of high pertinence Ready to use in standard database formats Selecting additional resour...

  17. Student participation in World Wide Web-based curriculum development of general chemistry

    Science.gov (United States)

    Hunter, William John Forbes

    1998-12-01

    This thesis describes an action research investigation of improvements to instruction in General Chemistry at Purdue University. Specifically, the study was conducted to guide continuous reform of curriculum materials delivered via the World Wide Web by involving students, instructors, and curriculum designers. The theoretical framework for this study was based upon constructivist learning theory and knowledge claims were developed using an inductive analysis procedure. This results of this study are assertions made in three domains: learning chemistry content via the World Wide Web, learning about learning via the World Wide Web, and learning about participation in an action research project. In the chemistry content domain, students were able to learn chemical concepts that utilized 3-dimensional visualizations, but not textual and graphical information delivered via the Web. In the learning via the Web domain, the use of feedback, the placement of supplementary aids, navigation, and the perception of conceptual novelty were all important to students' use of the Web. In the participation in action research domain, students learned about the complexity of curriculum. development, and valued their empowerment as part of the process.

  18. Research on Web-Based Networked Virtual Instrument System

    International Nuclear Information System (INIS)

    Tang, B P; Xu, C; He, Q Y; Lu, D

    2006-01-01

    The web-based networked virtual instrument (NVI) system is designed by using the object oriented methodology (OOM). The architecture of the NVI system consists of two major parts: client-web server interaction and instrument server-virtual instrument (VI) communication. The web server communicates with the instrument server and the clients connected to it over the Internet, and it handles identifying the user's name, managing the connection between the user and the instrument server, adding, removing and configuring VI's information. The instrument server handles setting the parameters of VI, confirming the condition of VI and saving the VI's condition information into the database. The NVI system is required to be a general-purpose measurement system that is easy to maintain, adapt and extend. Virtual instruments are connected to the instrument server and clients can remotely configure and operate these virtual instruments. An application of The NVI system is given in the end of the paper

  19. Interpretation of coagulation test results using a web-based reporting system.

    Science.gov (United States)

    Quesada, Andres E; Jabcuga, Christine E; Nguyen, Alex; Wahed, Amer; Nedelcu, Elena; Nguyen, Andy N D

    2014-01-01

    Web-based synoptic reporting has been successfully integrated into diverse fields of pathology, improving efficiency and reducing typographic errors. Coagulation is a challenging field for practicing pathologists and pathologists-in-training alike. To develop a Web-based program that can expedite the generation of a individualized interpretive report for a variety of coagulation tests. We developed a Web-based synoptic reporting system composed of 119 coagulation report templates and 38 thromboelastography (TEG) report templates covering a wide range of findings. Our institution implemented this reporting system in July 2011; it is currently used by pathology residents and attending pathologists. Feedback from the users of these reports have been overwhelmingly positive. Surveys note the time saved and reduced errors. Our easily accessible, user-friendly, Web-based synoptic reporting system for coagulation is a valuable asset to our laboratory services. Copyright© by the American Society for Clinical Pathology (ASCP).

  20. Accelerating cancer systems biology research through Semantic Web technology.

    Science.gov (United States)

    Wang, Zhihui; Sagotsky, Jonathan; Taylor, Thomas; Shironoshita, Patrick; Deisboeck, Thomas S

    2013-01-01

    Cancer systems biology is an interdisciplinary, rapidly expanding research field in which collaborations are a critical means to advance the field. Yet the prevalent database technologies often isolate data rather than making it easily accessible. The Semantic Web has the potential to help facilitate web-based collaborative cancer research by presenting data in a manner that is self-descriptive, human and machine readable, and easily sharable. We have created a semantically linked online Digital Model Repository (DMR) for storing, managing, executing, annotating, and sharing computational cancer models. Within the DMR, distributed, multidisciplinary, and inter-organizational teams can collaborate on projects, without forfeiting intellectual property. This is achieved by the introduction of a new stakeholder to the collaboration workflow, the institutional licensing officer, part of the Technology Transfer Office. Furthermore, the DMR has achieved silver level compatibility with the National Cancer Institute's caBIG, so users can interact with the DMR not only through a web browser but also through a semantically annotated and secure web service. We also discuss the technology behind the DMR leveraging the Semantic Web, ontologies, and grid computing to provide secure inter-institutional collaboration on cancer modeling projects, online grid-based execution of shared models, and the collaboration workflow protecting researchers' intellectual property. Copyright © 2012 Wiley Periodicals, Inc.